Corpus Linguistics Around the World
Transcription
Corpus Linguistics Around the World
&RUSXVOLQJXLVWLFV DURXQGWKHZRUOG /$1*8$*($1'&20387(56 678',(6,135$&7,&$//,1*8,67,&6 1R HGLWHGE\ &KULVWLDQ0DLU &KDUOHV)0H\HU 1HOOHNH2RVWGLMN &RUSXVOLQJXLVWLFV DURXQGWKHZRUOG (GLWHGE\ $QGUHZ:LOVRQ 'DZQ$UFKHU 3DXO5D\VRQ $PVWHUGDP1HZ<RUN1< &RYHUGHVLJQ3LHU3RVW &RYHULPDJH1$6$WKH9LVLEOH(DUWKKWWSYLVLEOHHDUWKQDVDJRY 2QOLQHDFFHVVLVLQFOXGHGLQSULQWVXEVFULSWLRQV VHHZZZURGRSLQO 7KH SDSHU RQ ZKLFK WKLV ERRN LV SULQWHG PHHWV WKH UHTXLUHPHQWV RI ,62,QIRUPDWLRQDQGGRFXPHQWDWLRQ3DSHUIRUGRFXPHQWV 5HTXLUHPHQWVIRUSHUPDQHQFH ,6%1836-4ERXQG (GLWLRQV5RGRSL%9$PVWHUGDP1HZ<RUN1< 3ULQWHGLQ7KH1HWKHUODQGV Contents Preface Andrew Wilson, Dawn Archer and Paul Rayson Methodology and steps towards the construction of EPEC, a corpus of written Basque tagged at morphological and syntactic levels for automatic processing Aduriz I., Aranzabe M.J., Arriola J.M., Atutxa A., Díaz de Ilarraza A., Ezeiza N., Gojenola K., Oronoz M., Soroa A., Urizar R. 1 The mood of the (financial) markets: in a corpus of words and of pictures Khurshid Ahmad, David Cheng, Tugba Taskaya, Saif Ahmad, Lee Gillam, Pensiri Manomaisupat, Hayssam Traboulsi and Andrew Hippisley 17 Towards a methodology for corpus-based studies of linguistic change: Contrastive observations and their possible diachronic interpretations in the Korpus 2000 and Korpus 90 General Corpora of Danish Jørg Asmussen 33 Synchronic and diachronic variation: the how and why of sociolinguistic corpora Kate Beeching 49 Statistical analysis of the source origin of Maltese Roderick Bovingdon and Angelo Dalli 63 Discovering regularities in non-native speech Julie Carson-Berndsen, Ulrike Gut and Robert Kelly 77 Tracking lexical changes in the reference corpus of Slovene texts Vojko Gorjanc 91 Relating linguistic units to socio-contextual information in a spontaneous speech corpus of Spanish José María Guirao, Antonio Moreno Sandoval, Ana González Ledesma, Guillermo de la Madrid, Manuel Alcántara 101 An analysis of lexical text coverage in contemporary German Randall L. Jones 115 Analysing a semantic corpus study across English dialects: Searching for paradigmatic parallels Sarah Lee and Debra Ziegeler 121 The curse and the blessing of mobile phones – a corpus-based study into American and Polish rhetorical conventions Agnieszka LeĔko-SzymaĔska 141 Using a dedicated corpus to identify features of professional English usage: What do “we” do in science journal articles? Judy Noguchi, Thomas Orr and Yukio Tono 155 Methods and tools for development of the Russian Reference Corpus Serge Sharoff 167 A profile-based calculation of region and register variation: the synchronic and diachronic status of the two main national varieties of Dutch Dirk Speelman, Stefan Grondelaers and Dirk Geeraerts 181 A multilingual learner corpus in Brazil Stella E. O. Tagnin 195 Quantitative or qualitative content analysis? Experiences from a cross-cultural comparison of female students’ attitudes to shoe fashions in Germany, Poland and Russia Andrew Wilson and Olga Moudraia 203 Survey and Prospect of China’s Corpus-Based Research Yang Xiao-jun 219 Corpus linguistics around the world Andrew Wilson, Dawn Archer and Paul Rayson Lancaster University Preface The scope of corpus-based research is becoming ever wider. Not so many years ago, the vast majority of corpus-linguistic research was concerned with the grammar and vocabulary of standard language varieties – the latter meaning, in many cases, British or American English. Whilst research on other topics, languages, and varieties was by no means completely absent from the scene, this was the general picture of the field which came across to the interested observer. Today, things have changed dramatically. As this volume shows, the range of languages, research questions, and, indeed, methodologies which are addressed by corpus linguists has diversified. It is probably true to say that none of the papers published in this volume focuses primarily on standard English as a general variety. Here we find work not only on English dialects (Lee & Ziegeler) but also on learner language (Carson-Berndsen, Gut & Kelly; Lenko-Szymanska; Tagnin) and on a wide range of world languages - Basque (Aduriz et al.), Chinese (Xiao-jun), Danish (Asmussen), Dutch (Speelman et al.), German (Jones), Maltese (Bovingdon & Dalli), Russian (Sharoff), Slovene (Gorjanc), and Spanish (Guirao et al.). In terms of the research questions addressed, the more ‘traditional’ areas of corpus linguistics are still well represented, with papers on vocabulary (Jones), spoken language (Carson-Berndsen, Gut & Kelly; Guirao et al.), synchronic and diachronic variation (Asmussen, Beeching; Bovingdon & Dalli; Gorjanc; Lee & Ziegeler; Speelman et al.), Languages for Special Purposes (Noguchi, Orr & Tono), tagging, and corpus development (Aduriz et al.; Sharoff). However, exciting new departures are also present, with corpus-based work now extending into areas such as cross-cultural rhetoric and social psychology (LenkoSzymanska; Wilson & Moudraia) and even economic forecasting (Ahmad et al.). The papers published in this volume are but a small selection from the many which were presented at the Corpus Linguistics 2003 conference, held at Lancaster University in March 2003. This was the second Corpus Linguistics conference which we hosted at Lancaster (the first was in 2001), and, like its predecessor, it truly amazed us with the range of corpus-informed work being carried out world-wide. Computer corpus linguistics continues to thrive and to extend into so many areas of inquiry, many of which would probably have been unimaginable for its pioneers in the 1960s and 1970s. We are sure that it will continue to hold surprises for us. In the meantime, perhaps this collection of recent corpus research around the globe will whet the reader’s appetite to follow up new developments in this fast-moving field. Andrew Wilson Dawn Archer Paul Rayson Lancaster, August 2004 Methodology and steps towards the construction of EPEC, a corpus of written Basque tagged at morphological and syntactic levels for automatic processing Aduriz I.*, Aranzabe M.J., Arriola J.M., Atutxa A., Díaz de Ilarraza A., Ezeiza N., Gojenola K., Oronoz M., Soroa A., Urizar R. University of the Basque Country * University of Barcelona. Abstract In this article, we will describe the different steps in the construction of EPEC (Reference Corpus for the Processing of Basque). EPEC is a corpus of standard written Basque that has been manually tagged at different levels (morphology, surface syntax, phrases) and is currently being hand-tagged at deep syntax level following the Dependency Structurebased Scheme. It is aimed to be a "reference" corpus for the development and improvement of several NLP tools for Basque. This corpus has already been used for the construction of some tools such as a morphological analyser, a lemmatiser, or a shallow syntactic analyser. 1. Introduction When specifying the strategic priorities for the development of language technology in minority languages, Sarasola (2000) stated: Language foundations and research are essential to create any tool or application; but in the same way tools and applications will be very helpful in research and improving language foundations. Therefore, these three levels [applications, tools, and language foundations] have to be incrementally developed in a parallel and coordinated way in order to get the best benefit possible. Moreover, Sarasola (2000) proposes five phases as a general strategy to follow in the processing of a language: (1) laying foundations, (2) basic tools, (3) tools of medium complexity, (4) advanced tools and multilinguality, and (5) general applications. In all the phases proposed, corpora, first raw and then tagged, stand out as an essential language resource. In this article, we will describe the different steps in the construction of EPEC (Reference Corpus for the Processing of Basque). EPEC is a corpus of standard written Basque that has been manually tagged at different levels (morphology, surface syntax, phrases) and is currently being hand tagged at deep syntax level. 2 Aduriz et al. It is aimed to be a “reference” corpus for the development and improvement of several NLP tools for Basque. In section 2, we explain how the raw corpus was compiled and we briefly describe the design of the tagset. In section 3, we account for the morphological disambiguation process carried out manually over the outcome of MORFEUS (the morphological analyser for Basque). The shallow syntactic tagging and phrase tagging are explained in sections 4 and 5 respectively. Finally, in section 6 we explain the tag system chosen for the dependency-based syntactic analysis and how the treebank is being tagged manually. In figure 1, we can see a diagram showing the different phases in the construction of EPEC, contrasting the manual tasks (right column) with the computer-based ones (left column) as well as the dependencies between them. 2. The tagged corpus 2.1 Compilation of the corpus EPEC is a 50,000-word sample collection of standard written Basque. It is a strategic resource for the processing of Basque and it has already been used for the development and improvement of some tools. Half of this collection was obtained from the Statistical Corpus of 20th Century Basque (http://www.euskaracorpusa.net). The other half was extracted from Euskaldunon Egunkaria (http://www.egunero.info), the only daily newspaper written entirely in standard Basque. The Statistical Corpus of 20th Century Basque is a reference corpus of Basque including 4,658,036 word-forms. It was created by UZEI (http://www.uzei.com), a non-profit organisation devoted to making the Basque language suitable for any specialised field. The corpus was constructed on the basis of an exhaustive inventory of 20th century Basque publications, from which a random sampling was extracted. This corpus has become an invaluable linguistic reference for written Basque of this period. It was classified taking into account the following criteria: the publications were divided into 4 periods (1900-1939, 1940-1968, 1969-1990, 1991-1999), 6 different dialects (Biscayan, Guipuzcoan, Souletin, Labourdin-Navarrese, Standard Basque, and non-classified), and 14 genres (literary prose, poetry, theatre, administration, newspapers...). Each book or article also contained information about the author(s) and its title. A subcorpus of about 25,000 word-forms was extracted from this corpus in order to build EPEC. Texts written in standard Basque, corresponding to the last period (1991-1999) and belonging to both literary and non-literary prose, were chosen for this purpose. The second part of EPEC consists of several articles extracted from the Euskaldunon Egunkaria written in the second half of 1999 and in 2000. The The construction of EPEC 3 articles were chosen so that they covered an assorted range of topics (economics, culture, entertainment, international, local, opinion, politics, sports...). PARSER CHUNKER SHALLOW MORPHOSYNTACTIC SYNTAX TAGGING TAGGING COMPILATION OF TEXTS MANUAL DISAMBIGUATION MORPHEUS CG DISAMBIGUATION ( LINGUISTIC KNOWLEDGE) STOCHASTIC DISAMBIGUATION COMPARISON CG DISAMBIGUATION ( LINGUISTIC KNOWLEDGE) MANUAL DISAMBIGUATION COMPARISON Future work CG CG CHUNKER ( LINGUISTIC KNOWLEDGE) MANUAL REVIEW PARSER ( LINGUISTIC KNOWLEDGE) MANUAL TAGGING COMPARISON EPEC Forthcoming Figure 1. Sketch of the different steps in the completion of EPEC . Figure 1: Sketch of the different steps in the completion of EPEC 4 2.2 Aduriz et al. Design of the tagset Choosing an appropriate tagset is a crucial task since the usefulness of further applications depends on it. The main problem we found while defining the tagset for Basque was the absence of an exhaustive one for automatic use. Moreover, printed dictionaries of Basque also lacked systematisation of categories. For the morphosyntactic treatment of Basque texts, the tag system we developed is a four level system, ranging from the simplest part-of-speech tagging scheme to full morphosyntactic information. At the first level, 20 general categories are included for lexical items (noun, adjective, verb, pronoun, conjunction...). At the second one, each category tag is further refined by subcategory tags. For instance, the category ‘pronoun’ has 6 subcategories: common, emphatic, interrogative, indefinite, reflexive and reciprocal. The third level includes some basic morphosyntactic information such as declension case, number, etc. This morphological information is carried by the dependent morphemes attached to the stem. The full output of the morphosyntactic analysis constitutes the fourth level of tagging. The only difference between this and the previous level is that, here, all the morphological information is considered along with the tags for syntactic functions. Morphology and syntax are closely related in Basque, so most syntactic functions are provided by the database, along with inflexional morphemes. For instance, the ergative case in Basque marks the subject in a clause (with transitive verbs) the absolutive case may either indicate the subject or the predicative (with intransitive verbs), or the direct object. The specification at this level is very detailed and constitutes the input for the morphosyntactic disambiguation process as well as for syntactic and other types of language processing. In addition to these four levels, further tags are added to mark verb chains, noun phrases, and postpositional phrases (see sections 4.3.1 and 4.3.2) Nowadays, we are involved in the syntactic tagging of the corpus, following the Dependency Structure-based Scheme (see section 5). About 31 syntactic tags are being used for this purpose. 3. Morphosyntactic tagging of the corpus A morphological analyser of words is an indispensable basic tool when defining a general framework for the automatic processing of agglutinative languages like Basque (Aduriz et al., 1998). However, prior to the completion of the The construction of EPEC 5 morphological analyser MORFEUS, the design of the tagset had to be accomplished (see section 2.2) and a lexical database developed. 3.1 EDBL, a lexical database for Basque EDBL (Aldezabal et al., 2001) is a general-purpose lexical database used in Basque text-processing tools. This large repository of lexical knowledge is the basis for many different NLP tasks, and provides lexical information for several language tools including, obviously, the morphological analyser. At present, it consists of nearly 80,000 entries divided into (i) dictionary entries (the same found in any conventional dictionary), (ii) inflected verb forms, and (iii) dependent morphemes, all of them with their respective morphological information. 3.2 MORFEUS, automatic morphological analyser MORFEUS is a robust morphological analyser for Basque. It is a basic tool for current and future work on NLP. The analyser is based on the two-level formalism proposed by Koskenniemi (1983), which has had widespread acceptance due mostly to its general applicability, declarativeness of rules and clear separation between linguistic knowledge and program. The architecture of the analyser was defined using three main modules: 1 The standard analyser that uses a general lexicon and a user’s lexicon. This module is able to analyse and generate standard language wordforms. In our applications for Basque, we defined more than 130 patterns of morphotactics and two-rule systems in cascade, the first one for longdistance dependencies among morphemes and the second one for morphophonological changes. These elements are compiled together in the standard transducer. 2 The analysis and normalization of linguistic variants (dialectal uses and competence errors). Due to non-standard or dialectal uses of the language and competence errors, the standard morphology is not enough to offer good results when analysing real text corpora. This problem becomes critical in languages like Basque in which standardisation is in process and dialectal forms are still in widespread use. For this process, the standard transducer is extended with new lexical entries and phonological rules producing the enhanced transducer. 3 The guesser or analyser of words without lemmas in the lexicons. In this case, the standard transducer is simplified removing the lexical entries and allowing the analysis of any string. Therefore, the standard transducer is substituted by a general transducer to describe any combination of characters. 6 Aduriz et al. The morphological analyser gives as a result all the possible analyses of each token in the text. 3.3 Manual disambiguation of the corpus The manual disambiguation of the corpus was performed on the output of MORFEUS. Thus, the whole corpus was morphosyntactically analysed giving to each word-form every possible analysis, without taking into account the context in which it appeared. Once each word-form in the corpus was morphosyntactically analysed, we carried out the manual disambiguation process. Two linguists independently assigned the correct syntactic tag to each word in the corpus, applying the “double blind” method described in Voutilainen & Järvinen (1995). In case no right tag had been automatically assigned, they typed it themselves. Both linguists’ answers were compared and, when differences occurred, they agreed a single tag. This manually disambiguated corpus was used both to improve a Constraint Grammar disambiguator and to develop a stochastic tagger. After the corpus was manually disambiguated, we started to construct a grammar of constraint rules that would automatically select the correct syntactic tags in any real corpus. For this purpose, we chose the Constraint Grammar (CG) formalism (Karlsson et al., 1995; Tapanainen & Voutilainen, 1994), which was designed with the aim of being a language-independent and robust tool to disambiguate and analyse unrestricted texts. The CG grammar statements are close to real text sentences and directly address some crucial parsing problems, especially ambiguity. The role of the CG system is to apply a set of linguistic constraints that discard as many alternatives as possible, leaving at the end the most fully disambiguated sentences possible. Each rule produced for this grammar was checked on the manually disambiguated corpus so as to test its goodness and improve it iteratively whenever necessary. Moreover, in the cases in which the analyser did not assign any correct analysis to a word-form in the corpus, the linguists contributed greatly to the improvement of the lexical database and the analyser itself. Besides this, we also developed a stochastic tagger. Statistical methods need little effort and obtain very good results (Church, 1998; Cutting et al., 1992), at least when applied to English. In our case, we selected the TATOO tagger based on Hidden Markov Models (Armstrong et al., 1995). TATOO was designed to be applied to the output of a morphological analyser and the tagset can be easily switched without changing the input text. However, because Basque is an agglutinative and free-order language, the stochastic tagger turned out to be much less accurate than for English when trained directly on the output of the morphological analyser. So, we performed a supervised training on the output of the CG grammar. Since the CG The construction of EPEC 7 disambiguator leaves a relatively low ambiguity rate, the results of TATOO were much better. Currently, we apply a combination of the CG disambiguator with the stochastic tagger and get good results (Ezeiza 2003). The CG disambiguator is first applied and then the remaining ambiguities are solved using the results of TATOO. 4. Shallow syntax tagging After disambiguating the morphological tags in the corpus, the next step was to assign the corresponding syntactic tag to each word-form. Syntactic function tags follow the philosophy of the Constraint Grammar (CG) formalism in the sense that they are based on a functionally labelled dependency syntax 1 . By adopting the CG formalism, we express the syntactic functions of words and the interdependencies that exist among them rather than deep structural relations. So the syntactic tags at this level refer to shallow syntactic functions, i.e. they may provide information about the surface structure of verb chains, noun phrases, or postpositional phrases. Therefore, this results in a shallow parsing of the corpus. As we mentioned before, most syntactic functions are added to the word-forms together with inflectional morphemes. Morphological suffixes and syntactic functions are closely related in Basque and both are included in the database. Thus, the output of the morphological analyser displays most of these shallow syntactic tags. However, some other syntactic tags that are not inherited from the database are added to the analysis through CG mapping rules. These functions are mostly attached to parts of speech, and they are generally assigned to word-forms provided that they comply with some given contextual conditions. Mainly, the syntactic function tags are divided into three groups: main functions (subject, object, indirect object…), modifiers (indicating the direction relative to their head), and verb functions (used to detect verb chains). This distinction of the syntactic functions is essential for the tagging of the different kinds of phrases (see section 5). The ambiguity rate related to the shallow syntactic tagging is over 22% 2 , that is, for each 100 word-forms 22 are assigned more than one syntactic tag. 4.1 Manual disambiguation and applications Once each word-form in the corpus was given at least one syntactic tag, we repeated the manual disambiguation process. This method was similar to the one used for the morphological disambiguation in the previous step. Two linguists independently assigned the correct syntactic tag to each word in the corpus or, in case no right tag had been automatically assigned, they typed it themselves. Then, both linguists agreed a single tag when differences occurred. After the corpus was manually disambiguated, we started to make up a grammar of constraint rules that 8 Aduriz et al. would automatically select the correct syntactic tags in any real corpus. Each rule produced was checked on the manually disambiguated corpus so as to test its goodness and improve it if necessary. 5. Tagging phrases At this stage we have the corpus manually tagged with surface syntactic tags following the CG syntax. No phrase units are marked yet, although based on this representation, the identification of various kinds of phrase units, such as verb chains, noun phrases, and postpositional phrases is reasonably straightforward. 5.1 Tags for verb chains In order to detect verb chains, we use the verb function tags (@+FAUXVERB, @-FAUXVERB, @+FMAINVERB, @-FMAINVERB 3 …) and some particles (the negative particle, modal particles…). Based on these elements we are able to detect not only continuous verb chains but also dispersed ones. So as to mark up continuous verb chains, the following tags are attached using again CG mapping rules: • %VCH: this tag is attached to a verb chain composed of a single element. • %VCHI: this is attached to the initial element of a complex verb chain. • %VCHF: this is attached to the final element of a complex verb chain. The tags used to mark-up the dispersed verb chains are: • %NCVCHI: this tag is attached to the initial element of a noncontinuous verb chain. • %NCVCHC: this tag is attached to the second element of a noncontinuous verb chain. • %NCVCHF: this tag is attached to the final element of a non-continuous verb chain. 5.2 Tags for noun phrases and postpositional phrases Our assumption is that any word having a modifier function tag is linked to some word with a main syntactic function tag. Moreover, a word with a main syntactic function tag can, by itself, constitute a phrase unit. With this in mind, we established three tags to mark up this kind of phrase units (noun phrases or postpositional phrases): • %PHR: this tag is attached to words with main syntactic function tags that constitute a phrase unit by themselves. The construction of EPEC 9 • %PHRI: this tag is attached to the initial element of a phrase unit. • %PHRF: this tag is attached to the final element of a phrase unit. In order to attach one of these tags to each word-form, we have simultaneously developed two subgrammars containing CG mapping rules. The first subgrammar is aimed at delimiting verb chains whereas the second one marks noun and postpositional phrases. 5.3 Manual tagging and applications At present, a linguist is checking the tags that the first set of mapping rules marked up in the corpus. Whenever necessary, she adds, removes, or changes the tags automatically assigned. Once this work is finished, the first set of mapping rules that were developed will be tested on the corpus and the results will be used to improve the rules iteratively as well as to develop new ones. 6. Treebank The next logical stage in the completion of the corpus is deep syntax tagging, in order to build a treebank (Aduriz et al., 2002.) Although manually tagging a treebank is an expensive and time-consuming task, it is also an essential step for the development of syntactic tools and applications for Basque. A group of linguists in our research group is currently involved in this arduous task 4 . dependant structurally_case- marked complements nc arg_mod clausal modifier detmod ncsubj ncobj nczobj finite_clause non-finite_clause ccomp_subj ccomp_obj ccomp_zobj xcomp_subj auxiliary conjunction nc clausal pred ncmod cmod xmod xcomp_obj ncpred xpred xcomp_zobj Figure 2: Hierarchy of grammatical relations. After considering a number of diverse choices (including Skut et al., 1997; Oflazer, 1999) we decided to follow a dependency-based procedure, for it was, in our opinion, the one that could best deal with the free word order displayed by Basque syntax. The dependency-based analysis describes the relations existing between components (i.e. word-forms). This way, for each sentence in the corpus 10 Aduriz et al. we explicitly determine the syntactic dependencies between the heads and the dependents. In order to define the syntactic tagging system, we adopted the framework presented in Carroll et al. (1998, 1999). By following this line of work, we developed a coding-system based on hierarchies of grammatical relations, both for lexical and empty elements, such as pro (see figure 2). As can be seen in figure 2, the hierarchy distinguishes between several general levels, which are further specified in subsequent levels. Thus, for instance, in the general level we find structurally case-marked complements, thematic roles (arg_mod), modifiers, auxiliaries and conjunctions. In turn, structurally casemarked complements, for example, are divided into noun phrases and clauses. Each continuous gradation achieves further specification by taking into account their grammatical function (e.g. ncsubj, ncobj, and nczobj). Next, we present an example showing some of the grammatical relations specified in the hierarchy: ncsubj (Case, Head, Head of NP, the Case-marked element within NP, subj ) ncobj (Case, Head, Head of NP, the Case-marked element within NP, obj) nczobj 5 (Case, Head, Head of NP, the Case-marked element within NP, ind.obj) These are examples of structurally case-marked complements when complements are nc (non-clausal, Noun Phrases, henceforth NP), as, for instance, in the sentence Aitak haurrari sagarra eman dio ‘Father has given an apple to the child’ (literally ‘Father to-child apple given has’): ncsubj (erg, eman, aitak, aitak, subj) nczobj (dat, eman, haurrari, haurrari, ind.obj) ncobj (abs, eman, sagarra, sagarra, obj) This description is extremely important, since it determines the number and type of tags needed for each relation (number of slots, the characteristics of each one, etc.). This formalisation will be very useful for future treatments, for example, to transfer all this information in to XML format (see section 7). Tagging the corpus manually has enabled us to find solutions to problems that emerge in the analysing process, such as discontinuous constituents, coordination, or comparative clauses. Moreover, it is not unusual that similar phenomena are treated as distinct by the different linguists tagging the corpus. In these cases, the group of linguists tries to agree a single analysis that will be regarded as correct thereafter. Consequently, as the tagging process goes on and we find new solutions to arising problems, accuracy, robustness, and speed will improve. Besides, we are The construction of EPEC 11 currently developing a computational tool designed to make the manual tagging easier and faster. All of this work is being carried out within a project that aims at constructing treebanks for Catalan, Spanish, and Basque (Civit & Martí, 2002). 6.1 Applications When the manual tagging of the corpus is finished, we plan to develop a tool based on linguistic knowledge that will be able to parse real corpora automatically. As in the previous steps of manual tagging, each rule produced for the parser will be tested on the manually tagged corpus in order to assess its effectivity and improve it accordingly. In the future, we also plan to apply machine learning methods to the corpus, in order to carry out automatic tagging. 7. Representation of the Corpus using XML Over the last three years much effort has been made in our research group (Artola et al., 2002) to integrate the NLP tools for Basque described in this chapter. Due to the complexity of the information to be exchanged among the tools, Feature Structures (FSs) are used to represent it. Feature structures are coded following the TEI’s DTD for FSs, and Feature Structure Definition descriptions (FSD) have been thoroughly defined. The documents used as input and output of the different tools, contain TEI-P4-conformant feature structures (FS) coded in XML. The use of XML for encoding the I/O streams flowing between programs forces us to describe the mark-up formally, and provides software to check that these mark-up hold invariantly in an annotated corpus. We could deeply analyse the framework for linguistic knowledge representation and integration developed in our group, but as it is not the goal of this paper, we will only show the output of the tools of the analysis chain (figure 3). Different representations of the sentence Noizean behin itsaso aldetik Donostiako Ondarreta hondartzara enbata iristen da (Once in a while, a storm arrives from high seas to the Donostia’s beach of Ondarreta) coded in XML are shown in figure 3. 12 Aduriz et al. <text id=‘T1> .. <P>Noizean behin itsaso aldetik Donostiako Ondarreta hondartzara enbata iristen da.</p></text> .xml ... </p> <p id='w'> <linkGrp type=‘MWLU‘ tagOrder='y'> <link ID=‘mwlnk1 targets='Xw1 Xw2'/> ... <w id='w1' sameAs='Xw1' type=‘BEG_CAP'>Noizean</w> <w id='w2' sameAs='Xw2'>behin</w> <w id='w3' sameAs='Xw3'>itsaso</w> <w id='w4' sameAs='Xw4'>aldetik</w> <w id='w5' sameAs='Xw5' type=‘BEG_CAP'>Donostiako</w> <w id='w6' sameAs='Xw6' type=‘BEG_CAP'>Ondarreta</w> <w id='w7' sameAs='Xw7'>hondartzara</w> <w id='w8' sameAs='Xw8'>enbata</w> <w id='w9' sameAs='Xw9'>iristen</w> <w id='w10' sameAs='Xw10'>da</w> <w id='w11' sameAs='Xw11' type='PUNCT_FSTOP'>.</w> </p> ... w.xml <p id='linkGrp'> <linkGrp type=‘mwlnk-lem‘tagOrder='y'> <link targets=‘mwlnk1 ADV'/> <linkGrp type='w-lem‘tagOrder='y'> <link targets='Xw3 COM-NOUN-1'/> Morphosyntactic Analysis <link targets='Xw4 COM-NOUN-2'/> <text id='LemDoc0001'> .... <fs id="COM-NOUN-2" type="Lemmatisation"> <f name="Form"><str>aldetik</str></f> <f name="Lemma"><str>alde</str></f> <f name="morphological-Features"> <fs type="Top-Features-List"> <f name="POS"><sym value="NOUN"/></f> <f name="SUBCAT"><sym value="COM"/></f> <f name="DET"><sym value="DET"/></f> <f name="NUM"><sym value="S"/></f> <f name="CASE"><sym value="ABL"/></f> </fs> </f> </fs> </p> <p> <fs id="PROP-N-LOC-1" type="Lemmatisation "> <f name="Form"><str>Donostiako</str></f> <f name="Lemma"><str>Donostia</str></f> <f name="Top-Features-List"> <fs type="upper-level-features"> <f name="POS"><sym value="NAME"/></f> <f name="SUBCAT"><sym value="PROP-LOC"/></f> <f name="DET"><sym value="DET"/></f> <f name="NUM"><sym value="S"/></f> <f name="CASE"><sym value="GEL"/></f> </fs> </f> </fs> </p> <p> .lem.xml ... <link targets='Xw5 PROP-N-LOC-1'/> <link targets='Xw6 PROP-N-LOC-2'/> ... .lemlnk.xml EUSLEM Chunker <text id=span0001>... <p id=“linkgrp”> <linkgrp type=“span” tagOrder='y'> <text id=cad0001> ... <fs id=COM-NOUN-1 type="phrase"> <f name="chain"> <str>itsaso aldetik</str></f> <f name="head"> <str>itsaso</str></f> <f name="POS"><str>NOM</str></f> <f name="SUBCAT"><str>COM</str></f> <f name="SFL" ORG=‘list’> <sym value = "NCMOD" > </f> </fs> ... Dependencies .sint.xml <text id=dep0001> ... <linkgrp type="dep" targorder=“y“ targFunc=“head dependant case-dep-unit" domains=???> <!-- ... --> <link id="dep1" targets="Xw9 span1 span1"> <link id="dep2" targets=“Xw9 Xw7 Xw7"> ... </linkgrp> dep.xml <link id=“span1” targets=“Xw3 Xw4”> <link id=“span2” targets=“Xw5 Xw6 Xw7”>... </linkgrp> </p> .spanlnk.xml </text> <text id=span-sint0001> ... <p id=“linkgrp”> <linkgrp type=“span-sint” tagOrder='y'> <link targets=“span1 NOUN-COM”> ... </linkgrp> </p> </text> .sintlnk.xml ... <linkgrp type="dep-deplib" targorder="yes"> <link targets="dep1 D-NCMOD-ABL"/> <link targets="dep2 D-NCMOD-ALA"/> ... </linkgrp>... Deplnk.xml <fs id="D-NCMOD-ALA“ type="dependency"> ... <f name=“nombre"> <sym value="NCMOD-ABL"/></f> <f name=“CASE"> <sym value="ALA"/></f> .. </fs> Figure3: Output of the different tools coded in XML. Figure 3: Output of the different tools coded in XML Deplib.xml The construction of EPEC 8. 13 Future work The lexical information gathered in the lexical database (EDBL), which is the basis for several NLP tools in our research group, is constantly being renewed. New entries from diverse sources are periodically added to the database. Moreover, new tools such as multiword units, named entities, or postposition recognisers have been developed. These changes must be reflected in the corpus, so we must review it regularly. Therefore, in the near future, we intend to update EPEC with new information. This will be done semiautomatically, so that only the new information needs to be reviewed. 9. Acknowledgements This research is supported by the University of the Basque Country (9/UPV00141.226-14601/2002), the Ministry of Industry of the Basque Government (project XUXENG, OD02UN52), the Interministerial Commission for Science and Technology of the Spanish Government (FIT-150500-2002-244), and the European Community (MEANING project, IST-2001-34460). Notes 1 The concept of dependency-syntax has a long tradition in grammatical analyses since the Greco-Roman era. More recently, within the application of formalisms to syntactic theory, among others we find Tèsniere (1959), Hays (1964) and Mel’cuk (1988), the ones who have recovered dependency-syntax in theoretical terms. 2 This ambiguity was estimated taking into account the syntactic functions of a subset of 200 common words. 3 Finite auxiliary verb, non-finite auxiliary, finite main verb, non-finite main verb… 4 In this research line, our group is taking part in the project entitled “The IXA group, tools for an automatic treatment of Basque: creating a database composed of syntactic-semantic trees” (See ‘acknowledgments’) 5 nczobj would be equivalent to the English nciobj (non-clausal indirect object). References Aduriz I., Agirre E., Aldezabal I., Alegria I., Ansa O., Arregi X., Arriola J.M., Artola X., Díaz de Ilarraza A., Ezeiza N., Gojenola K., Maritxalar A., Maritxalar M., Oronoz M., Sarasola K., Soroa A., Urizar R., Urkia M. 14 Aduriz et al. (1998), A Framework for the Automatic Processing of Basque. Proceedings of the First International Conference on Language Resources and Evaluation, Granada. Aduriz I., Aldezabal I., Aranzabe M., Arrieta B., Arriola J., Atutxa A., Díaz de Ilarraza A., Gojenola K., Oronoz M., Sarasola K. (2002), Construcción de un corpus etiquetado sintácticamente para el euskera. Actas del XVIII Congreso de la SEPLN, Valladolid, Spain. Aldezabal I., Ansa O., Arrieta B., Artola X., Ezeiza A., Hernández G., Lersundi M. (2001), EDBL: a General Lexical Basis for the Automatic Processing of Basque. IRCS Workshop on Linguistic Databases, Philadelphia (USA). Alegria I., Aranzabe M., Ezeiza A., Ezeiza N., Urizar R. (2002) Robustness and customisation in an analyser/lemmatiser for Basque. Proceedings of Workshop on "Customizing knowledge in NLP applications ". Third International Conference on Language Resources and Evaluation, Las Palmas de Gran Canaria (Spain). Armstrong S., Russell G., Petitpierre D., Robert G. (1995) An Open Architecture for Multilingual Text Processing. Proceedings of the 7th Conference of the European Chapter of the Association for Computational Linguistics, Dublin, Ireland, pp. 101-106. Artola X., Díaz de Ilarraza A., Ezeiza N., Gojenola K., Hernández G., Soroa A. (2002) A Class Library for the Integration of NLP Tools: Definition and implementation of an Abstract Data Type Collection for the manipulation of SGML documents in a context of stand-off linguistic annotation. Proceedings of the Third International Conference on Language Resources and Evaluation, Las Palmas de G 1 ran Canaria, Spain. Carroll J., Briscoe T., Sanfilippo A. (1998) Parser evaluation: a survey and a new proposal. Proceedings of the International Conference on Language Resources and Evaluation, Granada, Spain, pp. 447-454. Carroll J., Minnen G., Briscoe T. (1999) Corpus Annotation for Parser Evaluation. Proceedings of Workshop on Linguistically Interpreted Corpora, EACL´99, Bergen. Church K. W. (1998) A Stochastic Parts Program and Noun Phrase Parser for Unrestricted Text. Proceedings of the Second Conference on Applied Natural Language Processing, Québec, Canada, pp. 136-143. Civit M., Martí M. (2002) Design Principles for a Spanish Treebank. Proceedings of the Treebanks and Linguistic Theories (TLT2002), Sozopol, Bulgaria. Cutting D., Kupiec J., Pederson J., Sibun P. (1992) A Practical Part-of-speech Tagger. Proceedings of the Third Conference on Applied Natural Language Processing, Philadelphia, USA, pp. 133-140. Ezeiza, N. (2003) Corpusak ustiatzeko tresna linguistikoak. Euskararen etiketatzaile sintaktiko sendo eta malgua. PhD thesis, University of the Basque Country. Hays D. C. (1964) Dependency theory: a formalism and some observations. Language 40, pp. 511-525. The construction of EPEC 15 Karlsson F., Voutilainen A., Heikkila J., Anttila A. (1995) Constraint Grammar: Language-independent System for Parsing Unrestricted Text. Mouton de Gruyter, Berlin. Koskenniemi K. (1983) Two-level Morphology: A general Computational Model for Word-Form Recognition and Production. University of Helsinki, Department of General Linguistics. Publications 11. Mel’cuk I. (1988) A Dependency Syntax: Theory and Practice. State University of New York Press. Oflazer K., Zeynep D., Tür H., Tür G. (1999) Design for a Turkish treebank. Proceedings of Workshop on Linguistically Interpreted Corpora, at EACL, Bergen. Sarasola K. (2000) Strategic priorities for the development of language technology in minority languages. Proceedings of Workshop on "Developing language resources for minority languages: re-usability and strategic priorities". Second International Conference on Language Resources and Evaluation, Athens, Greece. Skut W., Krenn B., Brants T., Uszkoreit H. (1997) An Annotation Scheme for Free Word Order Languages. Fifth Conference on Applied Natural Language Processing, Washington, DC, USA, pp. 88-95. Tapanainen P., Voutilainen A. (1994) Tagging Accurately-Don´t guess if you know. Proceedings of the 4th Conference on Applied Natural Language Processing, Washington. Tesnière L. (1959) Eléments de Syntaxe Structurale, (2nd ed.) Paris, Klincksieck. Voutilainen A., Järvinen T. (1995) Specifying a shallow grammatical representation for grammatical purposes. Proceedings of the 7th Conference of European Association of Computational Linguistics, Dublin. This page intentionally left blank The mood of the (financial) markets: In a corpus of words and of pictures Khurshid Ahmad, David Cheng, Tugba Taskaya, Saif Ahmad, Lee Gillam, Pensiri Manomaisupat, Hayssam Traboulsi and Andrew Hippisley Department of Computing, University of Surrey Abstract Corpora of texts are used typically to study the structure and function of language. The distribution of various linguistic units, comprising texts in a corpus are used to make and test hypotheses relevant to different linguistic levels of description. News reports and editorials have been used extensively to populate corpora for studying language, for making dictionaries and for writing grammar books. News reports of financial markets are generally accompanied by time-indexed series of values of shares, currencies and so on, reflecting the change in value over a period of time. A corpus linguistic method for extracting sentiment indicators, e.g. shares going up or a currency falling down, is presented together with a technique for correlating the quantitative time-series of values with a time series of sentiment indicators. The correlation may be used in the analysis of the movement of shares, currencies and other financial instruments. 1. Introduction Financial markets are places where financial instruments are bought and sold. These instruments include shares, currencies, bonds: there are shares traded for individual organisations and traders take options – slang bet – on the aggregate value of key shares e.g. Financial Times Stock Exchange (FTSE). Some of these instruments are traded in millions, others in thousands and yet others in hundreds: the prices of instruments change frequently during single trading or over a longer trading horizon. A set of buying / selling prices of instruments, ordered in time, is usually referred to as a (quantitative) time series. In the financial pages of newspapers, and now on specialised web sites, these time series are either displayed independently or as graphical illustrations within (long) texts. The buying and selling of instruments in itself causes changes in their value: too many buyers for a share and its value goes up, too many sellers and the value goes down. The so-called efficient market hypothesis (EMH) suggests that the (trading in a) financial market is the sole arbiter of the price of an instrument. Despite the preponderance of the EMH, newspapers and financial web sites regularly report the reactions of individuals, acting on their own or on behalf of organisations or governments: some web sites display results of polls of financial experts. The experts report whether they are either ‘bearish’ – or shy to buy or 18 Ahmad et al. sell, or ‘bullish’ – too eager / aggressive to buy or sell, and indeed some of them are neutral. These polls are conducted at regular intervals and the sentiments of experts are displayed as a time series. And then there are some who ‘correlate’ the time series comprising the bearish / bullish / neutral voting figures with the time series of financial instruments. The correlation is then used as a cipher for buying or disposing of an instrument. The sentiment of the market traders – or market sentiment for short – is shaped by, and in turn shapes, the value of a financial instrument usually in the short term and perhaps in the long term as well. Much as the sentiment of a trader influences others, others also influence him. The view of the others is typically communicated through press statements. One can argue that (financial) news stories may affect the trader’s sentiment, or more precisely, his or her attitude towards an instrument. Ergo, positive news stories may persuade people to invest in the market thereby driving the prices of instruments up: conversely negative or gloomy stories force prices down. Note that the physical prepositions (up/down) used to describe the position/location of a physical object, are also used to describe the change in value of an abstract financial instrument during a fixed time period. News stories and people’s conversation about the financial markets extend these spatial metaphors further by talking in terms of state change – one sees a change in the value of an instrument in terms of rising or falling. The use of literary allusions, including bear/bull, vibrant/anaemic, and more colourful slang, including the phrase dead cat bounce, to describe that the upward movement of stock is much like a lifeless object merely moving because of the laws of gravity, shows a creative use of language in the specialist field of financial trading (Ahmad 2002). The sentiment of a trader toward the market may change by reading a news story in that bullish stories may cheer him or her up, and bearish stories may depress him or her and in turn depress the market (Knowles 1996). We wish to explore whether it is possible to extract the sentiment from a news story through linguistic analysis. It is possible to use the literature on buying/selling in semantic theory (Jackendoff 1991) as a framework for analysing the meaning of the news stories. The literature on natural language processing (Simmons et al 1984) and on knowledge representation suggests that frame semantics has been used to build systems that can, in principle, analyse, extract and disseminate the meaning (intent?) of a specialist news report. Frame semantics has a number of limitations, and a prominent one is the need for a lexicon that is rich and extensive in terms of meaningful data. What about a purely lexical approach? Recall, Quirk et al.’s (1985) observation that relates frequency of a lexical item to the acceptability of that item by an The mood of the (financial) markets 19 educated-native speaker of the language. Stretching this dictum to financial markets, one can argue that higher frequency of phrases that express positive sentiment suggests that all is well in the market. Similarly, a predominance of negative phrases might suggest that all is not well in the market and that it may have fallen or is about to fall. We have analysed three year’s output of Reuters financial news comprising over 10 millions tokens published during 2000-2002. 1.1 rose fell 1 FTSE 100 0.9 Normalised Value 0.8 0.7 0.6 0.5 0.4 0.3 0.2 Jan Feb Mar Apr May Jun Jul Aug Sep Oct Nov Dec Month Figure 1: Monthly variation of “rose”, “fell” and FTSE 100 Figure 1 shows the variation in the frequency of two verbs rose and fell over a one-year period (Ahmad et al. 2002). Also plotted is the value of FTSE-100 at the close of daily trading during 2002. There appears to be an encouraging numerical correlation amongst the sentiment verbs and FTSE-100. One further exploration of the above hypothesis that sentiment may correlate with frequency of phrases that may express positive and negative sentiment used to describe changes in the value of an instrument, requires an understanding of the following: How do we analyse texts such that frequent phrases hypothetically related to a sentiment are indeed used in a sentence or in many sentences to express the Ahmad et al. 20 sentiment? The ambiguity of language does play a significant part in confusing the sense of these phrases, for example, rose, as in something rising, with that of a flower or indeed the name of a person. We show how simple cues related to the grammatical categories of phrases in the neighbourhood of the potential sentiment laden word ensure that the word rise quite x suggests that the value of a financial instrument is rising. How do we use the time series of sentiment word frequency in conjunction with the time series of the values of an instrument either to predict stability or chaos in the market, or to discover the ‘turning point’ in the value of an instrument? The turning point is the point in time when the value of the instrument stops decreasing and starts to increase or vice versa. How do we organise texts in a diachronic corpus such that these texts may be added or omitted according to the pragmatic attributes of the texts? This is important if the market movements and sentiment analysis are required for a specific instrument (e.g. US$, Euro, BT stocks, FTSE derivative) traded in a certain country or group of countries. 2. Sentiment words and their frequency 2.1 A note on up-and-down phrases Amongst the many different phrases chosen to describe the changes in the actual or potential values of financial instruments the phrases up, down, rise and fell have an intuitive prominence. Other related phrases and synonyms are also used: growth, slump, jump and drop are good examples. The British National Corpus (BNC) shows the preponderance in the general language of these up-and-down phrases: Table 1: BNC frequencies obtained from the on-line version of the corpus; for rose and fell we restricted the query to verbs only (NBNC = 100,106,008) Token fBNC Token fBNC up growth rose 207709 12794 5566 Down Fell Slump 92285 9563 632 The BNC shows that there is more ‘positive’ sentiment in the corpus than, say, the ‘negative’. However a much more detailed analysis is required before any serious conclusion can be drawn from these literally raw figures. A similar analysis of financial news texts shows the dominance of the phrases up and down but with a somewhat different distribution of the other four phrases. Consider a The mood of the (financial) markets 21 sample from Reuters UK-financial News for the month of November 2002 (comprising 400,000 or so tokens): Table 2: Frequencies of ‘positive’ and ‘negative’ sentiment phrases based on Reuters UK-financial News November 2002. This is an untagged sample and homographic conflicts (e.g. rose as noun or verb) have not been resolved (NReuters=402089) Token fReuters Token fReuters up growth rose 1435 650 424 down fell slump 716 391 73 The rank order of the three ‘positive’ sentiment phrases and that of the ‘negative’ sentiments is preserved when the register changes from the BNC – a corpus which largely comprises general language texts (c. 62% texts are drawn from fiction, leisure, world affairs, the arts, belief and thought) – to the specialist financial news which wholly comprises, what the BNC compilers would call, commerce and finance texts (the BNC has just under 8% of such texts in its composition). However, the relative distribution of the different phrases within the ‘positive’ and ‘negative’ sentiment categories is substantially different: Table 3: Relative distribution of “positive” and “negative” sentiment phrases across BNC and Reuters UK-financial News November 2002. Token fBNC Up growth Rose Total 207709 12794 5566 Positive =226069 N BNC fBNC / 91.8 5.66 2.46 100% Positive N BNC fReuters 1435 650 424 Positive N Re uters = 2509 fReuters / Positive N BNC 57.2 25.9 16.9 100% The predominance of up and down is reduced by about one-third and the more domain-specific growth and rose increase dramatically when we change the register from general language to special language: from 6% in the BNC to about 26% in Reuters for growth and from under 3% to about 17% for rose. Similar results can be obtained for negative sentiment words. Financial journalists typically express rise and fall in percentage terms: this perhaps gives their story a quantitative and objective look and feel. A concordance of Reuters UK-financial News (November 2002) containing the phrases rose shows a varied usage of the pattern: Ahmad et al. 22 º ª » « by » X percent rose « «only/nearly» » « to ¼ ¬ Some examples of these patterns include: enterprise shares Volatile mortgage payments , home loan repayments , 152 pence, logica Property companies -last week house prices rose rose rose rose rose rose 2.83 by 0.1 to 2.3 over eight nearly seven only 1.4 percent percent percent percent percent percent to 584 pence in on the month to in the year to and cmg climbed 6 to 1,235 last month - - A further analysis of the stories published by Reuters, under the UK-financial rubric throughout the calendar year 2002, shows that not all the above patterns are used as frequently and that one pattern, rose X percent, tends to dominate – or, in other words was used preferentially by the journalists in the year 2002. Table 4: Variations of sentiment rose across year 2002, where fx denotes the raw frequency of the phrases Patterns frose X% by X % to X % over X % Nearly X % by over X % only X % frose [phrase] % Proportion of rose [phase] % Jan 417 58.0% 2.2% Feb 369 52.0% 7.0% 4.1% 0.5% 0.3% 284 238 Mar 245 49.4% 0.8% 1.6% 0.8% Apr 263 58.2% 1.1% 1.5% 0.4% 143 174 May 376 57.4% 1.9% 1.6% 0.8% Jun 245 53.1% 1.6% Jul 427 50.1% 1.9% 0.4% 1.4% 0.3% 0.4% 0.5% 0.3% 149 0.2% 254 0.6% 217 238 Aug 351 51.3% 2.8% 0.3% Sep 357 44.8% 3.6% 0.3% 0.6% 0.3% 190 Oct 342 59.4% 2.9% 2.3% 0.6% Nov 424 50.7% 3.8% 2.8% 1.4% 0.6% 0.7% 0.0% 0.0% 236 0.2% 0.7% 256 Dec 231 63.6% 3.5% 0.4% 175 68.1% 64.5% 58.4% 66.2% 63.3% 60.8% 59.5% 61.8% 53.2% 69.0% 60.4% 75.8% For example, in January 2002, of all the patterns rose [phrase] X percent, over 58% were rose X percent, followed by 2.2% comprising rose by X percent and just under 1% comprising rose over X percent. During the year 2002, the frequency of pattern rose by X % was above 50% for all months except March and September where it was 49.4 % and 44.8 % respectively (the average value was 54%). Table 4 shows the monthly variation in the various patterns comprising the verb rose. The rise/fall and up/down metaphor may be expressed through other related words and through synonyms. Textbooks on ‘report writing’ suggest that without The mood of the (financial) markets 23 losing accuracy, one may use related words/synonyms instead of repeating the same words. Indeed, slump instead of fall, and jump/climb instead of rise appear to be good candidates for not only improving writing style but also may make the news report more sensational. We looked into Roget’s Thesaurus (1980) and the WorldNet online to find related words/synonyms for the rise/fall we have discovered in Tables 2 – 4 thus far. It appears that the frequency of the related words/synonyms is much lower (Table 5). Table 5: Related words/synonym sets of rise and fall 3. Lemma fall, inc. fell, falling fReuters 875 rise, inc. rose, rising, risen 1004 Related words/Synonyms drop slump strike jump climb lift fReuters 155 69 53 133 75 33 Sentiments and lexico-grammatical patterns? The domination of one word form over others with related meanings shows perhaps the precision, which came from a lack of imagination some would argue, in the language of financial journalists. This is good news for people interested in information extraction – a branch of computing dedicated to the extraction of meaning from natural language text. The initial results related to the preponderance of one word form over others (Table 5) together with a preponderant lexico-grammatical pattern in which the word form is found (Table 4), suggests to us that when the word forms rise and fall, or rather rose and fell, are used, especially followed by a number and ‘percentage’, then they indicate an upward or downward movement in the value of some part of the financial market – this being true of shares, aggregate share index of one business section (cf. Property companies, telecommunications) and house prices. This still could be an ambiguity though: the report telling us that ‘mortgage defaulters fell by X percent’ is bad news except if you are in the debt collection business. Zellig Harris and his pupil Maurice Gross (1991) have been keen to suggest that in certain types of text one may find local grammars in operation: certain phrase structures that occur more frequently in one type of text, or one set of text fragments than in the language as a whole. Harris illustrated this point by citing examples of recursive noun-phrases used in biochemical literature to either refer to complex biochemical compounds or complex biochemical processes. Gross (1993) focussed on how we specify time and date and showed cardinal numbers used to denote time and calendrical expressions (day / month / year, century) embedded in their own local grammar. Barnbrook and Sinclair have used this notion to argue that dictionary definitions are also written in a local grammar (1993). Local grammars and the Hallidayan term ‘lexico-grammatical’ patterns 24 Ahmad et al. have a certain resonance, and this we have used to explore the grammatical environment of market-sentiment indicating words. The sentences comprising rose and fell embedded in the patterns shown in Table 4 were analysed using a reliable part-of-speech tagger – CLAWS. The local grammar of the two-sentiment indicators rose/fell comprises these patterns (also see): VVD ° ° ® ° ° ¯ Ø PRP ½ ° ° ¾ ° ° ¿ AV0 AVP DT0 CRD NN0 VVD AV0 ^ DT0 ` CRD VVD PRP ^ AT0 ` CRD NN0 NN0 Here VVD = verb; CRD = cardinal; NN0 = numeral; DT0 = determiner; AV0 = adverb; AVP = adverb particle; PRP = preposition. Table 6: Grammatical properties of the lexico-grammatical patterns Pattern rose X % rose by X % rose to X % rose over X % rose nearly X % rose by over X % rose only X % 4. Grammar VVD CRD NN0 VVD PRP CRD NN0 VVD PRP CRD NN0 VVD AV0 CRD NN0 VVD AV0 CRD NN0 VVD AVP AV0 CRD NN0 VVD AV0 CRD NN0 “Virtual Corpora” The designers of individual corpora have discussed the organisation of the text files within a corpus. The early corpora divided texts into informative/imaginative types (cf. Lancaster-Oslo Bergen and Brown Corpora, c.1960s), the texts were then divided into genre, topic and other pragmatic attributes. The subsequent ones took two different approaches: first, genre-based classification was used by some, where texts were classified into books, magazines, personal correspondence and so on (cf. Collins-COBUILD, c.197080s); second, topic-based classification was used by the designers of LancasterLongman corpus (c.1980s) wherein texts were divided into subject topics (science, world-affairs, news and so on). The other pragmatic attributes were included as well. The LOB and Brown corpora were developed for the study of The mood of the (financial) markets 25 (English) language in general and the Collins-CoBuild and Longman-Lancaster for lexicographical purposes. The more ambitious British National Corpus extended the list of pragmatic attributes, and perhaps made the attributes more explicit. The texts can be selected for analysis through a complex query language or by selecting the texts from a file store. For us, there is a hierarchical structure that drives the design of a given text corpus. The selection of the top-level features drives the selection of the other features – these are “doable” tasks if you know the organisation of the corpus you are using. The user of the computerbased corpus must know how the corpus is structured and what is more desirable is the use of ‘logical’ (attribute-oriented) features of the text. The organisation of a corpus of news reports suggests that sometimes there would be a need to analyse the corpus diachronically, focusing on a given individual/organisation, financial instrument, and at other times there is a need to conduct a synchronic analysis, for example, the analysis of news about all organisations within a given industry sector or all members of a political party. The analysis may be required based on a query comprising (a number of) keyword(s) that may be used in the indexation of a set of news reports. The permutations of the pragmatic and lexical attributes of a given text are numerous. It is possible that individual users of a text corpus may like to organise the texts according to their own needs. In order to have a user-configurable corpus, the notion of a virtual corpus was introduced (Holmes et al. 1994) for analysing technical and scientific texts. The notion of virtual corpus is similar to that of a virtual machine: there is in reality only one corpus, but the users can arrange the text attributes in a hierarchy of their choice based on a physically extant set of texts, for the duration of their use. This configurable hierarchy will have to be made available through the agency of a program, within a suite of corpus management programs, for producing this virtual corpus. The notion of virtual corpus introduces a shift from the usual pre-defined and explicit corpus hierarchical approach, in that it allows the definition of virtual hierarchies. Texts can then be retrieved by navigating through an organisation – the virtual hierarchy – specified by the user. We have designed a corpus for use in the automatic extraction of financial information from newspaper texts. The principal source of this corpus are the Reuters News Agency texts – in NewsML format , which is an extension to XML for enriching news stories, conceived by Reuters, developed and ratified by the International Press and Telecommunications Council (IPTC). Atkins, Clear and Ostler (1992) discussed criteria for corpus design. Based on this, and with attributes available in NewsML, news can be organized using the six major (pragmatic) attributes shown in Table 7 below. Ahmad et al. 26 Table 7: Pragmatic attributes used for organising NewsML texts Publisher Name; Place of Publication; Source of Publication; Date of Publication; Date of Origination Availability Copyright Status; Copyright Duration; Copyright Owner; Usage Restriction Text Title/Headline; Dateline; Text Type(book, newspaper, journal, etc), Text Mode (written, spoken?), Text Entry (electronic, transcribed?) Language Language Name; Regional Variant Author Byline; Reporter Nationality; First Language; Editor Category Industry Code; Topic Code From the above set of pragmatic attributes, a news database management system, Virtual Corpus Manager (VCM) was developed at Surrey. VCM can be used to organise texts; to share and retrieve texts; to navigate through (content of) the corpus and to impose integrity and security checks on texts. We have used six different types of constraints. Each constraint allows the user to choose one set of inter-related attributes at a time. For example, the users can choose as many major attributes and for each attribute can choose the sub-attributes. One example in the use of VCM is selecting UK-specific financial texts from Reuters daily news stream for a given year. 5. Visualising the mood of the market Time series analysis is one of the established and persuasive branches of statistics. This analysis is used extensively for analysing ‘a sequence of data indexed by time, often comprising uniformly spaced observations’ in science, engineering, economics, commerce, biology and in almost every subject. Financial news reports are usually illustrated with a time series of the instrument that is being reported: the share price of a company at the opening or closing of a day’s trading plotted over a financial (calendar) year shows the perception of the traders of the financial health (another metaphor) of the company. There are time series, which include both opening/closing and day’s highs/lows – the Japanese candlestick patterns as they are called in the trade. There are time series of the geometric mean of major organisations whose shares are traded in the market – FTSE-100 share index is the geometric mean of the share prices of 100 leading organisations at the close of daily trading, and there is FTSE index, which is the geometric mean of all-shares traded in the London Stock Exchange at the close of trading. We similarly have DAX-30 for Germany and CAX-100 for France, and there is Dow-Jones Industrial Average (DJIA) in the USA. Financial traders, having sought opinion from statisticians, tend not to deal with the ‘raw’ data value, but use other statistical measures related to the value of the instrument(s): typically used indices are those of volatility – a measure based on the standard deviation of closing price from its average value in The mood of the (financial) markets 27 the past few days/hours/minutes; the moving average; and return value, which is the (logarithmic) difference between the value of the instrument at time t-1 and at time t. The use of other statistical measures of the quantitative changes in the value of instrument(s) are important for us as we try to attempt to incorporate the use of sentiment indicators – or rather changes in sentiment – in an overall financial analysis framework. One such attempt has involved helping the traders to correlate the quantitative signal, either in its raw form or the derived forms (return and volatility measures), with the movement of sentiment indicating phrases. The traders typically use two sophisticated computer systems almost simultaneously during a trading session: one screen dedicated to the value of financial instruments, sometimes resolved at 50 values per minute, and the other screen dedicated to news streams supplied by Reuters, Bloomberg and others. A typical trader looks from one to the other and then makes his or her decision. This, rather simplistic view of financial trading has led to the development of SATISFI – which can simultaneously display, or help to visualise the news, the value of an instrument, and the changes in the frequency of sentiment indicators. Table 8 and Table 9 below show the two most frequent sentiment words used in generating the positive and negative sentiment time series respectively. Table 8: Dominant Sentiment words rose and up Rose Up Total Jan 87.80 117.48 205.28 Feb 79.81 135.81 215.62 Rose Up Total Jul 63.27 99.89 163.16 Aug 86.09 134.09 220.18 Relative Frequency (10-5) Mar Apr May 57.94 64.51 85.39 109.40 88.97 96.87 167.34 153.48 182.26 Relative Frequency (10-5) Sep Oct Nov 58.28 62.26 63.67 96.92 94.71 99.98 155.2 156.97 163.65 Jun 58.41 69.78 128.19 Dec 67.25 80.70 147.95 Table 9: Dominant Sentiment words fell and down Fell Down Total Jan 62.45 84.09 146.54 Feb 88.53 100.94 189.47 Fell Down Total Jul 63.02 91.92 154.94 Aug 75.38 87.28 162.66 Relative Frequency (10-5) Mar Apr May 69.28 27.06 68.89 68.88 74.89 61.71 138.16 101.95 130.6 Relative Frequency (10-5) Sep Oct Nov 51.53 48.81 51.73 101.22 81.78 67.65 152.75 130.59 119.38 Jun 69.78 92.51 162.29 Dec 68.40 85.31 153.71 28 Ahmad et al. Figure 2: SATISFI prototype shown with one-year FTSE index based on monthly data with upward and downward movement indicator series SATISFI has four major components that have been fully integrated as shown in figure 2. i) Time Series Display: SATISFI can display three time series at a time. These time series comprise of FTSE-100 close index values, upward movement indicators and downward movement indicators. As discussed above, upward and downward movement indicators are the quantification of the market sentiment expressed in financial news. Over 70 terms each have been identified for conveying ‘good’ and ‘bad’ news. For example upward movement indicators would contain terms like ‘up, rise, growth’ etc. while downward movement indicators would contain terms like ‘down, fall’ etc. The movement indicator time series are synthesized by counting these movement indicator terms within the financial news published for a particular day. Each time series is normalised for proper display purposes. SATISFI is capable of displaying the above time series in three forms: (1) Raw form denotes the original time series. (2) Return form refers to the logarithmic difference between two consecutive values. (3) Volatility (historical volatility) is the relative rate at which the time series moves up or down. ii) Time Series Correlation: Correlation is a measure of the degree of linear relationship between two time series. SATISFI provides the The mood of the (financial) markets 29 user the facility of cross-correlating two series in any form (raw, return, volatility). Any series can be shifted forward or backward and the cross correlation recalculated to determine whether the market is followed by the news or vice versa. iii) Document Display: This is comprised of two parts: (1) Document Titles: Clicking a dot (date) on any of the time series displays the corresponding date’s news titles. (2) Document Content: The content of any document title can be viewed by clicking that news title. iv) Document Analysis: Whenever a document title is selected from the news list, the extracted sentiment keywords along with the frequencies are displayed in “Document Keywords” area. Positive sentiment keyword analysis details appear under the title “Upward Movement Indicators” and negative sentiment keyword analysis details appear under the title “Downward Movement Indicators”. 6. Finding ‘meaningful’ patterns 6.1 A case study: Movements in 2002 The year 2002 has seen its ups and downs like many other years and the movements in financial markets worldwide have been in the downward direction. Or, at least, the geometric mean of the value of the shares of the major corporations in North America, Japan, and the European Union, with some exceptions, has reduced substantially. The UK FTSE-100 shows mainly downward movements interspersed with small periods of upward movements, which unfortunately, could not compensate for the previous reduction in the value of the index. In time-series analysis literature one finds techniques that tend to separate the so-called trends from cyclical movements in the series: the cyclical movements may be due to factors like holidays, trading patterns that may be seasonal and so on (Tino et al. 2001). The trends, it is claimed, show a change that is caused because the basic structure of the market has changed. Techniques like fractal analysis and the related chaotic systems methods help in disentangling the trends from the cycles. Another related and robust technique is the wavelet analysis (Rioul et al. 1991): a wave comprises oscillations of a number of different frequencies and trends and wavelet analysis suggests ways in which these could be disentangled. We have used wavelet analysis on the FTSE-100 data for 2002 together with the time series of upward and downward indicating phrases, in Reuters Financial News for the same year. Figure 3a shows the raw figures for the daily trading data for the FTSE-100. Figure 3b shows the longterm trend in the time series (downwards) while figure 3c shows the short-term trend and hence some cyclical behaviour. Note the turning points in the cyclical data (marked by arrows in figure 3c). There is, as noted earlier, a considerable 30 Ahmad et al. interest in identifying these turning points. The system SATISFI is being extended to generate a textual description of the turning points. Figure 33a shows a time series of the frequency of a number of upward movement indicators, including rise, growth and other less frequent phrases indicating upward movement. There is a corresponding long-term decay in the time-series of upward movement indicators (figure 33b) as found in the raw FTSE-100 data (figure 3b). The cyclical movements are much more pronounced (figure 33c) but a comparison with eye suggests that there is a similar pattern in figure 33c as in figure 3c. The downward movement indicators show that for the first six-months or so of 2002 the frequency of downwards indicators increases rapidly but shows a decay in the latter half of 2002 (figure 33b). The sentiment indicating time-series has to be refined and much more work is required before we may use it to predict the actual mood of the market. However, our approach is perhaps amongst the first of the explorations, which investigate how the quantitative movements in a financial market are influenced by the news stories, some influencing the market and others showing the influence of the market. Afterword We have attempted to build a system from various linguistic, visual and mathematical components that allows us to explore the behaviour of the traders, and to attempt to model this behaviour. The purpose of the system is to assist the trader by reducing the amount of textual and numeric data that the trader needs to assimilate to form a view of the market. In terms of corpus linguistics, the analysis of qualitative data available in collections of news texts organised by descriptive metadata (through XML and Reuters codes), combined with text processing techniques to determine key patterns, proper-noun analysis to determine key entities and the use of terminology collections for reducing the understanding overhead necessary for the text, and for automatic classification of the text, shows early benefits. This work is directly relevant to the potential for Information Extraction techniques to be adopted across domains, an issue that is being investigated also in a related EPSRC-sponsored project Scene of Crime Information System (SOCIS). The integration of techniques in corpus linguistics with other forms of analysis including mathematical analysis and image analysis provides a supporting environment for future projects that do not focus on a single medium of communication. Experts do not generally rely on the text medium alone; this work provides evidence of the kinds of information fusion that financial experts carry out many, many times each day. The mood of the (financial) markets 31 Figure 3: Acknowledgements The work described in this paper has been supported by the EU IST Programme’s Generic Information Decision Assistant (GIDA IST 2000-31123). Based on papers presented at two workshops: LREC Event Modelling Workshop (Spain 2002) and Financial News Analysis Workshop, 11th International Terminology and Knowledge Engineering Congress (France 2002). References Ahmad, K. (2002), Events and the causes of events: the use of metaphor in financial texts, in Proceedings of the workshop at the International Conference on Terminology and Knowledge Engineering. Nancy, France. Atkins, S., J. Clear and N. Ostler (1992), Corpus Design Criteria, Literary and Linguistic Computing 7(1): 1-16. Barnbrook, G. and J. Sinclair (1995), Parsing CoBuild Entries, in J. Sinclair; M. Hoelter and Peters, C (eds.), The Languages of Definition: The Formalization of Dictionary Definitions for Natural Language Processing. 32 Ahmad et al. Luxembourg: Office for Official Publications of the European Communities. pp. 13-58. Gross, M. (1993), Local Grammars and their Representation by Finite Automata, in M. Hoey, (ed.), Data, Description, Discourse: Papers on the English Language in Honour of John McH Sinclair. HarperCollins Publishers. pp. 26-38. Harris, Z. (1991), A Theory of Language and Information: A Mathematical Approach. Clarendon Press, Oxford. Holmes-Higgin, P., S. Abidi and K. Ahmad (1994), ‘Virtual’ Text Corpora and their Management. In Proc. of Sixth EURALEX International Congress on Lexicography, Amsterdam. Jackendoff, R. (1991), Semantic Structures. Cambridge (USA) & London: The MIT Press. Knowles, F. (1996), ‘Lexicographical Aspects of Health Metaphors in Financial Texts’, in: M. Gellerstam et al (eds.), Euralex96 Proceedings (Part II). Göteborg, Sweden: Göteborg University. pp. 789-796. Maybury, M. (1995), Generating Summaries from Event Data. Information Processing and Management 31(5): 735-751. Quirk, R., S. Greenbaum, G. Leech and J. Svartvik (1985), A comprehensive Grammar of the English Language. Longman. Rioul, O. and M. Vitterli (1991), Wavelets and Signal Processing, IEEE Signal Processing Magazine, pp. 14-38. Robert A. (eds) (1980), Roget’s Thesaurus. Great Britain, Longman. Simmons, R. F. (1984), Computations from the English: A Procedural Logic Approach for Representing and Understanding English Texts. Englewood Cliffs, NJ: Prentice Hall. Tino, P., C. Schittenkopf and G. Dorffner (2001), Volatility Trading via Temporal Pattern Recognition in Quantized Financial Time Series, Pattern Analysis and Applications. Reuters NewsML Showcase: http://about.reuters.com/newsml/ Tagger: http://www.comp.lancs.ac.uk/computing/research/ucrel/claws/trial.html System Quirk: http://www.computing.surrey.ac.uk/ai/SystemQ WordNet: http://www.cogsci.princeton.edu/~wn/ Towards a methodology for corpus-based studies of linguistic change: Contrastive observations and their possible diachronic interpretations in the Korpus 2000 and Korpus 90 General Corpora of Danish Jørg Asmussen 1 Society for Danish Language and Literature Abstract The Korpus 90 and Korpus 2000 Corpora of Danish were both designed and compiled at the Society for Danish Language and Literature (DSL). The joint web-based query interface of the two corpora enables immediate comparative studies. This paper focuses on examples of contrastive observations and their possible diachronic interpretations. It discusses whether observable differences of word frequency, inflection, collocation, connotation, and word order reflect real changes in the Danish language, or whether they reflect differently compiled corpora. Finally, the paper proposes prerequisites for a methodology of comparative corpus investigation and the determination of diachronic corpus similarity. In this context, the concept of invariant textual features is introduced. 1. Background: The Korpus 2000 and Korpus 90 General Corpora of Danish The Korpus 2000 and Korpus 90 General Corpora of Danish were both designed and compiled at the Society for Danish Language and Literature (Det Danske Sprog- og Litteraturselskab, DSL). DSL, which is a type of academy, was founded in 1911 with the aim of publishing scholarly editions of Danish works of linguistic or literary importance, including dictionaries. During the 1990s, the design and compilation of corpora and other electronic linguistic resources of the Danish language has become an important activity for DSL. Today, DSL is regarded as the main public institution 2 in Denmark within the fields of development and compilation of dictionaries and corpora 3 of the Danish language. DSL’s long-term goal is to combine dictionaries and corpora to yield a unique language information system – a bank of Danish language. Korpus 2000 (K2000) holds 28 million text words and is composed of text material from the years 1998-2002, covering a wide variety of genres. The objective of the Korpus 2000 project was to compile a general corpus in order to document written Danish language around the turn of the millennium. Korpus 34 Jørg Asmussen 2000 has been made publicly accessible on the Internet, one of the main purposes of the project being to increase laymen’s awareness of the advantages of a corpus as an extension to dictionaries. Korpus 90 (K90) is a 28-million-word subset of the 40-million-word Corpus of The Danish Dictionary (Den Danske Ordbog). Spoken material, as well as texts with certain copyright restrictions, has been excluded from this subset. The Corpus of The Danish Dictionary, holding text material from the decade 19831992, was compiled by DSL in the early 1990s. 4 It served as a major source in the compilation of The Danish Dictionary, an entirely new, written-from-scratch dictionary of contemporary Danish to be published in six volumes by DSL during 2003-2004. As part of the Korpus 2000 project, K90 was made publicly accessible on DSL’s website 5 , where it serves as an approximately 10-year-older counterpart to K2000. Both K2000 and K90 are morphosyntactically tagged with the Constraint Grammar-based DanPars-tagger 6 and are accessible through a joint query interface 7 that enables immediate comparisons of certain linguistic aspects. The following section will provide some examples of contrastive observations made through this interface, as well as possible diachronic interpretations – and misinterpretations – of them. 2. Problem: Contrastive interpretation observations and their diachronic Query results such as frequencies and collocates are presented in the K2000 interface in contrastive tables showing results for K2000 and K90. Hence the user can compare frequency figures and collocates immediately, and hopefully get an idea of what kinds of changes in vocabulary, inflectional forms, or collocates (semantics) have taken place in the years between the compilation of K90 and K2000. However, the contrastive presentation of query results has its shortcomings as well, since it may seduce some users into making faulty generalisations about language change. Below, we will present some examples of these pitfalls and sketch some possible ways to avoid them, some of which could be implemented in the user interface as a diachronic interpretation facility. 2.1 Vocabulary A comparison of the frequencies of all words in K2000 and K90 reveals, not surprisingly, that some words are significantly more frequent in one corpus than the other. Assuming that each of the two corpora reflects Danish language usage during a given period of time (1983-1992 and 1998-2002), we might interpret these frequency differences as changes in the incidence of words in the Danish language as a whole. Towards a methodology for corpus-based studies of linguistic change 35 Figure 1 (below) shows how the frequencies for the noun regn (rain) are presented in the user interface: the first column lists the possible inflectional (and orthographic) forms of the lemma, the second column indicates the frequency of each form in K2000, and the third column the frequency in K90. The form listed in the bottom row is the lemma itself with all its inflectional forms; i.e., it is the total of all of the forms listed above the bottom row. 8 Frequency is not shown in absolute decimal figures, but as a logarithmic score 9 represented by a number (07) of bullets. Experimentally, it has been established that this logarithmic score apparently follows general intuitions on word frequency quite well, such that words with 1-2 bullets are less frequent, e.g. entomologi (entomology), words with 6-7 bullets are very frequent, e.g. i (in) and og (and), and middle-range words with 3-5 bullets are of average frequency, e.g. regn. As the table shows, there do not appear to be any striking differences in the use of these forms between K2000 and K90, with one exception - the indefinite genitive form regns (rain’s), which does not occur at all in K2000, but scores one bullet in K90. The raised thumb indicates that this form is “interestingly” more frequent in K90 than in K2000 - which in this case means that it occurs at least twice as often as the corresponding form in the other corpus. 10 This does not necessarily mean that the difference is statistically significant as well. Especially with low frequency words, the significance may be skewed. However, the phenomenon may have some linguistic relevance, and our intention, therefore, was to highlight this in the frequency table - a practice which may have confused users, as it turns out. Clicking on one of the magnifying glasses yields the KWIC concordance for the corresponding form, together with the number of occurrences. For regns, we get three occurrences in the 28-million-word K90. Figure 1: Frequency table for the noun regn (‘rain’) The total frequency of all inflectional variants of the lemma mobiltelefon (mobile phone) is about 25 times higher in K2000 (1,486 occurrences) than in K90 (59 occurrences). Assuming that the vocabulary of a language reflects general changes in society, and keeping in mind technological changes from the eighties to the late nineties, it seems evident that this frequency observation could be interpreted as a general change in the vocabulary of the Danish language. Similar 36 Jørg Asmussen examples are biltelefon (car phone) and benchmarking; biltelefon, which is five times more frequent in K90 (51 occurrences) than in K2000 (9 occurrences), denotes a technical device that by and large has been replaced by mobile phones. Benchmarking does not occur at all in K90, whereas it occurs 34 times in K2000, which might indicate that this word is new to Danish vocabulary. Jarvad (1999), whose dictionary of new words in Danish is based on good old-fashioned “manual” language excerpts, dates the first occurrence of this word in Danish to 1996 - an observation which might support the assumption that this word is new. All these examples appear to indicate that the two corpora reflect changes in language use as a result of general changes in society. A less frequent word such as kambrium (Cambrium) occurs four times in K90, but not at all in K2000, and is therefore marked with a raised thumb in the K90 column. A somewhat naive, but not entirely impossible, interpretation of this could be that the incidence of this word in the Danish language is decreasing. However, a closer look at the word reveals that it only occurs in a single text on geology, a scientific domain which appears to have been covered unevenly in the two corpora. So, in certain cases, raw frequency comparisons turn out to be unreliable: they indicate different corpora compositions rather than diachronic changes in the vocabulary of the language in question. In the case of kambrium, the introduction of some measure of word dispersion in the corpus would probably mitigate against misinterpretation; thus dispersion could be used to correct the use of raw frequency to measure incidence. However, the disadvantage of this method is that it generally diminishes the weight of linguistically interesting, infrequent words, such as new words. Yet the implementation of this method in the K90/2000 user interface would prevent users who are unfamiliar with quantitative linguistic methods from misinterpreting words such as kambrium. 2.2 Inflection A comparison of the inflectional forms of particular words will at first glance show some striking differences between K2000 and K90. Looking at our previous example, regn, we already have noted some differences between K2000 and K90 for the indefinite genitive form regns (rain’s), which occurs three times in K90, but not at all in K2000. 11 For the definite genitive form regnens (the rain’s), the table does not indicate any quantitative differences, but if one looks at the corresponding concordances, one will observe that this genitive form occurs twelve times in K90 and nine times in K2000. The same picture emerges again: fewer genitives in K2000. Elbro 2002 reports on observations from K90 and K2000 that genitives of certain frequently used nouns appear to be less prevalent in K2000 than in K90. This, he suggests, might indicate a general tendency in Danish towards replacing genitive constructions by prepositional phrases, probably as a result of the influence of English. Thus, for example, bilens ejer (the car’s owner), which is considered the canonical norm in Danish, is often re- Towards a methodology for corpus-based studies of linguistic change 37 placed by ejeren af bilen (the owner of the car). This assumption is further supported by Elbro’s (2002) observation of increased frequencies of certain prepositions in K2000. Apparently, the absolute number of genitive forms of some nouns in K2000 is significantly lower than in K90. For instance, the noun bil (car) has a total of 393 genitive forms in K2000, as opposed to 586 in K90. Similar results are observed for other nouns denoting common objects, such as cykel (bicycle) and hus (house). A closer look at bil reveals that the lemma, including all its inflectional variants, occurs 10,360 times in K90, as opposed to 8,354 times in K2000, an observation which hardly - even among laymen - would yield an interpretation such as “the word bil is replaced by other words”, or even “the denoted object is about to disappear from our lives” - interpretations that appeared obvious for a word such as biltelefon, as discussed above. Furthermore, an examination of the relative numbers of genitive forms of the lemma bil in K90 and K2000 shows 5.7% and 4.7% respectively - a difference too weak to support conclusions about radical linguistic change. Other examples that may disprove the hypothesis of decreasing genitive forms include land (land, country) and the proper noun Danmark (Denmark). The lemma land occurs 28,222 times in K2000, the genitive proportion being 22.3%, against 21,478 times in K90, with a genitive proportion of 16.7%, thus showing the opposite tendency of what was reported above. Danmark occurs 30,730 times in K2000, against 22,243 times in K90, with 15.1% genitive forms in K2000 and 15.7% in K90. Once again, no alarming differences in the use of genitives can be observed. However, what might be alarming is the fact that some words, such as land and Danmark, which we intuitively would expect to have a constant prevalence in the language over ten years, show these astonishing frequency differences, despite their identical logarithmic score. This could be another indication of two differently-composed corpora, K2000 probably containing a larger amount of newspaper text than K90. The examples given above suggest that one presumably cannot account for general inflectional change merely by examining randomly-selected frequent words, since the result of the comparison will be too arbitrary to enable us to make any general conclusions. If one wishes to investigate this kind of linguistic change, one should examine the phenomenon in question throughout the corpus. Thus, in the case of the genitive forms, one should at least compare the total ratio of genitives forms of all nouns in each corpus, and probably also the relative number of nouns compared with other parts of speech in each corpus.12 2.3 Collocation The K2000 query system can display collocates as either frequently or typically co-occurring collocates. The former are mainly made up of co-occurring function words, whereas the latter are determined by means of mutual information, I(worda;wordb) = P(worda,wordb)/P(worda)P(wordb). 13 In order to prevent infrequent words from popping up among the typical collocates, some intuitively 38 Jørg Asmussen defined noise-reducing, and process time-reducing, conditions have been set up: the corpus frequencies of both examined words worda and wordb must be greater than 30, the collocate candidate’s frequency must exceed a threshold t = (log10(fk))2/10, where fk is the frequency of the collocator, i.e. the word for which we want to determine the collocates. Moreover, the collocation itself must occur at least twice, and the resulting (non-logarithmic) mutual information score must be greater than 20. The collocation window is set to two words to the left and two to the right of the collocator. The resulting collocates are presented in a table consisting of four lists, a left-hand and a right-hand list for each corpus. The lists are sorted in descending order according to their mutual information score, which is not explicitly stated in the list, but instead converted into a degree (1-5) of mutual attraction d = round(log10(I)-1), where I is the non-logarithmic mutual information score. The resulting degree of mutual attraction is expressed by a corresponding number of bullets in front of each collocate, where one bullet may be interpreted as a weak - but significant - mutual attraction between the given words, and the maximum of five bullets as a very strong attraction. Figure 2 shows, as an example, a collocation table for the lemma terrorist. Figure 2: Typical collocates for the noun terrorist In K2000, we find to the left of the noun terrorist adjectival forms that denote either a general quality or state (hensynsløse (ruthless), eftersøgte (wanted), fængslede (imprisoned), islamiske (Islamic), internationale (international)) or a (national) origin or location. On the right-hand side, we note a couple of proper nouns. In K90 we also find eftersøgte (wanted) and palæstinensiske (Palestinian) to the left, and furthermore vesttyske (West German); finally, we find dræbt (killed) to the right. The result may be interpreted as follows: one of the characteristics of a terrorist that remains constant over time is that he is wanted or Palestinian, whereas West German is no longer a significant attribute, but, instead, many other nationalities, a certain religious orientation, or even international. In K2000, terrorists furthermore have names or belong to certain organisations, whereas in K90 they Towards a methodology for corpus-based studies of linguistic change 39 either killed or were killed. The greater number of collocates in K2000 may indicate that terrorist has become a more frequently used word, and, in fact, the lemma occurs nearly twice as often in K2000 (477) as in K90 (253). Or perhaps this is just another indication that the corpora are composed differently, K2000 probably containing more newspaper material than K90. Yet the results and their interpretation seem reasonable to some extent in that they resemble general tendencies in Danish society related to this topic. Our historical knowledge helps us to understand these collocates and helps us to understand the constants and changes in the collocational behaviour of this word. A word which ought not to have changed its collocational behaviour over ten years is jul (Christmas) because its use is most likely embedded in a strong traditional context, and, accordingly, the majority of the listed collocates are to be found in both corpora, such as glædelig (happy), fejre (celebrate), or, on the right-hand side, nytår (New Year). However, what is conspicuous is that the number of collocates is somewhat greater in K90, indicating that jul occurs more often in this corpus (2,196) than in K2000 (1,275), which might be taken as another indication of compositional differences between the corpora. Thus, for example, hvid (white) does not occur as a collocate to jul in K2000, even if the collocation hvid jul occurs twice in K2000 and, of course, still is a valid collocation in Danish. Unfortunately it has been blocked by the noise-reduction measures. This example shows that incidental differences in frequency of a (normally) quite frequently used word, such as jul, may have a serious impact on the determination of collocates. Thus, comparing collocates derived from words with remarkably different frequencies in the two corpora does not necessarily give a correct impression of changes in their general collocational behaviour. Even if a word is used less frequently, it may still keep its well-established collocates, but they may not appear in collocational statistics any longer. What might appear, in some cases, are candidates that intuitively do not count as collocates. Looking up collocates for the word juletræ (Christmas tree), one will, as expected, find pynte (decorate) and danse [rundt=om] (dance [around]), but in K2000 we also find talende (speaking) - and this as the most significant collocate for this corpus, scoring four bullets on the degree of mutual attraction scale. On closer examination, it appears that all examples of the talende juletræ come from one single text, a strange story of a speaking Christmas tree. Thus, also when determining collocations, some kind of dispersion-based correction of the raw frequencies of the words involved would help diminish these odd findings. 2.4 Semantics Closely related to collocation is the phenomenon of semantic prosody, 14 the ability of words to establish certain flavours of meaning contextually, e.g. positive, negative, ironic, etc. Consider the word sideeffekt (side effect), an English loanword, which is not registered by Jarvad (1999), but occurs 11 times 40 Jørg Asmussen in K90 and 22 times in K2000. One might argue that, as a word, sideeffekt is superfluous in Danish because another word, bivirkning, already exists with exactly the same meaning. Approximately half of the examples in K90 are used in a clearly negative context, indicating that sideeffekt is something unwanted and harmful, and thus the semantics of this word appears to be very close to that of bivirkning. The picture has changed in K2000, where the majority of examples clearly show sideeffekt as something still unwanted, but quite good, and some examples are explicitly preceded by the adjective positiv (positive), as shown in Figure 3. Figure 3: Indications of changed semantic prosody for sideeffekt in K2000. The question is whether these examples empirically justify the conclusion that sideeffekt in Danish has indeed changed its semantic prosody and found a semantic niche. How many examples of a type of semantic change do we need in order to exclude corpus compositional bias and to be able to make conclusions about a language in general? 2.5 Syntax We will give one example of syntax that shows some topological differences for the negation ikke (not) which one can observe between K90 and K2000. In main clauses, the negation is placed after the finite verb: Peter drikker ikke te (‘Peter drinks not tea’), whereas the negation is placed before the finite verb in relative clauses: Anne serverer kaffe fordi Peter ikke drikker te (‘Anne serves coffee because Peter not drinks tea’). Main clause word order in subordinate clauses is generally considered substandard Danish. However, evidence of this usage can be found in both corpora. Intuitively, we would expect this non-canonical word order to be more frequent in current Danish, and we therefore would expect to find more examples of this in K2000 than in K90. However, a comparison shows more examples of the incorrectly placed ikke in K90, some of which are shown in Figure 4. Does this prove our intuition wrong, or does it indicate that K2000 contains more professional - and thus more ‘correct’ - language? Or is it the case that the corpus evidence is not convincing enough because of methodological shortcomings? Towards a methodology for corpus-based studies of linguistic change 41 Figure 4: Examples of non-canonical word order in K90. 3. Towards a methodology of corpus-based comparative studies The examples above demonstrate that our contrastive corpus observations in some cases may yield dubious interpretations and generalisations on linguistic change. In order to improve the quality of our comparative corpus investigations, we need to establish (i) a framework for the way we examine our data, i.e. a declaration of the analytical methods applied, and (ii) a method of describing and classifying our data, i.e. the corpus. To ensure that others can repeat our investigations on other data with analogous results, we would probably even want to standardise our procedures or create a general methodology for corpus-based studies of linguistic change. To be able to handle different scenarios of investigation, the analytical methods should at least cover the fields of vocabulary, morphology (inflection), collocation, semantic prosody, and syntax. Based on the examples given in section 2, we can list the following preliminary prerequisites for these analytical methods. 3.1 Prerequisites for a methodology of comparative corpus investigation The underlying word concept is important in comparisons based on vocabulary. In order to make these comparisons, we propose a word definition based on lexicalised units which include fixed multiword units, i.e. a word is defined as a lemma. Hence, what is comparable across corpora, when investigating vocabulary, are lemmas, not types. Furthermore, for vocabulary comparisons, measures other than raw lemma frequencies should be applied. A score based on different quantitative characteristics of a lemma in the corpus appears to be better suited to describing the prevalence of that lemma. These characteristics should at least include raw frequency and dispersion. To date, the field of vocabulary comparison appears to have attracted the most attention in corpus linguistic research, and a considerable number of statistical methods have been proposed.15 With regard to morphology, the comparison of the frequency of inflectional variants of a lemma must be handled differently from vocabulary comparison. As has been shown above, comparisons based on type can yield misleading results. First of all, one has to determine whether one wants to investigate a certain inflectional category within a lemma - for example changes 42 Jørg Asmussen in the quantity of plural forms of a noun - or if one wants to examine probable quantitative changes of a grammatical category as a whole throughout the corpus - for example the relative quantity of genitive forms of all nouns in the corpus. In the first case, the ratio of an inflectional variant of a lemma contrasted with all other inflectional variants should be compared with the ratio of the same inflectional variant of the same lemma in another corpus. To ensure that potential changes are lemma-specific, and not just part of a general diachronic shift, the observations should be contrasted with an investigation of the examined inflectional category as a whole throughout the corpus. In the second case, the inflectional category in question should be investigated throughout the corpus. In both cases, statistical methods of determining the significance of proportional changes must be developed and applied. The comparison of changes in the frequency of collocations in two different corpora appears to be more complex than plain lemma-based vocabulary comparison. Statistical methods of the mutual information type are probably not suitable in a comparison context; their strength is mainly found in detecting statistically salient collocations. Once a collocation has been detected, one can compare its number of occurrences across corpora. However, collocations should not be compared as if they were words by merely applying one of the methods for vocabulary comparison, since the frequencies of the collocations themselves also depend on the words involved. Hence, if one observes significant differences in the frequency of one of the words involved in a collocation when comparing two corpora, these differences will also be reflected in all collocations that involve this word. In these cases, it is not necessarily the frequency of the collocation, or its prevalence in the language, that has changed. Comparisons of collocations might therefore best be treated as composite vocabulary comparisons of the words involved in a collocation, in conjunction with the frequency scores of the collocations themselves. Within the field of semantic prosody, some kind of semantic tagging, or at least grouping of contextual words, is indispensable. Changes will often affect new loanwords, and thus relatively infrequent words that are semantically vague at first, but find a semantic niche after some years. The number of words that show these characteristics is normally quite limited, and hence the comparative statistical methods must be adjusted in these special circumstances. The number of observable instances will decrease as the complexity of a certain linguistic phenomenon increases, i.e. when it is composed of several words or linguistic constituents. This applies to most observable linguistic phenomena above the word level, e.g. collocations, semantic prosody, and - perhaps most saliently - syntax, and it will have implications on the statistics used in these cases. Towards a methodology for corpus-based studies of linguistic change 43 For other fields of linguistic comparisons between corpora, similar prerequisites have to be established. Common to all of these prerequisites should be the principles of (i) contrasting a phenomenon with its potential alternatives within one corpus, and comparing this contrast ratio with that of another corpus, and (ii) determining appropriate comparison statistics for linguistic phenomena of varying complexity. 3.2 Prerequisites for a methodology of measuring corpus similarities over time Apart from the development of analytical methods to be applied, one needs a method to describe and classify the data, i.e. the corpus. Although considerable efforts have been made to describe what is in a corpus and why, and although corpora usually are designed in accordance with some idea of representativeness, the problem remains that corpus descriptions are generally founded on relatively vaguely defined, intuitively applied textual categories. Classifying texts by intuition is necessarily imprecise because intuitions may differ from compiler to compiler, and a single compiler may even apply different criteria of classification according to extra-textual circumstances. However, this problem does not seem to be critical, as long as the majority of corpus-based linguistic research refers to the same well-established corpus, e.g. the BNC, or, in the case of Danish, K2000 or K90. Statistical methods for determining corpus differences (Garside and Rayson 2000) and corpus similarities have been proposed, mainly based on vocabulary (Kilgarriff 2001). But what happens if one wants to compile a corpus designed exactly like an existing one, e.g. a new BNC, holding a similar mixture of, e.g., texts that are ten years more recent? In this case, the definition of a concept of similarity might become quite intricate because differences are explicitly desired as long as they are the result of diachronic language change, but they must definitely not be the result of differently composed corpora. How can one compile an updated version of a well-established corpus that is absolutely comparable to the existing one in its composition, so that the linguistic differences detected by comparing the corpora truly and only can be explained diachronically - and not as a result of compositional bias? A solution to this problem of corpus composition is a measure, a quantitatively established figure, or - more likely - several figures, characterising the overall textual composition of a corpus. Corpora with the same characteristics, then, would probably hold the same mixture of text types and would hence be comparable. So far, this approach is not different from the (implicitly) synchronic similarity measuring approaches proposed by Kilgarriff 2001. The difference is that the required measure should be based on positively invariant features of language, i.e. features that do not change (significantly) over time. 44 Jørg Asmussen Within a synchronic context, the linguistic features to compare almost may be chosen by chance. They may be, and often are, based on vocabulary, but other features - such as part-of-speech, syntactic constituents, sentence length, word length, maybe even the distribution of character n-grams, etc. - may to some extent be useful as well. If statistics show that the two corpora have matching features, they may be expected to be of a similar type or composition. Within a diachronic context, however, many of these features may not be suitable because an ongoing general linguistic change might affect one or more of them; hence keeping them constant across corpora and over time might obscure the detection of possible changes. Linguistic features based on grammar or semantics have the disadvantage of requiring morphosyntactically or semantically tagged corpora. The task of tagging a corpus before one can determine its compositional characteristics introduces a certain amount of interpretation, as one has to apply a grammatical tradition or semantic framework that does not necessarily capture the possible, subtle linguistic changes that occur as a language develops over time. Furthermore, the task of tagging increases the degree of methodological complexity, since one does not only need a tool for determining the compositional characteristics of the corpus, but also tagging tools. To ensure that the measurement of the compositional characteristics can be easily achieved in different settings, all tools that are involved have to be well documented and maintained. Slight modifications in tagging algorithms or tag sets may have farreaching implications on the reliability of this approach. For these reasons, one should try to keep the number and complexity of the tools involved as low as possible, as well as the setting itself. We therefore propose that corpus similarity should be determined on untagged corpora only, and that any introduction of linguistic theory into this process should be avoided. The mere physical, orthographic surface of the corpus should suffice, since its recurrent symbols yield structures of varying complexity that resemble the overall textual composition of the corpus. It is beyond the scope of this chapter to try to determine the required invariant linguistic features. However, a couple of candidates will be briefly discussed. Belica (1998) gives an account of some statistical diachronic investigations of vocabulary and collocations in the German “Wendekorpus”. 16 He seems to be quite aware of the corpus compositional problem dealt with in this paper and points at the distribution of different collocations of selected function words 17 as a candidate for an invariant feature, but does not give any examples or further details. However, it does not seem evident that this type of collocation should remain stable over time. A more appropriate candidate for an invariant linguistic feature seems to be the frequency and dispersion of words belonging to a defined semantic core vocabulary. Ruus (1995) attempts to determine a core vocabulary for Danish which she believes constitutes the basic lexical norm of the language. The core vocabulary, which Ruus determines, includes semantically vacuous, but Towards a methodology for corpus-based studies of linguistic change 45 highly frequent function words, whereas the semantic core vocabulary, which we shall suggest as a candidate, excludes them. The remainder should be relatively frequent, semantically well-established words of the jul (Christmas) type mentioned above. This semantic core vocabulary can be expected to remain grammatically and semantically stable over time and thus may prove quite useful in our context. Further experimental work will show whether it is possible to derive such a constant semantic core vocabulary from a large amount of text material, 18 and whether it really can serve as an appropriate invariant linguistic feature. 4. Conclusion In this paper, we have discussed examples of comparative investigations on two general language corpora of Danish, K90 and K2000, the former reflecting Danish language usage around 1990, the latter one around 2000. We have argued that some of the interpretations of the observed differences in vocabulary, collocation, semantics, and grammar were not necessarily the result of general changes in language usage, but rather a likely consequence of differentlycomposed corpora. 19 This observation led us to the conclusion that a framework for comparative diachronic corpus investigation is needed, and we sketched some prerequisites for a methodology of diachronic corpus comparison. With regard to comparative corpus investigation, we found that standardisable approaches are required to account for the complexity of the observed linguistic phenomena, as well as the quantitative relationship between a linguistic realisation and its potential variants. In further work, such approaches should be concretised and evaluated. With regard to the data material, i.e. the corpora to be compared, we argued that in order to determine the similarity of corpora comprising material from different time periods, the approach must allow a certain degree of dissimilarity for those linguistic characteristics which are likely to change over time. Therefore, we proposed to base this approach on invariant textual features, the semantic core vocabulary of a language being an initial candidate for investigation in further work. Generally, we find that the determination of invariant features appears to be a conceptual, or hermeneutic, challenge rather than a mere statistical one. Notes 1 Thanks to my colleagues Britt Keson and Allan Ørsnes for their valuable comments on the manuscript. 46 Jørg Asmussen 2 Legally DSL is a semi-public institution under the jurisdiction of the Danish Ministry of Culture, and its activities are financed in part by the Danish Government and in part by private and public foundations. 3 Among other corpora compiled or distributed by DSL are the Danish PAROLE Corpus (compilation and distribution, cf. Keson 1998) and the Bergenholtz DK87-90 Corpus (distribution, cf. Bergenholtz 1988). An overview of Danish corpora is given in Asmussen (2001). 4 A comprehensive account of the background and the design of the Corpus of the Danish Dictionary is given in Asmussen & Norling-Christensen (1998). 5 Accessible at http://www.dsl.dk/korpus2000. 6 Developed by E. Bick under the VISL project, University of Southern Denmark, http://visl.sdu.dk. 7 The query tool being CQP (Christ 1994) in conjunction with a special web interface developed in the Korpus 2000 project. For details on Korpus 2000 and its web-based user interface cf. Andersen et al. (2002). 8 The smiling face indicates that the lemma is spelt in accordance with official Danish orthography. 9 The score s is computed as follows: s = round(0.5 + log10(f)), for f>0, and s = 0, for f=0, where f is the computed average number of occurrences in one hundred million words of text. 10 This rule has been arbitrarily defined. 11 Even if this is by no means statistically significant, a user who is not aware of this might make a comparison based on these raw frequencies. 12 As noted by Elbro himself. 13 Cf. Church & Hanks (1989); a modification is that their log2 function is not applied here. 14 For a brief description, some examples, and further references cf. Rundell (2002). 15 For a comprehensive overview and discussion cf. Kilgarriff (2001). 16 Institut für deutsche Sprache, Mannheim, http://www.ids-mannheim.de. 17 "die Verteilung von verschiedenen Funktionswörter" (p. 35). Kollokationen ausgewählter Towards a methodology for corpus-based studies of linguistic change 47 18 DSL stores approximately 400 million words of text material in its text bank. The material covers a time span of approximately 20 years and constitutes the raw material for other DSL corpora. A part of this material is - to date - seven years of newspaper data from a Danish daily newspaper (Berlingske Tidende), approximately 25 million words per year. A first step could be to determine the common vocabulary to be found in each year (or each month or day), and to use this as a basis for further investigations. 19 In fact, there is a compositional difference between these two corpora as K2000 holds approximately 2/3 of newspaper text as opposed to approximately 1/3 in K90. References Andersen, M.S., Asmussen, H., Asmussen, J. (2002), ‘The Project of Korpus 2000 Going Public’, in: A. Braasch and C. Povlsen (eds.) Proceedings of the Tenth EURALEX International Congress, EURALEX (2002), Copenhagen. Asmussen, J. (2001), Korpus 2000. Et overblik over projektets baggrund, fremgangsmåder og perspektiver. NyS 30. Nydanske studier & almen kommunikationsteori, Copenhagen. Asmussen, J. and O. Norling-Christensen (1998), ‘The Corpus of The Danish Dictionary’, in: Lexikos 8, Afrilex Series 8:1998, Stellenbosch, pp. 223242. Belica, C. (1998), Statistische Analyse von Zeitstrukturen in Korpora. Teubert W. (ed.): Neologie und Korpus. Tübingen. Bergenholtz H. (1988), DK87: Et korpus med dansk almensprog. Hermes, Journal of Linguistics 1, Århus, pp. 229-237. Christ, O. (1994), A modular and flexible architecture for an integrated corpus query system. COMPLEX’94, Budapest. Church, K. and P. Hanks (1989), Word association norms, mutual information and lexicography. ACL Proceedings, 27th Annual Meeting, Vancouver. Elbro, C. (2002), Ift, ifm, mht, mhp og andre uspecifikke præpositioner. Mål og Mæle 3:2002, Copenhagen, pp. 17-23 Garside, R. and P. Rayson (2000), Comparing corpora using frequency profiling. Proceedings of the workshop on Comparing Corpora, held in conjunction with the 38th annual meeting of the Association for Computational Linguistics (ACL 2000). 1-8 October 2000, Hong Kong, pp. 1 - 6. 48 Jørg Asmussen Jarvad, P. (1999), Nye Ord. Ordbog over nye ord i dansk 1955-1998. Copenhagen. Keson, B. (1998). Documentation of The Danish Morphosyntactically Tagged PAROLE Corpus. Society for Danish Language and Literature, DSL, Copenhagen, http://korpus.dsl.dk/e-resurser/parole-doc.rtf Kilgarriff, A. (2001), Comparing Corpora. International Journal of Corpus Linguistics, 2001, vol. 6, no. 1, 2001, pp. 97-133. Rundell, M. (2002), ‘Good Old-fashioned Lexicography: Human Judgment and the Limits of Automation’, in M-H. Corréard (ed.) Lexicography and Natural Language Processing. A Festschrift in Honour of B.T.S. Atkins. EURALEX 2002. Ruus, H. (1995), Danske Kerneord. Centrale dele af den danske leksikalske norm. Copenhagen. Synchronic and diachronic variation: the how and why of sociolinguistic corpora. Kate Beeching University of the West of England Abstract This paper aims to illustrate the potential of (spoken) sociolinguistic corpora for research studies in both synchronic and diachronic variation, with reference to French, and to suggest ways in which useful research corpora may be established for future generations of scholars. Spoken corpora and corpus tools are an excellent heuristic in charting distributional frequencies or probabilistic factors. Andersen (2000) suggests that the upsurge of innit and like in the COLT Corpus of adolescent English may be more than age-grading. The present paper will present broad-brush preliminary evidence with respect to the evolution of selected pragmatic particles in French. 1. Introduction The title of my paper on sociolinguistic corpora has put the how? before the why? This raises a very interesting methodological issue relating to corpus collection and corpus-based research. Generally speaking, linguistic research design which includes data collection focuses very precisely on the nature of the data required for the research project in question. You start with the why? and move on to the how? A hypothesis is formulated and the data collection is specifically designed to test that hypothesis. The nature of corpus design is quite the opposite: very large amounts of data are collected from a representative sample of the population, both written and spoken, and - as was the case for the Survey of English Usage - the corpus designer decides which genres are to be included as being those which are most typical of the language at any one time. Inevitably, the composition of the corpus influences the findings which may be made and bias may creep in. For a long time, for example, the spoken language was considered to have been underrepresented in the British National Corpus and, even now, if we compare how much time the average British English speaker spends speaking and listening as opposed to reading or writing, corpora are generally weighted in favour of the written language. The time-consuming nature of transcription plays a major role in the relative paucity of spoken material. Should researchers wish to chart distributional frequencies in the spoken as opposed to the written language, they are at liberty to select sub-corpora. Inevitably, too, the corpus may not include the precise register of the language which the researcher is interested in. In Beeching 50 Kate Beeching (1997), I provide an apology for the small home-grown corpus which focuses on, in this case, the type of spoken French most needed by horticulturalists (a group of learners I happened to be teaching at the time and for whose needs no specifically-designed corpus was available). At the University of the West of England, Bristol, we are currently engaged in a Joint Project collecting samples of learner English/French/German and Spanish at a number of levels. Already the methodological questions are being posed: what level of detail will be required in transcription? Does the method of data collection fit the individual research aims of the staff involved? How best can we collect data which it will be subsequently of interest to interrogate for studies which may be as diverse as morphological development, lexical richness and so on. The how? and the why? questions bedevil the collection of corpus material in a very large number of fields. The key factor about corpus investigations is size: millions of words are considered to be required to gauge the use of lexical items, even those which are not particularly rare. For studies focusing on variation, large is necessary but insufficient. In the case of spoken data requiring transcription, large means long hours of work by trained personnel - in other words, funding. Corpus Linguistics is a field where individual researchers, public and private funding, universities, libraries, publishers and government need to pull together, as has been most spectacularly demonstrated by the success of the BNC. A collaboration of this sort has, unhappily, not occurred for French. For studies focusing on variation, large is necessary, but insufficient, because samples must be accompanied by particular demographic data to allow investigations to be meaningful. And opinions differ on exactly how data should be collected and collated for reasonable conclusions to be drawn. In this paper, my remarks relate to the type of corpus which lends itself to the study of semantico-pragmatic change. On the one hand, the why? logically precedes the how? On the other hand, most empirical research papers place the methodology section before the results section. It is for this reason that I have positioned the how? before the why? in this paper, though the latter part of the paper serves as justification for the methods described in the former. The questions to be raised are: How reasonable is it for sociolinguists to ask for specific information about speakers to be included in the data set along with transcriptions when corpora are compiled? And what sort of information and what degree of detail is useful? 2. The how? Examining data collection methods which have been used in the past is one way to explore best practice for the future. Two outstanding examples of existing sociolinguistic corpora for French are the Etude Sociolinguistique d’Orléans (collected between 1966-1970) and the VALIBEL Corpus. The latter is the largest electronic corpus of spoken French, with 400 hours of recordings or about Synchronic and diachronic variation 51 4 million words, and hopes to be representative of the “communauté linguistique Wallonie-Bruxelles” (Francard et al. 2002). Each interview is accompanied by demographic information covering the following six parameters: sex, age, geographical location, place of birth, educational background and the socioprofessional situation of the informant. More information is accessible at http://valibel.fltr.ucl.ac.be. The VALIBEL centre focuses on variation studies (lexical, syntactic, phonetic and prosodic). Despite the excellence of this corpus, it is of little use for my exploration of hexagonal French, as its focus is entirely on Belgian French. As yet nothing similar is available for hexagonal French though the Orléans Corpus is an excellent model. My own very small Bristol Corpus is available on-line and the “reference corpus of spoken French” currently the responsibility of Mireille Bilger (Université de Perpignan: [email protected]) holds great promise. The Orléans Corpus, the collection and transcription of which was supported by grants from the DES, the French Embassy in London, BELC in Paris, the CRDP in Orléans, the CMPP in Orléans, INSEE in Orléans and the French Centre for European Sociology is an invaluable resource. As the authors explain in their catalogue, the origins of the ESLO (Etude Sociolinguistique d’Orléans) go back to 1966, the era of the ‘audio-visual revolution” in language teaching. There was an acute need for non-literary and everyday samples of authentic spoken French. The aim, then, of the corpus was to collect a body of recordings of spoken French from an urban environment, the choice of subjects being governed by sociological criteria in order to ensure that the corpus was representative of French speakers of the time. The corpus was to be transcribed and made available to researchers in linguistics, sociology and pedagogy and was to be used as the basis for teaching materials. The resulting electronic corpus comprises approximately 109 hours of spoken French, 902,755 words of which have been orthographically transcribed with a further 13 hours of phonetic transcription. These transcriptions are available at: http://bach.arts.kuleuven.ac.be/elicop. Each interview is accompanied (in the catalogue) with very full background information, including not only the six parameters favoured by VALIBEL but the initials of the interviewer, the date of the interview and date and place of birth of the interviewee, the INSEE coding of the interviewee, their marital status, political leanings, the length and quality of the recording and some indication of the topics covered. Most interviews covered a pre-determined set of questions, but the speakers appear relaxed and informal in their mode of response. This imaginative yet rigorously managed project is a treasure-house of historical evidence on the French language, the first and, as yet, virtually the only French corpus in existence which is freely available on-line. The Bristol Corpus also had its origins in language pedagogy. Supported by small grants and advances from OUP and CUP, I spent 10 years from 1980 to 1990 as a freelance French textbook writer, collecting, transcribing and preparing pedagogical materials for French (e.g. Beeching, 1985, 1986, 1989, Beeching & 52 Kate Beeching Page, 1988, Beeching & Le Guilloux, 1990, 1993). All were based on and had an accompanying cassette containing authentic taped spoken interviews with French people from all walks of life and from a number of urban and provincial parts of France. A selected sub-section of the interviews recorded during these years was transcribed as part of my PhD thesis, an abridged version of which was published in 2002. These interviews now make up the Bristol Corpus which comprises 17.5 hours of orthographically transcribed speech, or 155,000 words, involving 95 speakers, aged 7 to 90 years with a balanced range of educational backgrounds. The Corpus may be accessed at http://www.uwe.ac.uk/facults/hlss/languages/ research/staff/CORPUS.pdf. The file contains some demographic information about the speakers: their sex, age (in 3 age bands: 0-20; 21- 40; 41+) and level of education (in 3 bands: no bac, bac but no university degree, bac + university degree(s)). Bilger (2002) presented the “Reference Corpus of spoken French” the constitution of which was proposed in 1998 by the GARS teams and the CNRS ESA 6060 team directed by C.Blanche-Benveniste. Currently the DELIC team led by J. Veronis is completing the corpus. Bilger regretted that the corpus did not fulfil all of the criteria which Sinclair suggests characterise a ‘reference’ corpus as not all speech situations are included and a number of usages of the language are missing. She suggests that this is but a first step in a vaster project. An attempt has been made to capture a representative sample of common, current usage. The corpus is 400,000 words long, the interviews were recorded in 40 towns and three speech situations have been favoured: “private” speech, professional speech and public speech. Bilger did not mention in detail what background notes accompany the transcriptions. To sum up, a reference corpus which is useful for sociolinguistic research must attempt to provide samples which are representative of all members of the society in question talking at a similar level (or levels) of formality in order for comparisons to be made and variation discerned. A minimum of information should accompany each interview: the sex, age and educational background of the speakers. 3. The why? Reference corpora which are allied with demographic information can be used, not only to look at synchronic variation and the way that social hierarchies and power may be enshrined in language use (see Beeching, forthcoming a) but also to investigate diachronic variation (see Beeching, forthcoming b). This is the most exciting new development. Though change may conceivably occur because of the written language, it is generally recognised that most changes occur in face-to-face interaction. In the past, language change has been studied with some difficulty: written documents of various sorts have been scrutinised to find Synchronic and diachronic variation 53 evidence of what might have been happening in the spoken language at a particular stage (an excellent example of such an investigation at the phonological level is Sampson, 2002). The fact that we now have what amount to historical documents in the form of tape-recordings and, if we are lucky, transcriptions of those tape-recordings, means that, from now on, we will be able to monitor change as it occurs. Corpus evidence allows us to quantify shifts, gauge reversals, and check distributional frequencies. Electronic search facilities and statistical packages can alleviate much of the drudgery entailed. We can refine not only our descriptions of linguistic change but also our hypotheses about the causes and effects of such changes. Linguists have traditionally favoured language-internal causes of linguistic change. The advent of sociolinguistics and sociolinguistic evidence allows us to put forward a very strong argument for the importance of language-external, social factors. As Croft (2000: 4) remarks ‘the real entities of language are utterances and speakers’ grammars. Language change occurs via replication of these entities not through inherent change of an abstract system’. There is considerable evidence that synchronic and diachronic variation are inextricably linked and that language items which begin as one variant of a variable or one sense of a polyseme gradually gain the upper hand. A new formfunction configuration can eventually obscure the original one as, becoming more frequent, it is routinised. As Heine, Claudi and Hünnemeyer (1991:261) remark with reference to grammaticalisation, language change and synchronic variation are inextricably interlinked: Grammaticalization has to be conceived of as a panchronic process that presents both a diachronic perspective, since it involves change, and a synchronic perspective, since it implies variation that can be described as a system without reference to time. Jakobson (1952/1963:37) made a similar point when he said Pendant un certain temps, le point de départ et le point d’aboutissement de la mutation se trouvent coexister sous la forme de deux couches stylistiques différentes...Un changement est donc, à ses débuts, un fait synchronique 1 . Grammaticalisation is a term which is generally applied to nouns and verbs which come to serve grammatical functions and then continue to develop new grammatical functions. Classic examples include the development of LatinRomance articles, auxiliaries and indefinite pronouns. Whilst much focus has traditionally been placed on what Givón has famously referred to as ‘Today’s morphology is yesterday’s syntax’, it could be argued that the emphasis on morphology reflects the neogrammarians’ preoccupation with this area rather than giving a just view of a process or phenomenon which extends to semantic and pragmatic areas, not solely purely ‘grammatical’ ones. In order for a lexical 54 Kate Beeching or content word to gain in currency and become a function or grammatical word, it must of necessity lose the particular lexical meaning which ties it to specific contexts and become desemanticised. This ‘bleaching’ has been regarded as a somewhat negative fate for words though some researchers point up that we should not consider the desemanticisation as ‘loss’ but as an enriching or multiplying of meaning. In addition, desemantisation is a sine qua non for a term to become grammaticalised or gain in currency, to change from a ‘content’ to a ‘function’ word, as Haspelmath (1999:1062) explains: Semantic generalization or bleaching is usually a prerequisite for use in a basic discourse function, that is, for the increase in frequency that triggers the other changes. The semantic generalisation which is so often observed in the development of a lexical category to a functional one appears, then, to be not a consequence of routinisation but a pre-requisite for it. This “bleaching” and “gaining in currency” is not a phenomenon which restricts itself to morphology but is omnipresent, not least in cases of pragmatico-semantic change. In Beeching (forthcoming b) I argue precisely this and I posit that, along with the formal changes which appear to accompany grammaticalisation (desemantisation, fusion, coalescence), what we are witnessing in all cases of language change, is a process of pragmaticalisation which I define as follows: Pragmaticalisation is the manner in which words, used in context, shift in meaning or attract a new social semiotic, become habituated in that usage and are propagated because of the new fashion or prestige which is attached to them. Pragmaticalisation occurs during human interaction and human interaction is always heavily overlaid with connotations which may be most readily explored using politeness theory. Kasper (1990) distinguishes between two trends in the conceptualisation of politeness: politeness as strategic conflict-avoidance and politeness as social indexing. Both are involved in a consideration of pragmatic particles such as like and innit in British English or enfin, hein or quoi in hexagonal French. All are used to encourage speaker involvement, and they are part of the means whereby speakers downtone assertions, avoid presenting themselves as the expert and thus they are a means of avoiding conflict and managing face needs. But they are often words, too, which are stigmatised in “polite” circles, they are familiar, colloquial, only used in the spoken language, in informal contexts or by an easily identifiable subset of the population. Quotative like for example is much more frequent on the lips of younger speakers. Hein, and, to an even greater extent, quoi, are characteristically used to a much greater extent by working class speakers. Frequent use of such terms, then, contains a social semiotic, resulting in social indexation. Synchronic and diachronic variation 55 The underlying processes of language innovation and language diffusion appear to me to reside in the conventionalisation of generalised invited inferences, some of which may be cognitively motivated metaphorical or metonymic interpretations, but many of which are motivated by considerations of politeness in its broadest sense including both notions of sociability and face and of social indexation (see Eeelen, 2001, and Beeching, 2002: 25-27). On the one hand, Croft (2000: 178) intimates the usefulness of politeness theory in helping to explicate the directionality of the transmission of variants, an aspect not accounted for in the Milroys’ weak-tie model. On the other hand, drawing on Levinson (2000), Traugott and Dasher (2002) posit an Invited Inferencing Theory of Semantic Change (IITSC). This theory suggests that historically there is a path from coded meanings (Ms) to utterance-token meaning (IINs) to utterance-type, pragmatically polysemous meanings (GIINs) to new semantically polysemous (coded) meanings. Their arguments for the role of conceptual metonymy are persuasive. My theory of pragmaticalisation builds on the IITSC but places it firmly in the context of politeness theory and of Milroyan sociolinguistic theory and methodology. Both Croft (2000) and Givón (2002) draw out the parallels between adaptive behaviour in biology and linguistic phenomena, with variation seen as the indispensable tool of learning, change and adaptation and this in the context of the sociability, co-operation and communicative way of life rooted in a hunting-and-gathering society. The primacy of sociability, the bid for conflictavoidance in human interaction and the equally prevalent urge for social indexation promote not only a high rate of usage of politeness markers but also their differential usage in different identity groups. Moreover, because such particles are both frequent and highly polysemous, they are subject to change. Hence the fascination they hold for scholars of synchronic and diachronic variation. The techniques and tools of Corpus Linguistics are formidable allies when exploring the relative distributional frequencies of old and new forms coexisting in synchrony. Though linguists have generally focused on change and, in particular, on the beginning and end of a change, the far commoner and more ubiquitous situation is that of stable sociolinguistic and stylistic variation, out of which change may, or may not, emerge. As Hopper and Traugott (1993: 95) point out: Changes do not have to occur. They do not have to go to completion, they do not have to move all the way along a cline ..[....] the outcome of grammaticalization is quite often a ragged and incomplete subsystem that is not evidently moving in some identifiable direction. Polysemy and synchronic variation are far commoner than diachronic change and, coupled with social stratification, they constitute the raw material without which change is not possible. Distributional frequencies may fluctuate, depending upon fashion and prestige, a form-function association may dwindle and even die 56 Kate Beeching and, in those relatively rare cases, change may be said to have occurred. There appears to be a probabilistic aspect to the spread of a new form, the now famous S-curve, whereby after a slow start a new form surges forward at an accelerated pace and then falls backs as it stabilises. In Beeching (forthcoming a), I have attempted to capture the numerous complexities involved in the sociolinguistic promotion and demotion of a new form in the disarmingly simple formula l = p n, where l is the likelihood of spread, p are the positive aspects of the (changed) form-function configuration and n are the negative ones in the mind of the speaker involved. This formula allows for the fact that not all forms are positive for all members of a speech community and that attitudes to a form may change over time. In highly stratified societies, the n value of a stigmatised form entirely eclipses its p value. Only if strata become less rigid and identification with the life-style associated with the form rises can the n value drop. A measurement of the distributional frequencies of a form in speech samples from similar speech communities at two points in time may indicate the ‘l’ and help us to chart fluctuations over a large speech population. In this respect, the usefulness of large corpora of transcribed spoken language coupled with the demographic information classically associated with sociolinguistic studies - the age, sex and socioeconomic background of speakers - can hardly be exaggerated. In Beeching (forthcoming b), I survey the evolution of four pragmatic particles, as evidenced in a comparison of their usage in the (1966-1970) Orléans Corpus and the (1980-1990) Bristol Corpus. Many caveats must be issued concerning the conclusiveness of the data presented. The corpora are small by many standards. The Orléans Corpus is restricted to only one town in France, whereas the Bristol Corpus covers a number of different towns and regions. Although a similar amount of speech is examined in each corpus (around 155,000 words), ninetyfive speakers are surveyed in the Bristol Corpus while only twelve have been studied in detail in the Orléans Corpus, two men and two women from each of the three education groups which I have designated here as WC, LMC and MC. As a means of illustrating the ‘why?’ of sociolinguistic corpora, I wish now to focus on only two of the particles studied in Beeching (forthcoming b): hein and quoi. Hein and quoi both occur in an utterance-, or at the very least, tone-groupterminal position and are mildly stigmatised (they do not occur in formal written discourse and are highly unlikely to occur in formal spoken French). Quoi remains perhaps slightly more stigmatised than hein. Both, however, could be said to serve social/interactional purposes in maintaining hearer-involvement and in hedging or downtoning remarks which a speaker or hearer might consider over-assertive. Hein is generally translated into English by a tag question or by you know?, while quoi may be rendered ‘as it were’, ‘so to speak’, ‘know what I mean’, ‘like’ ‘sort of’, ‘kind of thing’, ‘you know’ or even ‘of course’. Oui, peut-être mais ça dépend aussi, hein? Yes, perhaps, but it depends, too, doesn’t it? Synchronic and diachronic variation 57 ne pas avoir que des contraintes dans dans sa vie, quoi, hein? not just to have obligations in in one’s life, as it were, you know? Hein and quoi appear, thus, to function in both of the manners in which politeness is conceptualised. They serve both to flag social indexation (they are nonstandard, stigmatised forms) and to mediate sociability. While quoi seems to have one core function or meaning (it flags tentativeness concerning the adequacy of one’s expression), Beeching (2002) charted two main functions or meanings of hein: a Hyperbolic function, where hein underscores an emphatic remark, and a Discoursal function, where hein is used as a backchannel device to maintain hearer involvement. Results in Beeching (2002) based on an analysis of the Bristol Corpus suggested a shift in distributional frequencies from the Hyperbolic to the Discoursal usage and, simultaneously a change in social semiotic: the Discoursal usage of hein was favoured by female speakers. 140 120 100 80 Mean HRANDQR 60 CLA SS 40 WC 20 MC 0 LMC Bristol Or léans CORPUS Figure 1: Mean rates of hein and quoi usage in the Bristol and the Orléans Corpus As Beeching (forthcoming b) shows, rates of hein usage are similar overall in both corpora. The class distribution of usage differs dramatically, however. In the later, Bristol, Corpus, the middle class speakers have adopted hein and working class rates are proportionately smaller. Rates of quoi-usage have, by contrast, doubled in the intervening years, with extremely high rates amongst working class speakers but much higher rates, too, amongst middle class speakers. 58 Kate Beeching Figure 1 charts the distributional frequencies of the sum of the means of hein and quoi rates per 10,000 words in the Bristol and Orléans Corpora subdivided for educational background. (It should be noted that the corpora had to be screened in the case of quoi to focus only on occurrences of it in tone-group-terminal function: quoi has of course a number of other more canonical pronominal uses in interrogative and relative constructions - de quoi parles-tu? il n’y a pas de quoi rire etc.) The most striking aspect of this figure is the increase in the usage of hein and quoi amongst middle class speakers. There appears to have been a slight dropping of in lower-middle-class usage of these stigmatised forms in the Bristol Corpus and a very slight increase in working class usage but the increase in usage overall is amongst middle-class speakers. Table 1 demonstrates the role played by the speaker’s sex in the increased rates of hein/quoi usage. Table 1: Rates of hein/quoi usage per 10,000 words in the Orléans and Bristol Corpora, subdivided by sex and class (N= the sum of the raw number of occurrences of hein and quoi). Male Orléans Corpus (1966) Bristol Corpus (1990) Female WC LMC MC WC LMC MC 65.61 N=200 57.91 N=271 62.54 N=191 27.69 N=53 15.03 N=37 24.33 N=22 22.94 N=69 46.89 N=150 13.92 N=51 33.96 N=108 15.84 N=29 35.98 N=54 In the Orléans Corpus, hein/quoi usage is a predominantly male WC and LMC phenomenon. The female speakers observe the same reticence concerning hein/quoi usage as the MC males. In the Bristol Corpus, however, rates are a great deal more evenly distributed. Though the highest rates belong to WC speakers, this is true of both the men and the women. Indeed, the female LMC and MC rates exceed those of their male equivalents. It seems that hein and quoi are becoming less stigmatised and that it may be women who are leading this change in social semiotic. In Beeching (forthcoming a), I discuss the relationship between politeness and power and the maintenance of hierarchies through asymmetrical language-usage. It is too early to make any conclusive remarks and replications of the study, drawing on more data from the invaluable Orléans Corpus and also from the more recent Reference Corpus, is advisable. However, it seems possible that, if middle class speakers are beginning to adopt stigmatised ‘working class’ speech forms, there has been a democratisation, a shift in the hierarchical nature of French society. Our study of distributional frequencies of pragmatic particles may inform Synchronic and diachronic variation 59 us not only about linguistic usage but also about the structure of society. It is the inter-relationship between the two which may bring about linguistic change. 4. Conclusion In this brief chapter, it has been impossible to do full justice to either of my main foci: the usefulness of sociolinguistic corpora and the way that diachronic change may be charted through differences in distributional frequencies (synchronic variation) noted at different points in time. As Macaulay (2002, 298) points out, there are many variables that affect samples of speech. He claims, however: Yet we need not despair. One way forward is in replication. As more studies are carried out, the influence of accidental factors may be easier to detect. Macaulay makes a number of very useful practical suggestions concerning the way that replication may be made more reliable. For one thing, researchers need to show rates of occurrence in a standardised way which will allow comparisons to be made e.g. how many occurrences per 10,000 words. He recommends the amassing of comparative data. When collecting corpus material, I urge fellow researchers to append as much information as possible concerning the date and place of collection, the sex, age, place of birth and educational and socioeconomic background of the speaker, the nature of the relationship between the interviewer and interviewee (or speakers, if a ‘fly-on-the-wall’ recording). Macaulay stresses, too, the usefulness of the computer, a point of view which I whole-heartedly support. One is forced, however, to recognise that linguists do not look set to be put out of their job by a computer: the polysemy and multifunctionality of certain terms, the subtlety of invited inferences and the continuum along which terms travel towards coded meanings pose a formidable challenge to the computer programmer. Finally, as any researcher who has collected and transcribed spoken material knows only too well, transcription is fantastically time-consuming. It has become obvious to me, however, that when we transcribe, what we are engaged in is not only a record and research base for the study of contemporary forms. What we are engaged in is the creation of a historical archive of spoken language which was impossible for our less technologically-advantaged forebears. To aid our understanding of the nature of language, of the way that it is structured and how that structure reflects social phenomena, of its variation, both synchronic and diachronic, every attempt should be made to take samples which are representative and to accompany our hard-won transcriptions with detailed demographic information. In this way, we allow others to stand on our shoulders and the fruits of our efforts are redoubled. 60 Kate Beeching Notes 1 For a while the departure and arrival point of the change coexist in the form of two different stylistic layers.... a change is thus, in its beginnings, a synchronic phenomenon. References Andersen G. (2000), Pragmatic Markers and Sociolinguistic Variation. A relevance-theoretic approach to the language of adolescents. Amsterdam/Philadelphia: John Benjamins. Beeching, K. (1985), Vrai de vrai! Oxford: Oxford University Press. Beeching, K. (1986), A vrai dire. Oxford: Oxford University Press Beeching, K. (1989), Actifrance. Oxford: Oxford University Press Beeching, K. (1997), French for specific purposes: the case for spoken corpora. Applied Linguistics. (1997). 18, 3, 374-394. Beeching K. (2002), Gender, Politeness and pragmatic particles in French. Amsterdam/Philadelphia: John Benjamins. Beeching K. (Forthcoming a), Pragmatic particles - polite but powerless? Tonegroup terminal hein and quoi in contemporary spoken French. To appear in one of the first two editions of Multilingua 2004. Beeching, K. (Forthcoming b), Pragmaticalisation, politeness and linguistic change: synchronic evidence from French. Beeching K., & I. le Guilloux. (1990), Ça se dit et ça s'écrit. Oxford: Oxford University Press. Beeching K. & I. le Guilloux. (1993), La passerelle. Cambridge: Cambridge University Press. Beeching K., & B. Page. (1988), Contrastes. Cambridge: Cambridge University Press. Bilger M. (2002.), Présentation du “corpus de référence de français parlé”. Paper given at the ATALA Conference, “Constitution et exploitation de corpus du français parlé” 25 May 2002. Croft, W. (2000), Explaining Language Change. An Evolutionary Approach. Harlow: Longman. Eelen, G. (2001), A critique of politeness theories. Manchester: St. Jerome Publishing. Francard M., G. Geron, V. Giroul, P. Hambye, A.C.Simon, and R.Wilmet (2002), Le centre de recherche Valibel: des corpus oraux au service d’un observatoire du français en Belgique. Paper given at the ATALA Conference, “Constitution et exploitation de corpus du français parlé” 25 May 2002. Synchronic and diachronic variation 61 Givón, T. (2002), Bio-linguistics. The Santa Barbara lectures. Amsterdam: John Benjamins. Haspelmath, M. (1999), Why is grammaticalization irreversible? Linguistics, 376: 1043-1068. Heine, B., U. Claudi, and F. Hünnemeyer. (1991), Grammaticalization: a conceptual framework. Chicago: University of Chicago Press. Hopper, P. J. & E.C.Traugott. (1993), Grammaticalization Cambridge: Cambridge University Press. Jakobson, R. ([1952] 1963), Essais de linguistique générale. Paris: Éditions de minuit. Kasper, G. (1990), Linguistic politeness: current research issues. Journal of Pragmatics 14: 193-218. Levinson, S. (2000), Presumptive Meanings: the theory of Generalized Conversational Implicature. Cambridge, MA: MIT Press, Bradford. Macaulay, R. (2002), Discourse Variation. In Chambers, J K, Schilling-Estes, N The handbook of language variation and change. Oxford: Blackwell, pp. 283-305. Sampson, R. (2002), A transient vowel in early modern French: i nasal, in: R. Sampson and W. Ayres-Bennett (eds.) Interpreting the history of French. A Festschrift for Peter Rickard on the occasion of his eightieth birthday. Amsterdam/New York: Rodopi. Traugott, E. and R. Dasher. (2002), Regularity in Semantic Change. Cambridge: Cambridge University Press. This page intentionally left blank Statistical analysis of the source origin of Maltese Roderick Bovingdon Angelo Dalli University of Sheffield Abstract This paper presents the results of the first ever large-scale statistical analysis of Maltese using the newly formed Maltilex Corpus. Traditional etymological and categorical analyses were supplemented with data mining techniques to provide accurate results with reduced effort. Statistics about the relationship between etymology and word classes were analysed from different viewpoints. Maltese grammar and morphology remain to this day largely Arabic, but with distinct Romance and English morphological accretions. Italian lexical influence upon present day Maltese has exceeded the Arabic content in a quantitative sense, enriching Maltese from a purely root based morphology with additional productive Romance features. 1. Introduction The most recent theory relating to the origin of the Maltese language points to a direct Sicilian-Arabic connection. This claim, by Dionisius Agius (Agius, 1990; Agius, 1993; Agius,1996) and Joseph Brincat (Brincat, 1994), although supported by significant research findings by other contemporary scholars from different disciplines 1 , remains inconclusive (Bonanno, 1988; Borg and Alexander, 1978; Borg and Alexander, 1994). In the light of the prevailing knowledge of his times, Joseph Aquilina was on the right track when he proposed the Arabic of the Muslim Aghlabids of the Maghreb, as the most likely original source for the Maltese language (Aquilina, 1959). Even at this early stage Aquilina does not rule out a possible SicilianArabic link (Aquilina, 1988). In contrast to Aquilina’s thesis, the current Siculo-Arabic school of thought presents a revolutionary twist. It repositions the original source of Maltese from a direct northern African origin to an offshoot of Sicilian Arabic. This theory reinforces the notion that the form of Arabic which was adopted in Malta, as the forerunner of today’s Maltese, was already tainted with non-Arabic Language influences from its very inception (Agius, 1996; Brincat, 1995). Whilst considerable language convergence exists between Maltese and Siculo-Arabic 64 Roderick Bovingdon and Angelo Dalli (Agius, 1996), there still remain significant unexplained aspects of inquiry into early Maltese 2 . The answers to most unexplained lines of inquiry, as sparse as the remaining evidence may be, may very well be laying dormant within the more subtle and distant aspects of early Maltese. The grammar and morphology of modern Maltese remain to this day Arabic, with distinct Romance and English morphological, and increasingly lexical, accretions (Aquilina, 1979; Mifsud, 1995). A thorough combing of European, North African and Turkish depositories as well as of private Maltese collections, has the potential of uncovering major revelations, as given in the relatively recent contributions in Wettinger and Fsadni (1968), Cassola (1992 and 1996), and Brincat (1995). Scholarship of Maltese has taken great strides forward by way of research, both within the spheres of linguistics as well as from a purely historical perspective (Cassola, 1992; Mifsud, 1995). Wettinger’s exposure of Vatican Manuscript 411 deserves closer scrutiny as a possible significant link between medieval Maltese and Arabic, as well as being a valuable pointer to the existence of other similar contemporary documents elsewhere, including possible links to Hebrew (Wettinger, 1979). The quantity and significance of recent documentary discoveries relating to medieval Maltese, are highly suggestive of even earlier records of written Maltese. Such exciting finds could provide scholars with the elusive missing linguistic links, bridging the gap between the pre-Arabic and Arabic beginnings. This need not necessarily conflict with Agius’ and Brincat’s postulations, as there is already ample evidence of the strong Siculo-Arabic connection. Agius openly claims, in his major work on Siculo Arabic, that Maltese is a direct offshoot of Siculo-Arabic with no connections to the north African littoral 3 . Brincat (1994) emphasises the lack of a substrate for the pre-Arabic period. The distinct possibility of remnants of a pre-Arabic substrate concealed within the later Siculo-Arabic strata is suspected. 2. Non-Arabic linguistic influences on contemporary Maltese Traditionally the Romance element in Maltese was thought to commence with the coming of the Normans from 1049 onwards. Nowadays the year 1127 is perceived as a more realistic date in historical terms (Cassar, 2000). The preNorman Arabic content appears to have been heavily Sicilian based (Brincat, 1995). Brincat’s contention that the lack of a substrate is the strongest pointer to a Siculo-Arabic genus 4 holds much merit along with Agius’ extensive approach to such origins. Statistical analysis of the source origin of Maltese 2.1 65 Knights of St. John of Jerusalem The more overt lexical and the later morphological Romance additions to Maltese, had to wait for the arrival of the religious Order of the Knights of St. John of Jerusalem, circa 1530, before any significant inroads began to become apparent. The Arabic-oriented populace of Malta, left to its own devices, continued to interact harmoniously for a long period of time, within a Muslim 5 , Christian 6 and Jewish milieu. Under the Knights, direct rule was introduced 7 , with the consequent imposition of the foreigners’ will and culture. During this long period in Malta’s history between 1530 and 1798, the Maltese came under more direct and imposing influence from the Romance element, due to the comparatively large numerical presence of Knights 8 , with an increasing quantity of Romance words and phrases from different regions of the Italian mainland (especially from the northern regions of Italy) being absorbed into Maltese. The increased social interaction with the local populace through this direct rule was an overriding force impinging on every aspect of the indigenous Maltese way of life, not least the language admixture. The Order’s rule over Malta lasted for over two centuries. During this vibrant period in the archipelago’s history, this considerably increased interaction between the rulers, the Knights, and the general populace, consisted for the most part in a master-subordinate relationship. It is important to note this point, as such social interaction between the overlords and the general populace meant that several aspects of the rulers’ culture, not least their predominantly Romance languages, imprinted their influences on the mainly peasant indigenous stock. After the Order was ousted from the islands by the French under Napoleon in 1798 – who in turn saw their demise after only two years’ occupation – the British rule took possession for the next century and a half. 2.2 Modern Italian Ironically, the greatest linguistic inroads from the Italian mainland occurred during the early days of the British rule 9 . Owing to the political turmoil aroused on the Italian mainland, during the unification of Italy many political refugees sought and were granted political asylum in the British protectorate of Malta. These émigrés, who included a number of prominent Italian intellectuals and politicians, banded together and formed a strong political lobby of their own. With Malta’s long tradition of strong political, cultural, administrative, religious and linguistic influence from the Italian mainland as well as from Sicily, these émigrés found many sympathisers amongst the local population (Friggieri, 1979). A considerable number of newspapers in the Italian language flourished on the island, stimulating the adoption and spread of modern Italian Language influence 66 Roderick Bovingdon and Angelo Dalli on both the Maltese idiom, as well as Maltese thought 10 . The infamous Language Question of the late 1920s and early 1930s, when the British colonial power squeezed out Italian from officialdom, saw the insertion of Maltese as the language of common parlance, while English took the coveted functionality of Italian. This political manoeuvring marked a phase when Italian linguistic influence suffered a temporary though prolonged lull (Fuccio, 1933; Hull, 1993). Modern Italian has since made vast inroads into contemporary Maltese, mostly through the influence of Italian television, which enjoys a wide following in Malta. So strongly has this new influx been felt that the large quantity of Italian lexemes entering Maltese has affected the morphology of the language – a phenomenon indicated by the statistics presented in this study. Contemporary Italian borrowings can be distinguished with ease from those previously adopted during the much earlier pre-British period 11 . These additional and more significant Italian influences have markedly shaped Maltese with distinctly European traits. Such trend emanates from the intellectual class which consistently borrows new terminology, mostly from Italian, with increasing inclusions from other European languages; these being the dominant language sources whence contemporary intellectual, scientific and technological innovations originate. This study clearly shows that Italian lexical influence on present-day Maltese, if only by way of numerical representation, has surpassed the Arabic content. Italian morphological influences have taken hold to such an extent that the updating of Maltese grammars to include this aspect is being considered. Such development has evolved Maltese from a purely root-based morphology, as a typological feature of its Semitic past (Schweiger, 1994), with the additional productive Romance feature of catenation (Mifsud, 1995). 2.3 English The most recent linguistic influence on Maltese is English. English has steadily and increasingly affected Maltese, adding another language facet to the overall structure of Maltese (Mifsud, 1995). Interestingly enough, despite these recent accretions from English and the vastly different morphological structure of the two languages, the assimilation of English lexemes into a Maltese mould, occurs with the least possible disturbance to Maltese morphological structure. By comparison, contemporary Arabic, under the influence of similar pressures of modern life, as well as the ever-changing world political scene, when borrowing from English and American-English, is adopting similar assimilative patterns as Maltese (Holes, 1995). In the case of Maltese, this is especially so with verbs, where the basic structure of the English word is left intact while other morphemes in the form of mainly prefixes and suffixes take over (i.e. the stem does not change its internal make-up). Statistical analysis of the source origin of Maltese 67 Other linguistic devices, also of a Semitic nature, such as gemination of the initial radical in the case of verbs that are frequently preceded with a euphonic i, are also at work (Mifsud, 1995). Furthermore, as a result of this linguistic enrichment and infusion, Romance affixes assimilate naturally with both Semitic and English lexemes as an intrinsic part of standard Maltese grammatical structure. 3. Statistical and analytical results The data used in this statistical analysis has been sampled from the Maltilex Project’s corpus as of November 2001. The Maltilex Project is a joint effort between the Department of Computer Science and AI and the Institute of Linguistics at the University of Malta. Its aim is to produce the first national collection of computerised language resources for Maltese (Rosner et al., 1999; Rosner et al., 2000). The Maltilex corpus is made up of a representative mixture of newspaper articles of different kinds, including local and foreign news coverage, sports articles, political discussions and others, together with transcripts of radio shows, official government publications and some novels. When this statistical analysis was performed the corpus had 1.8 million words consisting of almost 70,000 different word forms. 3.1 Selection methodology and analysis tools A representative sample of 1,000 words was needed for the purpose of this study. The sample was selected using a strictly random process to ensure the validity of our statistical results. A chaotic function was used to assign a random number to all 70,000 unique word forms in the corpus, permitting a randomly ordered list of words to be created. This list was then examined and the following modifications were made: 1. Spelling Mistakes – Spelling mistakes were immediately corrected. Words that were spelled ambiguously were excluded from the list. This kind of modification and filtering is statistically sound since no attempt is made to influence the contents of the list. During the cleaning process the data was also converted into Unicode format to be universally accessible by all analysis tools. 2. Hapax Words – The word occurrence frequency in the corpus was examined for every word and all hapax words were automatically removed. This process removed most superfluous and arcane words that were accidentally inserted in the corpus and that appeared in the sample. This process can be statistically justified since hapax words can be seen as 68 Roderick Bovingdon and Angelo Dalli outliers that can be safely removed without affecting the validity of the resulting analysis. The etymology and class was noted down for every word in the sample. When a word had more than one class, the word entry was duplicated and a single category was entered for every word. In order to maintain accurate statistics, a weight was added to every entry representing the number of classes associated with a particular word. Thus every entry of a word having n classes was assigned a weight of 1/n. A total of 1034 entries were thus obtained, representing all possible etymology/word class pairs for the sample word forms. The data matrix that was obtained was analysed using a custom-written data mining tool to extract statistics about the relationship between etymology and word classes in Maltese. Overall statistics about the source language origins of Maltese together with the most commonly occurring word classes was also extracted. The use of a data mining tool enabled us to analyse the data from two different perspectives – word class distribution for every etymological class and vice-versa. 3.2 Etymology In our data sample, most of the words derived from Italian, Arabic and English. Source Origin of the Maltese Language It 54% Other 5% Eng 4% Eng>Dutch 0% Ar 41% Eng>It 1% Ar Eng Eng>Dutch Eng>It It Figure 1: Source Origin of the Maltese Language There were also some isolated cases of Italian and Dutch etymons 12 finding their way into Maltese through English. In the analysis these languages are denoted by Statistical analysis of the source origin of Maltese 69 the codes It, Ar and Eng respectively. The isolated cases are denoted as Eng > It and Eng > Dutch respectively. Figure 1 shows the source origin of the Maltese Language, with Italian words forming the majority of the words at 54%, Arabic at 41% and English at 5%. Table 1 shows a summary of the etymological analysis data 13 , showing the exact number of word forms pertaining to every etymological class. Table 1: Maltese Language Etymology Summary Etymology Ar Eng Eng>Dutch Eng>It It Count 411 36 1 9 543 Following this summary analysis we then split up the data set according to the three main source languages and further analysed the source languages’ contribution in terms of word classes. The following abbreviations were used to denote different word classes: adj – adjective; adv – adverb; conj – conjunction; demon – demonstrative; interj – interjection; n – noun; pers – personal; poss – possessive; pp. – participle; prep – preposition; pron – pronoun; v – verb. The word classes were further annotated with m, f, pl and dual to denote masculine, feminine, plural and dual forms of word class respectively. Arabic Grouped Word Classes n 21% adv 2% pp 4% Other 9% v 66% Figure 2: Arabic Word Classes adj 2% pron 2% prep 2% conj 1% v n pp adv adj pron prep conj 70 Roderick Bovingdon and Angelo Dalli Figure 2 shows the summary results for Arabic word classes, showing this language source as a significant contribution in terms of verbs, and to a lesser extent, nouns. Italian Grouped Word Classes v 29% adj pp 11% 9% adv 3% interj 0% conj 0% n 48% Other 0% n v adj pp adv interj conj Figure 3: Italian Word Classes Figure 3 shows the summary results for Italian word classes, showing Italian’s significant contribution in terms of nouns, and to a lesser extent, verbs. This further mirrors the significant contribution of Arabic. Figure 4 shows the summary results for English word classes. The relatively low percentage of English words (5% of the total) is consistent with established Maltese literary convention. This result seems to underestimate the percentage of English words in common Maltese parlance. From a purely lexical viewpoint Standard Maltese comprises 41% from Arabic origins, 54% from Italian and 9% from English as illustrated in Figure 1. Linguistically this points to a predominantly Italian influence with English slowly edging its way into the Maltese mould. The Arabic content, on the other hand, at surface value appears to be waning to a marked degree, bearing in mind that Maltese is still looked on as belonging to the Semitic fold. A deeper analysis justifies the continued classification of Maltese with the Semitic family of languages. Table 2 presents percentage data formed from the comparison of the word classes in each of the three source language groups, illustrated in Figures 2, 3 and 4. Statistical analysis of the source origin of Maltese 71 . English Grouped Word Classes pp 7% n 63% v 22% adj n pp v adj 8% Figure 4: English Word Classes Table 2: Word Class Composition Class Verbs Adverbs Nouns Adjectives Participles Pronouns Prepositions Conjunctions Arabic 66% 2% 21% 2% 4% 2% 2% 1% Italian 29% 3% 48% 11% 9% 0% 0% 0% English 22% 0% 63% 8% 7% 0% 0% 0% The following insights can be obtained from Table 2: 1. Smaller word classes in the category of pronouns, prepositions and conjunctions are not favoured either from the Italian or the English lexicon. This trend attests to their stronger adherence to their older Arabic origins. Considering the sparse use of adverbs in the older (Arabic) Maltese, this word class appears to have no preference for English with a minor inclination towards Italian borrowings. 2. The major word classes impinging on Maltese from both Italian as well as English are the two main classifications consisting of the Verbs class together with the Nouns/Adjectives/Participles class. 72 Roderick Bovingdon and Angelo Dalli These latter results strongly suggest that Maltese morphology, when borrowing from Italian, exhibits a distinctly stronger preference towards the nominal lexicology than the verbal portion, while English displays quite the opposite. Italian borrowings from the Nouns/Adjectives/Participles class show 68% nominal borrowings as opposed to English’s 78%. On the other hand, the Italian verbal borrowings from the Verbs class are not that much higher at 29% than their English counterpart of 22%. Considering the far lengthier historical-political connection with Italian mainland compared to the relatively recent and much shorter British connection, this result shows that verbal borrowing from English is increasing. 3.3 Word Classes The contribution of the source languages to different word classes in the Maltese language necessarily entailed an etymological analysis. The data mining tool used for the analysis enabled us to split the data set according to word classes. This allowed us to analyse the composition of different word class groups according to the source languages. Figure 5 shows a summary of the word category classes in Maltese with verbs (43%) and nouns (37%) making up 80% of all words. Word Categories of the Maltese Language v 43% Other 5% adv 3% pp 7% conj 1% interj 0% n 37% adj 7% prep pron 1% 1% adj adv conj interj n pp prep pron v Figure 5: Word Categories of the Maltese Language Table 3 presents the normalized counts of all word classes in Maltese. The counts are not all integers since the actual counts are based on the 1,034 word sample that was created by duplicating word entries having multiple word classes, as previously explained. The count was then normalised to 1,000 words. Statistical analysis of the source origin of Maltese 73 Table 4 presents the percentage data for the word classes in standard Maltese, as illustrated in Figure 5. Table 4 shows that an obvious preponderance of verbs over nouns in Standard Maltese, exceeding them by 6%. As such variation is not strong enough to indicate a distinct characteristic, it suggests that Maltese is a less concrete and a more conceptual language than formerly assumed. It therefore appears that contemporary Maltese, under the influence of its two main language sources, especially of recent times, is developing more abstract means of expression than it was previously able to impart. Table 3: Maltese Word Category Classes – Normalized Counts Word Class adj adv conj interj n pp prep pron v Counts 73.5 26.83 5 1 368.5 68 6.83 8 442.34 A comparison of Table 2 and Table 4 points to a consistent trend towards the Maltese linguistic structure relying more heavily upon verbs and nouns than other word classes. The shifting from a purely root-based morphology to that of a more diversified form with the addition of concatenation is perhaps the most significant evolutionary device Maltese has adopted in recent times. Table 4: Maltese Word Category Classes – Percentages Word Class Verbs Adverbs Nouns Adjectives Participles Pronouns Prepositions Conjunctions Percentile 43% 3% 37% 7% 7% 1% 1% 1% 74 Roderick Bovingdon and Angelo Dalli Note 1 Including, amongst others, works by Arnold Cassola, Girolamo Caracausi, Adalgisa De Simone, Alberto Varvaro, Stanley Fiorini and Godfrey Wettinger. 2 A prime example is Alexander Borg’s investigation of the imaala phenomenon in Maltese, whose erratic behaviour is still left without a definitive explanation. 3 “..., my hypothesis is that contemporary Maltese, containing a mixture of Arabic and Romance, is directly linked with the Siculo-Arabic and not with North African dialects as has been so far believed.” (Agius, 1996, p.432) 4 There still lingers the remote possibility of linking Maltese to earlier origins than the current Siculo-Arabic claim. In this paper, the term Semitic includes, along with Arabic, the Berber element, owing to the long standing association of the Berber language with Arabic, both in Sicily as well as during much earlier times, along the North African littoral, with the considerable interchange between the two lexicons. Also such term is intentionally applied as an all-inclusive reference to any remote possible language influences from the wider Semitic language group. In similar manner, the terms Romance and Italian, for the scope of this study, include medieval and modern Sicilian and Italian, with all their dialectal and regional variations, as well as French and Andalusian Spanish with its own Arabo-Berber influences included. 5 The population of the Maltese Islands in 1240 consisted of the following family distribution: 836 Saracens (Muslims), 250 Christians and 33 Jewish (Wettinger, 1968, pg. 33). 6 According to the official report by the Apostolic Delegate Dusina in 1574, the Christian population of Malta, including the clergy, was lax in the extreme. Thus, at this relatively late date, one might be inclined to contemplate a religious belief and custom predominated by the numerically stronger Muslim presence. 7 Prior to the Knights, Malta’s affairs were handled from the occupying power's overseas quarters. 8 Mainly Italian, Portuguese, Spanish and French. The German Knights do not appear to have exerted any influence either upon the Maltese language or the culture of the populace (Aquilina, 1976). 9 Prior to British rule, Maltese had not acquired the status of a literary language. The main Romance linguistic influence made its entry mainly Statistical analysis of the source origin of Maltese 75 through the spoken idiom rather than literary texts or formal learning methods. Hence the Romance lexical material entering Maltese was of the most basic type, enabling it to become molded within Maltese Semitic morphology with relative ease. In contrast, during the British rule, the learning of Italian was formally imposed upon the populace through the educational system, as well as the general culture of the local ecclesiastical authorities, the administration, the law courts and the press. 10 Dormant notions of nationhood and nationalism were stimulated, resulting in the first formations of formal and popular political agitation with the resultant linguistic bent towards an Italianate mode of linguistic expression. 11 The early Romance element in Maltese became intrinsically integrated into the basic language structure, while the more modern and erudite forms of Italian, with increasingly less input from Sicilian and Southern Italian, tended to resist full assimilation. 12 Like fissuri (fissures) and jott (yacht). 13 In this case, word duplications due to a word having multiple word classes were not included, hence the sample size of 1,000 words. References: Agius, D. (1990), ‘Il-Miklem Malti: A contribution to Arabic lexical dialectology’, British Society of Middle Eastern Studies Bulletin, 17:2. Agius, D. (1993), ‘Reconstructing the Medieval Arabic of Sicily’, Languages of the Mediterranean, Msida: University of Malta. Agius, D. (1996), ‘Siculo Arabic’, Kegan Paul International, 12. Aquilina, J. (1959), The Structure of Maltese, Valletta: University of Malta. Aquilina, J. (1979), Maltese-Arabic Comparative Grammar. Libya: Socialist People’s Libyan Arab Jamahiriya Press. Aquilina, J. (1988), ‘Criteri di etimologia siculo-maltese’, Malta e Sicilia: Contiguita e Continuita Linguistica e Culturale. Catania: Gruppo Linguistico Catanese. Aquilina, G. (1988), ‘Il lessico agricolo e meteorologico nel maltese e le sue fonti arabe e siciliane’, Journal of Maltese Studies. Malta: University of Malta. 17-18. Bonanno, A. (1988), ‘Contiguita e continuita culturale e linguistica fra Sicilia e Malta in eta prearaba’, Malta e Sicilia: Contiguita e Continuita Linguistica e Culturale. Catania: Gruppo Linguistico Catanese. 76 Roderick Bovingdon and Angelo Dalli Borg, A. (1978), A historical and comparative phonology and morphology of Maltese. M.A. Thesis, Msida: University of Malta. Borg, A. (1994), ‘On some Mediterranean influences on the lexicon of Maltese’, Blaustein Institute for Desert Research and Ben Gurion. Israel: University of Beer Sheeva. Brincat, G. (1994), ‘Gli albori della lingua maltese: il problema del sostrato alla luce delle notizie storiche di al-Himyari sul periodo arabo a Malta (8701054)’, Languages of the Mediterranean. Msida: University of Malta. Brincat, J. (1995), ‘Malta 870-1054: Al Himyari's Account and its Linguistics Inplications’. Msida: University of Malta. Cassola, A. (1992), The Bibliotecha Vallicelliana Regole per la Lingua Maltese. Egypt: Said International. Cassola, A. (1996), Il Mezzo Vocabolario Maltese-Italiano del '700. Egypt: Said International. Cassar, C. (2000), Society, Culture and Identity in Early Modern Malta. Msida: Mireva Publications. Friggieri, O. (1979), Storja tal-Letteratura Maltija. Malta: Klabb Kotba Maltin. Fuccio, G. (1933). ‘Il Conflitto Anglo-Maltese’, Quaderni dell'Istituto Nazionale Fascista di Cultura, Treves-Treccani-Tumminelli, 3:8. Holes, C. (1995), Modern Arabic: Structures, Functions and Varieties. London: Longman Linguistics Library. Hull, G. (1993), The Malta Language Question. Egypt: Said International. Mifsud, M. (1995), Loan Verbs - A Descriptive and Comparative Study. Leiden: Brill. Schweiger, F. (1994), ‘To what extent is Maltese a Semitic Language?’, Languages of the Mediterranean. Malta: University of Malta. Rosner, M., R. Fabri, J. Caruana, M. Lougraïeb, M. Montebello, D. Galea and G. Mangion. (1999), Maltilex Project. Malta: University of Malta. Rosner, M., R. Fabri and J. Caruana. (2000), ‘Maltilex: A Computational Lexicon for Maltese’. Msida: University of Malta. Wettinger, G. (1973), ‘Arabo-Berber Influences in Maltese: Onomastic evidence’, Proceedings of the First Congress on Mediterranean Studies of AraboBerber Influence. Msida: University of Malta. Wettinger, G. (1979), ‘Late Medieval Judaeo-Arabic Poetry in Vatican MS411: Links with Maltese and Sicilian Arabic’, Journal of Maltese Studies. Msida: University of Malta. 13. Wettinger, G. and M. Fsadni. (1968), Peter Caxaro's Cantilena: A Poem in Medieval Maltese. Valletta: University of Malta. Discovering regularities in non-native speech Julie Carson-Berndsen1, Ulrike Gut2 and Robert Kelly1 University College Dublin 1 University of Bielefeld 2 Abstract This paper presents ongoing collaborative research which focuses on the application of computational linguistic techniques to the analysis of a corpus of native and non-native speech. The aim of this research is to use computational tools for modelling phonological acquisition and representation to identify regularities and sub-regularities between different speaker groups. The corpus is being collected and annotated at different levels as part of ongoing research into the acquisition of prosody by non-native speakers at the University of Bielefeld. The computational tools have been designed and implemented at University College Dublin as part of a development environment for modelling, testing and evaluating phonotactic descriptions of lesser-studied languages. 1. Introduction It is a well-known and easily observable fact that non-native speakers sometimes produce syllables in their speech which violate the phonological rules of the foreign or target language. This can have various causes: the simplest is that a speech sound is produced which does not exist in the language being learned. Alternatively, a sequence of sounds may be produced which is not permissible in the target language. This constitutes a violation of the phonotactic rules of the language. Many explanations have been put forward to explain the occurrence of these types of errors, of which the claim that speakers transfer phonological rules of their native language to their productions in the target language is the most popular one. The majority of studies on the acquisition of phonotactic rules are based on small numbers of participants, which reflects the time-consuming nature of a manual analysis of this kind of data. There are two specific aspects which motivate the work presented here. The first, a more computational linguistic motivation, is primarily concerned with the acquisition and evaluation of phonological structure that can be usefully employed in speech technology. The second motivation, a more theoretical linguistic motivation, is concerned with the application of the comparative methodology in the context of the phonological analysis of non-native speech. Each of these motivations is now addressed in turn in sections 2 and 3. Section 4 discusses a specific representation of phonotactic constraints and section 5 presents a set of finite state tools which are used to learn regularities from a phonological corpus. In section 6, a particular corpus containing data from non- 78 Julie Carson-Berndsen, Ulrike Gut and Robert Kelly native speakers of German with two different native languages, Italian and Polish, is used for a study of phonotactic errors found in non-native speech. Section 7 concludes with a discussion of future work. 2. Ubiquitous Acquisition of Phonotactic Constraints Ubiquitous language technology concerns the development of language technologies for different purposes on different platforms so that they can be made available to everybody at all times rather than to a select group for specific purposes. Clearly this is a long-term goal which involves a rethinking of current approaches to speech and language technologies combined with an enhancement based on information of varying levels of granularity paving the way for the development of robust multilingual applications which are easily scalable. One immediate prerequisite to this is the provision of a methodology which can be applied to numerous languages in order to accumulate descriptions which can be reused with varying speech technologies. One particular technology to which this approach is being applied is based on the Time Map Model (Carson-Berndsen, 1998). This model defines the constraints on permissible combinations of sounds in a language in terms of a phonotactic automaton. The advantage of this approach is that each sound in the language is described with respect to the structural context (both preceding and following) in which it can occur (see section 4). The phonotactic automaton thus models all possible syllables of a language and in this way caters for native speaker intuitions about the wellformedness of phonological representations. This paper presents an innovative methodology for automatic analysis of the phonotactics of spoken language based on a suite of finite state tools which have been developed primarily for use in multilingual ubiquitous speech technology. We are primarily concerned with the phonotactic level of description, which is employed in a computational linguistic approach to speech recognition and synthesis. We present an XML-based representation of various types of information defined with respect to the phonotactic context. This representation can be learned automatically from a phonemically labelled data set and can be processed to provide analyses of the phonological regularities which have been found. We demonstrate how this methodology can be applied to the task of phonotactic analysis of non-native speech. 3. Phonotactics in Non-Native Speech The term “phonotactics” refers to language-specific rules and constraints for how sounds can be combined in a syllable. A German syllable (ı) consists of three parts: an onset, a nucleus, and a coda (e.g. Wiese 1996). The nucleus constitutes the centre of all syllables and consists of one or two vowels (V). The onset comprises all prevocalic consonants (C) and can be filled with between zero and Discovering regularities in non-native speech 79 four consonants. All postvocalic consonants form the coda. In German, a sequence of up to five consonants can occur in this position. Figure 1 illustrates this with the word springt’ jumps’. It has the three consonants [6S^] in the onset position, the vowel [,] as the nucleus and the two consonants [1W] in the coda position. Figure 1: The syllable structure of the German word springt. Two kinds of syllable types can be distinguished in German (Carson-Berndsen, 1998): non-reduced and reduced syllables. Non-reduced have between zero and three consonants in the onset position, a short or a long full vowel (or diphthong) and up to four consonants in the coda position. Reduced syllables contain an optional single initial consonant, a weak vowel /l/ and an optional single final consonant. Reduced syllables are never accented. The occurrence and the ordering of consonants and consonant clusters in onset and coda is constrained by the phonotactics. An example of an occurrence constraint in German is for example that the phoneme /1/ is not permitted in the onset position and that voiced stops and fricatives such as /d/ and /v/ are not permitted in the final consonant position of the coda. Ordering constraints apply to the sequence of consonants within either onset or coda. The consonant cluster /lp/ for example is not permissible in the onset, but can occur in the coda position, as in the word Kalb [kalp] (see Kohler, 1995). It has often been observed that learners of a foreign language produce “illegal” consonant clusters in both the onset and the coda position. These differences between native and non-native speakers of a language are systematic and are therefore assumed to form part of the non-native speakers’ interlanguage, i.e. their current representation of the grammar of the target language. Several reasons for systematic phonotactic errors have been proposed: Carlisle (2002) claims that some onset consonant clusters in words are more marked (i.e. occur less frequently in all languages of the world) than others and that foreign language learners always acquire the less marked onsets before the marked ones. Similarly, Eckman (1991) proposed that the reduction of English final consonant 80 Julie Carson-Berndsen, Ulrike Gut and Robert Kelly clusters by native speakers of Cantonese, Japanese and Korean follows universal principles, that is, they are reduced from more marked to less marked forms. These approaches assume that the non-native speakers’ interlanguage is governed by universal principles. Other researchers find influence of native language phonotactic rules in the productions of non-native syllables. Broselow (1984) examined the consonant cluster simplifications in English produced by Arabic native speakers and concluded that they directly reflect the phonotactic rules of their native languages. The underlying assumption is that the interlanguage of non-native speakers contains phonological rules of their native language. Due to the time-consuming character of phonetic analyses, many of these studies are based on a small number of speakers only. However, in order to test hypotheses about the nature of interlanguage structure and the cause of systematic errors, large-scale studies are necessary (cf. Carlisle, 2002). The object of this paper is to present tools for automatic analysis of non-native speakers’ production of syllable types which allows processing of large data resources and suggests a model for interlanguage representations. 4. Representing Phonotactic Constraints The underlying assumption in this paper is that phonotactic constraints for any language can be represented in terms of finite state automata which define acceptable sequences of sounds within a syllable or word domain. A subsection of a phonotactic automaton describing CC-clusters in English syllable onsets is depicted in Figure 2. The complete phonotactic automaton for English syllables allows for a distinction to be made between well-formed and illegal structures, i.e. although the word blick does not exist in English, the sequence of sounds is permissible and would be recognised as well-formed by a native speaker of the language, whereas the form bnick will always be rejected by native speakers of English as a possible word. A phonotactic automaton is language-specific and can be developed for each individual language. A multilingual time map (Carson-Berndsen, 2002) extends the functionality of a phonotactic automaton by combining language-specific information at various levels of granularity represented as a multilevel finite state transducer. The advantage of this representation, as with the phonotactic finite state automaton, is that it is declarative, bi-directional, efficient to process, and can be easily learned. The multilevel finite state transducer can be viewed as an extension of the phonotactic automaton to include (at least) the following levels: graphemes, phonemes, allophones, canonical form, features, constraints on overlap relations, average duration, frequency and probability. Each arc specifies information on all of these levels (although some of this information may not be available in all cases; however, the transducer can be readily updated at any time). This representation serves as the basis for the learning, analysis and comparison tools Discovering regularities in non-native speech 81 which are presented in the next section. For the specific case study presented below, only the phoneme and canonical form levels are relevant to the discussion. Figure 2: Subsection of phonotactic automaton depicting CC- onset in English. A multilingual time map is represented in XML with a visualisation in terms of a directed graph. A subsection of a German multilingual time map depicting two consonant clusters in onset position is visualised in Figure 3. The full information is shown for only one arc. Figure 3: CC- onset in German. 82 5. Julie Carson-Berndsen, Ulrike Gut and Robert Kelly Finite State Tools This section presents a suite of finite state tools which have been developed for use in computational phonology applications motivated by the developments in linear and nonlinear finite state phonology (e.g. Kaplan & Kay, 1994; Bird & Ellison, 1994). The suite of tools is centred around two main programs GTI and PAL, which are described firstly in terms of their general functionality. At the end of this section the specific ways in which these programs underpin the finite state tools for phonotactic analysis are summarised. 5.1 The Generic Transducer Interpreter The Generic Transducer Interpreter, GTI, is a program designed to read the structure of a finite state transducer (FST) and using this structure test a set of sequences of tokens for acceptance by that FST. FST structures are stored in XML marked up form. A FST consists of a set of states and state transitions. A single start state is defined and a subset of the states is designated as final states. The state transitions of a FST may have any number of tapes defining different transductions between tapes. A sequence of tokens is accepted by a FST if a path can be traced from the start state of the FST to a final state of the FST for that sequence. Starting with the initial state as the current state a path is traced by examining each successive token in the sequence to determine if a state transition can be applied from the current state. A state transition can be applied at a given state for a given token if there is a state transition in the state transition set having a source state matching the current state and transition symbols matching the current token. Test sequences are currently stored in text files in which sequences must be specified in a particular format. For GTI to test a set of sequences it must have specific tapes of state transitions in the FST defined as input tapes. Any number of input tapes can be specified (up to the number of tapes in the state transitions). Input tapes are those tapes of state transitions used to match the tokens of sequences against in order to determine if a state transition can be applied. Thus, each token of a test sequence is an item of input for the FST. If L tapes of state transitions have been specified as input tapes then each token in a test sequence must have L tapes also. If GTI detects that there are an incompatible number of tapes in any token of any sequence in the test file then an error is flagged. It is also possible to specify which tapes of state transitions are to be treated as output tapes. Output tapes are those tapes for which the result of a transduction is required. When a sequence is accepted on the input tapes of a multi-tape FST there will be associated outputs, namely the concatenation of the sequences of symbols present on non-input tapes. It may be necessary to examine the output produced by any one of these output tapes. Again, any number of output tapes can be specified (up to the number of tapes in the state transitions). Discovering regularities in non-native speech 83 The names of both the FST file and the test sequence file together with input and output tapes are passed to GTI via the command line. If the FST file is in the correct format without error then GTI will create the FST as specified. As GTI tests each sequence, it reports whether the sequence was accepted or not when applied to the FST on the specified input tapes. Also, if output tapes are specified then the outputs are displayed for each output tape. Note that if the FST is nondeterministic that more than one acceptance path may be traced. In this case, the outputs for all successful acceptance paths are reported. It is also possible to output the actual trace(s) of state transitions for each accepted sequence. 5.2 The Phonotactic Automaton Learner The Phonotactic Automaton Learner, PAL, is a program that takes a set of training sequences of symbols and learns the structure of a stochastic finite state transducer (SFST) that accepts the training sequences. The training sequences to be used by PAL are stored in a text file in which sequences must be specified in a particular format. The training sequences may have multiple tapes and in this case all tokens of all sequences must have the same number of tapes specified or an error will be flagged. Once learned, the SFST representation is as an XML marked-up form. The PAL program uses the ALGERIA machine induction algorithm (Carrasco & Oncina, 1999) to learn the structure of the required SFST. The algorithm works in two stages. First, a prefix-tree SFST is built from the specified training sequences. A prefix-tree SFST has a single start state with a single path from the start state to a final state for each of the training sequences. For each token in each training sequence there is a state transition between states along the unique path for that sequence. Also, if two training sequences share a prefix then there is a single path in the prefix-tree SFST for that prefix. The path diverges into distinct state transitions for the distinct postfixes after the final state transition in the common prefix. Each state transition in the prefix-tree SFST has a frequency of traversal dependent on the number of training sequences that share that state transition. The frequencies of state transitions are then used to identify the canonical SFST based on state merging. The state merging process is the second stage of the ALGERIA algorithm. Two states are merged into a single canonical state if the language generated from that point on is statistically identical (i.e. for each continuation path emanating from the first state there is a matching continuation path emanating from the second, both of which have a statistically identical frequency). If two states are merged then all state transitions in the prefix-tree SFST that refer to either of the two merged states, now refer to the newly created merged state. By the process of comparing each state with each other state in the prefix-tree SFST the canonical SFST is identified. The ALGERIA algorithm has an associated confidence level. This confidence determines how rigorous the learning of SFSTs is. The lower the confidence the less likely it will be that the learned SFST is absolutely correct, that is, the less 84 Julie Carson-Berndsen, Ulrike Gut and Robert Kelly likely it will be that the learned SFST will accept only the training sequences. There is a default confidence which has been found by experimentation to be sufficient for effective learning, however, an alternative confidence can be specified. PAL accepts the name of a training sequence file and a destination for writing the learned SFST to as well as an optional confidence through a command line interface. In summary, GTI and PAL are used specifically in the context of the tools for phonotactic analysis of non-native speech as follows: The Learning Tool uses PAL to learn a deterministic finite state automaton from a phonemically labelled data set. This automaton defines the phoneme level of the multilingual time map and can be output as XML. A canonicalisation step allows each phoneme on each arc in the multilingual time map to be supplemented by its canonical form in terms of V or C. The final output is thus a two-level transducer. The Analysis Tool uses GTI to transduce from one level to another (e.g. phoneme to canonical form or vice versa). The Comparison Tool uses GTI to partition the data into accepted and rejected forms. An analysis summary shows how many forms were accepted or rejected. 6. The study of phonotactic errors in non-native speech This section is concerned with the application of the computational linguistic tools presented in section 5 to the corpus-based analysis of the phonotactics of non-native speech. The corpus used in this study is being collected and annotated at different levels as part of ongoing research into the acquisition of prosody by non-native speakers at the University of Bielefeld (see Milde & Gut, 2002). 6.1 Participants The data are taken from the LeaP corpus 3 , which consists of prosodically annotated non-native speech of more than 70 speakers. For the study, three data sets were used consisting of speech produced by German, Italian and Polish native speakers. The German natives are all speakers of Hochdeutsch (Standard High German). The Italian native speakers were between 21 and 31 years old when they were recorded and had been living in Germany for between two months and five years. They all studied German at school or at University level for up to six years prior to their arrival in Germany. The Polish speakers were between 22 and 29 years old at the time of recording and had been living in Germany for a period of between eight months and six years. They all have a University degree in German from their home country, where they studied German for between four and six years. All speakers were intermediate to advanced speakers of German and had no active knowledge of German phonotactic rules. Discovering regularities in non-native speech 6.2 85 The Data Recordings consisted of three parts. First, a short interview (approximately ten minutes) was conducted with the non-native German speakers, in which various questions about their language learning history such as age at first contact with German and length of formal instruction were asked. Second, the participants were asked to read out a short story. Third, they re-told the story in their own words and without reference to the written text. All recordings were carried out in Bielefeld in either a sound-treated or a quiet room. Only the readings were analysed for this study. The first set, German_Read, consists of three readings of a story by German native speakers. The second set, Italian_Read, consists of five readings (total of 624 syllables) of the same story by Italian speakers of German. The third set, Polish_Read, consists of five readings (total of 618 syllables) of the same story by Polish speakers of German. The acoustic analysis of the data was carried out using ESPS/waves+ and Praat and was done by one trained phonetician and four students with extensive training and experience in phonetic analysis. The data was annotated at a number of linguistic levels such as the level of the intonational phrase, the word, the syllable, and the skeletal (CV) structure, and comprises annotations of the prosodic structures of intonation and pitch range. The syllable level was then selected from the prosodic annotation as the relevant input for the finite state tools. All syllables were transcribed phonetically in SAMPA. Transcription was fairly broad but included processes such as nasalisation, unreleased stops and aspiration. In case of ambisyllabic consonants, half of the ambisyllabic consonants was considered to belong to the preceding syllable and half to the subsequent one. 6.3 Corpus-based investigation of non-native speech errors The finite state tools described in section 5 were applied to the data sets described above in order to identify regularities in the phonotactic errors produced by Italian and Polish speakers of German. In each case a specific phonotactic automaton for each data set was constructed using the learning tool, PAL. In order to be able to filter the phonotactic errors, it was necessary to identify which forms produced by the Italian and Polish speakers adhere to the phonotactics of German. For this a complete manually constructed phonotactic automaton for German (from Carson-Berndsen, 1998) was used; this automaton is referred to as Phono_German_Comp.fsm below. The procedure involved the application of the learning tool, the analysis tool and the comparison tool as follows. The Learning Tool was applied to the German_Read data set to generate a finite state representation of the phonotactic forms used in the read speech of the German native speakers. The output of this was termed Phono_German_Read.fsm. The phonemic forms contained in the 86 Julie Carson-Berndsen, Ulrike Gut and Robert Kelly Italian_Read data set were then compared against the Phono_German_Read.fsm using the Comparison Tool and the rejected forms were then compared against the Phono_German_Comp.fsm. This ensured that all forms which corresponded to the phonotactics of native speakers of German (the target language) were filtered and that any remaining forms in the data set corresponded to phonotactically ill-formed structures which had been produced by Italian native speakers of German. The Learning Tool was once again applied to these forms to produce an automaton representation of the ill-formed structures, termed Italian_Rejected.fsm. The Analysis Tool was applied to this data in order to map from the phonemic representations to the canonical CV form. This resulted in the data set, Italian_Read_Canonical, which characterised the canonical forms of all the phonotactically ill-formed structures in the Italian_Read set. Finally, the Comparison Tool was reapplied to associate all phonemic realisations with a canonical form so that particular phonotactic anomalies and regularities with respect to the target language could be identified. The flowchart for the application of the tools in this task is depicted in Figure 4 with respect to the Italian data. Figure 4: The application of the finite state tools to the Italian data. An analogous procedure was followed for the analysis of the Polish_Read corpus. The results are presented in the next section. One point to be noted here is Discovering regularities in non-native speech 87 that whilst the learning tool does, of course, generalise to some extent over these forms, there is no notion of complete coverage as defined in the data set, although currently we are investigating projection techniques to cater for gaps caused by the lack of data which would allow for a distinction to be made between idiosyncratic gaps and systematic gaps. This can be based on the notion of natural classes of sounds which function similarly in a particular phonotactic context. 6.4 Results Table 1 lists the percentage and absolute number of syllables in Italian_Read and Polish_Read which were rejected after comparing them first to Phono_German_Read and then to Phono_German_Comp. Only syllable types were considered, that is, if a rejected syllable occurred more than once it was only counted as one rejection. Thus, after two rounds of application, a total of 54% of the syllables produced by the Italian non-native speakers and 48.7% of the syllables produced by the Polish non-native speakers were rejected. Table 1: % of rejected syllables. % of syllable tokens rejected by Phono_German_Read % of syllable tokens rejected by Phono_German_Complete Italian Polish 67% (418 of 624) 59.5% (368 of 618) 80.6% 81.7% The results of the application of the Learning Tool to the remaining syllables were then classified manually into a) phonemic inventory errors (phonemes that do not occur in German were produced) b) onset consonant cluster errors c) coda consonant cluster errors Table 2 summarizes the results for the Italian and Polish speakers. Distinct differences between the phonotactic violations produced by the Italian and the Polish speakers can be seen. On the whole, the Polish speakers produce fewer errors; different types of illegal initial consonant clusters occur in the two speaker groups, and the postvocalic r (which is produced as an a-schwa in German) constitutes the major problematic area for the Italian speakers. Julie Carson-Berndsen, Ulrike Gut and Robert Kelly 88 Table 2: Error analysis of Italian and Polish speakers of German. Italian Polish 7. Phoneme inventory Errors 6 consonants, 5 monophthongs, 16 diphthongs Onset consonant cluster errors Coda consonant cluster errors 10 types, predominantly [nd], [mb] 2 types, 14 final voiced consonants, 27 postvocalic r 4 consonants, 3 monophthongs, 24 diphthongs 6 types, 2 initial consonants [x, N] 1 type, 11 final voiced consonants, 1 postvocalic r Conclusion In this paper we demonstrated how a suite of finite state tools can be applied to study the phonotactics of different varieties of spoken language. The results imply that research in second language acquisition can potentially benefit enormously from this methodology as it allows a rapid analysis of large corpora. While the corpus described in this study was relatively small, to perform a manual analysis of this data would have been a laborious task. Future work involves employing the tools to analyse phonotactic differences among other non-native speech groupings using the LeaP corpus. Currently this corpus consists of 253 annotated recordings of between 2 and 30 minutes’ duration by 88 different speakers with 21 different native languages. Since the corpus is annotated at a number of linguistic levels as described above, inter-level analyses are possible. For the application of the finite state tools, each level of annotation is viewed as a tier, analogous to the representations of autosegmental phonology (Goldsmith, 1990). Analysis can take place either with respect to individual tiers or with respect to an associated set of tiers. In the latter case, one tier is chosen as the primary tier and the others are associated with it in terms of overlap and precedence relations between the units as suggested in Carson-Berndsen (1998: 60). Using the computational linguistic tools, finite state automaton and finite state transducer representations of the tiers are extracted automatically from the annotated corpus. Regularities in the data are then identified either with respect to a single tier or with respect to an associated set of tiers. In addition to identifying phonotactic errors with respect to one non-native speakers of one particular language, we are also currently applying this methodology to the investigation of the phonotactics of typologically different languages, in particular spoken Yoruba and Igbo. Note 1 This work was part funded by an Enterprise Ireland Resreach Innovation Fund grant (no. IF/2001/021) and an Enterprise Ireland International Collabortaion grant (no. IC/2002/053). Discovering regularities in non-native speech 89 2 The LeaP project is funded by the MSWF (Ministry for Education of North-Rhine Westphalia, Germany). 3 http://www.spectrum.uni-bielefeld.de/LeaP/ References Bird, S. & T. M. Ellison (1994), One - level phonology: autosegmental representations and rules as finite state automata, Computational Linguistics 20: 55-90. Broselow, E. (1984), An investigation of transfer in second language phonology. International Review of Applied Linguistics 22 (4), 253-269. Carlisle, (2002), The Acquisition of Two and Three Member Onsets: Time III of a Longitudinal Study. Proceedings of New Sounds 2000, pp. 42-47. Carrasco, R. C. & Oncina, J. (1999), Learning deterministic regular grammars from stochastic samples in polynomial time, ITA, Vol.33, No.1, 1-19. Carson-Berndsen, J. (2002), Multilingual time maps: portable phonotactic models for speech technology. In Proceedings of the LREC 2002 workshop on Portability Issues in Human Language Technology. Carson-Berndsen, J. (1998), Time map phonology. Dordrecht: Kluwer. Eckman, F. (1991), The structural conformity hypothesis and the acquisition of consonant clusters in the interlanguage of ESL learners. Studies in Second Language Acquisition 13, 23-42. Goldsmith, J. (1990), Autosegmental and Metrical Phonology, Cambridge, Mass: Basil Blackwell. Kaplan, R. & M. Kay (1994), Regular models of phonological rule systems, Computational Linguistics 20:331-378. Kohler, K. (1995), Phonetik des Deutschen. Berlin: Erich Schmidt. Milde, J.-T. & Gut, U. (2002), A prosodic corpus of non-native speech. In: B. Bel & I. Marlien (eds.) Proceedings of the Speech Prosody 2002 conference, 11-13 April 2002. Aix-en-Provence: Laboratoire Parole et Langage, pp. 503-506. Wiese, R. (1996), German phonology. Clarendon: Oxford. This page intentionally left blank Tracking lexical changes in the reference corpus of Slovene texts Vojko Gorjanc University of Ljubljana Abstract The text focuses on lexical borrowings from English, introduced into Slovene in the last decade of the 20th century. Using the FIDA corpus, a reference corpus of Slovene, new lexical items can be tracked over the last decade: by means of corpus analysis, we can determine when a word entered in the Slovene language and how it established itself in the language. Corpus analysis reveals great creativity of Slovene language speakers; in addition to loan words, original Slovene coinages occur almost invariably. Corpus data shows a great deal of variability originating from the desire of the speakers of Slovene to coin new expressions, but after a few years, the data begin to reveal the prevailing variant. 1. Introduction With the help of a corpus, we can track lexical changes quickly and reliably, and also observe the response of a selected language to new lexical items introduced into it from other languages, e.g. English, or some other language with which the selected language has direct contact, in the case of Slovene, these are Italian, German, Hungarian or Croatian. This paper focuses on lexical items introduced into Slovene in the last decade of the 20th century. The starting point for the comparison with the state in the corpus of the Slovene language is John Ayto’s list of English lexical items from the 1980s and 1990s (Ayto 1999); the subsequent corpus analyses focus on lexical items from the fields of the Internet and computer science. With corpus investigations, we will try to determine how the Slovene language reacts to lexical items introduced into Slovene from English. Using the corpus, we can track the new lexical items through the last decade, and observe their characteristics in the corpus. 2. The Corpus The Corpus of the Slovene Language, called FIDA, is a reference corpus of Slovene compiled by a consortium of four project partners: University of Ljubljana (Faculty of Arts), Jozef Stefan Institute Ljubljana, DZS Publishing House and Amebis software company (http://www.fida.net). The corpus is composed of contemporary Slovene texts, the majority of which were published Vojko Gorjanc 92 in the 1990s. The corpus contains just over 100 million words, encompassing a broad variety of language variants and registers. It is composed of written texts and texts originally produced as written-to-be-spoken. The transcripts of Slovene parliamentary proceedings are the only spoken component of the corpus. The basic corpus characteristics according to taxonomic corpus parameters are (in %): Medium spoken electronic written Text type 1.97 0.03 98.00 literary technical other Linguistic proofreading 5.94 18.46 75.60 yes no unknown 63.92 3.13 32.95 The FIDA corpus is lemmatised and morpho-syntactically tagged, but all the tagging was done automatically without the possibility of disambiguation in cases where double or even multiple lemmas were possible. Since Slovene is a morphologically complex language, double or triple lemmas are frequent, which makes statistical data from the corpus unreliable to some extent. In the last few years, some significant steps were taken to solve the problem, both by testing the existing language-independent tools and by developing new ones (Džeroski and Erjavec, 2000; Mladeniü, 2002) but the situation is still far from ideal. Although less acute, the lemmatisation of non-lemmatised words is another problem waiting for the improvement of text processing tools for Slovene. The lemmatisation of the FIDA corpus was based on the lexicon developed by the software company involved in the project. Experiences show that in certain cases non-lemmatised words skew the results of statistical analysis, so all these have to be taken into account when interpreting the corpus data (Gorjanc and Krek, 2001). 3. Methodology Using a wordlist from the Slovene corpus, we will obtain information on the lexical items from Ayto's list relevant for the Slovene language. By means of corpus analysis, we will determine when a word occurs in the Slovene language and how it establishes itself in the language, and by means of statistical analysis, we will determine the possible collocations of the selected word and their changes from the first occurrence of the word until the end of the decade. Since pairs of synonyms or strings often occur with new expressions, we will try to determine how they occur in the corpus and how they disappear. With the help of markers of semantic relations already identified for the Slovene language by corpus analysis, we will identify pairs of synonyms and strings within the corpus. We will focus on the distribution of synonyms within the corpus regarding the time of their Tracking lexical changes in the reference corpus of Slovene texts 93 occurrence, and consider when and why one of the synonyms becomes dominant in the language while the other variants disappear. 3.1 Extracting collocations For extracting collocations, the MI3 value introduced by McEnery et al. with information on the probability of a word pair occurring together or separately will be used (McEnery, Langé, Oakes and Véronis 1997). The MI3 value has turned out to be sound information for content words in Slovene. On the other hand, with this value it is hard to determine function words as part of collocations; in particular, in the case of collocators of verbs and nouns with prepositional phrases. For instance, to detect propositional words, raw statistics provide more valuable information. After detection, a noun + propositional word pair, for example, the MI3 value for the whole pair is calculated to extract the string of collocators. The comparison of results between T, MI and MI3 values for the Slovene corpus shows that MI3 is the most effective of the three. MI introduced in Church and Hanks (1989) is far less effective, since the frequency of corpus elements is underestimated and a single co-occurrence of two elements in the corpus gives high scores which can diminish the importance of more frequent lexical units (Manning and Schütze, 1999). This fact is even more relevant in the case of the FIDA corpus, since specific forms of non-lemmatised words are attributed high MI scores. To some extent, MI3 neutralises the effects of low frequency of a corpus element which is why it gives better results for collocations. The noun raþunalnik (Engl. computer) with collocators according to MI and MI3 values in the FIDA corpus (frame 3) MI3 raþunalnik oseben (Engl. personal) (Engl. computer) prenosen (Engl. portable) na (Engl. on) vaš (Engl. your) biti (Engl. to be) z (Engl. with) in (Engl. and) v (Engl. in) poznavanje (Engl. knowing) za (Engl. for) delo (Engl. work) moj (Engl. my) uporabljati (Engl. to use) zmogljiv (Engl. high capacity) zagnati (Engl. to start) MI =uporablja =appleov =deskpro =skreširani =pomeþite =nnnn =gxi =upsajo =85prenosni =megatronski =macintosch =optiplex =sprojektiram =blagajniþarska =dlanþni 94 Vojko Gorjanc internet (Engl. Internet) ki (Engl. which) uporaba (Engl. useage) =windows* delati (Engl. to work) * = non-lemmatised =brskalnemu =vseuporabljajo =pc486 izklopitev =ignororajo By using MI3 values, we obtain relevant results for collocations, while the highest MI values are generally non-lemmatised words, mostly spelling errors. From the point of view of non-lemmatised words the results are interesting, since they reveal some of the problems the speakers of Slovene have with forming new words, e.g. the adjective derived from the name Apple; in Slovene there is an attempt to use the suffix -ov, appleov. 3.2 Identifying markers of semantic relations Semantically related words – synonyms, hyper- and hyponyms, abbreviations often collocate or appear in similar contexts and it is usually possible to identify the domain of a word on the basis of its textual environment. When we want to explain the relations between concepts within the reality portrayed, we often use explicit linguistic structures or phrases, such as X is defined as Y, X is an instance of Y, There are several types of X, for example A, B, C etc. So, if we, for example, identify the pattern “A, also known as B” to indicate complete or near synonymy, we can extract the noun phrases linked by the pattern as pairs of synonyms. Similar methods are proposed by several authors in the field of terminology, either for the purpose of an automatic construction of knowledge databases (Bowden et al., 1996), conceptual sampling for terminography (Meyer et al., 1999) or simply a search for synonyms in a corpus (Pearson, 1998). For Slovene the most frequent markers were identified by analysing the FIDA corpus. For synonymy, these markers are ali (Engl. or), ali tudi (Engl. also), imenujemo tudi (Engl. we also refer to it as), imenovan* tudi, (sinonim _) (Engl. also referred to as), je sinonim za (Engl. is a synonym for), znan* tudi kot (Engl. also known as), znan* tudi pod imenom (Engl. also familiar as), z drugim imenom (Engl. also called), ... (Vintar and Gorjanc 2000). For the purpose of this paper, these findings were used to extract pairs or strings of synonyms of selected words. Internet ali medmrežje, kot so ga izvirno poslovenili /.../, že dolgo ni veþ neznanka povpreþnemu Slovencu. Pair of synonyms: internet medmrežje Tracking lexical changes in the reference corpus of Slovene texts 4. 95 Corpus analysis In this analysis, we will focus on some key lexical elements, from the field of the Internet, which have undergone the process of being accepted into the Slovene language. Keeping the loan words unchanged, the most passive response of recipient-language users is confirmed as temporary and it leaves open the possibility for coining new terms; it turns out that if the new terms are formed in a way which is acceptable to users, no problems arise in introducing the Slovene variant. Let us illustrate this, using the example of the term World Wide Web, and its Slovene variant, svetovni splet (the graph below is in %). 100 80 60 40 20 0 1994 1995 1996 world wide web 1997 1998 1999 svetovni splet Figure 1: Occurrences of World Wide Web, and its Slovene variant, svetovni splet in the FIDA corpus between 1994 and 1999 In the two years after its first appearance, only the loan word occurs in the corpus, but when the Slovene variant appears, it immediately becomes a successful rival and the use of the loan word gradually decreases. In texts, the dominance of the Slovene synonym over the loan word is even more obvious in the case of another key word from the field of the Internet, i.e. home page. After eliminating corpus noise related to proper names of pages, it turns out that the Slovene term has dominated completely (91.8% of corpus occurrences). In addition to the calque domaþa stran (Engl. home page), there is also a new term predstavitvena stran (Engl. web presentation page) coined in Slovene, but it seems that the calque from English is more acceptable. The fate of the following term from the field of IT is quite different. The English term screen saver entered the Slovene language in the 1990s. Vojko Gorjanc 96 100 80 60 40 20 0 1 2 3 4 1 screen saver; 2 varþevalnik zaslona; 3 ohranjevalnik zaslona; 4 ohranjevalec zaslona Figure 2: Occurrences of screen saver, and its Slovene variants, varþevalnik zaslona, ohranjevalec zaslona and ohranjevalnik zaslona in the FIDA corpus After the loan word, the calque varþevalnik zaslona occurs next, but a later Slovene term formed by using attribute ohrajeva- (Engl. keep) turns out to be more acceptable. At first, there are two variants, but later the adjective with the suffix -ik dominates. While it is true that in Slovene a new word varþevalnik (Engl. saver) is derived from the verb varþevati (Engl. save), it seems that the semantic link is not strong enough for the speakers. In the corpus, the verb varþevati (Engl. save) tends to collocate with words such as banka (Engl. bank), denar (Engl. money); zaþeti (Engl. start), splaþati (Engl. worth). It thus covers a semantic field which the speakers do not associate with new terminology from the field of IT. The term Internet itself has become fully integrated into the Slovene language; this is partly due to its everyday use. As a noun, it occurs as a premodifier in noun phrases: e.g., internet storitev (Engl. Internet service), internet naslov (Engl. URL), internet povezava (Engl. Internet connection), internet ponudnik (Engl. Internet service provider), internet stran (Engl. Web page), internet raþun (Engl. Internet account), internet protokol (Engl. Internet protocol). In the Slovene language, the new type of noun phrase with a noun functioning as a premodifier is becoming increasingly common. This type of noun phrase is formed under the Tracking lexical changes in the reference corpus of Slovene texts 97 influence of the English language, in Slovene the premodifier in a noun phrase had to be an adjective. The noun as the premodifier appears frequently, even though there is also the possibility of forming the noun phrase with an adjective as the premodifier. The noun Internet happens to be extremely prolific in terms of word formation, since it forms adjectives using the suffixes -ni, -ski and -ov: internetni, internetski, internetovski, internetov in Adjective+Noun combinations; the adverb internetsko using the suffix -o in Adverb+Verb combinations, and newly derived nouns using the suffixes -ar and -ec: internetar; internetovec, as well as new compound nouns, e.g. internetdžanki (Engl. Internet junkie). In adjectives, the relatively high variability indicates that the newly formed words have not yet been fully accepted. The collocations of an individual adjective show that the collocators of the adjectives internetni, internetski and internetovski overlap /service, page, search engine, business, shop, bookseller, service provider.../, so that it is impossible to determine the specific dependent links between words. Therefore, it seems that the use is very optional and different variants of the adjective are possible with the same headword. In the case of the adjective internetov, which is the least common of the adjectives listed above, the link to the headword is completely dispersed; this indicates that the suffix variant -ov is not integrated and consequently inappropriate for the classifying character of the adjective, generally expressed by the adjectival suffixes -ni and -ski in Slovene. The frequent use of the classifying adjective with the suffix -ni shows a prevalence of this variant, and its only real rival is the classifying adjective with the suffix -ski. 60 50 40 30 20 10 0 1 2 3 4 1 internetski; 2 internetni; 3 internetov; 4 internetovski Figure 3: Occurrences of adjectives Internet using the Slovene suffixes -ni, -ski and -ov in the FIDA corpus On the other hand, Slovene term medmrežje, introduced when the term internet was already fully accepted, has very low corpus frequency (corpus occurrences 98 Vojko Gorjanc medmrežje : internet = 2.2% : 97.8%) and in terms of word formation it is not productive. Despite the fact that the speakers of Slovene did not accept the Slovene term medmrežje, there are still constant but unsuccessful attempts to force medmrežje instead of internet by normativist linguists. In assigning terms to concepts, the Internet has stimulated the formation of two strings of newly formed terms, which seem to be growing in prolificacy with the development of the Internet, i.e., terms of the type e-mail and terms of the type kiber- and cyber-. Among the latter, a great variability in spelling can be observed in Slovene: the new terms can be spelled as two words kiber prostor (Engl. cyberspace), cyber kavarna (Engl. cyber-café), or as one word, with or without a hyphen, e.g., kiber-kiþ (Engl. cyber-kitsch), cyber-kultura (Engl. cyber-culture) and kibersvet (Engl. cyberworld), cyberfolk (Engl. cyberfolk). The spelling of these terms is an important point of dispute among Slovene grammarians at the moment, with the question of the influence of the English pattern being particularly prominent. New terms of the type e-Noun, e-Adjective have already been presented and show a very open series (Jakopin, 2001); here, let us consider the trends in Slovene with kiber- and cyber-. At first glance, the results of the corpus are extremely dispersed; the process is very productive, and both the loan element as well as the Slovene element are prolific. With the loan element, the tendency to write the new terms as two separate words can be observed; the one-word spelling is reserved for a limited number of complete loan words, e.g. cyberspace, cybersex, cybercash. However, as there are so many new terms coined and cyber- generally occurs with a Slovene second element, the new term is generally spelled as two separate words, e.g., cyber otroški vrtec (Engl. cyber-kindergarten), cyber jasli (Engl. cybercréche), cyber kavarna (Engl. cyber-café), cyber umetnost (Engl. cyber-art). Consequently, the two-word spelling is used with loan words as well, e.g. cyber café, cyber space. The pattern is quite open and this can be seen in the possibility of the hyphenated spelling for both complete loan words as well as for the combinations of a loan element and a Slovene element, e.g. cyber-space, cyberpunk; cyber-gostilniþar (Engl. cyber-innkeeper), cyber-klobasica (Engl. cybersausage). The hyphenated spelling is particularly common with adjectives, e.g. cyber-totalitarni (Engl. cyber-totalitarian), cyber-kavbojski (Engl. cybercowboy). On the other hand, there is a tendency for one-word spelling with the Slovene version kiber, the spelling change thus only enables the formation process in Slovene. The one-word spelling is also a consequence of the fact that the adjectives kibernetiþni in kibernetski (Engl. cybernetic) assume the attributive function. The greater assimilation of kiber into the Slovene language is confirmed by the terms coined for the male and female representatives of the cyberworld: kibernetnica, kibernetniþar and kibernetniþarka, and it also occurs in the root of the verb kiberseksati (Engl. to have cybersex). Tracking lexical changes in the reference corpus of Slovene texts 99 In this case, the almost unbelievable range of terms reveals a dynamic process of assigning terms to concepts in Slovene, while the corpus indicates that the new terms with kiber- prevail (over 60%). The use also shows that the lexical elements kiber and cyber are extremely popular. In Slovene, it seems that the speakers accept the lexical term as semantically emptied, as something very fashionable at the moment. Thus, as we have seen, the collocations may be rather unusual. 5. Conclusion In its observation of the process of accepting new lexical items into Slovene, the corpus analysis reveals the great creativity of Slovene language speakers; in addition to loan words, original Slovene expressions occur almost invariably. Corpus data shows a great deal of variability, linked above all to the desire for original expressions. However, after a few years, the data begin to reveal the prevailing variant. The question of which variant is eventually fully accepted, though, remains open. Full acceptance of a variant is conditioned by a series of linguistic and even more non-linguistic factors. It is very important that the new terms emerge spontaneously in the language community and are not introduced into the language by linguistic intervention. If the speakers of Slovene even begin to suspect that the term is an attempt at linguistic intervention, they will probably reject it. In order for a variant to be fully accepted in Slovene today, it should have an appropriate semantic basis, so that the speakers of Slovene can identify it as “sexy” and “cool”. References Ayto, J. (1999), 20th Century Words. Oxford, Oxford University Press. Bowden P., P. Halstead and T Rose (1996), Extracting Conceptual Knowledge from Text Using Explicit Relation Markers. In Proceedings of EKAW-96, Nottingham, pp. 147-162. Church K. and P. Hanks (1989), Word association norms, mutual information and lexicography, in: Proceedings of the 27th Annual conference of the Association of Computational Linguistics, pp. 76-82. Džeroski S. and T. Erjavec (2000), Learning to lemmatise Slovene words, in: J. Cussens and S. Džeroski (eds), Learning Language in Logic. Berlin, Springer, pp. 69-88. Gorjanc V. and S. Krek (2001), A corpus-based dictionary database as the source for compiling Slovene-X dictionaries, in Proceedings of the COMPLEX 2001 6th Conference on Computational Lexicography and Corpus Research, Birmingham, pp. 41-47. Jakopin P (2001) Words and nonwords as basic units of a newspaper text corpus, in: Proceedings of the COMPLEX 2001 6th Conference on Computational Lexicography and Corpus Research, Birmingham, pp. 49-65. 100 Vojko Gorjanc Manning C., and H. Schütze (1999) Foundations of Statistical Natural Language Processing. Cambridge MA: The MIT Press. McEnery T., J. Langé, Oakes, M. and J. Véronis (1997), The exploration of multilingual annotated corpora for term extraction, in R. Garside, G. Leech, A. McEnery (eds.), Corpus Annotation. Linguistic Information from Computer Text Corpora. London, Longman. Meyer, I., Mackintosh K., Barriere, C. and T. Morgan (1999), Conceptual sampling for terminological corpus analysis, in Sandrini (ed.), Proceedings of TKE ’99. Vienna, TermNet, pp. 256-267. Mladeniü D. (2002), Automatic word lemmatisation, in: T. Erjavec and J. Gros (eds.), Jezikovne tehnologije, Language Technologies. Ljubljana, Institut Jozef Stefan, pp. 153-159. Pearson J. (1998), Terms in Context. Amsterdam, John Benjamins. Vintar Š, and V. Gorjanc (2000) Identifying markers of semantic relations in Slovene. http://www2.arnes.si/vinta/telri.rtf Relating linguistic units to socio-contextual information in a spontaneous speech corpus of Spanish José María Guirao Universidad de Granada Antonio Moreno Sandoval, Ana González Ledesma, Guillermo de la Madrid, Manuel Alcántara Universidad Autónoma de Madrid Abstract This chapter shows the application of statistical tests to a corpus of spontaneous spoken Spanish. Our goal is to find representative differences between different parts of the corpus. To this end, we tagged n-grams in the corpus with features related to the speaker (age, gender, etc.), or the context (dialogue, monologue, media, etc.), and applied the loglikelihood test (Dunning, 1993) in order to find the most distinctive lexical or grammatical items for each specific socio-contextual feature. This chapter is divided in three sections. In the first, the characteristics of the spoken corpus are shown. The second section is devoted to the explanation of the computational tool. In the third section, a first rough estimate of the results obtained is given, as well as possible applications of the model. 1. The Spanish corpus of the C-ORAL-ROM project. C-ORAL-ROM is a multi-lingual corpus of spontaneous speech for the main four Romance languages, French, Italian, Portuguese and Spanish (Cresti et al. 2002). The project is funded by the EU under the V Framework Programme (IST-200026228) and the consortium consists of 9 partners, co-ordinated by the University of Florence. The remarkable feature of C-ORAL-ROM is its spontaneity: texts have been recorded in their actual context and without any script. Each subcorpus is made up of 300,000 words with the same text distribution to assure comparability and sufficient register representation. The resource will be delivered in several formats: an orthographic transcription, an xml-tagged version, and the aligned audio source. Partial linguistic annotation will be provided, as well as some programs to handle the resources and quantitative studies. This paper shows preliminary results with respect to the Spanish corpus. 102 1.1 Guirao et al Differences between a speech database and a corpus of spontaneous speech When discussing spoken resources, a preliminary distinction has to be made. Most linguistic resources currently available are speech databases: collections of high-quality recordings and detailed phonetic transcriptions of speech set up in controlled environments (typically telephone services). These speech databases are mostly used for training and testing speech systems and they are developed by and for the language engineering industry. They aim to serve as a basis for recognizing and producing speech in restricted, predictable domains. In most cases, those databases contain many samples of the same word (that is, many tokens of the same type). Usually, the utterances are prepared and pronounced by professional speakers. The acoustic quality of the recording is essential. Speech databases usually provide detailed phonetic descriptions, including disfluencies, noises and other sounds. In general, those databases reflect the standard register, and distant variants (dialects, jargons) are poorly represented. Instances of those are SpeechDat (LRE-63314, Infrastructure for Spoken Language Resources), SpeechDat II (LRE2-4001, Speech Databases for the Creation of Voice Driven Teleservices), which have set up a standard for this type of resource. On the other hand, corpora of spontaneous speech are typically collections of a wide variety of spoken registers and non-scripted speech. Those corpora are collected mainly for linguistic analyses and applications (language teaching, grammars and dictionaries). In such corpora the acoustic quality is not essential. What is important is that the texts reflect as much variation as possible and the speaker behaves in a spontaneous manner. In some cases, those corpora are only concerned with a given register, for instance, a dialect or children’s speech. An important difference with respect to speech databases is the transcription: spontaneous spoken corpora usually are less precise in the acoustic and phonetic parts. On the contrary, they include detailed information about the context and the speakers. These corpora are used mainly for sociolinguistic, text-typologic, or psycholinguistic analyses. Examples are CHILDES and London-Lund. C-ORAL-ROM is a corpus of spontaneous speech, but it also shows some distinctive features: Multilingual: the main goal is to compare the four languages, on the same grounds, and provide comparative studies at different linguistic levels. Acoustic quality: in order to be re-usable by the speech industry, sufficient samples of digital recordings, media and phone conversations are included. Alignment of the transcription and the original sound: this is useful both to verify the accuracy of the transcription and for teaching and other applied investigation purposes. Relating linguistic units to socio-contextual information 103 The main limitation of C-ORAL-ROM is its size. 300,000 words per language is not a sufficient number for stating classifications and statistically significant analyses. We believe that this corpus will show the relevancy and usefulness of an approach that pays as much attention to the acoustic quality of the register as to the linguistic annotation. 1.2 Multi-lingual comparability Cross-linguistic comparison can provide two complementary perspectives: comparing a given feature or features across languages, and comparing a given register or text type across languages (Biber, 1995). On the C-ORAL-ROM project different traditions and experiences have interacted. On the practical side, the teams came to an agreement around two basic points: a text distribution (or sampling design) and a unified format for the transcription. 1.2.1 Text distribution In order to compare the linguistic features of the four languages, the same common sampling criteria and the same proportion of each type in the four subcorpora are needed. There is a long tradition in sociolinguistics and in corpus linguistics (Labov, 1966; Biber, 1988; Biber et al., 1999; Miller & Weinert, 1999) in determining the relevant non-linguistic parameters. Basically, authors agree with a series of socio-situational parameters, such as register and genre variation, sociological features of the speakers (sex, age, education, occupation, origin), and dialogic structure (monologue, dialogue, conversation). The disagreement is in how to combine these parameters. C-ORAL-ROM has chosen the design of the Spoken Dutch Corpus (http://lands.let.kun.nl/cgn/ehome.htm). The sampling design is different in both sub-corpora: Informal register is organised according to social context (familiar-private vs. public) and dialogic structure (monologue vs. dialogue-conversation). Formal register is organised according to channel (media, telephone, natural context). In addition, media texts and formal in natural context are grouped by genre (see table below). Sociological features have not been taken into account for text selection, but they are explicitly marked in the metadata section of the transcription. Male/Female distinction has been the only feature to be balanced. With respect to the text length, some decisions have been made. Only three types of size are allowed: short (1,500 words), medium (3,000 words) and large (4,500 words). Texts shorter than 1,500 words have allowed in genre types like meteorological reports, but always compounding segments of 1,500 words. 104 Guirao et al Tables 1 and 2 show the distribution design of the informal and formal subcorpora for each language. Table 1: Informal sub-corpus Private/Familiar Context 113,000 words Monologue Dialogue 33,000 words 80,000 words Public Context 37,000 words Monologue 6,000 words Dialogue 31,000 words Table 2: Formal sub-corpus Formal in Natural Context 65,000 words Political Speech Political Debate Preaching Teaching Professional Explanation Conferences Business Law Formal in Media Context 60,000 words News Meteo Interviews Reportage Scientific Press Sport Talk Show Political Thematic Explanation Talk Show Culture Talk Show Science Telephone 25,000 words Private Dialogues Phone to Call Services 1.2.2 Common format To ensure a valid comparison, it is also necessary to use a consistent annotation framework. The consortium developed the C-ORAL-ROM format, which is based in the known CHAT format. A conversion to XML is provided. The xmltagged version guarantees easy interpretation though the corresponding DTD. The combined use of XML and DTD ensures that every text in each corpus complies with the same requirements. In this way, textual uniformity are obtained throughout and between the four corpora. The format is divided into the header (with the meta-data) and the transcription. Most features in the header are compulsory, therefore a rich information is provided for every text. The transcription is divided into turns, where applicable. Each turn is marked by a three-letter code identifying the speaker. An orthographic transcription is provided, along with some tags marking disfluencies, noises, overlapping, and prosodic units. Morpho-syntactic tagging will be supplied in a separate tier. Figure 1 shows a fragment of a text. A large selection of fragments from the four languages can be consulted on the official webpage of the project, along with the sound source. Relating linguistic units to socio-contextual information 105 @Title: Raquel @File: efamdl04 @Participants: PAT, Patricia, (woman, B, 2, hairdresser, participant, Madrid) ROS, Rosa, (woman, B, 3, English teacher, participant, Madrid) @Date: 10/03/2001 @Place: Madrid @Situation: chat between friends at home, not hidden, researcher observer @Topic: friends, movies and future Use’s works @Source: C-ORAL-ROM @Class: informal, familiar/private, dialogue @Length: 7’ 58’’ @Words: 1509 @Acoustic_quality: A @Transcriber: Guillermo @Revisor: Manuel; Guillermo, Jesús and Manuel (prosody) @Comments: *PAT: si ya han [/] han decidido ir con ellos / y la conocen ... *ROS: ya / pero si yo no [/] si a mí me da igual / si yo no digo nada de Use y Nuria / yo digo que la peña es un / poco egoísta // *PAT: <no> // *ROS: [<] <y ya está> // Figure 1: A fragment of C-ORAL-ROM text 1.3 Other relevant aspects C-ORAL-ROM is compliant with the state of the art in spoken corpora. These aspects are briefly summarised in the following paragraphs. 1.3.1 The legal issue During the 1990s legislation on Copyright and Privacy changed in many European countries. In spoken language corpora, the law is applied when recording individuals or using sound documents from the mass media. In the first case, speakers retain their right to preserve privacy, and have to give their express authorisation in order to their speech will be transcribed and published. In order to preserve spontaneity, which is essential for our purposes, the procedure is to ask each participant to sign an authorisation after the recording. If a speaker refuses his/her consent, then the recording is discarded. The right to privacy applies to every recording in a private context, but not to ones in a public situation (a lecture, a political speech, a sermon). 106 Guirao et al On the other hand, many texts in the corpora are copyrighted, not only the media recordings but also those in which the speaker creates knowledge in the form of ideas or structure of contents. Typically, this is true of lectures and professional talks. We obtained the written authorisation from the authors or the copyright holders for all the texts included in the Spanish corpus. 1.3.2 The acoustic quality The Spanish corpus of C-ORAL-ROM has been collected from scratch, although other teams in the project have reused part of their previous texts. In our case, we preferred to make new recordings because, on one hand, we did not have the written consent for our previous texts and, on the other hand, the acoustic quality of the analogical tapes was poor. Most texts have been recorded with a DAT Tascam (model DA-P1) and two unidirectional microphones. The source has been converted into a WAV file, mono, 16 bit, 22.050 Hz, through a SPDIF port in a Sound Blaster Live Platinum 5.1, using the software Creative Recorder. In public places, when possible, the DAT recorder has been connected to the sound system. The media recordings either have been provided directly by the broadcasting station or recorded by a computer connected to the receiver. Acoustic quality is essential for the application in speech technologies and language engineering. 1.3.3 The linguistic annotation Corpora increase in value depending on the annotation layers provided. Tagging a spontaneous speech corpus is a task slightly different to the same for written corpora (Uchimoto et al., 2002). The difference is not in the tagged information but in the lesser efficiency of the taggers when applied to spoken corpora. For instance, some POS taggers are usually trained on written texts, which show a quite stable and determined word order. On the contrary, corpora of spontaneous speech are highly flexible in word order. In addition, they show repetition, restartings, overlapping, and other features of spoken syntax which have to be trained specifically. The lexicon is also different. One can find many words that are not included in printed dictionaries, because they are innovations, or belong to an informal register, or simply because they are mispronunciations. A complete lemmatization and POS tagging is provided. Moreno and Guirao (2003) report the development of a POS tagger and unknown words recogniser for morphosyntactic annotation of the Spanish corpus. The results provided by the lemmatizer are used in this paper (see section 3). Relating linguistic units to socio-contextual information 107 1.3.4 The validation To verify the reliability of data has become a fashionable topic in the recent years. Users of linguistic resources want to know how the resources have been collected and their accuracy. C-ORAL-ROM passes two types of evaluation. An internal validation is carried out by the team itself. Each text passes through five steps: transcription, first revision, prosodic tagging, second revision, and sound-text alignment. At least, three linguists transcribe/revise each text. A program verifies format errors, blanks, typos, badly formed tags, etc. Therefore, content and form have been validated exhaustively, guaranteeing that the transcription is accurate to the sound source. We want to stress that the alignment of sound and text is the best guarantee for validation of a spoken text: any discrepancy between the actual speech and its transcription will be easily detected. An external validation will be done by experts at the end of the project. 2. The computational tool We have developed computational tools for transforming the C-ORAL-ROM format into a more suitable tagging scheme in order to relate meta-data with lexical items, and compute the appropriate statistics. We will divide this section into three sections. 2.1 Using xml-tagged corpus for relating meta-data and linguistic features The original C-ORAL-ROM annotation has been designed for registering a wide range of features, including acoustic ones (prosodic marks, noises, etc.) which will be used by the speech technology community. An example of an xml-tagged file is shown: <Turn> <Name>PAT</Name> <Says> <Utterance Type= "interrogation"> y cómo está </Utterance> <Notes Type= "act"> cough </Notes> </Says> </Turn> <Turn> <Name>ROS</Name> <Says> <Utterance Type= "enunciation"> bueno </Utterance> <Utterance Type= "enunciation"> no está mal </Utterance> </Says> </Turn> 108 Guirao et al Our goal in this experiment is to seek out lexical units peculiar to each subcorpus. The first step was to remove the non-lexical information from the original xml tagging. In particular, we wanted to capture two types of information: i) The words that every speaker says, and ii) The split of every turn into utterances in order to prevent ill-formed word clusters. This task is similar to tokenisation in written corpora. This division into utterances is also needed for delimiting the context that the POS tagger uses for disambiguation. A Perl script generates a new tagged corpus with only two tags: one for TURN, with attributes for speaker and file, and another for UTTERANCE: <turn speaker="PAT" file="efamcv01"> <utt> y como está </utt> </turn> <turn speaker="ROS" file="efamcv01"> <utt> bueno </utt> no está mal </utt> </turn> By this means, every word in the corpus can be related with the speaker and the text. The file keeps in the header all the socio-contextual information. The corpus is partitioned in as many sub-corpora as different features appeared in the header. For instance, a male sub-corpus, an informal sub-corpus, a telephone sub-corpus, a meteo sub-corpus, etc. are all generated. After partition into sub-corpora, all occurrences (the tokens) for every lexical unit (the types) are counted in each subcorpus. The next table shows the distribution by sex of speaker. Table 3: Distribution by sex Sex Man Woman X Tokens for the category 182832 134693 9519 Total number of tokens 327044 327044 327044 Percentage 55.9 % 41.2 % 2.9 % The “X” value is assigned when the sex of the speaker is unknown (typically in a media recording). The procedure can be applied to any type of information derived from the corpus. For instance, we tagged it with POS and lemma, using a POS tagger for spoken Spanish developed in our laboratory (Moreno & Guirao, 2003). We show the previous example after lemmatization and POS tagging. Lemmas are shown in uppercase. Relating linguistic units to socio-contextual information 109 Lemmatisation <turn speaker="PAT" file="efamcv01"> <utt> Y CÓMO ESTAR </utt> </turn> <turn speaker="ROS" file="efamcv01"> <utt> BUENO </utt> NO ESTAR MAL </utt> </turn> POS tagging <turn speaker="PAT" file="efamcv01"> <utt> C </turn> P AUX </utt> <turn speaker="ROS" file="efamcv01"> <utt> MD </utt> ADV AUX ADV</utt> </turn> C= conjunction; P = pronoun; AUX = auxiliary; MD: discursive marker; ADV= adverb In summary, in this experiment we have considered three levels of linguistic data: words, lemmas and POS. 2.2 Extracting word clusters If we calculate statistics directly on every unit, the result will not be correct, since multi-words units will not be included in this count. Discourse markers as frequent as “por ejemplo” (for instance), “es decir” (in other words) or “o sea” (that is) will not appear if we work on single word units. To solve this, we developed an algorithm based on n-grams in order to extract multi-word candidates. We took out all n-grams with three or more occurrences, for n = 4, 3, and 2. Next, a filter is applied for discarding all n-grams that start or end with a determiner or auxiliary. Finally, multi-words are selected by hand. Every multiword is regarded as a lexical unit, equivalent to the simple/single words. 2.3 Applying the statistics of surprise In order to identify the distinctive words, lemmas or POS for a given sub-corpus, we have employed the log-likelihood ratio test proposed by Dunning (1993). This method does not assume normal statistical distributions of units in a corpus. Instead, the log-likelihood ratio O assumes a binomial distribution more appropriate for rare but distinctive words. “Texts are composed largely of such rare events” (Dunning, 1993). In addition, this test does not need balanced subcorpora for comparison. 110 Guirao et al This method has been successfully applied for finding collocations (Dunning, 1993) and terms (Daille, 1994). In order to test the method for finding distinctive units in specific domains, we can work on two hypotheses: i) Two registers (or sub-corpora) show no difference in distinctive units (Null hypothesis) ii) For a given sub-corpus, we can find out distinctive units (Alternative hypothesis). We applied the test to two well-defined sub-corpora, meteorological reports and law, in order to discard one of the hypotheses. Results are shown in Table 4. The critical value for one degree of freedom is 7.88. Table 4: Dunning Test applied to well-defined sub-corpora Meteo -2 log O 165 160 145 128 97 91 80 69 69 64 Freq 50 38 27 6466 16 19 19 12 12 92 Lemmas norte fuerza viento en componente temperatura noroeste oeste nube zona Law -2 log O 200 200 116 113 101 84 83 78 77 65 Freq 58 289 85 45 22 31 26 50 69 17 Lemmas policía persona derecho contrato judicial delito delincuente ley determinar cometer Results confirm the alternative hypothesis and the suitability of the Dunning test for the task. Most of the “top 10” lemmas in both domains have a low occurrence, but all are typical terms in its domain. 3. Preliminary results Our goal is to show a range of possibilities for the application of this method. We will show here a very incomplete set of data. Currently, there is a disproportion of social and register features with respect to annotated linguistic features. The linguistic annotation is being carried out this year (2003). Comprehensive results will be delivered with the final version of the four corpora, including a crosslinguistic comparison. In this paper, the only linguistic features that will be taken into account are: words and multi-words lemmas POS tags. Relating linguistic units to socio-contextual information 111 First, we show the 10 most frequent word types in our corpus of spontaneous speech for three professions: a consumers’ association manager, a football coach and a computer system administrator. Notice that words and multi-words are regarded as equivalent units. Table 5: Most characteristic words in three professions Consumers’ association manager -2 log O 110 establecimiento 107 establecimientos 106 encuesta 103 cesta 65 política de precios 45 marcas 44 precios 44 índice 42 consumidor 41 insisto Football coach -2 log O 31 30 28 25 18 17 16 16 16 16 Computer system administrator -2 log O directiva 91 tío club 48 grabando Real Madrid 41 web rueda de prensa 33 yo qué sé director general 33 no sé no 30 joder estabilidad 30 linux confianza 28 cabrón hombre yo creo que 28 detectan Y que 26 barato The procedure can be extended to any profession registered in the corpus, as a means of detecting sociolectal information. Now we will provide the results of the Dunning test on formal and informal registers, approximately 150,000 words each: Table 6: More characteristic words in formal and informal registers Formal -2 log O 134 91 62 61 60 55 49 49 47 46 de es decir su en gobierno en este momento desarrollo general nuestra países Informal -2 log O 422 279 238 231 231 222 182 179 173 171 si ah sabes claro me tía yo no sé no ya Another interesting comparison is to find out which POS are more typical in male and female registers. This table shows the results. In our corpus, men prefer to use nouns and women prefer clearly pronouns. Finally, after lemmatization, we can show the 10 most frequent verbs in general, male and female sub-corpora. 112 Guirao et al Table 7: POS in male and female registers General Total occurrences 47052 42531 38210 32284 31404 30737 25044 17418 12611 10112 Male V N PREP ART ADV C P AUX ADJ Q -2 log O 515 422 399 382 80 34 34 Female N ADJ ART PREP DEM Q REL -2 log O 1327 524 243 149 126 28 16 16 4 P ADV C MD INTJ V POSS AUX NPR Table 8: Most frequent verb lemmas in male and female registers General Total occurrences 3398 2973 2579 2067 1046 1026 995 802 779 577 4. Male ir tener decir hacer poder saber ver dar querer creer -2 log O 28 28 26 23 23 23 22 22 20 20 Female Escuchar Recordar Aparecer Llegar contemplar caminar intentar Amar juntar superar -2 log O 159 154 112 99 86 47 46 42 36 36 ir decir saber venir dar mirar comprar gustar quedar contar Conclusions and future work Here, we have shown the relevance of this procedure as an empirical method for the validation of sociolinguistic hypotheses in spoken language, as well as for determining register typology. The method correlates linguistic with socio-contextual data applying Dunning’s Statistics of Surprise. In order to achieve this, a rich linguistic-tagged corpus and the use of xml have been essential. The preliminary results are promising and have not been shown previously for Spanish. However, extracting conclusions and interpretations for these figures is premature, since the corpus is clearly not sufficient. For this reason, we will apply the method to CORLEC corpus, also developed by the LLI-UAM (see Moreno 2002 for an overview). The combination of C-ORAL-ROM and CORLEC corpora will contain over 1,500,000 words of spontaneous spoken Spanish. Relating linguistic units to socio-contextual information 113 We also wish to find out more about the correlation between linguistic and sociocontextual features, when complete morphosyntactic annotation will be finished. For instance, verb tenses, number, gender, persons in pronouns, etc, will be tagged. Biber (1988, 1995) provides a rich catalogue of linguistic features that can be traced. Finally, a cross-linguistic comparison between the four Romance languages in CORAL-ROM will be made, based on the same text distribution. References Biber, D. (1988), Variation across speech and writing. Cambridge: CUP. Biber, D. (1995), Dimensions of register variation. Cambridge: CUP. Biber D., S. Johansson, G Leech, S. Conrad and E. Finegan (eds.) (1999), The Longman grammar of spoken and written English. London: Longman. Cresti, E. et al. (2002) `The C-ORAL-ROM project. New methods for spoken language archives in a multilingual romance corpus’, in: Proceedings of LREC 2002. Las Palmas de Gran Canaria. Daille, B. (1994), Combined approach for terminology extraction: lexical statistics and linguistic filtering. Ph.D. Thesis, Paris 7. Dunning T. (1993), ‘Accurate methods for the statistics of surprise and coincidence’. Computational Linguistics 19(1): 61-74. Labov W. (1966), The social stratification of English in New York City. Washington: Center for Applied Linguistics. Miller J. and R. Weinert (1999) Spontaneous spoken language. Oxford: Clarendon. Moreno A. (2002), ‘La evolución de los corpus de habla espontánea: la experiencia del LLI-UAM’, in: Proceedings of II Jornadas de Tecnologías del Habla, Granada, Spain. Moreno A. and J. M. Guirao (2003), ‘Tagging a spontaneous speech corpus of Spanish’, in: Proceedings of Recent Advances in NLP (RANLP-2003) Borovets, Bulgaria. Uchimoto K., C. Nobata, A. Yamada, S. Sekine and H. Isahara, (2002), ‘Morphological Analysis of the Spontaneous Speech corpus’. In: Proceedings of Conference of Computational Linguistics (COLING 2002) Taipei, Taiwan This page intentionally left blank An analysis of lexical text coverage in contemporary German Randall L. Jones Brigham Young University Abstract One of the many practical applications of corpus studies is the generation of word frequency information. It makes sense that for the teaching of vocabulary in a second language, lexical frequency should play a significant role in the selection of vocabulary to be included in pedagogical materials. Using the concept of “lexical text coverage”, a study based on the BYU/Leipzig Corpus of Contemporary German has shown that a basic vocabulary of 3,000 high frequency words can account for between 75% and 90% of the words in the text, depending on the register. 1. Vocabulary and second language learning How many words must a learner know in order to read and understand a German novel, a newspaper, an academic text, or to understand a German television broadcast or a conversation? Which words are most important, i.e. which ones should be learned first? Is the vocabulary used in the above-mentioned registers about the same, or are there substantial differences? And what is a word and what does it really mean to know a word in a second language? These questions lie at the very root of second language vocabulary learning. To begin with the question of which words should be learned first, it is important to look at vocabulary frequency, as the frequency of the various words in a typical text differs significantly. Some occur numerous times and some occur only once. By focusing first on learning the most frequently occurring words, it would seem that the process of learning to read or comprehend speech in a second language would become more systematic and therefore more efficient. But again the question arises, what are the most frequently occurring words in German and how many does one have to learn? 2. The concept of lexical text coverage Work by scholars such as Nation (2001) and others has helped us immeasurably in understanding the nature of vocabulary frequency and its relation to vocabulary learning in a second language. For a learner to read and have some comprehension of a German text, a minimum or threshold vocabulary is 116 Randall L. Jones necessary. Nation suggests that an understanding of a 2,000 to 3,000 word family level is a minimum for reading an unedited English text (2001: 146). He has introduced the notion of “word token coverage” and applied it to various types of English texts (2001: 17, 147). The notion of “word token coverage” means the degree to which a defined level of high-frequency vocabulary from a text covers or accounts for all the words in the text. For example, what percentage of a given text is covered by the 1,000 most frequently occurring words? What about the next 1,000, and how many words does it take to achieve a coverage of, say, 90%? The procedure for determining this is quite straightforward. i) Generate a word frequency list for a text. ii) Construct word families from the n most frequent words. iii) Match that word family frequency list against the original text to determine what percentage is covered by the words in the list. Nation uses what he calls word families instead of simple words, i.e. all words that are related to and recognisable from a base word. For example, the base word agree also includes inflected verb forms as well as associated nouns and adjectives: agreed, agreeing, agreement, agreements, agrees, agreeable, disagree, disagreements, disagreeable, disagreed, disagreeing, disagreement, disagrees, a total of 14 word tokens in three word classes. It is assumed that by understanding the verb agree, one will also understand other morphological permutations. The decision is, of course, somewhat arbitrary, i.e. it is assumed that at an early stage of vocabulary learning one understands regular verb conjugation and has a knowledge of the meaning of simple affixes such as dis-, -ment, -able. Table 1 shows Nation’s analysis of text coverage by the 1,000 and 2,000 most frequent words in four English texts registers. The highest level of coverage for 1,000 words was for conversation (84.3%), while the lowest was for academic texts (73.5%). An additional 1,000 words increases the coverage by only a few percentage points. Each additional 1,000 words results in an increasingly smaller addition to the coverage. It is important to stress that we are dealing with coverage of word tokens, not word types. In a running text of 100,000 words of the conversation sub-corpus for example, about half of the words will occur more than once, and some of them numerous times. If the 2,000 most frequently occurring words in the text were compared with each running word in the text, about 90% of the running words would be accounted for or “covered”. These statistics are highly interesting and even a bit surprising. The 1,000 most frequently-occurring words in the English conversation text covered 84.3% of the words in the text. This means that by knowing this subset of vocabulary, a reader will recognise and have some knowledge of any given word in the text 84.3% of the time. It is also surprising that, by doubling the number of words to 2,000, only An Analysis of Lexical Text Coverage in Contemporary German 117 6% additional coverage is achieved. How many words are necessary in order to achieve 100% coverage? It is also interesting to speculate on the difference in coverage among the four registers. One would expect that conversation would represent the highest coverage, but why would newspaper English be ten full percentage points lower? Table 1: Text type and text coverage (tokens) by the most frequent 2,000 words of English in four different text registers (Nation, 2001: 17). Levels Conversation Fiction Newspapers Academic texts 1st 1000 2nd 1000 84.3% 6.0% 82.3% 5.1% 75.6% 4.7% 73.5% 4.6% Total 90.3% 87.4% 80.3% 78.1% These statistics may convey the message that by learning a relatively small number of words – a small number at least when compared to the total number of words in the English language – one can read and comprehend an average English text. Caution is in order here. We still do not understand the cognitive process of vocabulary learning and it is difficult to define what it means to know a word. In addition, if the reader understands 87% of the words in a text, it also means that he or she does not understand 13% of them, and these may be the most important words for understanding the full meaning. Nevertheless, a threshold vocabulary level is a useful concept and a good beginning for continued learning. 3. Lexical text coverage in German What happens when we apply this procedure to a language other than English, e.g. German? As is well known, German has a much more complex morphology than English and uses far more compounding of nouns, adjectives and verbs. For example, the German base word arbeit (“work”) could include not only three nouns (Arbeit, Arbeiter, Arbeiterin) and eight verb forms, but also literally hundreds of compound nouns (e.g. Arbeiterfamilie, Arbeitgeber, Schichtarbeit, etc.), verbs (e.g. ausarbeiten, bearbeiten, verarbeiten, etc.), and adjectives (arbeitslos, arbeitsfähig, etc.). Some of these might be recognised on the basis of an understanding of the root word, but many of them would not. It would seem that for German only the most basic morphological forms should be included in the word family. In spite of the lexical complexities of German, it seems to be possible to calculate lexical coverage in very much the same way as for English. An analysis was made from a subset of approximately 10% of the BYU/Leipzig Corpus of Contemporary German, using each of four registers in the corpus. The total BYU/Leipzig German Corpus consists of the following: Randall L. Jones 118 Spoken German: 1 million words. 700,000 conversation + 300,000 television Literature: 1 million words. Seven literature genres Newspapers: 1 million words. Twenty regional and national newspapers Academic prose: 600,000 words. Six academic areas, secondary & postsecondary Gebrauchstexte: 400,000 words. Instructions, advice, advertisements, etc. The texts are taken from the three major German-speaking countries and represent a variety of styles and levels. Most of the texts are from the years 20002002. A subset was used for this study because the full corpus was not yet complete. The 400,000 word German sub-corpus was processed using the RANGE software (available from Paul Nation, School of Linguistics and Applied Language Studies, Victoria University of Wellington in Wellington, New Zealand). First a raw frequency list was generated, then word family lists were constructed for the first 1,000 and 2,000 most frequent words. The results are shown in Table 2. Table 2: Text type and text coverage (tokens) by the most frequent 2000 words of German in four different text registers. Levels Conversation Literature Newspaper Academic 1st 1000 2nd 1000 82.6% 4.4% 72.0% 5.4% 64.0% 6.5% 65.4% 7.8% Total 87.0% 77.4% 70.5% 73.2% By comparing the German study with Nation’s English analysis one can make several interesting observations. First, the results are really not significantly different for conversation and academic text but quite different for literature and newspaper text. There could be a variety of explanations for this, including external factors relating to the choice of texts used for the respective studies. German national newspapers tend to be erudite and use more compound words. It is also interesting to note the respective differences among the four registers. Conversation represents the highest coverage in both languages, but whereas academic is the lowest for English, newspaper shows the lowest for German. Again, this difference may be the result of a number of factors. An attempt to account for these differences would be beyond our scope here but it is an intriguing cross-language phenomenon. A final curious statistic is the fact that the percentage difference for the second thousand words was almost exactly opposite of the first one thousand, i.e. lowest for conversation (4.4%) and highest for academic (7.8%), thus slightly decreasing the spread among the four registers. For the German study an additional 1,000 words were added. The results are shown in Table 3. Again we see that, by increasing the number of words by 50%, An Analysis of Lexical Text Coverage in Contemporary German 119 the coverage is increased by only 2.5%, 3.4%, 4.2% and 4.6% respectively, but that the percentage is lowest for conversation and highest for academic, thus lowering the gap even more. Is it possible that, at a certain point, the percentages would all be the same? Table 3: Text type and text coverage (tokens) by the most frequent 3000 words of German in four different text registers Levels 1st 1000 2nd 1000 3rd 1000 Conversation 82.6% 4.4% 2.5% Literature 72.0% 5.4% 3.4% Newspaper 64.0% 6.5% 4.2% Academic 65.4% 7.8% 4.6% Total 89.5% 80.8% 74.7% 77.8% One of the questions posed at the beginning of this paper was about the difference in vocabulary among the four registers represented. It is obvious that there is a quantitative difference, as is evidenced in the different percentages of coverage. But what degree of lexical overlap exists? In analysing the range of the 1000 most frequent words in a combined list of all four registers, it was interesting to observe that 89.1% of the words occurred in all four sub-corpora, 7.6% occurred in three of the four sub-corpora, 2.4% occurred in two, and only 0.8% occurred in just one. Table 4a: Words (partial listing) occurring in three of the four sub-corpora NA JEDOCH EURO MONTAG UHR GESICHT KRANKHEIT ZUDEM SOZIAL ZUGLEICH LEUTE URLAUB REGIERUNG Total 162 149 96 81 77 75 65 62 60 59 56 56 55 Academic 2 65 19 0 3 3 60 18 55 35 0 0 11 Spoken 141 0 1 1 0 0 0 0 1 0 35 48 1 Newspaper 0 62 76 75 53 6 2 41 4 13 5 2 43 Literature 19 22 0 5 21 66 3 3 0 11 16 6 0 This would suggest a high degree of commonality among the four registers, and has important implications for second language vocabulary learning. While there are language-learning programs that emphasise a specific register, in most cases vocabulary is learned without regard to how it might be encountered at a later time. 120 Randall L. Jones Table 4b: Words (partial listing) occurring in two of the four sub-corpora SOWIE MEDIZIN DOLLAR JÄHRIG MITTWOCH DIENSTAG QUARTAL PFERD ERSTMAL GARTEN 4. Total 108 105 68 51 51 47 45 43 42 41 Academic 58 102 0 18 0 0 0 0 0 0 Spoken 0 0 0 0 4 2 2 3 41 23 Newspaper 50 3 67 33 47 45 43 0 1 0 Literature 0 0 1 0 0 0 0 40 0 18 Conclusion Vocabulary learning in a second language can be difficult and time consuming, but language educators can contribute to the ease of learning by sequencing new lexical material based on frequency information. It would seem that in German as well as in English a threshold vocabulary of approximately 3,000 words would be suitable for the typical academic learning experience. There are, of course, thorny issues that still need to be addressed such as what it means to learn and know a word and how new vocabulary is best taught. But by being able to establish a frequency-based vocabulary pool, students can benefit from focusing on learning words that will most likely be of the greatest benefit to them. References Nation, I.S.P. (2001), Learning Vocabulary in Another Language. Cambridge: University Press. Analysing a semantic corpus study across English dialects: Searching for paradigmatic parallels 1 Sarah Lee and Debra Ziegeler University of Manchester Abstract In this investigation, we conduct a contrastive corpus analysis into the usages of the ‘get’ periphrastic construction focusing on semantic variation. Our primary interest is in standard Singapore English (SE), and British English (BE) and New Zealand English (NZE) were used for comparative purposes. The investigation found that, generally, SE use of the ‘get’ periphrastic construction was similar to that for BE and NZE. However, after conducting a search for paradigmatic parallels, we also found that in certain functional environments, typically filled by the ‘get’ causative in the other two dialects, SE may have gone further in evolving a competing form - the use of speech act verbs, especially ‘ask’ used in a causative sense. 1. Introduction This paper details the corpus-based approach we adopted to analyse the usages of a grammatical construction, the get periphrastic causative (henceforth getcausative) in an English dialect, and some interesting results of the investigation. We aimed to discover if standard Singapore English (SE) has developed linguistic alternatives to fill the causative function(s) of the get-causative, and/or has derived any indigenised usages of the construction, particularly cognitivelyinduced ones. As an English dialect, SE is interesting in that it exists in an ethnolinguistically diverse ecology wherein a number of genetically-unrelated languages compete with English for various functions. Feature selection as a result is at least potentially subject to influence by cognitive and situational factors that are much less likely to permeate largely monocultural monolingual dialects like BE and NZE. Keeping this important factor in mind, a comparative approach was chosen, and BE and NZE used as standards of measure. Important aspects of the approach included a prior analysis of the construction to determine the prototypical semantics associated with the construction as used in BE and NZE, and to focus the corpus study on finding quantitative and qualitative evidence that might indicate idiosyncratic behaviour related to the get-causative paradigm in SE. An example of the get-causative is given in (1): (1) No no they did Terry O’Neill’s session and it was such garbage they got him to reshoot (ICE-GB s1a-052) 122 Sarah Lee and Debra Ziegeler A particular construction such as the get-causative has, of course, a unique profile comprising a set of functions and semantic features, and a grammatical analysis at a very fine-grained level will reveal that neither the construction nor any of its key components can be replaced by other linguistic item(s) without some loss of information. In a case of paradigmatic replacement, the information that is lost must not include loss of salient function, otherwise, instead of a paradigmatic alternate, the replacement in effect signals an instance of general language change process. The idea of parallels to a well-defined paradigm is perhaps best represented by the variation shown by sociolinguistically and dialect-defined groups in their patterning preferences in a particular function. Cognitive factors can influence grammatical preferencing strategies. Corpus studies can be very useful in pointing out such systematic associations in dialectal variation by providing a massive amount of natural language for quantitative and qualitative analyses. However, data of significance may not be easy to find. Linguistic items in association patterns (‘the systematic ways in which linguistic features are used in association with other linguistic and non-linguistic features’ (Biber 1998:5)) that are also grammatical may not be arranged syntagmatically as grammatical parallels are often aligned paradigmatically. As well as quantitative means, this study uses clues from structural constraints of a construction to retrieve grammatical association patterns in the use of the get-causative. 2. Background to methodology Corpus studies have developed association-based methods to harvest instances of pattern recurrence in language use. More established methods such as collocation (Sinclair, 1991) exploit pairings and groupings of lexical items. Inroads made into the study of language use has shown that words very often do not co-occur at random in a text; rather, pairs or groups of words will consistently appear in the same or similar linear arrangements (though not necessarily directly adjacent) because there is a strong preference in language production to select ‘semi-preconstructed phrases that constitute single choices’ at least to the extent that they are non-random co-occurrences (see, for example, Sinclair’s 1991: 110 ‘idiom principle’). Harvested collocations can be measured and analysed in terms of the degree of association that ranges from a cohesive strength analogous to morphological compounding, e.g., of course (see Sinclair, 1991: 111) to randomness, where the co-occurrence is analysed to have been arbitrarily induced. Collocations can only treat items in essentially syntagmatic arrangements, i.e., collocates must be linearly-patterned words. More recently, the focus in corpus linguistics has also turned to association patterns of grammatical constructions, (e.g., Biber, 2000; Hunston & Francis, 1999), that is, association patterns that are instances of grammatical preferencing (Biber, 2000). There are (at least) two considerations that make retrieval and Analysing a semantic corpus study across English dialects 123 analysis of such data difficult. First, grammatical association patterns do not necessarily syntagmatically align. This means that the goal of a search is not for linear patterns, making the search for grammatical association patterns much more difficult than locating collocations. Second, some types of parallelisms, if cognitively or situationally-induced, are not transparently obvious. Transparent associations will be structurally similar, for example, the difference between overt use of that and the omission of it in clause structure preferencing, e.g. (Biber et al., 1998: 103): (2) a. b. I thought that she only wore jeans. She thinks he’s sweet. The two clause structure types are grammatical patterns and while serving analogous functions, there is an obvious structural difference that can be isolated and studied. But grammatical patterns are not always so easy to detect. A particular grammatical structure is very often the result of cognitive or communicative function, and possible structural representations of a function can vary, i.e., the paradigm is not equivalent to any particular linguistic form sui generis. In (3), for example, (3) Person A: How did you go to the airport? Person B: I ______ Jane (to) drive me. a paradigmatic focus is interested in finding alternatives either for the cause verb, let us say, get, or for the construction as a whole (perhaps a resultative construction like I got driven there rather than another causative one, for instance). So, paradigmatic parallels of the get-construction may not necessarily have the same structural form as the source construction. At best, a primarily structural approach may yield some alternatives but, more likely, to approach the investigation on the basis of structural similarity would harvest too immense a number of possible alternatives, 2 rendering the analysis too broad to be interesting. The central issue involves how one can predict identifiable structural alternatives for a cognitive or communicative function that are not always realised in structurally similar ways. Automated processes that can trawl a corpus for items that fit a semantic profile are not yet available commercially. In Section 4, our low-tech approach to finding parallels for the get-causative is detailed. 2.1 Corpora Consulted The corpora examined were the SE (ICE-SE), NZE (ICE-NZ) and BE (ICE-BE) sub-corpora from the International Corpus of English (ICE) Collection. 3 Each sub-corpus contains approximately one million words of the standard variety of English in each geographic location, with the ratio of the number of spoken to 124 Sarah Lee and Debra Ziegeler written words at 3:2. The number of texts and registers represented in each ICE sub-corpora is standardised, the consistency enabling greater accuracy for frequency counts in cross-dialectal studies. (For more information about the ICE Collection, see Greenbaum, 1996.) As ICE-SE is not tagged for grammatical information, one search issue that arose was related to occurrences of ellipted linguistic items that prevented classification of the token as an instance of causative periphrastic. Zero categories such as prodropping are found in informal registers in SE, e.g., private dialogues (ICE-SG s1a-001 to s1a-100), as a feature that has transferred over from the influence of Chinese (Bao, 2001). Where instances were found, we needed to look at the extended context to determine whether the construction was actually a periphrastic. If any ambiguity could not be resolved, we did not add the instance to our frequency count, as in this example: (4) The only problem is that you must get __ to allow you to tape the voices… (ICE-SG s1a-070) In (4), the ellipted item could have been an instance of pro-dropping (of the causee) or it may be a word like ‘permission’ (i.e., a transitive use of the verb get, plus a verbal complement beginning with the to infinitive serving as an adjunct to the main clause. 3. The periphrastic get-causative construction Here are some examples of the get-causative: (5) Get-causatives a. Like the doctor in Bristol, she looked at my eyes, she got me to touch fingers and noses, to hop on one leg and saw how I coordinated… [ICE- GB w2b 001] b. What we are trying to get you to realise is … [ICE-GB s1b 011] c. Yah or else … how you’re gonna get those middle age people to answer them? (ICE-SG s1a-070) d. With that kind of problem I got my chief engineer to go out and start the company making test equipment that will solve this type of problem. (ICE-SG s2a-043) e. Workflow simply means getting the computer to pass information from one person in a work chain to another. (ICESG w2c-004) f. The forecast of rain for the following week finally got him to fix the roof. (Talmy, 1976:106) Analysing a semantic corpus study across English dialects 125 The get-causative is only one of many constructions of the highly polysemous verb get. With respect to usage, the lexical item get historically has had a reputation of being appropriate for use only in informal or colloquial contexts. The prescriptive bias is mentioned in the Collins Cobuild English Usage (1992), American English Usage (1957), 4 and Fowler (1996). Whether or not this stigmatisation extends to all registers, is observed equally across dialects, 5 or applies only to some get constructions, has not been studied as far as we know. It is also not known how strongly the bias has affected today’s speakers. 6 This lack of knowledge meant that it was possible that the stigmatisation is still active today, so that usage preferencing differences across the dialects and across registers, and skewing, both of frequency distribution totals for registers and overall, may result. 3.1.1 Semantic profile of the get-causative In the literature, there are a few notes about typical usages of the construction but these have seldom been confirmed by studies that are based on high volume empirical data, and also do not provide a complete semantic profile of the construction. However, knowledge of individual essential semantic features can assist pre-analysis. Animacy is an important semantic variable in linguistic representations of causativity. For the get-causative, both Causer and Causee are prototypically sentient beings, viz. human.7 The sense of human agency can be extended metaphorically and metonymically, to animals, natural forces and the class of machines that can metaphorically be conceived as a system comprising parts working together like a computer, machine or car, as in (5e) (Givon, 1976: 348). We included extended animacy as instances of human agency in our frequency counts. The get-causative has also been associated with events viewed by the speaker to be difficult to achieve. This is possibly because of the history of the lexical meaning of get, i.e., the association with obtaining a goal via physical means. Although some analyses, e.g. Hollmann (m.s.) and Wierzbicka (2000: 118) have suggested that the difficulty or effort is associated specifically with the construction, it is not clear whether the association is fundamentally a lexicogrammatical one – i.e., a collocation with the lexical item get. In some dialects though, particularly Australian and New Zealand Englishes (from our personal observations), the construction seems to be used just as much in contexts which do not involve difficulty or effort, (see, e.g., 9a). Another semantic note about the get-causative is that it can be used for contexts allowing the achievement of the causative goal via verbal and non-verbal means, and perhaps usefully, without the actual means of conduit being specified. In (6) for instance, 126 Sarah Lee and Debra Ziegeler (6) With that kind of problem I got my chief engineer to go out and start the company making test equipment that will solve this type of problem (ICE-SG s2a-043) the speaker does not have to specify whether he verbally instructed his chief engineer or conveyed what he wanted indirectly, by simply exposing him to the problem. This is especially the case for non-human Causers, (see, e.g. 5e-f). 3.2 Relation of get-causative to other periphrastic causative constructions The get-causative can be viewed as a member of the class of directive causative constructions (Shibatani, 1976) in English, in which the doer of the event is overtly specified, typically realised periphrastically. Structurally, these constructions involve the use of an auxiliary cause verb such as cause, get, have, make, get, force, and in which the Causer engages an intermediary agent (doer), the Causee, to effect the event of the verb complement. The form of the periphrastic causative constructions is shown in (7): (7) [NPCAUSER-CAUSE COMPLEMENT]] VERB-NPCAUSEE-[(to)-VERB Examples of periphrastic causatives are listed in (8): (8) a. b. c. d. No no they did Terry O’Neill’s session and it was such garbage they got him to reshoot (ICE-GB s1a-052) hm you know some of them bring a client in from a workplace and then have the counsellor sit in and coach them. (ICE-GB s1a-060) The presence of steam at the top probably caused the beads to fuse and over-expand and subsequently led to shrinkage of the cushion after ejection. (ICE-SG W3a-038) Inconvenience and loneliness, but mostly loneliness, made me think of home. (ICE-SG W2f-014) At this level of granularity, the various cause verb causative constructions can be analysed as transparent alternatives in the periphrastic causative paradigm. While the structure remains constant, each construction has a unique semantic profile, a configuration of a number of salient semantic features and specific communicative functions, such as that detailed for the get-causative above. Variation in semantics can relate to animacy preferences (see Talmy 1979 [2000] for analyses of a number of periphrastics), verb complement type constraints, e.g. constraints on the make causative depending on the directness relation between the two causing and caused events (Kemmer & Verhagen, 1994; Kemmer, 2002), and appropriate communicative situations (register preferencing). Analysing a semantic corpus study across English dialects 127 While the known members of the periphrastic causative class are certainly possible candidates for paradigmatic alternatives, 8 it is also true that structurally different forms may also be competitors. In a new study, Ziegeler & Lee (m.s.) argue that the conventionalised scenario construction (9b) is evolving to become a paradigmatic alternative to some functions currently filled by the resultative construction environment (9a) in some dialects of English, especially SE. (9) a. I got my hair cut. b. I cut my hair. Such functional alternatives are problematic for the use of corpora in extracting frequency data, as the construction is formally indistinguishable from an alternative transitive construction involving no indirect causativity at all. Furthermore, no specific grammatical marking is used to express the implied indirect causativity, which would mean that any corpus retrieval must involve searching for haphazard lexical instances, and some methodology would need to be developed to delimit the pool of likely lexical candidates for retrieval. Such a methodology may involve closer field observation of language items in regular use than is possible through the constructed database of a corpus, no matter how representative it might be. There are clearly problems to be overcome and methods to be developed in searching for items which are considered to be openclass functional replacements of former closed-class grammatical categories, as they cannot even represent paradigmatic substitutes within a particular construction. In this study, we searched for members of the periphrastic class and not openclass replacements. A pre-analysis of the semantic profile of the construction was an important aspect of the methodology we used. 4. Approach used in the study The search for paradigmatic parallels is a particularly relevant strategy to studies that involve language change in dialects situated in a complex ecology. As noted, standard SE coexists with a number of languages genetically-unrelated to English (various Chinese dialects, Malay, Tamil, Baba Malay) as well as various Englishes including a creoloid form heavily influenced by Chinese. Most standard SE speakers are bilingual from infancy, and a number of Singaporeans are L2 speakers of English, increasing the possibility of ‘alien’ cognitive-induced features entering the dialect. 9 Previous studies have shown that grammatical features may be sourced from a background language, for example, elements of the aspectual system from Chinese (Bao, 2001) or that wider distribution of grammaticalised features in L1 dialects (hyper-grammaticalisation) may develop 128 Sarah Lee and Debra Ziegeler as a result of differences in the paths of ontogenetic grammaticalisation (Ziegeler 2000). The basic approach used here is a comparative one. Quantitative information can be an important indicator of how widespread the use of the get-causative construction is in SE, however, the figures are meaningless without a standard measure. BE and NZE were enlisted to serve as the contrasting dialects. BE was chosen partly because of historical reasons connected to the development of certain (e.g., more formal) registers in SE, i.e., styles for more formal registers followed an exonormative standard, and partly because it is a variety developed within a largely monocultural, monolingual environment. However, since BE is possibly much slower in dispensing with prescriptive bias against the use of get than in other dialects, 10 assuming the get-causative is affected, it seemed methodologically sounder to introduce a third dialect into the analysis. NZE, like BE, keeps to the monolingual, monocultural constraint but may be less held to the prescriptive bias. 11 NZE would therefore provide a suitable control quantitative measure for frequency information in this respect, lest SE be found to contain significantly more instances of the get-causative construction than in BE. Another important aspect of our methodological approach was to systematically account for each instance of distributional variation found for the construction and each association pattern found for SE, across registers as well as across dialects. We thought that to do this was crucial because the usage of a particular periphrastic causative construction is at least partly influenced dialect-internally due to differing causal exertions on usage in the respective dialects. Influences can be situationally-induced. It has been shown that linguistic units, whether a lexical item or construction, can often be shown preferential bias according to certain register-related variables such as formality and (im)personal style (Biber 1999; 1995, and references therein). The ICE corpora contains a number of texts standardised according to register so conceivably, distributional variation can arise from register bias. For our frequency counts, it was crucial that we recognise any skewing due to this factor so that semantically-induced variation can be identified. To find instances of qualitative variation, we manually analysed each instance found for animacy and overt marking of difficulty. In the case of the latter, as we found that words denoting difficulty or effort were frequently associated with the get-causative, we conducted a supplementary search for periphrastic causative constructions that were possibly paradigmatic alternatives for the same functional environment (see Section 5 below). Analysing a semantic corpus study across English dialects 129 5. Distribution patterns for the get-causative in ICE-GB, ICE-NZ and ICE-SG 5.1 General frequency of occurrence Table 1 below provides a summary of the frequency of occurrences of the getcausative in each corpus. 12 The information provided is the absolute number of instances found in each corpus (n), as well as the totals for the spoken and written modes. As ICE-NZ contained about 30% more words than in either of the other two sub-corpora, the table provides totals normed to 100,000 words as well as a ratio based on the ICE-GB results. The last may be a more accurate indicator of frequency across dialects generally. Table 1: Frequency of occurrence of the get-causative in ICE-GB, ICE-SG and ICE-NZ (in total instances found, per 100,000 words, and relative occurrence in other corpora compared with ICE-GB) Spoken Written Total (Average) ICE-GB n Per 100000 31 4.9 / 1 11 2.6 /1 42 3.8 /1 ICE-SG n Per 100000 42 6.6 /1.3 18 4.5 /1.7 60 5.6 /1.5 ICE-NZ n Per 100,000 68 9.3 /1.9 24 4.0/ 1.5 92 6.7 /1.8 Table 1 shows that the comparative frequency count of instances found in the three corpora support our personal observation that the get-causative may be used more frequently in NZE than in BE. The frequency count for ICE-SE is higher than for ICE-GB but not ICE-NZ. In terms of mode, the ICE-SG figure for written is slightly higher than in ICE-NZ. However, mode per se does not provide an accurate measure of language use, 13 thus the separated figures for mode in Table 1 provide a skewed view of the quantitative results. Additionally, the frequency count comparison merely indicates an impression that frequency distribution of the use of the construction in ICE-SG is somewhere in between BE and NZE, and is thus not helpful in isolating any deviation from the other two dialects. Table 2 breaks down the frequency distribution according to register (normed per 100,000 words). The striking aspect of comparison here is that all dialects are fairly similar in the way in which the get-causative is distributed across registers. Note that registers of the spoken mode do not necessarily contain more instances of the get-causative than the written ones. 130 Sarah Lee and Debra Ziegeler Table 2: Frequency distribution of get-causative across registers in ICE-SG/ICEGB/ICE-NZ per 100,000 words. ICE-SG ICE-GB ICE-NZ Legal cross-examinations (S) 25. Business Transactions (S) 0 9.7 Social Letters (W) 19.0 Persuasive Writing (W); Social Letters (W) Broadcast Talks (S); 12. Non-Academic Writing 1 (W) 6.9 Private Dialogues (S) 16.8 Private Dialogues (S); Reportage (W) 9.6 Private Dialogues (S) / Broadcast Discussions (S) 6.8 Business Transactions (S) 15.7 Broadcast Interviews (S) 9.0 Business Correspondence (W) 6.7 Instructional Writing: Skills & Hobbies (W) 10.5 Parliamentary Debates (S) 8.0 Public Dialogues (S) 5.0 Unscripted Speeches (S) 9.7 Unscripted Speeches (S) 6.3 Legal CrossExaminations (S) / Class Lessons (S) 4.7 Demonstrations (S) Broadcast Talks (S) 4.8 Non-broadcast Talks 4.6 Spontaneous Commentary (S) 8.4 Instructional Writing (both types) (W) 4.7 Broadcast Interviews (S) 4.5 Class Lessons (S); Creative Writing (W) 8.0 Broadcast Discussions (S) 4.5 Business Letters (W) Instructional Writing: 3.3 Administrative Writing 7.1 (W) Class Lessons 4.4 Reportage (W) 2.4 Business Transactions (S) 4.2 Creative Writing (W); Academic Writing (W) 2.3 Reportage (W) Creative Writing (W) 2.4 Legal Cross Examinations (S) 4.5 Non-academic Writing (W) 1.2 Non-academic Writing 4.0 Academic Writing (W) 1.1 Parliamentary Debates (S) 4.1 Broadcast Discussions (S) 3.7 Non printed Student Essays (W) 3.1 Academic Writing 1.6 Legal Presentations (S); Non-broadcast talks (S); Demonstrations (S); Spontaneous Commentary (S); Broadcast News (S); Student Essays (W); Exam Scripts (W); Business Letters (W); 0.0 (W)=Writing; (S)=Speech; Class Lessons (S); Parliamentary Debates (S); Broadcast News (S); Spontaneous Commentary (S); Demonstrations (S); Legal Presentations (S); Social Letters (W); Instructional Writing (W); Persuasive Writing (W) 0.0 8.7 Non academic Writing: 6.6 Tech (W) 6.4 Broadcast Interviews (S); Exam Scripts (W); 0.0 Business Letters (W); Persuasive Writing (W) Analysing a semantic corpus study across English dialects 131 The data appears to suggest that registers which inherently contain more events with directives and (inter)personal situations (perhaps personal style) are found likely to contain more of the constructions. For example, Business Transactions (which contain many situations where A wants B to do C, e.g., server/customer in a shop) ranked highly for ICE-GB and ICE-NZ, and Private Dialogues in all three corpora. It also may explain why Academic Writing, 14 which avoids personal, involved styles, ranked so low across all dialects. The stigmatisation of get may also be a factor contributing to the observation that more formal registers did not contain many instances. There may also be idiosyncratic preferences, for example, perhaps an explanation of why the register Legal Cross-Examinations in ICE-SG was ranked so high. But this could also be due to skewing due to low absolute numbers as word numbers in some registers were much higher than in others. For example, word numbers for Private Dialogues were around the 250,000 word mark but Legal Cross-Examinations were only represented by 20,000 words. Across the dialects, it is clear that the get-causative is more widespread across registers in ICE-NZ. The data may also suggest that style is more informal and personal generally across registers in NZE; however, further study is required beyond the scope of this one. The results show that accuracy of analysis would profit from adopting a register variable-focused approach (Conrad & Reppen, 1998), as there are strong regularities across registers in all dialects that appears to be consistent across dialects. 15 For our purpose, higher degree of accuracy and more information about register-related preferences is not required. We merely noted that the environments where paradigmatic parallels are likely to be found are in registers that contain a higher rate of directives, (inter)personal relationships and more informal styles of communication. 16 We also concluded that there were no significant differences between SE and the other two dialects in the influence of discourse factors on the use of the construction. 5.2 Register-related patterns On a related note, there is more evidence that supports the above conclusion about the parity of SE in relation to BE and NZE. We found that all three corpora were consistent in the way in which they used the get-causative in registers where style was impersonal and formal. While instances with human agents in Causer and Causee positions were found, in more impersonal and formal communicative styles, agents can be absent (unspecified), as in (10a), generic (10b) or nominal clausal or phrasal causers (10c). (10) a b. Workflow simply means getting the computer to pass information from one person in a workchain to another. (ICE-SG w2b-039 ‘Non-academic Writing’) So far, the EDB has played the role of growth advocate to the hilt, pushing hard to get the entire 132 Sarah Lee and Debra Ziegeler c. population to go for growth and helping the Government to focus all its policies on promoting growth. (ICE-SG w2e-005d ‘Press Editorials’) The matchmakers’ persuasive powers getting clients to lower their expectations are also being tested. (ICE-SG w2c-010) In general, frequency information did not reveal any significant variation in the use of the get-causative across the three corpora, and, significantly, did not demonstrate any usages uniquely SE. We now turn to the qualitative information the corpus analysis revealed. 5.3 Semantic features of the get-causative 5.4 Agency typology & distribution While agency is prototypically attributed to sentient entities, (usually humans), causative verbs can show a preference for the type of entity preferred. We found, for instance, that in all three corpora, cause is used in impersonal registers, mostly in scientific/technical documents, and causer and causee tend to be inanimate entities or clausal/phrasal nominal phrases. It is rare to find an example such as (11c), especially in impersonal registers. (11) a b. c. The presence of steam at the top probably caused the beads to fuse and over-expand and subsequently led to shrinkage of the cushion after ejection. (ICE-SG W2a038 ‘Academic Writing - Technology’) A precaution though; fertilizer granules should not be applied too near to its tree trunk as it may cause the stem to rot. (ICE-SG W2d-013 ‘Instructional Writing’) I caused John to go. (Shibatani 1976:3) By contrast, the configuration of participants in get-causative constructions is to have animate causers and causees, typically human beings, or metaphoric extensions of essentially sentient attributes (e.g., volition), supporting earlier observations about this (see Section 3.1.1.). Table 4 shows the prototypical animacy profiles of the get-causative in all three corpora (note then that an example like (5f) would be a very rare use of the construction). Table 4 shows that ICE-SG contains no significant variation from the other subcorpora with respect to animacy configurations. At least in terms of animacy association patterns, SE appears to show no variation from BE and SE. Analysing a semantic corpus study across English dialects 133 Table 4: Prototypical animacy profile of causers and causees in ICE-GB, ICE-SG and ICE-NZ (in %). ICE-GB Get ICE-SG Get ICE-NZE Get H+H H + NH NH + H NH + NH 85.7 7.1 4.8 0 91.7 3.3 1.7 1.7 84.8 3.3 12.0 0 (H=human; NH=nonhuman. H includes extended animacy.) 5.5 Association Patterns We noted in Section 3.1.1. that the get-causative contains the aspectual suggestion of difficulty, effort or the focus of success in achieving the causative event. We found that the get-causative does occur systematically with linguistic items denoting difficulty and effort but that this differed across dialects. Linguistic items marking the difficulty or effort included modals (e.g., must) and catenative verbs, such as try to/and. Try to/and is the most common collocate and indeed, in all three corpora, the lexical item get is the most frequent verb collocation for try to/and. The list of collocations found for the get-causative are given below: (12) Collocations denoting difficulty/effort found in ICE-GB, ICESG & ICE-NZE (only the lemmas of a verb is given) try to/and; manage to; it would be very hard to; difficult to; such an effort to; been unable to; cannot; could not; wouldn’t be able to; it is possible to; must; perhaps...can; succeed in; finally; eventually; In some instances, extended discourse context was consulted, to determine if difficulty or effort is suggested, as in (13). (13) a. b. like you’d get him to do things he wouldn’t usually do but then, he’d kind of give you some security… (ICENZ s1a-056) Mister Warburton says national insurance doesn’t provide a rebate or any financial incentive to get people to install detectors (ICE-NZ S2b-007) Neutral usages, e.g., (14), where no difficulty or effort was marked were also found. 134 Sarah Lee and Debra Ziegeler (14) a. b. I rung this morning. I rung your mother and she was out… and he said ‘can I get her to phone you’ and I said ‘oh no’. (ICE-NZ s1a-007) With that kind of problem I got my chief engineer to go out and start the company making test equipment that will solve this type of problem (ICE-SG s2a-043) In (14a), there is no indication that the speaker thinks that it will be difficult to get the mother to telephone the other person, similarly in (14b). The use of the getcausative is merely for instructional purposes. Table 5 shows the frequency of effort-coded in relation to neutral instances across the dialects. Table 5: Frequency of collocation of get-causative with linguistic items suggesting difficulty or effort in achieving the causative event in ICEGB, ICE-SG and ICE-NZ (in %). ICE-GB ICE-SG ICE-NZE Effort-coded Neutral 64.2 41.7 24.0 35.8 58.0 76.0 As Table 5 shows, ICE-GB had the highest number of instances involving difficulty or effort (64.2%), and ICE-NZ the lowest. The results for ICE-SG were again in between the two other dialects so, once again, ICE-SG shows no sign of deviation from expected usages of the get-causative. While other cause verbs can and do co-occur with items denoting effort and difficulty, none produced quantitative results that indicated a degree of association as cohesively as for get-causatives. We concluded that there was a pattern of grammatical association between the get-causative and events in which the causer is viewed as having to exert some effort in obtaining the caused event. 5.6 Finding paradigmatic parallels to the get-causative A supplementary search using the effort-coded association pattern was conducted to determine whether other possible cause verbs were also associated with difficulty and effort. We found no deviation from expected usages in ICE-GB and ICE-NZ. In ICE-SG, an instance (15) was found where the speech act verb ask was used in what appeared to be a function typically filled by the get causative. (15) T: I didn’t know that last time I bought so such a small policy C: Oh yes I am trying to ask you to buy a bigger policy you say no no no (ICE-SG S1b-075) Analysing a semantic corpus study across English dialects 135 C is an insurance salesman and T a customer. During the exchange, C is attempting to sell T a more expensive life insurance policy by giving the latter reasons for doing it. T keeps asking questions about the offer. C says ‘…I am trying to ask you to buy…’. There are two interesting points about this instance of ask. The utterance reports on an indirect speech act which was uttered with some apparent effort, and not the effort involved in effecting an indirect causative act. The same use of ask in (16) refers only to the difficulty of producing the question verbally: (16) My Italian is not good, I tried to ask you to pardon me. Another point about this example is that C may have been referring to his efforts in the past and not in this conversation. At least in the segment recorded for use in ICE, there is no record of T having said ‘no no no’. Some SE speakers do not make use of tense-marking if time is marked in some other way, e.g., (17). (17) That day you know you’re trying to tell me to do both by December…I mean I know it’s necessary but it’s (ICE-SG S1b029a) In (17), the event happened in the past marked by that day, but the present progressive tense form are trying is used. If indeed (15) is another instance of this, the evidence is even stronger that the use of ask is developing another function in SE. What is significant is that no instance involving the use of speech verbs as neo-causatives was found in ICE-GB and ICE-NZ. 17 However, since only one instance of causative ask was found, we need to do further work to determine whether this is a development for ask or whether other speech verbs are involved (some instances of causative call were also found). 6. Conclusions In our investigation of the get causative in SE, we exploited grammatical association patterns to determine whether paradigmatic parallels to the getcausative can be found in that dialect. As the results showed, the quantitative analysis did not indicate any variation of significance of SE use of the getcausative. Register-related usages of the construction in SE also showed no deviation from the findings for BE and NZE. We are cautious in our conclusion that there are no register differences in SE as the sampling size (each corpus is one million words) may have been too small to expose any subtle variations. As our primary interest is in semantic variation, we do not believe this point damages our account of the use of the construction in standard SE. The more interesting methodology point concerns the paradigmatic parallel we may have possibly found in the speech verb ask. If this is really an instance of paradigmatic 136 Sarah Lee and Debra Ziegeler replacement, then frequency of occurrence is not necessarily always a true indication of what features are used in a dialect. 18 It was noted by Biber (1995) that a feature that is distinct to a particular dialect need not occur frequently in a corpus to be significant. However, this raises the issue of how one can distinguish a significant occurrence from merely a non-patterned one. Clearly, more work needs to be done on this methodological issue, and in finding paradigmatic parallels in general. With respect to the methodology we did use, we found our objective to account systematically for the features found in usages of the get-causative to be a useful approach to cognitive (including semantic-based) investigations using corpora. Of course, much more could be done to fine-tune the methodology we used; however, we hope to have raised some issues relating to the difficulties of finding paradigmatic parallels in corpus study. Note 1 The authors would like to acknowledge the UK Economic and Social Research Council for sponsorship of the research project upon which this study is based (grant no. R000223787; PI: Debra Ziegeler). 2 A search for the form [NP1CAUSER-CAUSE VERB-NP2CAUSEE-[(to)VERB COMPLEMENT]]. 3 We used the Wellington Corpus (spoken and written) at the presentation given at the Corpus Linguistics 2003 Conference, in Lancaster, UK because ICE-NZE was unavailable to us at the time. 4 This book is based on Fowler’s Modern English Usage. 5 Hundt (cf. Denison 1998) mentions the possibility that the get passive may be more stigmatized in BE than in American English, hence the lower number recorded in her corpus study. 6 Subsequent surveying and comments by speakers of BE suggest that age may be a variable in determining frequency of use. Informants who are generally younger are less likely to observe the prescriptive rule, or even are unaware of its existence. 7 Talmy (2000:531) includes the get-causative as one example of ‘caused agency’ or inducive causatives. He Talmy (2000:534) also notes that in the use of the get-causative, the final event which the Causer induces is considered to be desirable to some involved entity but we did not use this point in our corpus study. Analysing a semantic corpus study across English dialects 137 8 A list is given in the Collins COBUILD grammar patterns. - 1: Verbs. (1996). 9 See Ho & Platt (1993) amongst others, for an introduction to Singapore English. 10 Hundt (2001) for example, cites Denison (1998) claiming that the get passive occurs more frequently in AE then in BE perhaps because of less pressure from prescriptive bias in the former dialect. 11 From personal observation. If true, this observation may hold for Australian English as well. Both claims need substantiation. 12 The total number of instances in each sub-corpora is not high so one can question whether the quantitative differences can support any statement about cross-dialectal variation. However, fewer instances of other periphrastics (with the exception of make-causative) were found in all corpora. It may mean that the sample size of one million should be increased for studies involving periphrastic causatives. 13 “Speech and writing are not homogeneous types but stereotypes. Speech is generally informal (face-to-face) and written, (information exposition) more formal.” (Biber 1988:36) 14 For Academic Writing and Non-academic Writing registers, the four subtypes, Humanities, Social Sciences, Natural Sciences and Technology, were collapsed. 15 For the multiple-dimensional approach, see Biber (1995, 1988); Biber et al (1998), which we could refer to if we were interested in register distribution. However, this investigation does not target situationallyinduced variation. 16 In subsequent research for instance, we examined Internet-based communications such as emails and newsgroups, in which participants generally use informal styles. 17 (We use the term neo-causative to refer to new members which may be used within the causative verb paradigm that we have described.) We have found some of these in Internet registers, in informal styles, typically by writers in younger age groups and American English. References Bao, Z. (1995), Already in Singapore English. World Englishes. 14, 2, July, 181188 138 Sarah Lee and Debra Ziegeler Bao, Z. (2001), The Origins of Empty Categories in Singapore English. Journal of Pidgin and Creole Languages. 16, 2, 275-319 Bao, Z. (2002), The aspectual system of Singapore English and systemic substratist explanation. (unpublished) Barlow M and S. Kemmer (1999), Usage-based models of language. Cambridge: Cambridge University Press. Biber D. (1999), Language use through corpus-based analyses, in: Barlow M and S. Kemmer (eds.) Usage-based models of language. Cambridge: Cambridge University Press, pp. 287-313. Biber D. (1995), Dimensions of Register Variation. Cambridge: Cambridge University Press Biber D. (1988), Variation Across Speech and Writing. Cambridge, Cambridge University Press. Biber D, S. Conrad and R. Reppen (1998), Corpus Linguistics. Investigating Language Structure & Use. Cambridge, UK: Cambridge University Press. Collins COBUILD grammar patterns. - 1: Verbs. (1996), London: HarperCollins [for] the University of Birmingham. Fowler, H.W. (1996), The New Fowler's modern English usage / first edited by H.W. Fowler. - 3rd ed. - Oxford: Oxford University Press. Givón T. (1979), On understanding grammar. New York: Academic Press. Greenbaum S. (ed). (1996), Comparing English worldwide. Oxford: Clarendon Press. Ho, M. and J. Platt (1993), Dynamics of a contact continuum. Singaporean English. Oxford: Clarendon Press Hollmann W. m.s. Towards a truly dynamic usage-based model: The case of periphrastic causative get. Hundt M. (2001), What corpora tells use about the grammaticalisation of voice in get-constructions. Studies in Language. 25:1 Hunston S. and S. Francis (1999), Pattern grammar. A corpus-driven approach to the lexical grammar of English. Amsterdam/Philadelphia: John Benjamins. Kemmer, S. and A. Verhagen (1994), The Grammar of Causatives and the Conceptual Structure of Event. Cognitive Linguistics. 5-2, pp.115-156 Nicholson, M. (1957), A dictionary of American-English usage, based on Fowler's Modern English Usa. - New York Shibatani M. (1976a), The Grammar of Causative Constructions: A Conspectus. Chapter in Shibatani M. 1976b. Shibatani M. (ed.) (1976b), Syntax and Semantics. The Grammar of Causative Constructions. New York: Academic Press. Sinclair J. (ed.) (1993), Collins COBUILD English language dictionary. Sinclair J. (1991), Corpus concordance collocation. Oxford: Oxford University Press. Analysing a semantic corpus study across English dialects 139 Talmy, L. (2000), Towards a cognitive semantics. Cambridge, Mass.; London: MIT Press. Talmy L. (1976), Semantic Causative Types. In Shibatani 1976. pp.43-116. Ziegeler, D. (2000), Hypothetical modality: grammaticalisation in an L2 dialect. Amsterdam: Benjamins This page intentionally left blank The curse and the blessing of mobile phones – a corpus-based study into American and Polish rhetorical conventions Agnieszka LeĔko-SzymaĔska University of àódĨ Abstract The study reported in this chapter applies a corpus-based methodology to investigate rhetorical strategies used by American and Polish apprentice writers in their argumentative essays. The data was collected from 79 Polish first-year students of English, and their 80 American counterparts, freshman non-English majors. The simplest method was chosen to analyse the discrepancies between the two groups of essays: the comparison of wordlists also known as the keyword analysis. The study revealed interesting textual differences between American and Polish essays which pertain to such rhetorical strategies as the choice of general versus experience-related arguments, the level of formality and the use of structuring devices. 1. Introduction Learner corpora have recently become an important source of data in second language acquisition studies. Samples of learners’ written and spoken L2 production are collected with the aim to describe as accurately as possible various characteristics of interlanguage. The main interests of researchers revolve around the differences between native and non-native linguistic systems. Thus, investigations focus on such linguistic features as lexical, grammatical and syntactic factors (Meunier, 1998). However, the differences between native and non-native production also occur at the macrolinguistic level and are related to the ways discourse is structured by both groups of language users. Such discrepancies reflect broadly understood cultural differences and are studied within the framework of contrastive rhetoric. The claims of contrastive rhetoric can be summarised as follows: Contrastive rhetoric maintains that language and writing are cultural phenomena. As a direct consequence, each language has rhetorical conventions unique to it. (Connor, 1996: 5) When writing in a foreign language learners show a tendency to transfer not only the linguistic features of their native tongue but also its rhetorical conventions. These conventions pertain to such factors as the structure or units of texts, explicitness, information structure, politeness and intertextuality (Myers, 2002). 142 Agnieszka LeĔko-SzymaĔska As a result, native speakers of a language may find learners’ written discourse ineffective or even incomprehensible. So far, the research in contrastive rhetoric has rarely applied the corpus-based and quantitative methods to support its claims (see Anderson, 2001 for a counterexample which, however, is the exception that proves the rule). Its findings are usually derived from the qualitative analysis of exemple texts. For example, Duszak’s (1994) study of the differences between English and Polish intellectual styles as reflected in the introductions to academic papers is based on a selection of 20 Polish and 20 English articles in the field of linguistics, which are exploited only as a source of examples to support the points the author makes, with no attempt to quantify the results. The qualitative approach undoubtedly has its merits in allowing a detailed scrutiny of the exact rhetorical strategies applied by writers. However, supplementing such methods with quantitative data can make the analysis more reliable and justifiable. The aim of this paper is to demonstrate that research in contrastive rhetoric can benefit from the application of the corpus-based and quantitative methodology. In the same fashion, adopting the text approach rather than the language approach to the analysis of corpus data (Scott, 2000) can lead to totally unexpected and very revealing insights. In the study reported here such an approach and such a methodology are applied to investigate the differences in the rhetorical strategies used by American and Polish apprentice writers in their argumentative essays. 2. Data The research reported in this chapter in fact represents a by-product of a large scale project, whose aim is to compile and explore a Polish learner corpus of written English. The corpus consists mainly of argumentative (but also narrative, descriptive and quasi-academic) essays written by Polish advanced learners of English at varying stages of proficiency in L2. As is the case with most learner corpus studies the explorations within the project focus on the areas of lexical, grammatical and syntactic characteristics of learners’ interlanguage. For one of the studies within the project, whose initial aim was the investigation of the breadth of learners’ lexical knowledge, data was collected from 79 Polish first-year students of English at the Institute of English Studies, University of àódĨ, and their 80 American counterparts, freshman non-English majors at South West University in Marshall, Minnesota. In order to ensure a close comparability of the samples, both groups of students were asked to write an essay in almost identical conditions (during one of their first composition classes at the university) and on the same topic – “The mobile phone – the curse or the blessing of the end of the 20th century”. The choice of this topic was fairly arbitrary, as the research questions were not related in any way to the study of attitudes or The curse and the blessing of mobile phones 143 lexis pertaining to this particular area. However, during the process of coding the data it was noticed that the American and Polish essays differed greatly in their treatment of the topic. Thus, the decision was made to pursue this observation in a more rigorous way. 3. Method Since the initial observation indicated that while writing on the same topic both groups of students wrote about different issues and problems, it seemed appropriate to turn to the field of content-analysis for the appropriate methodology to process the data. From among a range of tools widely used by researchers in this area, the simplest method was chosen to analyse the differences between the two groups of essays: the comparison of wordlists also known as the keyword analysis. According to Scott (2000) keywords are a good indicator of the ‘aboutness’ of a text, thus the procedure seemed an appropriate first step to pinpoint the discrepancies between the samples. Instead of using the Wordsmith’s tool for the keyword analysis, the decision was made to process the essays with Wmatrix, a corpus-analysis tool developed at Lancaster University. The advantage of this software over the Wordsmith Tools is that it detects recurring lexical phrases and treats them as single units in the analysis, so the final list of key items consists of both individual words and fixed phrases. Such a list is likely to give a better picture of the main topics surfacing in both groups of essays. Moreover, the generation of a keyword list in Wmatrix is based on a more reliable statistical procedure, log-likelihood chi-square, which has been proved to produce better results in establishing keywords (Rayson, 2003). For the purpose of the keyword analysis, first a wordlist was created for each group of essays. Next, Wmatrix compared the two wordlists against each other and produced a list of overused words and phrases in both subcorpora arranged according to the log-likelihood (LL) coefficient. Only the items whose LL coefficient was above 6.6 (p<0.01) were chosen for a detailed scrutiny. They made a total of 321 items which were further sorted into those overused by the American students (145 words and phrases) and those overused by the Polish students (176 items). The complete lists of the key items in the American subcorpus and the Polish subcorpus are in the Appendix. The cut-off point applied in this study needs some justification. Rayson (2003) claims that since the generation of a list of key items involves multiple statistical procedures the probability of error multiplies. He postulates the adaptation of a much higher cut-off point which should be located at the level of 15,13 (p<0.0001) for the results to be valid and reliable. However, since the list will not be further analysed quantitatively, this argument is not relevant here. The cut-off 144 Agnieszka LeĔko-SzymaĔska point chosen for the purpose of this study is fairly arbitrary and instead a decision could have been made to analyse the first 300 items on the list. Both lists were examined in search for those key items which could point to the recurring themes in the subcorpora (and thus the ‘aboutness’ of the texts) as well as to other characteristic features of the two groups of essays. The examination was supported by the scrutiny of the concordance lines of the key items. Whenever a specific theme could be identified, it was recorded next to the item in the list. In some cases the same key item could belong to two different themes. For example, the examination of the word plans (Figure 1), which was key in the American corpus, revealed that the item is related to two themes of COST or CONTACT. 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 ith a company that sell rate y all talk about their great es allow teenagers to set up h cell phones are the lso in case of a change in r cell phone the most. These nce charges. Many cell phone eight o'clock, when all you used for a simple change in ith friends in order to make urn it off after you've made inute appointments, and make make calls for work, dinner reless companies offer great nt of travel. There are also y of useful ways. Many phone phone calls. Most cell phone nes and lastly the different f they get the right kind of e the ability to get certain ave with cell phones are the your family. Many cell phone plans. The usual rates range from t plans. Oh yes, everything sounds gr plans with their friends and allows plans that are offered. Some of tho plans with where your family went w plans do not cost a tremendous amou plans include free long distance to plans for the evening have been mad plans. If you didn't have a cell ph plans. With all the things teens ha plans & enjoy the company of the pe plans while on the road. There a plans, or having important conversa plans' which are both convenient an plans that allow you to have free m plans available are so inexpensive plans have a cheaper rate per minut plans that can be used that will fi plan. The plans keep getting greate plans for your phone. Let's say tha plans that are offered. Some of tho plans now have free nationwide long Figure 1: Concordance lines of the word plans in the American subcorpus In some other cases the themes could overlap. For example, the keywords related to the motif of DRIVING in the American corpus at the same time pointed to the themes of EMERGENCY or HAZARD. In addition to the thematic grouping, other clusters of key items were also observed and recorded. They were related to the purely linguistic properties of the texts and included pronouns, linking expressions, fixed phrases and the target language variety. The following two tables present the first 15 key items in the two sorted lists. The curse and the blessing of mobile phones 145 Table 1: The first 15 key items in the American subcorpus word raw Year 1 % Native raw % 1 2 3 4 cell phones phone driving 46 81 104 5 0.17 0.30 0.38 0.02 758 475 516 106 2.30 1.44 1.57 0.32 5 6 1 0 0.00 0.00 44 36 0.13 0.11 7 8 9 10 11 12 13 minutes get_hold _of while have someone ways get you car 24 183 25 2 30 342 15 0.09 0.67 0.09 0.01 0.11 1.25 0.05 108 390 99 38 106 595 71 0.33 1.18 0.30 0.12 0.32 1.81 0.22 14 15 if road 104 1 0.38 0.00 232 30 0.70 0.09 LL THEMES and other groupings 639.53 American English 242.25 229.04 95.70 DRIVING (EMERGENCY, HAZARD) 45.36 COST 43.67 CONTACT 35, EMERGENCY 1, fixed phrase 43.67 43.66 34.85 33.37 32.35 30.98 pronoun (you) 30.17 DRIVING (EMERGENCY, HAZARD) 29.56 29.13 DRIVING (EMERGENCY, HAZARD) Table 2: The first 15 key items in the Polish subcorpus word 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 4. mobile_ph one of it we our nowadays mobile which is kind very fact moreover invention SMS Year 1 raw % Native raw % LL 321 1.17 98 0.30 168.92 800 608 263 170 45 55 107 587 43 124 42 23 42 20 2.91 2.21 0.96 0.62 0.16 0.20 0.39 2.14 0.16 0.45 0.15 0.08 0.15 0.07 502 363 101 46 2 6 35 468 5 55 5 0 8 0 1.53 1.10 0.31 0.14 0.01 0.02 0.11 1.42 0.02 0.17 0.02 0.00 0.02 0.00 133.55 114.90 107.06 99.99 56.80 54.73 52.48 43.70 41.75 41.28 40.40 36.24 31.92 31.52 THEMES and other groupings British English, fixed phrase pronoun (we) pronoun (we) GENERAL British English linking expression GENERAL TEXTING Results and discussion The analysis led to interesting findings on the content level of the texts as well as on the purely linguistic level. Some of these differences are not surprising and can easily be explained, but some reveal unexpected discrepancies between the groups. 146 Agnieszka LeĔko-SzymaĔska On the content level, the following thematic groupings were recorded for the American and Polish key items (Tables 3 and 4 respectively). Table 3: The readily-identifiable themes in the American subcorpus THEME Key items pointing to the themes DRIVING driving, car, road, drive, driver, on the road, pull over, vehicle EMERGENCY driving, get hold of, car, road, drive, stranded, 911, emergencies, accidents, on the road, help, ditch, weather, emergency, winter driving, road, car, drive, driver, distraction, on the road, security, pull over, vehicle, hazardous HAZARD COST minutes, plan, distance, service, plans, charges, free, month, roaming, rates, cards, cost, long CONTACT get hold of, drive, on the road, plans, teens, store, kids, reaching MANNERS distraction, service, class, movies, movie, movie theatre TECHNICAL NETWORK PROBLEMS service, area, reception, static, tower FAMILIARITY home, family, house, Minnesota, dad, family, mom, Americans IMAGE teens, cool GADGETS id, weather, mail, store, entertainment, games, directions, features GENERAL 20th Table 4: The readily-identifiable themes in the Polish subcorpus THEME Key items pointing to the themes GENERAL nowadays, invention, inventions, situations, human, influence, life, world, 20th century, civilisation, modern, 21st century, phenomenon, development, cultural, progress, social TEXTING SMS, sending, send HEALTH health, harmful, scientists, harm, diseases, researches CONTACT contact, contacts, holidays, in touch with MANNERS cinema, switch off, theatre IMAGE show off, newest, youngsters EMERGENCY mountains, ambulance Tables 3 and 4 show some variation between the groups in the choice of arguments. This variation can be attributable to the cultural differences between Poland and the U.S. as well as the differences in the life style of the two groups of students. Thus, the theme of low COST comes up as one of the advantages of mobile phones in America, especially for making long-distance calls, whereas such an argument does not surface in the Polish data, since mobile phone calls are generally expensive in Poland. For the same reason the theme of TEXTING appears only in the Polish data, as it is a popular method of using mobile phones The curse and the blessing of mobile phones 147 due to the relatively high costs of mobile-phone calls. TECHNICAL NETWORK PROBLEMS constitute an important issue for the American students living in a rural and scarcely-populated part of Minnesota, where the data was collected. The same issue does not surface in the Polish essays, as the problem is much less acute in Poland, especially in the urban areas where most of the students come from. Also for the reasons related to the density of population in the students’ immediate surroundings, the theme of EMERGENCY use of a mobile phone is much more strongly emphasised in the American data and is almost always connected with the emergency on the road. The prototypical emergency use of a mobile phone for the Polish students is associated with the accidents in the mountains. The differences in life styles are reflected by the fact that in the American data the most frequent theme is DRIVING (the experience that most Polish 19-year-olds do not have yet, at least not on regular basis), whereas the keyword mountains connected with the theme of EMERGENCY in the Polish subcorpus can be explained by the fact that trekking in the mountains is a popular holiday activity among Polish students. Although the findings mentioned above specify the differences between the two groups of essays, they hardly reveal any interesting facts about the subcorpora, since such differences could only be expected. There is nothing surprising in the fact that the students’ essays reflect the characteristic features of the setting in which they were written. However, further scrutiny of the keyword lists brings to light a deeper and more intriguing variation between the two subcorpora. While the American keyword list contains many items pointing to several readilyidentifiable and recurring themes which can be easily explained by the factors related to the culture, setting and life-style, the Polish list contains very few of these identifiable and explainable themes. With the exception of the theme of using TEXTING as a convenient way of contacting people, there are few items in the Polish list whose keyness can be attributed to cultural or environmental factors. On the other hand, one of most recurring themes in the Polish data is the concept of HEALTH represented by items such as health, harmful, and diseases. Although there is considerable research evidence alerting mobile-phone owners to the potential health risks related to the use of this device, it is hard to assume that the students themselves or the people in their immediate surroundings actually suffer from such health problems. Another frequent theme recurring in the Polish data is labelled GENERAL and is composed of items such as inventions, human, and development. It shows that the Polish students tend to discuss the problem in fairly general terms. On the other hand, the American list includes the names of family members (father, dad, and mom), and other items pointing to FAMILIARITY such as house, home, Minnesota and Americans. The variation in the choice of general versus experience-based arguments seems to be the most striking difference between the Polish and the American data. While the American students answered the essay question by making reference to 148 Agnieszka LeĔko-SzymaĔska their own life experience such as driving or technical network problems, the Polish learners talked about civilisation and health hazards. The discrepancies discussed in the previous paragraphs are further supported by the examination of another group of keywords found in the subcorpora: pronouns. Table 5 lists the key pronouns occurring in the two lists. Table 5: Pronouns in the American and Polish keyword lists. pronouns in the American keyword list you, my, I, they, she, her pronouns in the Polish keyword list we, our, us The American key pronouns indicate that the students talk about their own experience (I, my) or refer to real people (she, her), whereas the Polish students make a frequent use of the generic WE (we, our, us) to support the generalisations they make. All these findings point to the fact that the Polish students approach the topic on a more general level, and the American students relate to their own experience in tackling the problem. One explanation of such variation could be that the Polish learners, contrary to the American students, did not own mobile phones at the time of writing. Unfortunately, no data is available on this issue, since the essays were collected with an entirely different purpose; yet this explanation seems highly unlikely, because the mobile phone is a common-place device among Polish teenagers. Thus, it can be claimed that the observed differences reveal the discrepancies in the rhetorical strategies used by American and Polish apprentice writers. These discrepancies pertain to the choice of general versus experiencebased arguments. Such an explanation accords with with Kaszubski’s (1997) findings. He demonstrated that Polish apprentice writers tend to overuse abstract nouns of reference, which, he concluded, may imply that Poles are “particularly prone to make sweeping generalisations when writing in English” (1997: 155). Another interesting observation can be made based on the examination of the keyword lists. The keyness of the pronoun you in the American data indicates that the American students adopt an informal style in their essays which allows them to address the reader directly. On the other hand, the Polish keyword list abounds in linking expressions (Table 6), which could indicate that the Polish students use a more formal style and pay more attention to the structure of their essays. However, there is also an alternative explanation of the high number of linking expressions in the Polish data. Milton (1998) and de Cock (2000) observed that in structuring their essays learners of English tend to over-rely on a small set of linking expressions promoted by textbooks and teachers. Thus, the keyness of linking expressions can also be attributable to the learning strategies rather than the rhetorical strategies employed by the learners. At the moment, it is impossible to establish which explanation – an L2 learning strategy or transfer of an L1 The curse and the blessing of mobile phones 149 rhetorical strategy – is responsible for the high representation of linking expressions in the Polish data. The examination of comparable essays written by Polish students in their native tongue could be revealing in this respect, but unfortunately such data is unavailable at the moment. Table 6: Linking expression in the Polish keyword list. Linking expression 5. moreover, on the other hand, however, apart from, first of all, for instance, thus, furthermore, nevertheless, sum up, what is more, not only, firstly, on the one hand, as far as, for example Other findings The examination of the keyword lists on the linguistic level also reveals some interesting differences between the subcorpora. One of the discrepancies is related to the standard of English used by both groups of students. The Polish list of keywords contains items such as mobile and cinema, whereas the American list contains items such as cell, cellular and movie-theatre. This points to the fact that Polish learners of English tend to use the British variety of English, which can be explained by the fact that the vast majority of EFL materials available in Poland are produced in Britain and obviously promote the British English standard. Perhaps it can be interesting to point out that teaching materials have more influence on learners’ language than other sources of authentic language available outside the classroom, such as films or music, dominated by the American variety of English. A further interesting observation is related to the native/non-native language characteristics expected to surface in the keyword lists. It was assumed that the degree of idiomaticity (understood in the broader Sinclairian terms) would be higher in the native data, and that the American keyword lists would contain more fixed phrases than the Polish one. However, quite the opposite was found to be true (Table 7). Table 7: Fixed phrases in the American and Polish keyword lists. phrases in the American keyword list get hold of, on the road, look at, having to, pick up, pull over, some one, movie theatre, going off, too many, going to, so many phrases in the Polish keyword list mobile phone, on the other hand, mobile phones, apart from, first of all, for instance, point of view, by means of, sum up, what is more, in my opinion, switch off, 20th century, not only, 21st century, phone box, no longer, show off, so called, turn out, more and more, of course, such a, on the one hand, thanks to, as far as, in touch with, for example Such unexpected results in fact stay in tune with de Cock’s (2000) findings. Contrary to the popular belief that one of the characteristic features of 150 Agnieszka LeĔko-SzymaĔska interlanguage is its lack of idiomaticity, she found that learners overuse two-, three- and four-word expressions both in writing and speech. Moreover, her further qualitative analysis of learners’ fixed phrases revealed that ... advanced learners’ use of frequently recurring sequences of words displays a complex picture of overuse, underuse, misuse and use of idiosyncratic sequences, which may well play a significant part in the foreign-soundingness of their speech and writing (de Cock, 2000: 65) A similar complicated pattern of overuse can be observed in the two keyword lists. The Polish list consists mainly of linking expressions whereas most of the American key phrases are phrasal verbs. Thus, this study supports the claim of the complex nature of learners’ use of idiomatic expressions. 6. Conclusions The study has demonstrated that the corpus-based and quantitative methodology can be of value to the field of contrastive rhetoric producing interesting results and justifiable claims. It has also been shown that adopting the text rather than the language approach to the analysis of corpus data can bring totally unexpected, but very revealing insights. Specifically, it has been demonstrated that the two subcorpora exploited in the study contain more differences than could be anticipated before collecting the data. The study has also revealed important textual differences between American and Polish argumentative essays written in English by apprentice writers. These differences pertain to such rhetorical strategies as the choice of general or experience-related arguments, the level of formality and the use of structuring devices. These differences are not a result of ‘nativeness’ and ‘non-nativeness’ in language use but represent deeper rhetorical conventions existing in the two cultures. References Anderson, W. (2001), Discourse-based diversity: a corpus analysis of collocation in European Union and national French Administrative Language. Paper presented at the annual meeting of the British Association for Applied Linguistics. September. de Cock, S. (2000), Repetitive phrasal chunkiness and advanced EFL speech and writing, in: C. Mair and M. Hundt (eds.) Corpus Linguistics and Linguistic Theory. Amsterdam-Atlanta, GA: Rodopi, pp. 51-68. Connor, U. (1996), Contrastive Rhetoric: Cross-cultural Aspects of Second Language Writing. Cambridge: Cambridge University Press. The curse and the blessing of mobile phones 151 Duszak, A. (1994), Academic discourse and intellectual styles. Journal of Pragmatics 21:191-313. Kaszubski, P. (1997), Polish student writers – Can corpora help them?, in: B. Lewandowska-Tomaszczyk and P.J. Melia (eds.) PALC’97. Practical Applications of Language Corpora. àódĨ: àódĨ University Press, pp. 133158. Meunier, F. (1998), Computer tools for the analysis of learner corpora, in: S. Granger (ed.) Learner English on Computer. London and New York: Longman, pp. 19-38. Milton, J. (1998), Exploiting L1 and interlanguage corpora in the design of an electronic language learning and production environment, in: S. Granger (ed.) Learner English on Computer. London and New York: Longman, pp. 186-198. Myers, G. (2002), Contrastive rhetoric and academic discourses: an institutional view. Paper presented at the 2nd International Contrastive Linguistics Conference. October. Rayson, P. (2003), Matrix: A statistical method and software tool for linguistic analysis through corpus comparison. Unpublished PhD thesis, Lancaster University. Scott, M. (2000), Focusing on the text and its key words, in: L. Burnard and T. McEnery (eds.) Rethinking Language Pedagogy from a Corpus Perspective. Frankfurt: Peter Lang, pp. 103-122. Software: Wordsmith Tools, Scott, M. Wmatrix, Rayson, P. 152 Agnieszka LeĔko-SzymaĔska Appendix A: Key items in the American corpus 1. 2. 3. 4. 5. 6. 7. 8. 9. 10. 11. 12. 13. 14. 15. 16. 17. 18. 19. 20. 21. 22. 23. 24. 25. 26. 27. 28. 29. 30. 31. 32. 33. 34. 35. 36. 37. 38. 39. 40. 41. 42. 43. 44. 45. 46. 47. 48. 49. 50. 51. 52. cell phones phone driving minutes get_hold_of while have someone ways get you car if road plan my call drive i driver stranded on now used 911 distance emergencies they could out accidents are believe person good then distraction talk a on_the_road service around when use go hassle look_at home plans she there 53. 54. 55. 56. 57. 58. 59. 60. 61. 62. 63. 64. 65. 66. 67. 68. 69. 70. 71. 72. 73. 74. 75. 76. 77. 78. 79. 80. 81. 82. 83. 84. 85. 86. 87. 88. 89. 90. 91. 92. 93. 94. 95. 96. 97. 98. 99. 100. 101. 102. 103. 104. family help area caller security today class house reached convenience seen Minnesota calls charges loved movies positives reception teens many into having_to pick_up been free along college cool ditch downside id location missed month overall pull_over purchase roaming static three vehicle weather talking emergency movie here things another 20th blessings causing contract 105. 106. 107. 108. 109. 110. 111. 112. 113. 114. 115. 116. 117. 118. 119. 120. 121. 122. 123. 124. 125. 126. 127. 128. 129. 130. 131. 132. 133. 134. 135. 136. 137. 138. 139. 140. 141. 142. 143. 144. 145. dad mail movie_theatre rates some_one store women else needs father away cellular entertainment mom though games kids her problem after americans cards cost directions going_off hazardous item negatives reaching responsibility too_many tower winter long allow going_to something regular down features so_man Appendix B: Key items in the Polish corpus 1. 2. 3. 4. 5. 6. 7. 8. 9. 10. 11. 12. 13. 14. 15. 16. mobile_phone of it we our nowadays mobile which is kind very fact moreover invention SMS mobiles 17. on_the_other_hand 18. 19. 20. 21. 22. 23. 24. 25. 26. 27. 28. 29. 30. 31. 32. 33. 34. 35. 36. 37. 38. 39. 40. 41. 42. 43. 44. 45. 46. 47. 48. 49. 50. 51. 52. us mobile_phones matter however apart_from inventions some situations as health comfortable imagine useful human such small first_of_all for_instance whether short influence mentioned owners contact quite life owner sides addicted point_of_view thanks often necessary world difficult 53. 54. 55. 56. 57. 58. 59. 60. 61. 62. 63. 64. 65. 66. 67. 68. 69. 70. 71. 72. 73. 74. 75. 76. 77. 78. 79. 80. 81. 82. 83. 84. 85. 86. 87. 88. 89. 90. 91. 92. 93. 94. 95. 96. 97. 98. 99. 100. 101. 102. 103. 104. real aware thus unfortunately simply by_means_of cinema concerned furthermore nevertheless sum_up undoubtedly what_is_more businessmen in_my_opinion harmful sending switch_off 20th_centuryusing device scientists its addiction civilisation contacts harm importance not_only lots especially telephone arguments claim advantages modern cells control theatre 21st_century firstly phenomenon phone_box regarded surely whose that send development no_longer living rather under 105. 106. 107. 108. 109. 110. 111. 112. 113. 114. 115. 116. 117. 118. 119. 120. 121. 122. 123. 124. 125. 126. 127. 128. 129. 130. 131. 132. 133. 134. 135. 136. 137. 138. 139. 140. 141. 142. 143. 144. 145. 146. 147. 148. 149. 150. 151. 152. 153. 154. 155. arises basic crucial effect groups hardly irreplaceable knowledge posses present private show_off so_called somehow turn_out youngsters more_and_more of_course not such_a happy nobody normal ordinary mention consider among constant cons pros opinion want appears consideration cultural diseases divided exist famous gadget holidays mountains newest on_the_one_han d opponents precious produce producers regard researches thanks_to 156. 157. 158. 159. 160. 161. 162. 163. 164. 165. 166. 167. 168. 169. 170. 171. 172. 173. 174. 175. forget say against disadvantage possessing progress social various ambulance as_far_as mean in_touch_with makes latest situation important no for_example advantage availabl This page intentionally left blank Using a dedicated corpus to identify features of professional English usage: What do “we” do in science journal articles? Judy Noguchi, Thomas Orr and Yukio Tono Mukogawa Women’s University University of Aizu Meikai University Abstract This chapter addresses the problem of inadequate educational materials for the effective training of non-native speakers in the professional English of the scientific community. We claim that one cause of this inadequacy is lack of proper linguistic research based on suitable linguistic research tools. The development of a major international corpus of professional English in the sciences and other fields is described as a significant resource for solving this problem, illustrated with a specific research project examining the use of “we” in professional English from the Corpus of Professional English. Results reveal the value of well-designed dedicated corpora for addressing the specific English instructional needs of non-native speakers of English in the professions. 1. Introduction Although English is the language of preference for science in international contexts, the majority of the world’s scientists are not native speakers of English and, therefore, they can be said to be considerably disadvantaged in professional English communication. Language teachers may have introduced non-native speakers (NNS) to the sounds, symbols, and structure of English for general purposes such as asking for directions, reading a menu, and writing classroom essays. However, few NNS of English have ever enjoyed the luxury of specialised training in the spoken and written discourse of their profession. In general, this means that they have had to pick this English up on their own, at considerable effort, with disappointingly limited results (Coates et al. 2002). This situation is terribly unfortunate, for it not only severely marginalises NNS in their scientific communities, but it also prevents the world from benefiting from the creative potential of those who cannot disseminate their ideas persuasively because of poor English. Basically, there are two reasons for this unfortunate situation. One is the scarcity of language teachers who possess expertise in the English of science, and the second is the scarcity of instructional materials that adequately explain professional English text in scientific contexts. The English of research articles and technical specifications is very different from the English of storybooks and street signs, and mastery of this English does not come easily. Native-speaker 156 Judy Noguchi, Thomas Orr and Yukio Tono intuition or professional work experience alone are not enough to generate proper linguistic insight, and training materials based on “research” of this kind tend to be disappointingly vague and fraught with inaccuracies. Teachers and students of English in the sciences require more substantial materials based on far more rigorous research, enabled by far better research tools. The development of computer-based corpora, along with sophisticated software tools to analyse them properly, has now started to make this kind of research possible. 2. The value of specialised corpora The advent of computers, computer-based corpora, and the whole new field of corpus linguistics is now beginning to provide some satisfactory solutions to the problems mentioned above. By collecting discourse samples in the linguistic domains requiring study, corpus linguists can now begin to identify features of language that were beyond the scope of thorough observation in the past. This development is welcome news for teachers and students of scientific English, for scientific genres, with all of their peculiarities, can now be studied on grander scales to generate far more objective and reliable data. General corpora, such as the British National Corpus (BNC), provide the means for studying how English is used in general contexts; while the development of specialised corpora provide the means for studying how English is used in special contexts. For rigorous investigations of English in the theoretical and applied sciences, a very special corpus is required, dedicated to the language of these professional communities. Gathering a large enough collection of suitable discourse to develop a dedicated corpus of this kind, however, has not been an easy task. The value of corpora of scientific English on a small scale has been suggested in the past (Johns, 1991; Bondi, 2001), one good example being a corpus of physics research articles and their parallel academic conference presentations developed by Umesaki (2000). Using this corpus, Umesaki explored the variety of referents to the writer in academic papers, which she identifies as “one of the difficulties for non-native speakers of English in writing academic papers” requiring greater educational attention based on findings from corpus research. Although work of this nature is being conducted in various disciplinary fields on small scales, there remains a need for a major dedicated corpus of English for studying professional English in the theoretical and applied sciences. This would provide a referent corpus against which these smaller corpora could be compared and also contribute to a better understanding of how English is used by professionals internationally when they communicate with each other. Using a dedicated corpus to identify features of professional English usage 3. 157 The construction of the Corpus of Professional English, CPE The creation of a major international corpus of English in the sciences and other professions to aid research in professional discourse is a serious undertaking, and a new organisation has recently been established to take on this challenge. Called the Professional English Research Consortium (PERC), this non-profit academic organisation headquartered in Tokyo, was established in April 2002 to create such a corpus and generate research in professional English that might be of particular benefit to NNS in the professions as well as to the educators and material developers that support them. The corpus, named simply the Corpus of Professional English (CPE), aims to become a 100-million word balanced corpus of Professional English. It is designed for various research purposes, ranging from pure research on Professional English (e.g. variations of usage in lexis, syntax, semantics, and discourse across text types) to applied research such as lexicography, language testing, educational materials and program development. The CPE will be serviced by a sophisticated web-based query system so that those who are not familiar with corpora can extract linguistic features they need in a user-friendly manner. The design scheme of the CPE is shown in Table 1. The ultimate goal of the CPE is to achieve a 100-million word written corpus, composed from a reasonable balance of text types. The balance of each text type will be further examined in the future, but tentatively it was decided that the following ratio of text types seemed most suitable for representing professional English: academic journals (30%), legal/workplace documents (20%), trade journals (10%), reference books (10%), websites (10%), newsletters (5%), correspondence (5%), manuals (5%) and ephemera (5%). At present, the project team is focusing on the collection of academic journal articles, for this is the text that frequently proves most difficult when obtaining copyright permission, and yet is needed most by researchers in PE. Table 1: Tentative balance of text types for the CPE Academic journals Legal/workplace docs Trade journals Reference books Websites Newsletters Correspondence Manuals Ephemera 30% 20% 10% 10% 10% 5% 5% 5% 5% In order to select the journal texts in an objective way, the project team decided to base content decisions on data obtained from Journal Citation Reports (JCR), which presents quantifiable statistical data for an objective and systematic 158 Judy Noguchi, Thomas Orr and Yukio Tono approach to determining the relative importance of journals within their subject categories. At present, the JCR contains 5,700 journals in the Science Edition and has a unique measure called “Impact Factor”, which provides a way to evaluate or compare a journal’s relative importance to others in the same field. Employing this data, the top 20% of the journals with the highest impact factor in each field were selected for inclusion in the CPE. JCR classifications were also used to define the subject fields. Acquiring texts for a corpus is always difficult but not impossible. Following procedures established by the creators of the BNC, the CPE project team sent letters to major journal publishers in order to obtain copyright permission for more than 1,500 journals. As is often the case, the first letter did not yield a sufficient response; however, much more favourable responses were obtained after the launching of the PERC website which explained the CPE in greater detail than the letters. At present, copyright clearance has been obtained for almost 300 journals from over 50 publishers. To facilitate research on the CPE, a web-based corpus query system with a 3level interface has also been developed in collaboration with the major PERC member institution, Shogakukan, Inc. For elementary users, the tool provides simple word search and KWIC displays only. For intermediate users, independent word/POS/lemma queries can be made with detailed collocation statistics (raw frequency/T-score/MI-score/log-log). And for experienced researchers, complex search (word/POS/lemma and combinations of the three) is made available with more sophisticated output. Shogakukan plans to launch a portal site for megacorpora such as the BNC, the ANC, and COBUILD-Direct in the future so that ordinary language teachers can access these large corpora without worrying about the installation of a complicated interface. Currently, the CPE project team is in the process of acquiring the targeted texts as well as cleaning and formatting the existing texts that have already been obtained. A prototype version of the CPE (c.2 million words of copyright-cleared academic journal text) has been set up on this new query system and is beginning to provide an impressive wealth of data that contrasts significantly with that which can be obtained from general corpora such as the BNC. 4. The research project and its results To demonstrate the value of a corpus specifically dedicated to professional English, we have chosen for this chapter a simple analysis of current usage of the pronoun we as it appears specifically in professional scientific texts in comparison with similar data gathered from corpora devoted to English on a broader, more general scale. We will first explain the methods we employed in this research, then follow this with specific results and a discussion of their significance. Using a dedicated corpus to identify features of professional English usage 4.1 159 Methodology As stated above, the CPE is currently under construction. The research for this paper was conducted on a prototype corpus from the CPE of 260 journal article texts from 18 journals in fields ranging from multidisciplinary agriculture, biochemistry and molecular biology, cell biology, developmental biology, plant science and forestry to multidisciplinary materials science, mineralogy, oceanography, general and internal medicine, health care sciences and service, orthopaedics medicine, pharmacology and pharmacy, and psychiatry. The prototype corpus included 1,787,484 tokens of almost 60,000 types. The corpus was examined for verbs used after we using Wordsmith Ver. 2 (Oxford University Press). The concordance lines were rearranged with the word to the right of we in alphabetical order, and then the verbs were classified according to the seven major semantic domains described in the Longman grammar of spoken and written English (Biber et al., 1999): “activity verbs, communication verbs, mental verbs, causative verbs, verbs of simple occurrence, verbs of existence or relationship, and aspectual verbs.” A summary of these verb features and some examples are presented in Table 2. Table 2: Verb semantic domains based on the LGSWE (Biber et al., 1999: 360364) Semantic Domain Features Examples Activity Actions and events associated with choice; subject is semantic role of agent Communication Subcategory of activity verbs; associated with activities for communication Activities and states experienced by humans; including cognitive, emotional, and perception bring, buy, carry, come, give, go, leave, work ask, explain, say, suggest think, know, love, hate, see, taste, read A new state of affairs is brought about by a person or an inanimate entity Events happening without volitional activity State of relationship between entities allow, cause, enable, require, permit become, change, happen, develop be, seem, appear State of progress of an event or activity begin, continue, keep, stop Mental Facilitation or causation Occurrence Existence or relationship Aspectual According to the LGSWE (Biber et al., 1999), the most frequently used verb type is the activity verb. The LGSWE bases its findings on corpus studies which identified the four registers of conversation, fiction, news, and academic prose. Examination of the distribution of commonly-used verbs according to the four registers revealed that activity verbs ranked the highest in three of the four registers. It was only in academic prose that existence verbs display almost Judy Noguchi, Thomas Orr and Yukio Tono 160 equivalent frequency. Another feature of academic prose overall is the use of more causative verbs and occurrence verbs, compared to that of other registers. With respect to the usage of the personal pronoun, the LGSWE (Biber et al. 1999: 329-330) recognises its common use “to refer to a single author, a group of authors, to the author and the reader, or to people in general.” Biber et al. also state that “In some cases, academic authors seem to become confused themselves, switching indiscriminately among the different uses of we.” This highlights the problem pointed out by Umesaki (2002) faced by the NNS scientist who often finds author-referent conventions confusing. In this work, we focused on identifying the type of verb following we in the prototype CPE corpus. Knowing what types of verbs are commonly used in academic papers should offer help to the NNS scientist in deciding when and how to use we when writing up research. 4.2 Research data from the CPE A total of 3,401 instances of we followed by a verb were identified in the CPE prototype corpus. The main verb was identified and classified according to the semantic domain classifications in the LGSWE. Its tense and aspect were also noted. If the verb occurred at least twice, it was included in the count for the distribution of verb types presented in Tables 3a and 3b. Table 3a: Distribution of verb types used with we according to semantic domain and tense/aspect Verb type Total Men 1433 45.13 668 46.62 472 32.94 119 8.30 Act 1038 32.69 612 58.96 187 18.02 160 15.41 Com 440 13.86 71 16.14 290 65.91 45 10.23 Exi 168 4.66 63 31.76 70 44.59 6 4.05 Cau 40 1.89 23 65.00 3 11.67 11 18.33 Asp 32 1.01 8 25.00 29 90.63 6 18.75 Occ 24 0.76 14 58.33 3 12.50 5 20.83 1459 45.95 1054 33.20 352 11.09 3175 % past % pres % prep % Using a dedicated corpus to identify features of professional English usage 161 Table 3b: Distribution of verb types used with we according to semantic domain and tense/aspect Verb type mod % prec % pstp % Men 122 8.51 11 0.77 16 1.12 Act 49 4.72 18 1.73 10 0.96 Com 33 7.50 3 0.68 0 0.00 Exi 20 13.51 5 3.38 0 0.00 Cau 3 5.00 0 0.00 0 0.00 Asp 7 21.88 2 6.25 0 0.00 Occ 1 4.17 0 0.00 1 4.17 235 7.40 39 1.23 27 0.85 Verb semantic domains: men = mental, exi = existence, act = activity, com = communication, exi = existence, cau = causative, asp = aspective, occ = occurrence. Verb tense and aspect: past = past tense, pres = present tense, prep = present tense perfect aspect, mod = modal auxiliary, prec = present tense continuous aspect, pstp = past tense perfect aspect The present tense continuous aspect occurred in 5 instances, two with mental verbs and three with activity verbs, but these data are not included in the table. As can be seen from Tables 3a and 3b, the most frequently used verb type after the personal pronoun we was the mental verb accounting for 45.13% of the total. This was followed by activity verbs at 32.69% and communicative verbs at 13.86%. The most frequently-used verb tense was the past tense, accounting for 45.95% of all instances catalogued. However, this tense was the predominant one only for the mental, activity, causative, and occurrence verbs. The present tense form was more commonly used for the communicative, existence and aspectual verbs. A closer examination of the verbs in their semantic domain classifications reveals a more complex picture. As can be seen from Tables 5a and 5b, while the mental verbs find, observe and examine overwhelmingly occur in the past tense, conclude occurs more than 80% as the present tense form. In the case of activity verbs with the highest frequencies, use, analyse and test are predominantly used in the past form, but show is used in the past form in only 14.29% of the instances observed while it appears more frequently and almost equally as the present tense perfect aspect (41.27%) and the present tense (40.48%). 162 Judy Noguchi, Thomas Orr and Yukio Tono Table 4: Number of verbs in each semantic domain and some examples Verb semantic domain No. of verbs observed Mental 97 find (163), observe (108), examine (103), conclude (58), identify (54), know (47), compare (46) , see (45), determine (41), investigate (39) Activity 110 Communicative 34 use (183), show (126), analyse (51), test (48), demonstrate (46), perform (46), measure (33), calculate (24), obtain (23), do (21) thank (101), report (49), present (34), describe (30), propose (29), note (19), ask (18), suggest (18), acknowledge (14), discuss (11) Existence 12 have (71), be (28), exclude (14), include (11), stand (5), have to (4) Causative 5 be able to (28), be unable to (20), allow (6), require (3), subject (3) Aspect 8 Occurrence 6 begin (10), continue (5), start (4), undertake (4), initiate (3), achieve (2), enter (2), keep (2) develop (12), fail (4), modify (2), increase (2), change (2), become (2) Examples (No. of instances) Verbs which appeared after we two or more times were counted. Table 5a: Verb tense and aspect distribution for most frequently used mental and activity verbs Verb Mental find observe examine conclude Activity use show analyse test Total past % pres % prep % 163 108 103 58 129 83 78 6 79.14 76.85 75.73 10.34 11 16 3 47 6.75 14.81 2.91 81.03 19 7 17 11.66 6.48 16.50 0.00 183 126 51 48 131 18 43 39 71.58 14.29 84.31 81.25 24 51 2 1 13.11 40.48 3.92 2.08 17 52 5 6 9.29 41.27 9.80 12.50 Using a dedicated corpus to identify features of professional English usage 163 Table 5b: Verb tense and aspect distribution for the most frequently used mental and activity verbs Verb Mental find observe examine conclude Activity use show analyse test mod % 3 3 5 1.84 0.00 2.91 8.62 9 4 1 1 4.92 3.17 1.96 2.08 prec 1 % pstp % 0.00 0.00 0.00 0.00 1 1 0.61 0.93 0.00 0.00 0.55 0.00 0.00 0.00 1 1 0.55 0.79 0.00 0.00 Not including one instance of the past perfect for observe Verb tense and aspect: past = past tense, pres = present tense, prep = present tense perfect aspect, mod = modal auxiliary, prec = present tense continuous aspect, pstp = past tense perfect aspect Table 6: Top twenty clusters in the vicinity of we N Cluster Freq. 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 we found that in this study we conclude that we examined the we have shown acknowledgments we thank we used the have shown that in the present we did not we do not found that the we show that we have previously the effect of the present study in this paper this study we we have found we used a 76 50 39 39 35 32 32 28 28 28 25 23 22 21 20 19 18 18 18 18 164 Judy Noguchi, Thomas Orr and Yukio Tono The clusters presented in Table 6 show that the top three clusters involve mental verbs, found, conclude, and examined. This suggests that cognitive activities of dynamic nature (find and examine) are expressed using the past tense, while the more stative conclude appears most frequently in the present tense. The sixth cluster is from the acknowledgements section and indicates an almost formulaic usage of we thank. The repeated references to the work being presented in the paper (in this study, in the present, the present study, in this paper) in the vicinity of we indicate that direct reference to the author(s) appears particularly when attention is being drawn to the study under discussion. Interestingly, of the 260 texts examined, we was used at least once in 223 texts. The highest number per 1,000 instances was 13.94, with the total overall average being 2.12 per 1,000. The average for the top ten texts was 9.74 instances per 1,000 words. Table 7: Texts with high frequency usage of we No. File Words Hits per 1,000 Field 1 2 3 4 5 6 7 8 9 10 079.txt 078.txt 072.txt 070.txt 047.txt 035.txt 195.txt 076.txt 058.txt 314.txt 1,865 5,549 2,956 6,292 4,821 7,748 4,668 2,693 7,696 7,921 26 70 34 68 50 73 38 21 50 50 Ave 13.94 12.61 11.5 10.81 10.37 9.42 8.14 7.8 6.5 6.31 9.74 Health services Health services Health services Health services Psychiatry Plant sciences Cell biology Health services Plant sciences Oceanography Table 7 suggests that some journals or fields may display a higher frequency of first-person pronoun usage than others. Also revealing were the negative data of texts in which we was not used even once. Such texts occurred across all fields, but all sixteen texts from one journal in multidisciplinary materials science had no instances of we at all. The only instance that was detected was the abbreviation of WE for working electrode. 5. Present applications and future research The above analyses on the usage of we in the 260 texts in the prototype CPE corpus of scientific academic journal articles from a range of fields reveal the following: i) The use of we is rather common, occurring at least once in 85.77% of the texts examined. Using a dedicated corpus to identify features of professional English usage 165 ii) Some journals and fields tend to display more we usage that others. On the other hand, all texts coming from one of the journals had no instances of we. Thus, there seems to be a need for even further specialization of corpora to illuminate differences among journals and/or fields of study. iii) The verb type most commonly used with we is the mental verb (45.13%) followed by the activity verb (32.69%). iv) The most frequently used tense is the past tense, but the distribution of past and present tense usage is reversed for some verb types and even within verb type categories. v) High-frequency clusters in the vicinity of we tend to be related to mental activities and references to the work under discussion. The findings overall point to the need for even further refining of corpus studies of specialised texts in order to reveal features which can be used when planning course materials for English for specific purposes classes. Such comparative studies of texts from different research fields and genres must await the completion of the CPE corpus; however, this research conducted from a small sample of dedicated texts already reveals some interesting things that differ from earlier findings based on general English corpora. Even in its present state, the CPE corpus prototype dedicated to text in scientific disciplines can prove very helpful as reference material for postgrad students or NNS scientists who are at the stage of writing up their research. If they are given background instruction in the genre-analysis approach to understanding the framework of moves and steps that compose the research journal article (Swales, 1990; Weissberg and Buker, 1990), a data-driven learning approach to concordancing (Johns, 1991a and b) can serve as a valuable tool for aiding NNS with their professional writing, when it comes to word choice or other writing issues (Hunston, 2002). As the CPE continues to develop, we envisage a system by which NNS scientists can access a website for online support when creating professional documents, which would include access to selections of dedicated corpora in the specific fields for which they are writing. References Biber, D., S. Johansson, G. Leech, S. Conrad, and E. Finegan (1999), Longman grammar of spoken and written English. Harlow: Longman. Bondi, M. (2001), ‘Small corpora and language variation’, in M. Ghadessy, A. Henry, and R. L. Roeberry (eds.), Small corpus studies and ELT. Amsterdam: John Benjamins. Coates, R., B. Sturgeon, J. Bohannan, and E. Pasini (2002), ‘Language and publication in Cardiovascular research articles’, Cardiovascular research, 53: 279-285. 166 Judy Noguchi, Thomas Orr and Yukio Tono Hunston, S. (2002), Corpora in applied linguistics. Cambridge: Cambridge University Press. Johns, T. (1991a), ‘From printout to handout: Grammar and vocabulary teaching in the context of data-driven learning’, in T. Johns and P. King (eds.), Classroom concordancing. Birmingham, UK: Centre for English Language Studies, The University of Birmingham, 27-46. Johns, T. (1991b), ‘Should you be persuaded: Two examples of data-driven learning’, in T. Johns and P. King (eds.), Classroom concordancing. Birmingham, UK: Centre for English Language Studies, The University of Birmingham, 1-16. Johns, T. (2002), homepage http://web.bham.ac.uk/johnstf/timeap3.htm Swales, J. (1990) Genre analysis: English in academic and research settings. Cambridge, UK: Cambridge University Press. Umesaki, A. (2000), ‘Syntactic differences in the discourse of oral and written papers’, English corpus studies, 7: 39-59. Umesaki, A. (2002), ‘Reference to the presenter in academic papers’. Paper presented at AILA 2002, Singapore International Convention and Exhibition Centre, Dec. 21, 2002. Weissberg, R. and S. Buker (1990). Writing up research: Experimental research report writing for students of English. Englewood Cliffs, NJ: PrenticeHall. Methods and tools for development of the Russian Reference Corpus 1 Serge Sharoff University of Leeds Abstract The paper discusses the history of development of Russian corpora and presents methods and tools that are used in the ongoing development of the Russian Reference Corpus. Development of the corpus follows the key design principles of the BNC and extends them further by introducing an elaborate model of text typology and by adding lemmatisation and morphosyntactic annotations to POS tagging. The paper also discusses problems in development of the corpus that are related to the Russian language and culture. 1. The history of development of Russian corpora It is not too big a generalisation to say that development of Russian computer corpora followed the pattern established by English corpora. The Brown Corpus (Kucera & Francis, 1967) set up the standard for the design, size and coverage of general-purpose corpora in other languages, including Russian. In the 1970s a corpus of 1 million words was developed by Zasorina and her colleagues; it consisted of 500 samples of 2,000 words each and covered four genres: mass media, fiction, science (including humanities) and drama (as an attempt to cover the spoken language). The study resulted in a frequency dictionary (Zasorina, 1977), but not in a publicly available resource. The best-known comprehensive Russian corpus was developed in the 1980s in Uppsala, Sweden; it also resulted in a frequency dictionary (Lönngren, 1993). The Uppsala Corpus (UC) consists of 1 million words in 600 samples equally divided between fiction and non-fiction texts. UC is popular for various reasons, partly because it can be freely accessed via the Internet, but for modern standards it is too small and restricted in genre coverage. It also lacks morphosyntactic annotations and lemmatisation. The lack of lemmatisation hinders the search of multiple word forms, which often cannot be found using regular expressions, e.g. the verb vyjti (to leave) in Russian has about 40 forms, including many dissimilar forms like vyjdu, vyshla, vyshedshij. The lack of morphosyntactic annotations hinders even simple searches of grammatical relations, for example, searching for uses of the partitive case or for complements of a particular verb in the dative case. Another attempt to develop a comprehensive corpus was made in the Soviet Union in the mid 1980s. It is known as the Computer Fund of Russian Language (CFRL). Its aims were similar to those of the British National Corpus (BNC), which was to be developed a few years later. The main goal was to create a very large corpus of general language and subcorpora for various genres that would 168 Serge Sharoff help in the development of NLP applications. The set of corpora would also provide resources for studying and teaching the Russian language, including development of dictionaries, grammars, textbooks, etc (Andryuschenko, 1989). It was also expected that the corpus would include an historical component to cover the development of the Russian language from the earliest available sources (10th century AD). However, the project did not produce the expected outcome: no representative corpus has been collected. Resources available from the CFRL now include Russian literature of the 19th century and samples of newspapers from 1997. The progress in development of OCR software resulted in multiple ad hoc collections of Russian fiction and reference texts, for instance, Moshkow’s Library (ML), but such collections are not balanced and representative. The same applies to collections of newspapers available online. Currently, corpus studies of Russian are based mostly on the Internet. The Internet can be regarded as the largest Russian corpus, because the amount of Internet documents available for Russian search engines can be estimated at about 250 billion words (1.5 TB of unique texts indexed by Yandex), much larger than any conceivable corpus. However, there are three types of problems that hinder its use for corpus studies. First, it cannot be claimed that the material is representative and that there is a balance of text types. Texts presented on the Russian Internet are chaotic: their set depends on preferences and interests of a very specific group of Russian language speakers that are active on the Internet. The recall of search results also cannot be evaluated, because it depends on unknown parameters: which texts are available or not available on the Internet; which texts available on the Internet were not found by the search engine used for the query, etc. Second, search engines address the needs of information retrieval, rather than linguistic search. Even though search engines provide lemmatisation, so that one can search for all forms of a word, a query cannot be formulated in terms of grammatical features, including tenses, cases or word classes. As for lemmatisation performed by search engines, it is not designed to handle the queries of (corpus) linguists. For example, normal users, who are interested in information retrieval, pay no attention to the aspect of verbs used in their queries and want to get pages corresponding to the verb irrespective of its form. Search engines anticipate the need and index verbs of the perfective and imperfective aspect under one lemma. However, this technique drastically decreases the precision of linguistic searches and leads to some funny results, when pomni used in a query leads to pages with myatyj, because pomyat’ and myat’ form an aspectual pair. Third, search engines present search results in a way that also does not correspond to the needs of a linguist. The pages are ordered in terms of their information rank that has nothing to do with linguistic criteria. The output also Methods and tools for development of the Russian Reference Corpus 169 does not form a concordance, because pages in the output are separated by documents, rather than by contexts of their uses. Finally, search results are based on words occurring in titles of pages or keywords or even in other pages that refer to the link being displayed as relevant. 2. The content of the Russian Reference Corpus From the viewpoint of corpus linguistics, Russian is one of few major world languages that lack a comprehensive corpus of modern language use. However, the need for constructing such a corpus is growing in the corpus linguistics community both in Russia and in the rest of the world. The objective of the project presented in this chapter is to develop the Russian equivalent of the BNC, namely the Russian Reference Corpus (BOKR, BOljshoj Korpus Russkogo yazyka). It is designed as a corpus of 100 million words with the proportional coverage of major varieties of texts in modern Russian, with POS annotation and lemmatisation. The annotation scheme (which is based on the TEI) also marks noun phrases and prepositional phrases, because they are important for the resolution of ambiguity and can be reliably detected. The corpus consists of texts originally written or uttered in Russian by native speakers 2 in recent years. The exact diachronic sample depends on the text type and is discussed below. Table 1: Corpus composition Russian Standard BOKR quantity 10 million words (500 texts) 100 million words (10,000 texts) quality a representative sample of Russian fiction written between 1960 and 2002 a representative corpus of modern Russian, balanced according to a text typology annotation POS tags, morphological and partial syntactic properties with manual disambiguation POS tags, morphological and partial syntactic properties with automatic disambiguation access public Internet access with a query interface shared between the two corpora (Russian Standard is a subcorpus of BOKR) BOKR will include the Russian Standard, a subcorpus of 10 million words of modern fiction representative of the standard literary language. The relationship between the two corpora is described in Table 1. The two corpora differ mostly in their foci: on the large size, wide coverage and the balance of genres in BOKR and on selection of culturally-salient modern literary works and manual disambiguation of morphosyntactic annotations in the Russian Standard. The latter aspect is similar to the design intentions of the hand-corrected core BNC subcorpus (Leech, 1997). The Russian Standard is aimed to be the basic source of information for the development of corpus-based Russian grammars for 170 Serge Sharoff academic and teaching purposes, while BOKR will provide a complementary source of grammatical information and will be the basic source of lexical information. In one respect, the design of the Russian Standard is remarkably different from the design of the core BNC subcorpus. The core BNC is based on a proportional selection of texts from the whole set of the BNC files, while the Russian Standard is based on literary texts. This reflects the difference in the cultural status of the language of imaginative writing in British and Russian cultures: in Russian the literary language is treated as the authoritative source, which effectively defines the language used by native speakers. This fact is also the reason for the higher proportion of fiction in the Uppsala Corpus and the corpus used by Zasorina (1977): fiction texts covered about the half of their content, much higher than the proportion of fiction in the Brown Corpus (25%) and the BNC (17%), cf. also the balance of genres proposed for BOKR in the discussion below. 2.1 The typology of texts The balance of genres in BOKR is based on a text typology that is more sophisticated than that of the BNC. The basic principles for describing texts in BOKR follow the EAGLES guidelines (Sinclair, 1996), which distinguish between text-external (E) and text-internal (I) parameters in text classification: 1. 2. 3. 4. 5. ȿ1 (origin) - parameters concerning the origin of the text, i.e. the creation date, the author’s age and sex, the place of his/her origin, other circumstances of text creation that can affect linguistic parameters; E2 (state) - the appearance of the text, in particular, the distinction between written and spoken text modes (including written-to-be-spoken and electronic communication as the two border cases), and between published sources (books, magazines and newspapers), ephemera and correspondence within the written mode; ȿ3 (aims) - matters concerning the reason for making the text and the intended effect it is expected to have, including (1) the size of the audience (and subclasses for private and public speech) and (2) the communicative function of the text, i.e. discussion, information, recommendation, instruction or recreation. I1 (topic) - the main topic of the text, following a shallow classification of knowledge domains similar to classes used in the BNC, e.g. natural sciences, applied sciences, life or politics; I2 (style) - “the patterns of language that are thought to correlate with external parameters” (Sinclair, 1996), such as formal or informal, one-way or interactive, etc. The changes in the finer classification of parameters in comparison to Sinclair (1996) are based on the experience in development of other representative corpora, such as the Brown Corpus, BNC, and the TEI guidelines (Sperberg- Methods and tools for development of the Russian Reference Corpus 171 McQueen & Burnard, 2001), as well as considerations from Russian texts. This concerns, for example, the use of an additional mode (written-to-be-spoken), which is borrowed from the BNC (E2), the intended audience age (E3.1), a classification of fiction genres (E3.2) and styles (I2). It was considered helpful to extend the classification of text styles with separate subclasses for fiction and non-fiction texts. The patterns of language detected for fiction include the following styles (some better-known writers that often use the style are also indicated): 1. 2. 3. 4. 5. neutral, — the style characteristic for standard literary texts in Russian, regional, derevenskaja proza — an imitation of regional, mostly rural, language varieties, e.g. Astafiev, Rasputin, lowly, snizhennyj — an imitation of the spoken language used by a “lesser educated” population, often slang, e.g. Ju. Aleshkovskij, Limonov, official, socrealism — the official style of the Soviet literature, e.g. Dangulov, Markov, individual, — a marked way of language use with significant deviations from the neutral style, this style is typically the result of linguistic or stylistic experiments, e.g. S. Sokolov. Each style in the list instantiates a specific set of implications on lexicogrammatical properties (with the exception of the individual style, which is often author-specific, but this is exactly the reason to classify a text in this way). Nonfiction is classified according to the following styles: neutral, formal, informal, and academic writing. Since the project is aimed at a representative sample of modern Russian, all meaningful combinations of parameters should be represented in the corpus by at least a handful of texts, though the number of texts in each group depends on the estimated number of respective texts in the Russian discourse and the availability of their electronic copies. Text length is another important technical parameter. It is easier to develop a large corpus using longer texts. However, this means that the corpus contains fewer texts, so an idiosyncratic use of language in each text significantly influences lexicogrammatical properties that can be described using the corpus. This is the reason for the balance of texts of various sizes in the two corpora, i.e. both shorter and longer texts should be included in each category with a greater number of shorter texts to alleviate the influence of longer ones. The intended coverage of knowledge domains (I1) roughly follows the proportion used in the BNC. The comparison is shown in Table 2 (the data are from the BNC Index by David Lee). Since the typology of texts in the BNC is based on other principles, the comparison presents the content of texts in BOKR, as if they were described in terms used by the BNC. For instance, spoken language is treated as a domain in the BNC, so the figures in Table 2 also include it, even 172 Serge Sharoff though a spoken discourse can be devoted to any other topic in the list of domains, so it is described as the mode of speech in BOKR (E2). It would be desirable to increase the proportion of spoken language in BOKR at least to the coverage of the BNC, if not to 50% of the total corpus, but the small amount of available transcribed recordings make the ideal target impractical. The major departure from the BNC is the (already discussed) higher proportion of fiction texts, which are not considered in our scheme as a knowledge domain of its own (similar to the spoken domain), but as the most important component of the knowledge domain “Life” (cf. respective sections in newspapers, which in the Russian context often include short fiction stories). Note that the corpus is currently under development, so the figures in the third column in Table 2 are approximations for the expected coverage. Table 2: The proportion of knowledge domains Domains as in the BNC BNC BOKR Spoken (not a domain in BOKR) 10.7 % 5% Imaginative Texts (Life in BOKR) 16.7 % 30 % Natural Sciences 3.8 % 5% Applied Sciences 7.2 % 10 % Social Sciences 14.2 % 12 % World Affairs (Politics in BOKR) 18.9 % 15 % Commerce 7.6 % 5% Arts 6.8 % 5% Belief/Thought (Religion and philosophy in BOKR) 3.1 % 3% Leisure 11.2 % 10 % Currently tools and techniques for working with BOKR and the Russian Standard are tested using a corpus of 40 million words. Its subcorpus of about 1 million words of fiction texts (corresponding to the Russian Standard) has POS annotations that have been automatically assigned and manually inspected. It is also used for correcting the POS tagger used for processing the larger corpus. It is expected that the final release of the corpus will be available by the end of 2004. 2.2 The methodology for achieving the proportional coverage The costs of compiling a representative corpus now are lower than 10 years ago, when the BNC was collected. Many types of source texts are readily available in electronic form, in particular, fiction and news texts are widely accessible via the Internet and can be legally available for the corpus. Other types of discourse, like business or private correspondence, are harder to obtain and deposit in a corpus Methods and tools for development of the Russian Reference Corpus 173 because of legal obstacles. Yet other types of sources, like samples of spontaneous speech, are rare for technical reasons. The proposed solution is to increase the amount of ephemera (including leaflets, junk mail and typed material), correspondence (business and private) and spoken language samples whenever possible, because they reflect everyday language produced and reproduced regularly in the discourse. Anyway, various types of published texts will take the rest of the share. In this respect, the situation is similar to the early time of the BNC: the amount of texts from unpublished sources in the written part of the BNC is about 4.5%. It is unlikely that in BOKR we will have significantly more; even though the majority of source texts are available in electronic form now, their holders are unwilling to share them. For the reasons of protection of privacy, personal and business letters are subjected to an anonymisation procedure with respect to names of persons and companies. Person names are replaced with MX, FX or CX tags (for male, female or child participants respectively) and names of companies with CoX (X is the identification number of a participant in the text; the same practice is also used in the Bank of English). In some cases, text providers manually replace names with codes. In other cases, they provide original texts, but when texts are stored in the corpus, names are replaced automatically using the lists of known given names and surnames of persons and names of companies. Care has been taken so that names of prominent figures and characters from popular books and films have not been replaced, for instance, even though Karamazoff and Putin are valid Russian names, it is much more likely they are not participants in the exchange, so their names are left as they are in texts (given that the corpus lacks private letters from or to prominent figures). We understand that the anonymisation procedure is not completely satisfactory, cf. analysis in (Rock, 2001). First, it does not lead to complete anonymity: contextual clues are left in texts and allow the detection of participants. Second, a text in which names are replaced with codes looks less natural. Third, errors in setting identification numbers of participants are possible; they can lead to problems in discourse studies based on such texts. Finally, the anonymisation removes the possibility of studying the frequency of personal names, discourse patterns of their uses, as well as phonological patterns. However, we regard this as the best possible practice for storing private and business letters without violating the privacy of their authors and addressees. The text description framework is much more elaborate than a list of domains, so the balance of texts should be achieved on the basis of the text typology described above. The typology can be represented in terms of the systemic network of interrelated choices (Martin, 1987). For instance, when a text is described as fiction, it can be described in terms of the style of fiction, such as, stylistically neutral, low or regional, and in terms of the genre of fiction, such as general, historical, science fiction, etc., but not in terms of the interaction between the 174 Serge Sharoff author and the audience, because it is not produced spontaneously in the presence of the audience. The network of options is traversed using the Systemic Coder. A person that encodes metatextual information about a document has access to its record, including the author, title, year of creation, location of the file and its size in words. Encoding options are selected from a list of categories, for instance, the age of the intended audience is selected from adult, child, teen or x-age, and the age of the author at the time of text creation is selected from child, teen, young, mid, senior (mid-aged authors is the broadest category that covers ages 22 to 55). Even though the typology is elaborate, experience shows that most texts can be described in just a few seconds. Some combinations of features are logically impossible, for instance, a personal letter aimed at a very large audience or a private discussion on TV. Some other combinations are very unlikely, for instance, books written in the domain of natural sciences in formal style, aimed at a very large female audience for entertainment: the combination of formal style and entertainment or natural sciences and a sex-targeted audience is unlikely. However, if a combination of parameters is meaningful, an effort should be made to cover it in the corpus by, at least, several texts. As an example, consider the set of texts within the knowledge domain “Politics”, subdomain “home affairs” (the parameter I1 in our typology). Variation over other parameters involves selection of texts written in neutral, formal, academic and informal styles (I2), texts created by male or female authors or texts with corporate authorship, texts written within the period of 1990-2000 in different regions of Russia (E1), texts printed in newspapers, magazines, or books, as well as letters and reports, or spoken discussions, on site, on TV and radio (E2), texts aimed at different audiences (general vs. informed vs. professional, public vs. private, etc.), and aimed at various communicative functions, e.g. discussion, recommendation, instruction or entertainment (E3). Each text in the corpus should be described by this set of options, for instance, the Russian Constitution is a text written in formal style (I2) in 1993 in Moscow, the authorship is corporate (E1), it is written material printed as a book of 9,500 words (E2), aimed at a very large audience (even if it is rarely read by the majority of population), with the intended function of recommendation, as a legal document. The typology ensures that every text to be included in BOKR can be described in terms of the parameters listed above. Since texts aimed at more public audiences are easier to obtain, extra efforts are taken to cover texts aimed at more private audiences. The text collection activity could lead to a corpus significantly larger than 100 million words. The next step is to balance the collection. The balance takes into account the proportion of basic genres according to Table 2, as well as the proportion of texts within each parameter. For instance, the classes of the intended outcome of a text include discussion, recommendation, instruction, Methods and tools for development of the Russian Reference Corpus 175 recreation and information. The exact proportion of intended outcomes in the corpus can hardly be determined, e.g. the allocation of 25% for discussions or 10% for instructions looks fairly arbitrary. However, if texts classified as a recommendation take 90% of the total corpus, this is a clear sign of disproportionate coverage, which should be corrected. The balance will be monitored using statistical tools available in the Systemic Coder, such as Cell Analysis or Significance Tests. Another problem with sources concerns the choice of diachronic sampling, because the turbulent history of Russia in the 20th century radically affected the language. For instance, according to the frequency list (Zasorina, 1977), which was compiled on the basis of texts from 1930-1960, such words as sovetskij (Soviet) and tovarishch (comrade) belonged to the first hundred of Russian words on a par with function words, but this is no longer true of modern texts. The language of fiction has not been so radically affected. The decisions on the chronological limits of the study are different for different text types, for instance, fiction texts are taken from 1960, scientific texts from 1980, political texts and ephemera from 1990 (earlier ephemera texts are also hard to obtain), while news texts from 1997. 3. The principles of morphosyntactic annotation English corpora, including the BNC, are annotated with complex tags, like NNS for plural common nouns – a technique that is impossible in a highly inflective language, such as Russian. For instance, an adjective inflects for case, gender and number, giving 36 basic adjectival categories in total, while a verb in addition to its own 14 basic categories has up to 4 participles, each of which declines for adjectival categories. This leads to thousands of separate tags that cannot be searched effectively. For this reason, each word is annotated with a list of features, which provides a unique identification of their type according to the morphosyntactic codes from EAGLES (Calzolari and McNaught, 1996), for instance, bylo (was) is annotated as “verb,ifve,int,act: n,sg,past”, which stands for verb, imperfective, intransitive, active voice (the features describe the lexical item byt’), followed by a colon with features that describe the form: neutral gender, singular number, past tense. Separate features from a feature bundle associated with each word can be selected in a window of the query interface (Figure 1). 176 Serge Sharoff Figure 1: The query interface 3.1 How to resolve the ambiguity The POS class and the morphological properties of a word are reflected in the flexion and the probability of deciding the lemma and the POS class from a word form is higher than in, for instance, English. As a result, there are many morphological analysers for Russian, which make decisions on the basis of word forms, but virtually no Russian taggers that take into account the local context. However, if real Russian texts are to be tagged with the intention of making a corpus that includes lemmatisation and morphosyntactic annotation, the level of ambiguity is high: very frequent word forms like stali, shli, or ego can correspond to several lemmas, i.e. stali – stat’ (verb, to become) vs. stal’ (noun, steel), shli – idti (to go) vs. slat’ (to send), ego – ego (possessive) vs. on vs. ono (both are personal pronouns). The ambiguity of POS classes is relatively rare, but many noun forms have multiple readings, for instance, the word form pole is an instance of four different nouns pol (floor), pole (field) and pola (lap) and Polya (a person name, when it is capitalised). Also, in many cases the ambiguity concerns the set of morphological features of the same lemma, e.g. knigi is the singular form in the genitive case or the plural form in the nominative or accusative case, while znakomoj is the singular form in the genitive, dative or prepositional case of either a noun or an adjective. According to initial experiments, the frequency of ambiguous detection Methods and tools for development of the Russian Reference Corpus 177 of lemmas and POS classes in running text is about 25%, while the frequency of ambiguous morphological properties is about 55%. The values are too high for a corpus with morphosyntactic annotations and so the ambiguity should be reduced. Currently, there are no tools available that allow reliable parsing in a corpus of this size for Russian or any other language. Use of language-independent POS taggers based on statistical models cannot improve the quality of the output, because they are typically based on considerations relating to the word order, which is flexible in Russian, and because the genuine ambiguity between POS classes is relatively rare, and the most frequent type of the ambiguity concerns different readings of a word form (cf. the example with pole) and morphological features (cf. the example with knigi). If word forms or sets of morphological features (e.g. plural+dative+feminine) are treated as POS classes, then their number increases and the quality of language-independent POS tagging declines. However, partial parsing that detects nominal and prepositional phrases is reliable enough and can be used for deciding the reading of ambiguous forms. In BOKR and Russian Standard we use Dialing, a morphological analyser with simple mechanisms for syntactic and semantic analysis. Since two analyses of the same word form have distinct morphological properties (case, number and gender), the agreement of the noun and the adjective in noun phrases removes some types of ambiguity, e.g. the word combination znakomoj knigi from Otkroem stranitsy etoj horosho znakomoj knigi (Let’s open pages of this wellknown book) can be parsed only as the genitive singular form of both words, and the first word is an adjective. Another simple mechanism that requires only partial parsing is the agreement of the subject and the predicate; it can remove, for instance, the plural reading of a noun in the nominative case (knigi), when the predicate is singular, as in Knigi na polke ne bylo. Some types of ambiguity of lemmas and POS classes are left after partial parsing: 12% of forms are ambiguous with respect to lemmas, and 22% with respect to morphological properties. Since the BOKR corpus is annotated without human intervention, ambiguous analyses are subjected to further filtering according to statistical heuristics. For instance, two nouns spina (back) and spin (spin) have several identical word forms, which cannot be separated by means of parsing. However, spin is a term in theoretical physics, so the reading can be ignored in normal texts. Few other cases resist even complete parsing and statistical consideration, for instance, the ambiguity between the two readings of the word form banke (bank vs. banka) in Xranite svoi denjgi v banke (keep your money in a bank/in a jar) can be resolved only on the basis of semantic and pragmatic constraints. Such cases of ambiguity are retained in BOKR. The same applies to the ambiguity of morphological properties left after the syntactic filter. Currently, 3.6% of forms remain with ambiguous lemmas. The ambiguity in the Russian Standard is corrected manually. 178 3.2 Serge Sharoff How to store annotations The design of the annotation format of the two corpora follows the best practices in corpus development established in the 1990s, namely EAGLES (European Advisory Group on Language Engineering Standards) and TEI (Text Encoding Initiative). Even though XCES (XML Corpus Encoding Standard) is expected to become the international standard for language resources (Ide and Romary, 2002), the annotation scheme based on it is extremely verbose (the size of an annotated file is a hundred times larger than a plain text file and, for a corpus of 100 mln words, the size really matters). The XCES scheme is also not suited for querying word uses in the corpus, because information on similar properties is represented at different levels of the XML structure. Thus, we have postponed the use of XCES until the standard is established and there are publicly available software tools for working with the format. The format of BOKR is based on the TEI scheme and uses standard tags, like <phr>, <s>, <w> for representing phrases, sentences and words. Morphological annotation is stored in <ana> tags that describe word properties in lemma and feats attributes. Ambiguity is represented using multiple <ana> tags. The following is an example from the beginning of a sentence: Mne bylo ochen’ zhalko svoih chasov, … (I was very sorry about loosing my watch, …) <s id="kozlotur.1476"> <w n="1">Ɇɧɟ<ana lemma="ɹ" feats="pron,sg,1: dat"/></w> <w n="2">ɛɵɥɨ<ana lemma="ɛɵɬɶ" feats="verb,ifve,int,act: n,sg,past "/></w> <phr type="ADV+ADV"> <w n="3">ɨɱɟɧɶ<ana lemma="ɨɱɟɧɶ" feats="adv"/></w> <w n="4">ɠɚɥɤɨ<ana lemma="ɠɚɥɤɨ" feats="adv"/></w> </phr> <phr type="ADJ+NOUN"> <w n="5">ɫɜɨɢɯ<ana lemma="ɫɜɨɣ" feats="pron,poss: pl,gen"/></w> <w n="6">ɱɚɫɨɜ<ana lemma="ɱɚɫ" feats="noun,m,in: pl,gen"/> <ana lemma="ɱɚɫɵ" feats="noun,pl: gen"/></w> </phr> </s> The parser was able to resolve the ambiguity between two analyses of mne (the dative or prepositional case), svoikh (the genitive, accusative or prepositional case of either a personal pronoun or a possessive pronoun), and zhalko (an adverb or an adjective). However, the ambiguity between two readings of the word form <w n=“6”> is left in the output in the two <ana> tags. It can be read as chas (an hour) and chasy (a watch), in the latter case it is pluralia tantum, which is reflected in the position of the number value before the colon. Metainformation about a document is stored in the header and is also based on the TEI format. In some cases, TEI provides tags for encoding the contextual Methods and tools for development of the Russian Reference Corpus 179 settings used in the text typology, for instance, <creation> or <textClass>. In other cases, the information on the expected outcome or the size of the audience is expressed using the general framework of taxonomy specifications by means of <catRef> (category reference) tags. Notes 1 The research presented in the paper has been supported by the Alexander von Humboldt Foundation, Germany, when the author was affiliated with the University of Bielefeld and the Russian Research Institute for Artificial Intelligence. I am grateful to my Russian colleagues, in particular, to Vladimir Plungian and Katia Rakhilina, who took the leadership in the ongoing development of the Russian Reference Corpus. 2 Development of a translation corpus is considered to be a separate task. References Andryuschenko, V.M. (1989). Konzepziya i arhitectura Mashinnogo fonda russkogo jazyka (The concept and design of the Computer Fund of Russian Language), Moskva: Nauka, 1989 Calzolari, N., McNaught, J. (eds.) (1996). Synopsis and Comparison of Morphosyntactic Phenomena Encoded in Lexicons and Corpora. EAGLES document EAG-CLWG-MORPHSYN/R http://www.ilc.cnr.it/EAGLES96/morphsyn/morphsyn.html Ide, N., Romary, L. (2002). Standards for language resources. In Proc. of Language Resources and Evaluation Conference (LREC02). May, 2002, Las Palmas, Spain. 59-65. Leech, G. (1997). A brief users’ guide to the grammatical tagging of the British National Corpus, UCREL, Lancaster University. http://www.hcu.ox.ac.uk/BNC/what/gramtag.html Lönngren, Lennart (ed.) (1993). Chastotnyj slovar’ sovremennogo russkogo jazyka. (A Frequency Dictionary of Modern Russian. With a Summary in English.) Acta Universitatis Upsaliensis, Studia Slavica Upsaliensia 32. 188 pp. Uppsala. 180 Serge Sharoff Martin, J.R. (1987). The meaning of features in systemic linguistics. In M.A.K. Halliday, R.P. Fawcett (eds.) New Developments in Systemic Linguistics. Vol. 1. London: Pinter Publishers. 14-40. Rock, F. (2001). Policy and practice in the anonymisation of linguistic data. International Journal of Corpus Linguistics, 6(1). Sinclair, J. (1996). Preliminary recommendations on text typology. EAGLES Document EAG-TCWG-TTYP/P. http://www.ilc.pi.cnr.it/EAGLES96/texttyp/texttyp.html Sperberg-McQueen, C. M., Burnard, L. (eds.) (2001). Guidelines for Electronic Text Encoding and Interchange. http://www.hcu.ox.ac.uk/TEI/P4X/index.html Verbitskaya, L.A., Kazanskij, N.N., Kassevich, V.B., (forthcoming). Nekotorye problemy sozdanija natsional’nogo korpusa russkogo jazyka. NTI, Series 2. (in Russian) Zasorina, L.N. (ed.) (1977). Chastotnyj slovar’ russkogo jazyka. Moscow: Russkij Jazyk. Internet links BNC Index: http://www.comp.lancs.ac.uk/ucrel/bncindex/ BOKR: Boljshoj Korpus Russkogo yazyka (the Russian Reference Corpus, a description of the project), http://bokrcorpora.narod.ru/ CFRL: the Computer Fund of Russian Language, http://irlras-cfrl.rema.ru/ Coder, a markup and classification tool: http://www.wagsoft.com/Coder/ Dialing, the morphological analyser: http://www.aot.ru/download.htm Moshkow’s Library: http://lib.ru/ RS: the Russian Standard (online access), http://corpora.yandex.ru/ UC: the Uppsala Corpus, available from the University of Tübingen, http://www.sfb441.uni-tuebingen.de/b1/en/korpora.html Yandex, the search engine: http://www.yandex.ru/ A profile-based calculation of region and register variation: the synchronic and diachronic status of the two main national varieties of Dutch Dirk Speelman, Stefan Grondelaers and Dirk Geeraerts University of Leuven Abstract In this paper we present a profile-based method for analysing regional and register variation, and we explain how we apply it to the study of Belgian and Netherlandic Dutch. A ‘profile’ comprises frequency information about the set of alternative synonymous terms for naming a particular concept. Our comparison of subcorpora is based on differences in naming preferences, as revealed by such profiles. Apart from the actual algorithm for the comparison of subcorpora this paper also addresses some practical issues. In particular the preliminary step of profile selection is addressed: in this context the concept of ‘stable lexical markers’ is introduced. 1. Introduction 1.1 Broader context of the study 1 In recent years our research unit has investigated the internal stratification of Belgian and Netherlandic Dutch. Although officially the same language, these two national varieties do differ to some degree, especially, but not exclusively, in the context of more informal language use. One important difference in the evolution of both varieties lies in their standardisation: due to a history of foreign occupations Belgian Dutch has known an interrupted, and therefore slower standardisation process - which may even be considered to be not fully completed yet, in the sense that the standard variety is not that well established and by many speakers is not felt to be the natural variety of choice in all that many different situations. The overall model of the current internal stratification of Belgian and Netherlandic Dutch that we take as our point of departure can be summarised in one synchronic hypothesis and one diachronic hypothesis. The synchronic hypothesis is that if we look at different registers within the two contemporary national varieties, we will see that the linguistic difference between language use in less formal and in more formal registers is larger in Belgian Dutch than in Netherlandic Dutch. In other words, if we compare situations where informal language use can be expected to situations where more formal language use can be expected, and we repeat this exercise for Belgium and for The Netherlands 182 Dirk Speelman, Stefan Grondelaers and Dirk Geeraerts (looking at the same situations in both cases), we will measure more important differences in Belgium. The diachronic hypothesis is that in the last decades, i.e. in the second half of the twentieth century, Belgian Dutch has been moving closer to Netherlandic Dutch and in doing so has been using the latter as its reference point for (further) standardisation. In Geeraerts, Grondelaers and Speelman (1999) we report on a study in which a first body of empirical evidence was found to confirm both the synchronic assumption of a different internal stratification, with Belgian Dutch having a more outspoken informal register, and the diachronic assumption of a convergence characterised by Belgian Dutch moving towards Netherlandic Dutch. The dataset used in this study consists of some 40000 naming events of items related to clothing and football concepts. With respect to the synchronic hypothesis the naming of clothing terms was investigated for two different registers. The more formal situation was the naming of garments in magazines. The more informal situation was the naming of garments in shop windows (which can be expected to target a more local audience and hence to be more informal). The data used to investigate the diachronic hypothesis were the football terms and clothing terms encountered in Belgian journals and magazines form three different points in time (1950 - 1970 - 1990). In subsequent research, a dedicated corpus was compiled to replicate this study for other registers and other linguistic variables. This corpus is the so-called ConDiv corpus, a 40 million token corpus created for measuring convergence and divergence patterns (cf. Grondelaers, Deygers, van Aken, van den Heede and Speelman, 2000). The corpus consists of a diachronic component, which contains newspaper material from 1950, 1970 and 1990, and a more elaborate synchronic component from 1990, which contains texts from 5 different registers: quality newspapers, nation-wide popular newspapers, regional popular newspapers, texts from Usenet newsgroups, and texts from IRC chat channels. Several replication studies based on this corpus corroborated the original findings from Geeraerts, Grondelaers and Speelman (1999). For instance, Grondelaers, van Aken, Speelman and Geeraerts (2001) reports on two replication studies, one in which the original clothing terms study was applied to the ConDiv corpus, and another in which variation in the choice of prepositions was used as linguistic evidence. 1.2 Formal onomasiological variation This paper focuses on some key methodological choices made in our investigation of the stratification Belgian and Netherlandic Dutch. The first and most important methodological choice is that among the different types of variation that exist we choose formal onomasiological variation as our basis for measuring linguistic distances. According to the definition we adhere to (cf. Geeraerts, Grondelaers and Bakema, 1994), formal onomasiological variation is that type of onomasiological variation, in other words is that type of variation in the terms one uses to name an item, that is not motivated by semantic differences, and therefore can be said to be purely ‘formal’. An example would be the A profile-based calculation of region and register variation 183 variation between the terms ‘underground’ and ‘subway’ to refer to the concept UNDERGROUND. An example of onomasiological variation that is not purely formal would be the variation between the terms ‘garment’ and ‘pants’ to refer to a pair of trousers. In the latter example the choice for a different term is related to a different conceptualisation of the thing that is referred to, which is not the case in the UNDERGROUND example. The motivation for our choice to measure linguistic distance on the basis of formal onomasiological variation is not that this type of variation is more important than other types of variation, but rather that it exceptionally useful for detecting the type of regional and stylistic differences we are interested in our study. The structure of the remainder of this paper is as follows. In section 2 we introduce profiles and profile-based calculations of linguistic differences as our primary technique for quantifying formal onomasiological variation. After explaining the actual calculations we focus on the most salient feature of this technique, which is the use of profiles. We argue that profiles are a straightforward technique for measuring regional and stylistic differences, while controlling other types of variation. We subsequently illustrate the different outcomes of actual profile-based calculations and non-profile-based calculations (the latter being exemplified by calculations based on key words). The calculations are applied to the synchronic Belgian Dutch part of the ConDiv corpus. In section 3 we introduce a complementary technique, the method of ‘stable lexical markers’, which is a technique for exploring and charting a more complete set of noteworthy, different types of variation in the comparison of corpora. Rather than neutralising certain types of variation in advance by our choice of technique (which is what we do in section 2), we now aim to obtain an overview of the different types of variation, of their relative importance, and of the actual linguistic variables (i.e. the actual terms) that are representative of these different types of variation. In section 4 we summarise the most important characteristics of profile-based calculations and of the technique of stable lexical markers. 2. Profile-based register analysis of Belgian Dutch 2.1 Profiles A profile is an exhaustive set of synonymous terms for naming the same concept, together with information about the frequency with which each term is used to name the concept 2 . Table 1 shows three profiles, more specifically the profiles for the concept UNCLE in three different subcorpora of the Belgian Dutch (B) synchronic part of the ConDiv corpus. The subcorpora contain chat material (ircVL), material from a regional popular newspaper (regL1) and material from a quality newspaper (quaSR) respectively. In this case there are two different terms 184 Dirk Speelman, Stefan Grondelaers and Dirk Geeraerts for referring to the concept at hand, namely ‘oom’ and ‘nonkel’ (note that in other profiles there are often more than two alternatives). Table 1: The profiles for UNCLE in three different subcorpora of the ConDiv corpus Oom Nonkel ircVL (B) regL1 (B) quaSR (B) 0 (0 %) 55 (100 %) 25 (61 %) 16 (39 %) 15 (94 %) 1 (6 %) . This example illustrates how profiles are sensitive to stylistic variation. The term ‘nonkel’ is a somewhat more colloquial term to refer to UNCLE. The term ‘oom’ is the more formal, standard variant. As can be seen in the table, the chat subcorpus shows an outspoken preference for the colloquial variant ‘nonkel’ (all 55 occurrences), the regional paper shows a moderate preference for the standard variant ‘oom’ (25 out of 41 occurrences), and the quality paper shows an outspoken preference for the standard variant ‘oom’ (15 out of 16 occurrences). It is obvious that in a similar vain profiles are sensitive to regional variation, but we will not give examples of this because in this text we will only deal with Belgian material. Linguistic distance between two corpora, on the basis of their profiles for one particular concept is calculated as a city block distance on the basis of the relative frequencies 3 . For example, the distance between regL1 and quaSR, on the basis of their UNCLE-profiles (cf. Table 1), is calculated as 0.5 * (|0.61 – 0.94| + |0.39 – 0.06|). In other words, we obtain the linguistic distance by summing up the row-wise differences in percentages, and by subsequently dividing that sum by two. The division by two is a normalisation step that guarantees the result is a value between 0 and 1 (i.e. between 0% and 100%). After calculating the distances between two subcorpora on the basis of individual concepts, we want to summarise this information in a global distance between the two subcorpora. Linguistic distance between two corpora on the basis of their profiles for a whole range of concepts is calculated as the average of the distances based on a single concept 4 . After calculating these global distances for all couples of subcorpora, the final step is to feed these distances into a multidimensional scaling analysis, in which we try to plot all subcorpora on a low-dimensional space in such a way that distances in the plot reflect linguistic distances as well as possible. We illustrate the procedure by means of an example. The 20 subcorpora we compare in the example are listed in Table 2. They are all part of the Belgian Dutch synchronic part of the ConDiv corpus. This section of the ConDiv-corpus consists of material from one quality newspaper (qua), one national popular newspaper (nat), two regional popular newspapers (regL and A profile-based calculation of region and register variation 185 regA) and several Usenet newsgroups (use) and IRC chat channels (irc). All of these categories are further subdivided, sometimes with (irc, use, nat, qua) and sometimes without (regL, regA) a further topic-wise differentiation. Table 2: The 20 subcorpora of the synchronic Belgian Dutch part of the ConDiv corpus Name nr tokens register topic IrcRE IrcVA IrcLE IrcVL IrcBE UseTE UseSP UseSR regL1 regL2 regL3 regA1 regA2 regA3 NatRE NatSP NatSR QuaSP QuaTE QuaSR 205560 1182849 1784084 2736111 1686571 2486797 117195 2376788 1561362 1450968 1666916 1563799 1504606 1810548 1945461 427280 518670 994867 1431786 3607513 IRC-material IRC-material IRC-material IRC-material IRC-material Usenet Usenet Usenet regional popular newspaper regional popular newspaper regional popular newspaper regional popular newspaper regional popular newspaper regional popular newspaper national popular newspaper national popular newspaper national popular newspaper national quality newspaper national quality newspaper national quality newspaper regional chat channels varia chat channel “Leuven” chat channel “Flanders” chat channel “Belgium” technical topics sports supra-regional topics (no differentiation) (no differentiation) (no differentiation) (no differentiation) (no differentiation) (no differentiation) regional sports supra-regional interest sports technical topics supra-regional topics Table 3: Terms and concepts for 10 different sorts of profiles CONCEPT Terms A MIND TO IF FOR THE PRICE OF (+ amount) UNCLE IN A MOMENT ONCE AGAIN TO BE READY EACH OTHER MOMENT TO CONTRIBUTE TO goesting, zin als, indien aan (+ amount), tegen (+ amount), voor (+ amount) oom, nonkel seffens, dadelijk weeral, alweer gereed zijn, klaar zijn elkaar, mekaar, elkander, mekander moment, ogenblik bijdragen aan, bijdragen bij, bijdragen tot The terms and concepts of the 10 different profiles type we use in the example are listed in Table 3. The result of the MDS-analysis is shown in Figure 1. One can see an axis IRC-Usenet-newspapers running from left to right. Moreover within the newspaper corpora most subcorpora that belong to the same newspaper are rather close to each-other (two notable exceptions are ‘natSP’ and ‘quaTE’). In summary, the recognition of the registers is not impeccable, but seems acceptable in general. 186 Dirk Speelman, Stefan Grondelaers and Dirk Geeraerts Figure 1: Example of MDS-plot based on profile-based linguistic distance 2.2 Profiles compared to key words An obvious alternative to working with profiles would be to calculate differences on the basis of isolated individual terms. In fact, this is common practice, and many methods that are currently used are based on individual terms (for an overview, see Kilgarriff, forthcoming). The main reason why we refrain from doing this is that we want to avoid thematic bias in our measurements. Let us illustrate what we mean by this with an example. Suppose that in corpus A ‘oom’ and ‘nonkel’ are used 20 times each. Also suppose that in corpus B ‘oom’ and ‘nonkel’ are used 100 times each. And let us, for the sake of simplicity, assume the corpora are equal in size. Now, if we would compare the corpora on the basis of isolated individual terms, there is a clear risk of misinterpreting the data. One approach would be to look only for items that are known to be colloquial, and to count their occurrences. In this case the colloquial item is ‘nonkel’, so we would A profile-based calculation of region and register variation 187 count the number of occurrences of ‘nonkel’. We would conclude that corpus B contains more colloquial material than corpus A. Another approach would be to list all items that are more typical of one corpus than of the other. If we would do that, we would conclude that both ‘oom’ and ‘nonkel’ are more typical of B than of A. However, if we now move to profiles, we can see that the actual difference between A and B lies in the frequency of the concept UNCLE, and that the actual naming preferences are identical in both corpora (in both cases 50% ‘oom’ and 50% ‘nonkel’). So for our purpose, which is the investigation of regional and stylistic differences, we want to measure no difference in this example. To put it another way, the high frequency of ‘oom’ and ‘nonkel’ in corpus B is ambiguous in terms of popularity of terms and popularity of concepts. The profile-based method disambiguates these two levels 5 . Figure 2: MDS-plot based on number of key words In Figure 2 we give an example of a comparison of corpora based on individual isolated terms. We used the log likelihood ratio test for binomial distributions (cf. 188 Dirk Speelman, Stefan Grondelaers and Dirk Geeraerts Dunning, 1993) to calculate which terms, in the comparison of two corpora A and B are used significantly more often in A (positive key words 6 of A) or significantly less often in A (negative key words of A). We then used the total sum of key words (both positive and negative) as a dissimilarity measure, and compared all corpora pairwise 7 . Finally, we fed these dissimilarities to a MDSanalysis. The result is shown in Figure 2. Apart from the fact that the axis IRCUsenet-newspapers is a bit more fuzzy 8 than in Figure 1, the most interesting observation is that the three sports-related corpora (in the centre of the figure) are very close to each-other, in spite of their belonging to different registers. This seems to illustrate that this type of measurement is so sensitive to topical bias that it becomes less useful for our purpose, i.e. the measurement of regional and stylistic variation. 3. Stable lexical markers 3.1 Purpose of using stable lexical markers Although the plot in Figure 2 indicates that the key words method as applied in 2.2 is less suited to our needs, we believe that a further analysis of the results of this method is worthwhile. Being based on all terms in the corpora, this method contains a wealth of information. If we can obtain a clearer picture of the different sources of variation (such as style, topic, etc.) that interact in the constellation we see in Figure 2, and if we can obtain information about their relative importance, and if we can also discover which linguistic variables (i.e. which terms) are representative of these different sources of variation, then there are several possible ways to benefit from this knowledge in our analysis, both directly and indirectly. Directly, because it helps us to construct a representative set of profiles for measuring regional and stylistic variation 9 . Indirectly, because it serves as an additional autonomous analysis of variation in the corpora. Unfortunately a direct analysis of the key words is cumbersome, because in each of the 190 pair-wise comparisons of corpora thousands or even tens of thousands of key words show up 10 . In order to overcome this practical problem 11 , we want to simplify the picture and look at a subset of key words that according to some criteria can be called a particularly salient subset. The criterion we use is consistency in being typical of specific groups of corpora. More specifically, we introduce the concept of stable lexical markers or stability as a marker. The concept is straightforward 12 . If we compare two sets of corpora S1 and S2, then the stability with which a term is a marker for set A is equal to the number of couples {X, Y}, with X a member of S1 and Y a member of S2, for which the term is a positive key word for X. For instance, if S1 contains three corpora, say S1 is {regL1, regL2, regL3}, and S2 also contains three corpora, say S2 is {regA1, A profile-based calculation of region and register variation 189 regA2, regA3}, then there are 9 different couples {X, Y} with X a member of S1 and Y a member of S2 , so the maximum stability would be 9. 3.2 Applications of stable lexical markers We illustrate two ways the concept of stable markers can be applied. In a first application we consider all partitions of our set of 20 corpora into two subsets S1 and S2 with subset S1 having size 3 and subset S2 having size 17. There are 1140 such partitions. We sort this list by number of maximally stable markers of S1. The top of this list is shown in Table 4. Right after the last item displayed here the number of maximally stable markers drops to 18, and then the number rapidly declines to hardly any maximally stable markers. The reason we look at subsets of size 3 is that most groups of corpora with the same source type are of size 3, with the exception of IRC, of which there are 5 items. As we see in Table 4, all these groups appear at the top of the list. Additionally, three mixed-group subsets also appear. Those are the ones with the grey background. Table 4: all subsets of size 3 with 30 or more maximally stable markers nr maximally stable markers subset S1 1 202 {natRE, natSP, quaSP} 2 187 {ircRE, ircVL, ircBE} 3 186 {regA1, regA2, regA3} 4 185 {useTE, useSP, useSR} 5 74 {regL1, regL2, regL3} 6 66 {regL3, quaTE, quaSR} 7 55 {natRE, natSP, natSU} 8 54 {ircRE, ircVA, ircBE} 9 48 {ircVA, ircLE, ircVL} 10 36 {quaSP, quaTE, quaSR} 11 34 {ircRE, ircLE, ircBE} 12 30 {natSP, regA3, quaSP} The second application of stable makers then of course is to look at the actual list of markers. As a first example, we look at the set of maximally stable markers of {natRE, natSP, quaSP}, the triplet at position 1 in Table 4. In Table 5 we show the 20 13 first items from the list of 202 maximally stable markers. It is clear that the lexicon that sets this triplet apart is topic-related – more precisely sportsrelated. Applying the same procedure we found that the same is true for triples 6 and 12 in Table 4. The markers in triplet 6 are related to ‘hot news’ topics. The markers in triplet 12 are once again sports-related. 190 Dirk Speelman, Stefan Grondelaers and Dirk Geeraerts Table 5: The first 20 items in the list of maximally stable markers of {natRE, natSP, quaSP} aanvoerder (captain), achillespees (Achilles tendon), affluiten (to whistle), aftrap (kick-off), beker (cup), beloften (junior team), beslissende (decisive), bezoekende (visiting), coach (coach), competitie (competition), competitiestart (competition start), dij (thigh), doelkansen (opportunities to score), doelman (keeper), doelpunt (goal), doelpunten (goals), eindstand (final score), finales (finales), fit (in shape), forfait (walk over), … As a second example of the second application, we look at one very specific subset S1 of size 8; we look at the subset S1 of all 8 computer mediated communication subcorpora, and compare it to the subset S2 of all 12 newspaper subcorpora. The results are shown in Table 6. Maximum stability is 96 in this case. There are 158 maximally stable markers, which we have manually classified in different categories in order to distinguish between different sources or types of variation. The categories, most of which correspond to dimensions of variation as described in Biber (1995), are shown in the first column. The second column shows some examples for each category. The examples are taken from the set of 279 terms that have a stability of 92 of more (i.e. that are at least 95% stable). The last column shows the absolute and relative frequencies of the different categories for the 158 fully stable terms (assigning each term to exactly one category). Table 6: Stable lexical markers of the computer mediated communication corpora category examples of 95% stable markers nr of 100% stable markers 1st + 2nd person jij (you), uw (your), ik (I), denk (think, 1st person singular), bent (are, 2nd person singular), … after, again, sorry, stuff, … chat (chat), email (email), comp (computer), … snap (understand, 1st person singular), tis (it is), niks (nothing), ne (a), zijt (are, 2nd person singular), … groetjes (greetings), greetz (greetings), bedankt (thanks), hmm (hmm), oeps (oups), ... redelijk (reasonably), … 1 (0.63%) English thematic specificity colloquial style conversational elements downtoners 94 (59.49%) 47 (29.75%) 4 (2.53%) 11 (6.96%) 1 (0.63%) At first sight, one would conclude that the type of variation that interests us most, that of stylistic variation (colloquial style), is relatively rare: 2.53% of the cases, as opposed to 59.49% for English, 29.75% for thematic specificity (i.e. topic) and A profile-based calculation of region and register variation 191 6.96% for conversational elements. However, it is obvious that assigning each term to one single category is an oversimplification. For instance, the examples ‘greetz’ in conversational elements is also related to English. Also, the colloquial style examples ‘snap’ and ‘zijt’ also are examples of 1st + 2nd person. In a similar way we find items that are not classified as colloquial style but that are related to colloquial style, for instance ‘comp’ and ‘greetz’. In summary, sources of variation are often hard to separate, and more in particular, although purely stylistic variation is rare in these stable markers, stylistic variation is often superimposed on other types of variation. 4. Conclusions In this chapter we have argued that key-word-based methods are less suited for the analysis of stylistic and regional variation, because they are too sensitive to topical bias. We further argued that profile-based calculations are more suited because they neutralise this topical bias. In order to illustrate this we have used both methods to calculate linguistic distance between 20 subcorpora that represent different registers of Belgian Dutch, but some of which are further differentiated by topic. The analysis confirmed that the key word method is too sensitive to topical bias. Next we introduced the concept of stable lexical markers, and showed two applications in which stable markers are used to discover noteworthy patterns in the multitude of information that resides in a classical keyword-based analysis, but is hard to be extracted from it. On the basis of these applications we illustrated that stable markers are a useful device to disentangle different sources of variation, to learn about the relative importance of these different sources, and to discover which terms are representative of these different sources. Notes 1 The research reported on in this paper was supported by VNC-grant 205.41.07 3 as well as by OT-project OT 01/05. For the corpus-linguistic procedures and some of the statistical analyses the tool Abundantia Verborum was used [http://wwwling.arts.kuleuven.ac.be/genling/abundant]. For the multidimensional scaling analyses the environment R was used [http://www.r-project.org/]. 2 This definition of profiles clearly targets lexical variation. It is possible to broaden the definition in such a way that other types of variation (morphological, syntactic, ..) also fit in. Profiles would then group together 192 Dirk Speelman, Stefan Grondelaers and Dirk Geeraerts competing constructions. But since in this paper we only deal with lexical information, we will stick to the more restricted definition. 3 In practice there is one additional step. We first apply a log likelihood ratio test for multinomial distributions (cf. Dunning 1993) to test whether the two profiles are significantly different from each other. If they are, we use city block distance to calculate linguistic distance. If they are not, we use (a constant very close to) zero as linguistic distance. 4 In practice there is one more subtlety. Rather than simply averaging over the concept-specific distances, we use a weighted sum of the conceptspecific distances. The weight of a concept-specific distance is determined by the sum of al naming events in the union of the two corpora under comparison that pertain to this concept. The purpose is to give more weight to naming choices for concepts that on average are referred to more frequently. To give an extreme example from our analysis of clothing terms, we want to prevent that variation in the naming of a rarely referred to concept such as BOWLER HAT has as much influence on the global linguistic distance as variation in the naming of a frequently referred to concept such as TROUSERS. 5 In theory, one could claim that such thematic bias should be avoided in the design of the corpus. In practice, however, this is virtually impossible. 6 We borrow the term key words from Scott (1997). Actually, the term is older, but initially the procedure was quite different (cf. Williams, 1976, and, for a discussion, Stubbs, 1996). We use the term key word as it is used in Scott (1997). By a positive key word of corpus A (when compared to B) we mean a key word that, in terms of relative frequency, is used more often in A than in B. By a negative key word we mean a key word that, in terms of relative frequency, is used less often. 7 The pair-wise comparison of all corpora is not the most typical use of key words, which normally involve comparing corpora to a reference corpus. However, we believe that in the context of our design (analysing clusters through MDS) a direct comparison of the corpora is more straightforward. 8 It also should be mentioned that in this MDS-analysis the stress is much higher than in the profile-based analysis. In fact, in the key word based calculations, one actually needs three dimension to obtain an acceptable fit. Nevertheless, we chose to show the two-dimensional plot because the plot is easier to interpret, and because the main observations are not affected by this simplification. For a more detailed analysis of these key word based calculations we refer the reader to Speelman, Grondelaers and Geeraerts (forthcoming). A profile-based calculation of region and register variation 193 9 Currently our compilation of profiles is mainly based on synonym lists that are derived from dictionary information. Automatic or semi-automatic derivation of potentially interesting variables from the corpora serves as a complementary strategy. 10 We applied the method to word forms, not to lemmata. 11 It could be called more than a practical problem. It is often claimed that many methods that are currently used to detect significant differences in the use of a term (most notable chi square based methods), signal a counter-intuitively large number of significant difference. One of the sources of the problem, as mentioned in Kilgarriff (forthcoming), is the distribution of the lexicon: high frequent terms are so frequent, and therefore produce so many observations, that even subtle, often linguistically less relevant or distinctive patterns reach significance. 12 The concept is related to the key key concept (cf. Scott, 1997), but stable markers are more generic, because they involve the comparison of two sets of corpora, and are used for different purposes. 13 After filtering out a few proper names. References Biber, D. (1995), Dimensions in Register Variation, Cambridge University Press. Dunning, T. (1993), ‘Accurate Methods for the Statistics of Surprise and Coincidence’, Computational Linguistics, 19(1): 61-74. Geeraerts, D., S. Grondelaers and P. Bakema (1994). The Structure of Lexical Variation. Meaning, Naming and Context. Berlin: Mouton de Gruyter. Geeraerts, D., S. Grondelaers and D. Speelman (1999). Convergentie en divergentie in de Nederlandse woordenschat. Een onderzoek naar kledingen voetbaltermen. Amsterdam: Meertensinstituut. Grondelaers S., K. Deygers, H. van Aken, V. van den Heede and D. Speelman (2000), ‘Het ConDiv-corpus geschreven Nederlands’, Nederlandse Taalkunde, 5: 356-363. Grondelaers S., H. van Aken, D. Speelman and D. Geeraerts (2001), ‘Inhoudswoorden en preposities als standaardiseringsindicatoren. De diachrone en synchrone status van het Belgische Nederlands’, Nederlandse Taalkunde, 6: 179-202. Kilgarriff, A. (forthcoming). ‘Comparing corpora’. International Journal of Corpus Linguistics. Scott, M. (1997). ‘Pc analysis of key words – and key key words’. System, 25:233-245. Stubbs, M. (1996), Text and Corpus Analysis, Oxford: Blackwell. 194 Dirk Speelman, Stefan Grondelaers and Dirk Geeraerts Speelman, D., S. Grondelaers and D. Geeraerts (forthcoming), ‘Profile-based Linguistic Uniformity as a Generic Method for Comparing Language Varieties’, Computers and the Humanities. Williams, R. (1976), Keywords, London: Fontana. A multilingual learner corpus in Brazil Stella E. O. Tagnin University of São Paulo Abstract A learner corpus can provide useful data to detect specific difficulties of language learners and consequently inform the production of pedagogic material to address these problem areas. The USP Multilingual Learner Corpus will initially be built in English, German and Spanish. A heading with detailed information about the student (course level, age, sex, mother tongue, etc.) will allow for different types of research. The corpus’s multilingual character will make it possible to look into difficulties common to all Brazilian learners, irrespective of language. Its varied content may also provide insights into the effectiveness of different methodologies. 1. Introduction One of the problems with textbooks used in Brazil for teaching a foreign language is that most are written by foreign authors unacquainted with Brazilian students’ difficulties. It is a known fact that a learner corpus can provide useful data to detect such specific difficulties and consequently inform the production of pedagogic material to address these problem areas (Leech, 1998). Until recently, the only learner corpus under construction in Brazil was the BrIcle (Berber Sardinha, 2001), the Brazilian Portuguese part of the ICLE project (Granger, 1993, 1994), which at the time of writing contains 40,000 words of argumentative texts collected at the Catholic University of São Paulo (PUC-SP). In early 2002 the University of São Paulo (USP) joined the project. The PUC-USP partnership in the Br-Icle project triggered an interest in extending the collection of texts to include all genres in which there is student production at our Department of Modern Languages, which is composed of five areas: English, German, Spanish, French and Italian. This chapter will give an overview of foreign language teaching in Brazil, with a special focus on the state-of-the-art at the University of São Paulo. It will discuss the PUC-USP partnership and how this project motivated teachers from other languages to build their own learner corpora, which will all be brought together in the Multilingual Learner Corpus (MLC). This corpus is part of a larger project – COMET, a Multilingual Corpus for Teaching and Translation, which is also being built at the University of São Paulo under my co-ordination (Tagnin, 196 Stella E. O. Tagnin 2002a, 2002b). Next, it will discuss the design and structure of the corpus. The last part will focus on the research possibilities envisaged by the MLC. 2. Foreign language teaching in Brazil One of the problems with textbooks used in Brazil for teaching a foreign language is that most are written by foreign authors unacquainted with Brazilian students’ difficulties. The other is that they take no account of Brazilian culture or students’ interests. 2.1 The curriculum at the University of São Paulo The Department of Modern Language is divided into five areas, one for each “modern” foreign language taught at the undergraduate level: English, French, German, Italian and Spanish. The curriculum is composed of eight semesters during which students follow courses addressing both the language and the literature components. With the exception of English, where it is taken for granted that students have a fairly good command of the language on entering the course, the other languages start teaching “from scratch” as these languages are not taught regularly in secondary school whereas English is part of the compulsory curriculum. Due to the high demand for foreign language courses in general, the Department also offers extracurricular courses, mainly aimed at the academic community. These go by the name of English on Campus, Español en el Campus etc. and extend from five to ten semesters, depending on the language. They are taught by postgraduate students under the supervision of a co-ordinator, who is a regular teacher in that language. In section 4.1 we will go into more detail about these courses and how they can contribute to the corpus. 3. Learner corpora in Brazil To our knowledge, the only learner corpus under construction in Brazil is the BrIcle. To date it is composed of 40,000 words compiled at the Catholic University of São Paulo under the co-ordination of Tony Berber Sardinha. In line with ICLE requirements it is restricted to argumentative texts and should reach 200,000 words on completion. The project was joined by USP in early 2002, and is due to be completed by the end of 2004. Although there is no notice of any other “formal” learner corpus, several materials involving the teaching of FL have been assembled by various researchers. A few are in electronic format (diskettes, CD-ROMs), but most are probably not. In any case, the material is not prepared for investigation with the aid of electronic search tools. A multilingual learner corpus in Brazil 197 The German area has been working on a Contrastive German-Portuguese Grammar project for which it has collected different types of student production. Most of this material is recorded on CD-ROMs or diskettes but is only available for internal use: x Verbs of transportation. Vol. 1. CAPLE - Corpus of German and Portuguese as Foreign Languages (Blühdorn et al., 1997). This material was collected in several schools engaged in foreign language teaching. In Brazil it came from third and fourth year undergraduate students at USP and intermediate and advanced students at the Goethe Institut in São Paulo; in Germany from undergraduate German students and learners of Portuguese at the University of Erlangen-Nürnberg. It consists of three types of production: a) sentences in which students were required to use verbs of transportation, b) translations of 17 sentences into the foreign language, and c) description of the stories presented in six different sequences of cartoons. x Compositions in German – Vol. 2 CAPLE ದ Corpus of German and Portuguese as Foreign Languages. (Blühdorn et al., 1999) This material was collected between 1996 and 1998 from 342 informants at three German-Brazilian secondary schools and is divided into three categories: a) Brazilian learners of German, b) Brazilian learners who have both German and Portuguese at home, and c) German native speakers living in Brazil. x Contrastive Analysis Corpus of Mistakes in Portuguese and German as Foreign Languages (Glenk & Stanich, 2000). This material was collected from second to third year undergraduate learners of Portuguese at the Universities of Erlangen-Nürnberg (1997) and University of Vienna (1999 and 2000), and from third and fourth year undergraduate learners of German at USP (1998 and 1999). It consists of descriptive texts based on the cartoons used in the research referred to above, narrative texts and essays. Other non-contrastive materials are: x x Corpus of letters exchanged between learners of German in Fès (Morroco) and São Paulo (University of São Paulo) (Blühdorn, 1997). Studentenzeitung (Blühdorn, 1999). A newspaper written by third year German learners at USP during the second semester 1999. 198 Stella E. O. Tagnin x Student writing, different typologies: criteria for text production (Nomura, in preparation). Material produced by second year German learners at USP. In English and Spanish there are scattered collections of texts as a result of individual research by post-graduate students, mainly intended for contrastive studies. However, they are not in a format that makes them searchable by corpus tools. 4. The Learner Corpus at USP When the PUC-USP partnership was established, several teachers at USP became interested in corpora studies. Because the Br-Icle requires only argumentative texts, the English teachers decided to build a corpus with the other types of texts produced by their undergraduate students, mainly narratives and essays. However, once the goal of 200,000 words of argumentative texts has been reached for the Br-Icle, this type of text will also be included in the USP Multilingual Learner Corpus. The German and Spanish areas have already joined the MLC project. French and Italian have shown some interest but no official contact has been made as yet. Nevertheless, the project is underway, and it will also be fed with texts from the on campus courses. 4.1 The on campus courses The participation of the on campus courses at USP opens up other possibilities, such as including other genres of texts and gathering the production of another type of students. As these courses are aimed mainly at the academic community, that is, students, teachers and other employees, one gets a fairly varied audience, both in terms of age and cultural background, which is certain to affect the content of their production. Quite a few undergraduate students attend these courses to “catch up” with the rest of their class, that is, as remedial work to improve their linguistic performance. The teaching is more informal than in the regular undergraduate courses and students feel they have less responsibility when it comes to passing or failing the course. The English on Campus (EOC) course has a total of ten semesters: Basic 1, 2 and 3, Pre-Intermediate 1 and 2, Intermediate 1 and 2, Advanced 1 and 2 and a semester of Conversation. The course books used are New Interchange: from Introduction up to volume 3, Part B at the Basic, Pre-Intermediate and Intermediate levels, and Passages Parts A & B at the advanced level. Topics for the written assignments are suggested according to the grammar points addressed. For instance, Write a conversation with a friend in which you describe your A multilingual learner corpus in Brazil 199 apartment or house and ask about his or her living place when the focus is on the Simple Present, short answers; questions with how many and answers with there is, there are. Due to the high number of students at the English on Campus courses – approximately 700 – and considering that two assignments are submitted per student, it would be possible to collect around 1,000 texts per semester, that is, as long as most learners agree to sign to give their permission to have their assignments included in the corpus. The picture is slightly different for the German on Campus (GOC) course. They offer a five-semester course: Basic 1, 2, 3 and 4 and a semester of Conversation. A German course book Moment mal! is used at the first three levels. Basic 4 and Conversation use material prepared by the teachers and based on other German books. The content of the Conversation course varies each semester and students may take it more than once as it is mainly aimed at giving undergraduate students an opportunity to exercise their oral production. As there are approximately 100 students enrolled, the number of possible texts for inclusion each semester would be around 200, again with two texts per student. The levels at Spanish on Campus (SOC) are Basic 1, 2, Intermediate 1, 2 and Advanced 1, 2. Three enhancement modules are also offered, each one semesterlong: a) Culture, b) Literature, and c) Conversation. A Grammar module is in preparation. As opposed to the previous courses, Spanish relies on material prepared only by their own teachers and is based on the profile and a needs analysis of their students. There are currently about 460 students enrolled in their courses, which would give a total of 920 texts per semester. The course has also been a source for postgraduate research on exclusion and self-exclusion factors; methodological and ideological analyses of textbooks; error analysis; evaluation of assessment procedures; and theories of language acquisition focusing especially on how or how much learning can contribute to acquisition. One such research took a contrastive approach comparing how the simple and compound past tenses are used by learners of Spanish and of English. This study was informed by data from both the English on Campus and Spanish on Campus courses. 5. The USP Multilingual Learner Corpus To date the USP Multilingual Learner Corpus (MLC) will be composed of texts produced by their undergraduate (UG) and on Campus learners in the areas of English, German and Spanish. As mentioned above, English undergraduate texts Stella E. O. Tagnin 200 of the argumentative type will be fed into Br-Icle until it has reached its goal of 200,000 words, (see diagram below). USP Learner Br- English German UG UG Spanish GOC EO UG SO Figure 1: The composition of the USP Learner Corpus Each student will be identified by a code and a profile with basic information as to his course, level, year of attendance, age, sex etc. Each text will be stored in its full form and preceded by a header with information as to text type, grammar point covered, topic of assignment, course book (or other materials) in use, etc. 6. Possible areas of research Learner corpora in various parts of the world have already produced a wealth of research (Granger, 1998b; Granger, 2002; Granger et al., 2002) but to our knowledge there is no multilingual learner corpus, that is, learners with a common mother tongue learning different foreign languages. This is in contrast with the ICLE project in which one has learners with different mother tongues learning a common language. With this design, and given that each subcorpus is also a stand-alone contrastive learner corpus, in that it allows comparison between productions originating in the two distinct courses offered at USP, it is envisaged that the corpus will not only allow for horizontal studies, comparing student production originating in the same class or in the same level, but also studies on the vertical axis, assessing student development over a period of time, either individually or collectively (cf. Kaszubski, 2000; Lenko-Szymanska, 2000; among others). Research on student writing strategies like paraphrasing, the (over/under) use or avoidance of certain syntactic structures, vocabulary items, collocations and formulas (cf. Altenberg & Tapper, 1998; Altenberg, 2002; Berber Sardinha, 2001; De Cock, 1998; Granger, 1998a, 1998b; and many others) will also be possible. More interesting perhaps is the possibility of cross-linguistic studies, like detecting problems common to learning a foreign language or problems common to Brazilian learners. Another contrastive area made possible by the design of the A multilingual learner corpus in Brazil 201 corpus lies in the field of methodology as it will enable researchers to evaluate the effectiveness of different methodologies or materials at both the undergraduate and/or the on campus courses. 7. Conclusion The Multilingual Learner Corpus under construction at the University of São Paulo is currently being fed with student production from two types of courses: the regular undergraduate courses and the extracurricular on campus courses offered by the areas of English, German and Spanish, at the Department of Modern Languages. The MLC is not only a promising project in terms of the array of possible research areas, but it has also integrated the different languages taught at the Department by bringing them together to work under a common project. References Altenberg B. (2002), Using bilingual corpus evidence in learner corpus research in: S. Granger, J. Hung and S. Petch-Tyson (eds.) Computer Learner Corpora, Second Language Acquisition and Foreign Language Teaching. Amsterdam and Philadelphia: Benjamins, 37-54. Altenberg B. and M. Tapper (1998), The use of adverbial connectors in advanced Swedish learners’ written English, in S. Granger (ed.) Learner English on Computer. London and New York: Addison Wesley Longman. 80-93. Berber Sardinha T. (2001), O Corpus de Aprendiz Br-Icle, http://lael.pucsp.br/~tony/2001bricle-interc.pdf. Blühdorn H, G. Evangelista and M. C. Reckziegel (eds.) (1999), Redações em alemão. Vol. 2. CAPLE - Corpus em alemão e português como línguas estrangeiras. São Paulo, USP. Blühdorn H. (ed.) (s/d), Sagen und Legenden, Tänze, Festtagsbräuche und Kochrezepte aus Brasilien. Eine interkulturelle Korrespondenz zwischen Deutsch-Studenten in São Paulo, Brasilien, und Fès, Marrocos. Blühdorn, H. (ed.) (1999), Studentenzeitung, São Paulo, USP. Blühdorn, H., L. F. Dias Moreira and R. F. Silva (eds.) (1997), Verbos de transporte. Vol. 1. CAPLE - Corpus em alemão e português como línguas estrangeiras. De Cock, S (1998), A Recurrent Word Combination Approach to the Study of Formulae in the Speech of Native and Non-Native Speakers of English. International Journal of Corpus Linguistics, vol. 3(1): 59-80. Glenk, E. and K. Stanich (eds.) (2000), Corpus de Análise Contrastiva de Erros em Português e Alemão como Línguas Estrangeiras. São Paulo, USP, setembro de 2000. 202 Stella E. O. Tagnin Granger, S. (1993) The International Corpus of Learner English, in: Aarts, J., P. de Haan and N. Oostdijk (eds.) English Language Corpora: Design, Analysis and Exploitation. Amsterdam: Rodopi: Amsterdam. 57-69. Granger, S. (1996), From CA to CIA and back: An integrated approach to computerized bilingual and learner corpora, in: Aijmer, Karin, Bengt Altenberg & Mats Johansson (eds). Languages in Contrast – Papers from a Symposium on Text-based Cross-linguistic Studies. Lund 4-5 March 1994, Lund: Lund University Press. 37-51. Granger, S. (1998a), Prefabricated patterns in advanced EFL writing: collocations and formulae, in: Cowie, A. (ed.) Phraseology: theory, analysis and applications. Oxford: Oxford University Press. 145-160. Granger, S. (ed.) (1998b), Learner English on Computer. , London & New York: Addison Wesley Longman. Granger, S. (2002), A Bird’s-eye View of Computer Learner Corpus Research, in: S. Granger, J. Hung and S. Petch-Tyson (eds.) Computer Learner Corpora, Second Language Acquisition and Foreign Language Teaching. Amsterdam and Philadelphia: Benjamins. 3-33. Granger S., J. Hung and S. Petch-Tyson (eds.) (2002), Computer Learner Corpora, Second Language Acquisition and Foreign Language Teaching. Amsterdam and Philadelphia: Benjamins. Kaszubski P. (2000), Lexical profiling of English (learner) corpora: can we measure advancement levels?, in: B. Lewandowska-Tomaszczyk and P. Melia (eds.) àódĨ studies in Language. Vol. 1: PALC’99: Practical Applications in Language Corpora [Papers from the International Conference at the University of àódĨ, Poland, 15-18 April, 1999]. Frankfurt am Main: Peter Lang,. 249-86. Leech, G. (1998), Learner corpora: what they are and what can be done with them, in S. Granger (ed.) (1998b), Learner English on Computer. , London & New York: Addison Wesley Longman. xiv-xx. Lenko-Szymanska, A. (2000), How to trace the growth in learner’s active vocabulary. A corpus-based study. Paper presented at the 4th International Conference on Teaching and Language Corpora. Graz, 19-23 July 2000. Nomura, M. (org.) (in preparation). Redação de textos de diferentes tipologias: critérios de produção de texto. Tagnin, S E O (2001), COMET – A Multilingual Corpus for Teaching and Translation, In Lewandowska-Tomaszczyk, B. (ed.) PALC 2001 – Practical Applications in Language Corpora, Frankfurt am Main: Peter Lang, 535-540 Tagnin, S E O (2002), Taking off in Brazil: COMET – A Multilingual Corpus for Teaching and Translation. Paper presented at ICAME 2002 – The Theory and Use of Corpora – The 23rd International Conference on English Language Research on Computerized Corpora of Modern and Medieval English, Gothenburg, Sweden, May 22- 26, 2002. Quantitative or qualitative content analysis? Experiences from a cross-cultural comparison of female students’ attitudes to shoe fashions in Germany, Poland and Russia Andrew Wilson and Olga Moudraia Lancaster University Abstract In order to examine differences in attitudes to shoe fashions between women in Germany, Poland and Russia, we asked three samples of advanced female students of English to write a short English composition in response to the stimulus: “Tell us a little bit about the footwear (shoes, boots, etc.) you own and when you wear it”. We analysed the results using a manual qualitative content analysis and two forms of quantitative computer content analysis: one using project-specific categories developed from the qualitative content analysis and previous theory, the other using general semantic field categories. Both techniques were successful in highlighting similar between-group differences, suggesting that qualitative content analysis and project-specific categories can largely be dispensed with. Some issues in using non-native student English compositions as data in cross-cultural studies are also considered. 1. Introduction Appearance is an integral part of communication and miscommunication. If we want proof of this claim, we need cast our minds back no further than the British Conservative Party’s annual conference in 2002, when its new chairperson, Theresa May, took to the platform in kitten-heeled leopardskin shoes. This deviation from a traditional “business” shoe received huge media attention, with close-up photographs in many national newspapers, but it ultimately overshadowed what she had to say: it has been remembered long after the content of her speech has been forgotten. Thus, something as everyday as her choice of shoe can be said to have led to an overall communication failure 1 . It is clear from this example that the messages sent out by clothing and footwear (non-verbal communication) deserve as much attention as those messages transmitted by words (natural language), and it is from this position that we began to work on apparel-based non-verbal communication, and, in particular, on footwear. 1.1 Cross-cultural differences in non-verbal communication We know that differences in non-verbal communication (hereafter “NVC”) exist between nationalities and cultures (Andersen, 1988). However, although a 204 Andrew Wilson and Olga Moudraia number of comparative studies in NVC have been carried out (e.g. Remland et al., 1991), these have tended to focus on the more bodily aspects, such as gesture, facial expression, touch and proxemics. By contrast, a very large proportion of the empirical work on apparel has been carried out in the USA, focussing primarily on American subjects. Comparatively little empirical work of this kind has been carried out in Europe (especially eastern Europe), and even less has had a contrastive orientation. An important consequence of this is that the claims of popular “dress for success” manuals, whatever their empirical and theoretical foundations in one culture, cannot simply be transferred by translation to other cultural contexts. In order to test how, and to what extent, perceptions of apparel differ cross-culturally, we have been collecting data from female students in three European countries: Germany, Poland and Russia. 1.2 Approaching NVC in the respondents’ own words Although some of the early work on NVC made use of open-ended qualitative data (e.g. Stone, 1962), much recent research in the field has been carried out in an experimental scale-based paradigm (Davis & Lennon, 1988). However, this latter approach, although readily quantified, has the problem of imposing the researchers’ pre-defined categories and interpretive schemata on the experimental subjects: for instance, they have to rate the image of a variously clothed figure as being more or less “friendly and approachable”. A notable exception to this experimental paradigm is the small-scale study by Golliher (1987), who employed a form of open-ended depth interviewing in order to re-construct the categorisation and interpretation of clothing items in his informant’s own terms. This qualitative ethnomethodological approach is valuable, because it can provide much richer data than the experimental approach: it taps into the actual attitudes and perceptions of the informants rather than having them agree or disagree with the researchers’ attitudes. In this study, we have applied (for logistical reasons) a written variant of open-ended interviewing. 1.3 Analysing open-ended responses For the analysis of open-ended response data, there are two main paradigms available to the researcher: qualitative analysis and quantitative analysis. One of our goals in this pilot study was to examine the relative merits of a qualitative and quantitative content analysis of our data. Qualitative content analysis, as we understand it, is a variant of the grounded theory approach to text analysis – in other words, it is a bottom-up approach to identifying the main ideas within people’s discourse. In the course of careful reading, linguistic units within the text (words and phrases) are annotated and aggregated into broader categories as themes begin to emerge from the data. As its name suggests, qualitative content analysis is not, in itself, a quantitative approach; however, as well as functioning as a stand-alone methodology, it can also feed into dictionary construction for quantitative content analysis. Quantitative or qualitative content analysis? 205 In contrast to qualitative content analysis, quantitative content analysis attempts to provide a numerical measure of mention which can then be used to make statistical comparisons between respondent groups. Quantitative content analysis is typically fully automated (or at least computer aided) and can take two main forms: a form that relies on the classification of text words according to a predefined set of categories (dictionary-based content analysis) and a form that makes use of the multivariate analysis of word contiguities to derive themes from the set of texts being examined (correlational content analysis) (Hogenraad, McKenzie & Péladeau, 2003). In dictionary-based content analysis – the method which we have used in this paper – a dictionary of words and content categories is constructed, and words in the running text are then automatically matched with the dictionary categories. The content analysis program then provides a frequency of use for each category in the study across each group of texts indicated by the researcher. The categories in dictionary-based content analysis may be quite general, semantic-field-type categories, which can be used for a wide range of studies, or they may be specially designed for a particular study, typically on the basis of previous theory and a prior qualitative analysis of at least some of the data. In this study, we have undertaken two quantitative content analyses – one with general semantic field categories and one with project-specific categories – and we have compared both these sets of results with each other and with the outcome of the qualitative analysis. 1.4 Language issues in cross-cultural survey research One of the major difficulties in conducting cross-cultural research lies in language. If material is collected in several languages (e.g. German, Polish and Russian), it can then prove rather difficult to quantify and compare the results. This is certainly the case if one is working with word-level data within a methodology such as correlational content analysis, since differences in vocabulary range and grammatical structure can themselves lead to differences in the relative frequencies of comparable words. If dictionary-based content analysis is used, then vocabulary range and grammar may prove rather less of a problem, but another problem then arises: the availability of exactly comparable content analysis dictionaries and related processing tools for all the languages in question. In an attempt to avoid both these sets of problems, we have been collecting our data in a single language – English – from advanced learners of that language in each of the countries covered by the project. In principle, this should enable us to make direct comparisons between national groups. However, at the outset, it was not clear what other issues non-native Englishes might pose for the content analyst, and so this is an issue which we kept in focus during this pilot study and to which we return later in the paper. 206 2. Andrew Wilson and Olga Moudraia Data Our data consisted of compositions written by advanced female students of English at universities in Germany, Poland, and Russia. These were all collected during the summer of 2002 and were provided in computer-readable form. All respondents were natives of the country in question: any students not conforming to this criterion were dropped from the subsequent analysis. In this pilot study, there were 11 compositions from Germany, 34 from Poland, and 13 from Russia. 2 The students were asked to write on each of the following open-ended stimulus questions: x What sort of things would you wear to a job interview and why? x What sort of things would you wear to a friend’s birthday party and why? x Tell us a little bit about the footwear (shoes, boots, etc.) you own and when you wear it. In the present study, we focus only on the last of the three questions – the one about footwear. 3. Qualitative analysis Close reading of the material helps us understand what the writers are telling us in their own words. A special coding frame was constructed to list and classify the words and phrases in our shoe data. We aimed to find the answers to three Whquestions: What? Why? When? Qualitative content analysis of the data revealed the following main themes: x What – shoe types and their physical attributes; x Why – reasons for wearing different types of shoes such as comfort, practicality including matching outfits, suitability, design, colour, attractiveness and price; x When – occasions when particular types of shoes are worn: related to weather, outdoor and indoor activities, lifestyle, formal and informal social events, etc. The examination of Becker and Lißmann’s (1973) two levels of content, primary content (themes and main ideas of the text) and latent content (context information), lead us to the following research conclusions on the similarities and differences in the subjects’ classifications of shoes. Firstly, the Poles described the widest range of shoe types with main emphasis on comfort and elegance, while occasions when particular types of shoes are worn were mainly related to outdoor activities, and formal and informal social events. Secondly, Russian respondents, quite unsurprisingly, had a bigger variety of boots, and generally Quantitative or qualitative content analysis? 207 referred to weather-related occasions for wearing particular types of footwear with a focus on cold weather conditions; they placed major emphasis on comfort and attractiveness of shoes. Thirdly, German subjects wrote less about high heels, placing more emphasis on practicality; occasions when particular types of shoes are worn were varied, including outdoor activities, formal and informal social events, as well as weather-related occasions with a focus on warm weather conditions. 4. Quantitative analysis For the quantitative content analysis, we used the USAS suite of programs, developed over the past 13 years at Lancaster University (Wilson & Rayson, 1993, Rayson & Wilson, 1996). USAS is a software package for dictionarybased content analysis: we hope to apply correlational content analysis to these data in a future study. Using a comprehensive dictionary and multi-word-unit list (which we updated to cover the vocabulary of the present data), USAS assigns a basic semantic field code (or “SEMTAG”) to each lexical item or phrase within a text – for example, “Colour”, “Body and Bodyparts”, “Power”, “Similar/Different”, etc. USAS performs this task with approximately 92% accuracy. We shall return to these SEMTAGs later in the paper. A further feature of USAS is the module called MAPPING. This enables researchers to create their own, project-specific category system (or “CONTAGs”) by conflating, subdividing, or ignoring various SEMTAG categories. For our data analysis, we used MAPPING to create categories based both on previous work on the perception of shoe fashions and on the results of the qualitative analysis. To examine differences between the three groups of respondents, we used the TMATRIX module of USAS. This module enables the researcher to see the frequencies of words and categories for each group, examine concordances, and carry out log-likelihood or chi-squared tests on word and category frequencies. 4.1 Project-specific category construction Although shoes have been said “never to lie” about the wearer’s personality (Pond, 1986), in the literature on NVC and first impressions they have received comparatively little attention when set alongside other apparel items such as dresses or suits. Shoes have occasionally been included in broader studies (e.g. Lennon & Miller’s (1984/85) study included a pair of brown boots), but, at the time of carrying out the research, the only detailed study known to us was that of Kaiser et al. (1987). 3 It was thus from Kaiser et al.’s study that we set out to create CONTAGs for our project. These CONTAGs would enable us to see Andrew Wilson and Olga Moudraia 208 whether the European nationalities used the same kinds of criteria as Kaiser et al.’s American sample when writing about footwear and whether or how they differed amongst themselves. Prior to carrying out a semantic differential (questionnaire) study with a larger sample of respondents, Kaiser et al. undertook a focus group with five men and five women living in California. The respondents were shown a range of shoe styles, which they then discussed. From these discussions, the following primary dimensions were extracted: x x x x x x x x x x old-young liberal-conservative work-leisure comfortable-uncomfortable unsexy-sexy formal-casual high-status-low-status inexpensive-expensive dislike-like fashionable-unfashionable These dimensions were relatively unproblematic to operationalise as content categories, with the exception of liberal-conservative. As an approximation to this category, therefore, we made the distinction classical-modern. In addition to the categories based on Kaiser et al’s dimensions, we also created categories to represent the main types of footwear mentioned by our respondents. Most of these are self explanatory; the category “elegant shoes” contains references to stilettos, court shoes, pumps, etc. Finally, we created categories to measure references to the main additional issues that emerged from the qualitative content analysis. Table 1 details the set of content categories that we worked with. Table 1: Content categories used in the analysis Shoe styles: BOOT CLOG DOCS ELSH FLTS SLPR SNDL TRNR OTHE Boots Clogs Doc Martens (and similar shoes) Elegant shoes Flat shoes Slippers Sandals Trainers Other shoe styles Quantitative or qualitative content analysis? 209 Kaiser et al dimensions: AGES Young-old CLAS Classical-modern COMF Comfortable-uncomfortable FASH Fashionable-unfashionable FORM Formal-casual LIKE Like-dislike PRIC Expensive-inexpensive SEXY Sexy-unsexy STAT High-status-low-status WORK Work-leisure Additional categories from qualitative analysis: ATTR Attractive-unattractive PHYS Physical attributes PRAC Suitable-unsuitable SEAS Seasons TEMP Temperature WEAT Weather 4.2 Results All of Kaiser et al.’s dimensions were used by our respondents, though to varying degrees: for example, high-status-low-status was relatively little used, whilst work-leisure was a very widely used category. Table 2 shows the results for the Kaiser et al. categories aggregated for the whole sample. Table 2: Frequencies of Kaiser et al dimensions (whole sample) WORK LIKE COMF FORM FASH CLAS PRIC SEXY AGES STAT Freq. per million words 21,483 9,645 7,453 2,255 1,754 752 626 626 564 376 Table 3 shows the between-groups differences for all the content categories. For three groups (2 d.f.), log-likelihood (LL) values of 9.21 or greater are significant at p < 0.01 and values of 5.99 or greater are significant at p < 0.05. In terms of between-groups differences, only one of the Kaiser et al. categories (fashionable-unfashionable) distinguished between groups at p < 0.01, with the Russian sample writing the most about this dimension. A further category – classical-modern – distinguished at the lower probability level (p < 0.05), again with the Russian sample writing the most about this dimension. 210 Andrew Wilson and Olga Moudraia Table 3: Between-groups differences on content categories Category freq. per million words PHYS BOOT SEAS TEMP WEAT FASH SLPR ATTR FLTS SNDL CLAS DOCS OTHE PRIC COMF STAT ELSH CLOG WORK LIKE FORM SEXY PRAC AGES TRNR LL German Polish Russian 46,415 7,682 13,444 7,682 3,201 1,280 1,601 4,161 0 6,722 0 320 2,241 0 4,802 320 5,442 320 19,846 7,682 2,561 320 2,241 960 640 29,434 2,725 7,849 2,834 2,180 981 436 8,721 1,744 2,725 654 872 2,180 763 8,285 545 7,304 0 23,111 10,575 2,507 545 3,270 545 327 63,778 13,628 16,626 7,086 7,086 4,088 2,726 4,906 1,090 3,543 1,635 0 545 818 7,632 0 9,267 0 18,806 8,994 1,363 1,090 3,816 273 545 75.0 47.9 19.9 16.7 15.7 12.4 11.5 10.4 9.5 8.8 7.7 5.9 5.4 4.4 4.2 3.4 3.4 3.3 2.8 2.3 1.9 1.7 1.4 1.4 0.6 Taking into account the further content categories developed on the basis of the qualitative content analysis, the following picture of the three national groups emerged. We discuss here only those categories which showed significant between-groups differences. The Russian sample wrote most about the physical attributes of their shoes, such as size, shape, material and colour. By contrast, the Poles wrote the least about these dimensions overall, although this was still the most frequent category within the Polish sample. The Russian sample also wrote the most about seasons and the weather as determiners of when certain styles of shoes are worn. The German sample also wrote substantially about seasons; they wrote rather less about weather conditions, but had the highest frequency of references to temperature (these being a mixture of weather references and references to shoes being warm or cold to wear). Again, the Polish sample seemed somewhat less concerned about these three dimensions. Interestingly, seasons, weather and temperature were not included as dimensions of usage in Kaiser et al.’s study, but, at least for some of our sample, they seem important determiners of shoe style choice. Quantitative or qualitative content analysis? 211 The only non-shoe-style category on which the Polish sample had the highest frequency of use was attractiveness. It seems that the Polish respondents, as a whole, were more concerned with this dimension than the other two national groups. Attractiveness was not included explicitly in Kaiser et al.’s dimensions, but it could be considered to be implied by categories such as like-dislike and fashionable-unfashionable. We treated attractiveness as a separate category, since we felt that many of its component words (such as smart and elegant) did not fit any of the Kaiser et al. categories particularly well. However, some might argue that these words might be classified elsewhere. In terms of shoe styles, boots were referred to most often by the Russian sample, then by the German sample. The Russians also made the most references to slippers. By contrast, the German sample referred the most often to sandals. This may be a seasonal effect, since the data were collected during the summer, but, if so, it does not explain why this style was not also prominent in the writings of the other two samples. The Poles had the highest frequency of references to flat shoes and “Doc Marten”-style boots. 5. Discussion 5.1 Qualitative or quantitative content analysis? Both qualitative and quantitative content analysis of the shoe data yielded similar results revealing the same main themes. Qualitative content analysis can provide an accurate overall picture of a text corpus, although, being interpretive and subjective, it may overlook some specific details. As Mayring (2001) remarks, qualitative content analysis remains an act of interpretation since “relating categories and parts of the material is no automatic technique but a creative act of interpreting meanings in the text” by the content analyst, who puts into the process of analysis all his/her competencies, pre-knowledge and empathic abilities. Qualitative content analysis defines itself within this framework as an empirical, methodical and controlled approach to the analysis of texts within their context of communication and without precise quantification. In contrast to qualitative analysis, quantitative content analysis (as its name suggests) provides an exact quantification by counting the category frequencies in the texts. In our data, although most themes were identified with equal success by both the qualitative and quantitative analyses, the relative emphasis placed on them by the different nationalities became much clearer in the quantitative analysis: for example, the qualitative analysis identified comfort as a major emphasis of the Polish sample in particular, and, in the quantitative analysis, this category showed a statistically significant between-groups difference. 212 5.2 Andrew Wilson and Olga Moudraia CONTAGs or SEMTAGs? Although we used project-specific content categories (CONTAGs) for our main analysis, the data had initially been tagged, as mentioned earlier, with more general semantic field categories (SEMTAGs). We therefore asked ourselves the question: Would we have obtained the same results if we had not used CONTAGs but had conducted the analysis only using the pre-existing SEMTAGs? If we would, this provides a useful argument for future studies’ taking the SEMTAGged data “as is” and thus reducing the manual effort needed in developing new CONTAGs for different research projects. Table 4 shows the SEMTAG frequencies per million words for categories that were significant at p < 0.01 or p < 0.05. As there are more SEMTAGs than CONTAGs, it will be seen that there are a greater number of significant categories in this table. However, the same themes still predominate: physical attribute categories such as materials, colour, temperature and physical properties, as well as weather and time periods (used for seasons), are more frequent in the Russian and German samples, whilst the evaluative category “judgement of appearance: positive” is more frequent in the Polish sample. In looking at SEMTAG profiles, it is worth making a distinction between analysis categories and retrieval categories. As SEMTAG attempts to tag all the words and phrases in a text, rather than a selection, not all the tags relate to content items (nouns, verbs, adjectives, and lexical adverbs): many relate to word classes such as degree adverbs, numbers, modal verbs, pronouns, and so on, which are not key concepts within the content analysis and for most purposes can be disregarded. Other categories – such as similar and different – are not very meaningful in themselves as analysis categories but can be useful as retrieval categories. For instance, when examined in context, the category “similar” can be seen to be referring mainly to another key reason for wearing particular shoes: because they go with or match particular clothes. An advantage of using SEMTAGs is the greater precision that they can provide: for instance, here we are able to see that the Russian sample writes most about cold temperatures whilst the German sample writes most about warm temperatures. On the other hand, when using SEMTAGs, it is sometimes necessary to revert to the lexical frequency list in order to examine the use of detailed, project-specific concepts: for example, in our study, we were interested in different kinds of footwear, but, in the SEMTAGs, all kinds of footwear receive a single tag (clothing and accessories), which is applied to any apparel item. Perhaps a compromise can be found in developing detailed subcategorisations of key, project-specific categories but leaving the remainder of the SEMTAGs unaltered. Quantitative or qualitative content analysis? 213 Table 4: Between-groups differences on SEMTAGs Category description Substances and materials: solid Colour and light Numbers Pronouns Buying and selling Temperature: cold Clothing and accessories Time periods Weather Degree: maximisers Judgement of appearance: positive Continuous Possible Difficult Long Short Arts and crafts Grammatical words Unfriendly Young, new Degree: approximators Friendly Interest Dry Exclusivizers/particularizers The same Using Similar Showing Past Negative Seem, appear Heavy Location and direction Transport by land Phoney Large Different Physical properties: general Putting and placing Temperature: hot Point in time Important Intimate/sexual relationship Unusual Substances and materials: liquid Touch Games Relationship: general Body and body parts General, unspecific Category freq. per million words German Polish Russian 11,204 27,209 16,325 147,567 960 1,601 101,152 19,206 3,201 1,921 7,682 320 10,563 0 1,280 960 0 287,452 640 640 960 1,601 0 0 8,323 0 1,601 1,601 2,881 960 10,883 2,881 0 3,841 3,201 320 0 9,603 1,280 640 5,122 3,201 2,241 0 0 640 640 640 320 6,082 960 5,669 14,281 7,522 167,230 6,323 327 93,208 12,101 2,180 4,470 16,025 2,180 12,973 1,635 2,289 654 327 255,096 1,417 1,417 1,308 1,308 872 0 4,579 218 1,090 3,816 763 1,308 14,935 4,252 763 3,161 981 0 1,090 8,503 1,744 2,943 2,071 4,797 3,488 654 654 0 0 0 1,744 7,631 109 18,534 26,983 17,171 128,373 1,635 3,543 120,741 21,532 7,086 818 13,900 273 6,269 545 5,179 2,998 1,635 267,920 0 0 3,816 0 0 818 3,271 1,363 0 1,363 2,180 0 9,267 1,363 0 6,814 2,453 818 273 4,633 4,088 2,726 3,271 1,908 1,090 0 0 0 0 0 545 11,447 0 LL 42.0 32.0 28.9 27.6 27.4 19.4 19.0 17.6 15.7 15.2 13.2 12.2 12.0 10.2 10.0 9.8 9.5 9.3 9.2 9.2 9.1 9.0 8.9 8.8 8.6 8.5 8.3 8.3 8.2 8.1 8.0 7.8 7.8 7.7 7.6 7.6 7.3 7.3 7.1 6.9 6.9 6.8 6.7 6.7 6.7 6.5 6.5 6.5 6.5 6.4 6.4 214 Andrew Wilson and Olga Moudraia Giving Degree: compromisers Usual Liking: positive (+++) 5.3 1,280 1,921 3,841 0 763 2,943 6,759 763 0 818 3,816 1,363 6.4 6.4 6.3 6.2 Using non-native English compositions as data Returning to our secondary research question – the issues posed by using English as a survey response medium with non-native speakers – we are able to make three main observations. Firstly, it is possible that, when working with students as respondents, the nature of the exercise may have some effect on its outcome. In the case of the present survey, the Russian exercise was also used as a compulsory graded exercise in an EFL class; the German and Polish exercises, by contrast, were intended as compulsory but did not form part of the assessment for the respective courses, so students would not be penalised for not doing them (or for doing them “badly”). This may account for the fact that the Russian responses were in many ways more detailed, in several cases involving numbered lists of footwear items owned by the writer. An interesting alternative – or perhaps supplementary – explanation for some between-group differences is that they may arise in part from different “rhetorical strategies” in approaching the compositions. 4 Aside from this consideration, two more specifically lexical issues arise: the use of (often very culture-specific) L1 vocabulary in the English compositions and the non-standard use of English vocabulary by particular groups of respondents. Examples of the first issue in our data were the term glans in the Polish compositions, about which we had long discussions with our Polish colleague (it appears to refer to a kind of “punk rock” style involving lots of black leather and metal studs), and the term kapron in the Russian data, a kind of material. An example of the second issue was the use by the Russian respondents of the term top-boots. They were the only group to use this term, and they used it rather frequently (10 times across 12 respondents). This is a relatively dated expression in native-speaker English and, according to the Oxford English Dictionary, is properly used for the style of riding boot that has a cuff of differently coloured leather at the top of the boot shaft: these are the style of boots worn by jockeys and show-jumpers, also commonly seen in a female fashion version at the turn of the 1980s/90s. However, the Russian respondents appear to be using this term to refer to any boot with a shaft reaching to the knee of the wearer. The Oxford English Dictionary also recognises this sense, though castigates it as “incorrect”. These lexical issues are perhaps not a large problem in a dictionary-based content analysis, providing the respective words and phrases are properly categorised, but Quantitative or qualitative content analysis? 215 they can potentially be problematic for those approaches to content analysis that work primarily on vocabulary items, such as correlational content analysis or keyword analysis (Scott, 1997, 2001). In such cases, a partial synonymisation process, although frequently dispensed with as uneconomical, might be advisable. 6. Conclusion Our pilot study has revealed interesting differences between the three nationalities in the way they write about shoe fashions. Although they all made use of the “American” dimensions delineated by Kaiser et al. (1987), they did so to different degrees, and they also made use of other conceptual dimensions not mentioned by Kaiser et al. In general, the Russians seemed to place the most emphasis on the physical and practical characteristics of their shoes. They also seemed to be the strongest “trend-followers”. By contrast, the Poles seemed to emphasise their judgements of the attractiveness of shoes. The Germans tended to fall midway between the Russians and the Poles on most dimensions. As this was a pilot study, these results should for the moment be considered suggestive rather than conclusive. However, if further work (which we are planning to carry out) does support these findings, then they may have important implications both for crosscultural apparel-based NVC and for footwear marketing. On a methodological level, we have shown that a general content analysis dictionary based on semantic fields can deliver the same substantive results as a specially constructed project-specific dictionary. We have also shown that such a dictionary provides broadly the same results as a manual qualitative content analysis, which replicates the findings of Thomas and Wilson (1996) on doctorpatient interaction. We therefore suggest that qualitative analysis and special dictionary construction are costly stages of analysis which can be dispensed with without much effect on the outcome of a study. However, it may be that such general dictionaries require some minor modification (mostly the insertion of more detailed sub-divisions) to cover the fine detail of specific topic areas under examination. Notes 1 Unless, of course, one of the aims had been to project a trendier image for the party, regardless of the speech content. 2 We are grateful to our colleagues in this project – Amei Koll-Stobbe in Greifswald, Germany; Agnieszka Lénko-SzymaĔska in Lodz, Poland, and Tatyana Astafurova in Volgograd, Russia – for collecting these data. 216 Andrew Wilson and Olga Moudraia 3 We have since become aware of further research on shoes carried out by Marianne Herzog in Germany – e.g. Herzog (1995). 4 See Lénko-SzymaĔska (this volume) for a detailed discussion of this idea in the context of Polish and American English compositions written in response to the same question (about mobile phones). References Andersen, P.A. (1988), Explaining intercultural differences in nonverbal communication, in: L.A. Samovar and R.E. Porter (Eds.) Intercultural Communication: A Reader. 5th ed. Belmont, CA: Wadsworth, pp. 272281. Becker, J. and H.-J. Lißmann (1973), Inhaltsanalyse – Kritik einer sozialwissenschaftlichen Methode. Arbeitspapiere zur politischen Soziologie 5. München: Olzog. Davis, L.L. and S.J. Lennon (1988), Social cognition and the study of clothing and human behavior. Social Behavior and Personality 16(2): 175-186. Golliher, J.M. (1987), The meaning of bodily artefacts: variation in domain structure, communicative functions, and social contexts. Semiotica 65(1/2): 107-127. Herzog, M. (1995), Auftreten... Mensch und Schuh, in: D. Grünewald (ed.) "Was sind wir Menschen doch!..." Menschen im Bild. Analysen. Festschrift für Hermann Hinkel. Weimar: Verlag und Datenbank für Geisteswissenschaften, pp.105-114. Hogenraad, R., D.P McKenzie, and N. Péladeau (fc.) Force and influence in content analysis: The production of new social knowledge. Quality & Quantity 37(3): 221-238. Kaiser S.B., H.G. Schutz and J.L. Chandler (1987), Cultural codes and sex-role ideology – a study of shoes. American Journal of Semiotics 5(1): 13-33. Lénko-SzymaĔska, A. (this volume), The curse and the blessing of mobile phones – a corpus-based study into Polish and American rhetoric strategies. Lennon, S.J. and F.G. Miller (1984/85), Attire, physical appearance, and first impressions: more is less. Clothing and Textiles Research Journal 3(1), 18. Mayring, P. (2001), Combination and integration of qualitative and quantitative analysis. Forum Qualitative Sozialforschung / Forum: Qualitative Social Research [On-line Journal], 2(1). URL: http://www.unituebingen.de/qualitative-psychologie/t-ws01/Mayring_en.htm. Pond, M. (1986), Shoes Never Lie. London: Grafton. Rayson, P. and A. Wilson (1996), The ACAMRIT semantic tagging system: progress report, in: L. J. Evett and T. G. Rose (eds.) Proceedings of the AISB Workshop on Language Engineering for Document Analysis and Quantitative or qualitative content analysis? 217 Recognition. Brighton: Faculty of Engineering and Computing, Nottingham Trent University, pp. 13-20. Remland, M.S., T.S. Jones and H. Brinkman (1991), Proxemic and haptic behavior in three European countries. Journal of Nonverbal Behavior 15: 215-232. Scott, M. (1997), PC analysis of key words - and key key words. System 25(1): 113. Scott, M. (2001), Comparing corpora and identifying key words, collocations, and frequency distributions through the WordSmith Tools suite of computer programs, in: M. Ghadessy, A. Henry and R.L. Roseberry (eds.) Small Corpus Studies and ELT: Theory and Practice. Amsterdam: Benjamins, pp. 47-67. Stone, G.P. (1962), Appearance and the self, in: A. Rose (ed.) Human Behavior and Social Processes. Boston: Houghton Mifflin, pp. 86-118. Thomas, J.A. and A. Wilson (1996), Methodologies for studying a corpus of doctor-patient interaction, in: J.A. Thomas and M.H. Short (eds.) Using Corpora for Language Research: Studies in the Honour of Geoffrey Leech. London: Longman, pp. 92-109. Wilson, A. and P. Rayson (1993), Automatic content analysis of spoken discourse: a report on work in progress, in: C. Souter and E. Atwell (eds.) Corpus Based Computational Linguistics. Amsterdam: Rodopi, pp. 215226. This page intentionally left blank Survey and Prospect of China’s Corpus-Based Research1 Yang Xiao-jun Hunan University of Science & Technology Beijing Foreign Studies University Abstract This chapter conducts a survey of China’s corpus-based research, focusing on the following aspects: i) the history and development of China’s corpus-based research; ii) corpus compilation; iii) leading research figures and institutions; iv) major academic publications on these research. Then, the author predicts the prospect of this research, focusing on the developing trend of China’s corpus-based research, corpus annotation problems, corpus processing tools and how to apply the corpora to language teaching and translation studies. 1. Introduction Corpus linguistics is playing a more and more important role in linguistic research and lexicography, etc. Tognini-Bonelli (2001: 1) points out that what we are witnessing is the emergence of a new research enterprise and a new philosophical approach to linguistic enquiry that has a theoretical status and, because of this, is in a position to contribute specifically to other applications such as lexicography, language teaching, translation, stylistics, grammar, gender studies, forensic linguistics and computational linguistics, etc. Hunston (2002: 1) notes that it is no exaggeration to say that corpora, and the study of corpora, have revolutionised the study of language, and of the applications of language over the last few decades, and that the improved accessibility of computers has changed corpus study from a subject for specialists to something that is open to all. In China corpus linguistics has led to a qualitative change in our understanding of language. In this paper the author first conducts a survey of China’s corpus-based research, focusing on the following aspects: i) the history and development of China’s corpus-based research; ii) corpus compilation; iii) leading research figures and institutions; iv) major academic publications on this research. Then, the author predicts the prospect of this research, focusing on the developing trend of China’s corpus-based research, corpus annotation problems, corpus processing tools and how to apply corpora to language teaching and translation studies. 220 2. Yang Xiao-jun Survey of China’s Corpus-based Research According to Huang Changning and Li Juanzi (2002: 4), China’s earliest corpusbased research on Chinese dialects has a history of nearly 3,000 years. Even the computer corpus-based research on the English and Chinese Languages has a history of over 20 years. Wang Jianxi (2001) stated that, in recent years, corpus linguistics in China has made considerable progress in the compilation, annotation and analysis of Chinese corpora, and in the compilation and studies of corpora of English as a Foreign Language. 2.1 The history and development of the research The history of China’s earliest corpus-based research may date back to the Zhou Dynasty (1100-221 B.C.) and the Qin Dynasty (220-206 B.C.). It is reported that Yang Xiong, who showed great interest in studying dialects, had spent 27 years interviewing some scholars and soldiers who were summoned from different districts to the capital. He had collected enough data and resources manually to compile the first book on Chinese dialects (Huang Changning & Li juanzi, 2002: 4). Since the compilation of China’s first English corpus—JDEST in 1982, many achievements have been made in China’s contemporary corpus-based research. First, fifteen Chinese corpora, eight English corpora and five bilingual corpora (including parallel corpora) have been compiled or are being compiled (to be described in detail in the next section). Secondly, four books on corpus linguistics and four dictionaries on Modern Chinese have been published. Thirdly, eighteen universities and research institutions have been carrying out research programmes on corpus linguistics. Fourthly, there are more and more regular contributors to major national journals of foreign language research and linguistics year by year, covering topics from the application of corpus linguistics to the techniques for compiling and exploiting corpora. Searching papers on corpus linguistics from www.cnki.net, we can find out that there are altogether 107 academic papers on corpus-based research in Chinese academic journals (CAJ), and up to now, the number is about 140 adding the papers searched from www.baidu.com. Besides, there are a number of MSc/MA dissertations and several doctoral theses in corpus-based work. 2.2 China’s Corpora compiled and being compiled It is during these 20 years that about 15 Chinese corpora, 8 English corpora and 5 bilingual corpora (including parallel corpora) have been compiled or being compiled. Some of them are used for general purposes; some are used for specific purposes. 2.2.1 Chinese Corpora (1) The Modern Chinese Language Corpus (MCLC). The biggest representative written Chinese corpus to date, compiled at the Research Institute of Survey and Prospect of China’s Corpus-Based Research 221 Language Application in Beijing in 1995. It aims to contain 70 million contemporary Chinese characters, systematically sampled from among 1.4 billion characters of original texts covering the period 1919 to the 1990s. With regard to contents, 59.6 percent of these texts belong to the humanities, 17.24 percent to natural science, 13.7 percent to newspaper material and 9.37 percent to the miscellaneous class. This MCLC serves the potential needs in five areas: (1) information processing in the Chinese language; (2) the unification and standardisation of the Chinese language; (3) academic research; (4) language education; (5) application of the Chinese language, (see Wang Jianxin, 2001). (2) The Si Ku Quan Shu. (This refers to the Complete Library in Four Branches of Literature completed in 1782, which has the world’s longest series of books. The work, comprising four traditional division of Chinese learning (classics, history, philosophy, and belles-letters), contains 3,503 titles bound into more than 36,000 books with a total of 853,456 pages.) This is the largest electronic Chinese text database so far. Composed of the most comprehensive and outstanding classic Chinese works in the past 3,000 years before the end of the 18th century. This database of 800 million Chinese characters in 4.7 million pages was published by the Digital Heritage Publishing Ltd. in Hong Kong in two versions. One is the complete facsimile version of 183 CD-ROMs with characterretrieving tools with a total of 4.7 million pages of the original books and was put into the market in 1998; the other one has 168 CD-ROMs with title-retrieving tools with a total of 800 million characters and was put into market in 1998. The database has greatly facilitated the study of the ancient Chinese language, literature, culture and history. (3) Corpus of the Chinese Language as Interlanguage. This corpus was compiled by the Research Institute for Computational Language of Beijing Language and Culture University with more than 3.5 million characters in 1995. This corpus consists of 5,774 written texts in Chinese as interlanguage of 1,635 overseas students learning Chinese. The students are from 96 countries and regions studying in nine universities. This corpus is used to enhance the teaching of Chinese to foreign students. (4) The Academia Sinica Balanced Corpus (version 3.0). This corpus was compiled at the Academia Sinica in Taiwan. It is one of the largest annotated Chinese corpora and has been put on the web with 5 million Chinese characters tokenized, tagged and parsed. This grammar tree bank and the statistics will be very useful for the processing system of the Chinese language. (5) The Corpus of Chinese Phonology. This corpus was set up by the Institute of Applied Linguistics of the Chinese Academy of Social Sciences (CASS) with 45,000 Chinese words and 1,200 syllables with tones, numeral strings, lightly pronounced words, words pronounced with a rising “/r/”ending sound, bi-syllables, tri-syllables, separate sentences, short passages and dialogues. This corpus has been annotated manually. An 222 Yang Xiao-jun automatic phonetic-analysis system of Mandarin Chinese, based on the English phonetic analytical system ToBI, is in progress. (6) Contemporary Beijing Spoken Chinese Corpus. This corpus was compiled by Lu Bisong, Ren Yuan and other six scholars in 1992. They have interviewed 500 people from six districts of Beijing talking on about 28 topics and sampled 378 recording materials. This corpus gives a vivid description of Beijing Spoken Chinese in the 1980s with a total of 1.7 million characters. (7) Situated Discourse for Spoken Chinese corpus. This corpus is being compiled at the Research Department for Contemporary Linguistics of the Research Institute for Languages and Linguistics of the Chinese Academy of Social Sciences under the guidance of the doctor supervisor Gu Yueguo. As part of the workplace-related discourse, more than 269 hours of recordings are planned for this corpus, which are being transcribed and annotated to be made available on the Internet in a multimedia form. (8) Corpus of Chinese Textbooks for Primary and Middle Schools. This corpus was compiled by Beijing Normal University with 1 million characters and composed of middle-school teaching material on Chinese literature and language. (9) The People’s Daily Annotated Corpus (PFR). This corpus has been segmented and annotated jointly by the Fujitsu Research Institute of Japan and the Institute of Computational linguistics (ICL) of Peking University with about 200 million Chinese characters. (10) Huayu2. This is a balanced corpus of 2 million Chinese characters, which is entitled Huayu and compiled by both the State Key Laboratory of Intelligent Technology & System (Key Lab), Tsinghua University, and the Language Information Processing Institute, Beijing Language and Culture University on Mainland China. This corpus has been tokenised and tagged with Cseg & tag 1.0, a segmentation and tagging system developed by them, and then manually proofread, from which a grammar tree bank of 10,000 Chinese sentences is being built to be used as a testbed for Chinese parsers. (11) Linguistic Variation in Chinese Communities (LIVAC Synchronous Corpus). This corpus entitled LIVAC Synchronous Corpus was compiled by the City University of Hong Kong and is a representative computerised text corpus from Chinese newspapers and electronic media in Mainland China, Hong Kong, Macau, Taiwan and Singapore. LIVAC aims to cover a ten-year period starting from July 1995. This corpus will be very useful for comparative studies and can provide quantitative data for language engineering. (12) Statistical Corpus of Chinese Word Frequency. This corpus was designed for the specific purpose of the statistics of Chinese word frequency and compiled jointly by 10 universities and research institutes. It has several subcorpora, such as: (a) 20-million character corpus of modern Chinese Survey and Prospect of China’s Corpus-Based Research 223 was compiled by the Beijing University of Aeronautics and Astronautics in 1983, (b) 5.27-million character corpus of modern Chinese corpus was compiled by Wuhan University in 1979, (c) 5-million-character Chinese corpus was compiled by Peking University in 1992, (d) 100-millioncharacter corpus of ancient and modern Chinese was compiled by Shanghai Normal University. A modern Chinese corpus of 66,186,297 Chinese characters was compiled by Shandong University, (e) 2.5 million-character Chinese corpus of news was compiled by Shanxi University in 1988. (13) The Corpus of the Contemporary Chinese Language. This corpus was compiled by the Department of Chinese and Bilingual Linguistics at the Hong Kong University of Science and Technology (HKUST) and comprises 6 million characters of contemporary Chinese used in mainland China, Hong Kong and Taiwan. It has been segmented with an automatic algorithm, and considerable research has been conducted using this corpus. (14) The Hong Kong Cantonese Child Language Corpus (CANCORD). This corpus grew out of the project “The development of grammatical competence in Cantonese-speaking children” funded by the Hong Kong Research Grants Council from 1991-93, which is a joint effort of three local universities: The Chinese University of Hong Kong, the Hong Kong Polytechnic University, and the University of Hong Kong. This database contains 171 files coded according to the internationally accepted CHAT format and tagged with 33 part-of-speech labels. The data should be of use to any one interested in early language development, be they linguists, psychologists, philosophers or educationalists. Queries about the corpus should be directed to Thomas Lee (e-mail: [email protected]). (15) The Electronic Database of the Chinese Documents (Scripta Sinica). This corpus is probably the most empirically sound one in China today and fully available on the Internet, which is entitled Scripta Sinica and compiled collectively by ten research institutes of the Academia Sinica in Taiwang. The 2.0 version of this database contains 139,940,071 Chinese characters, and each year 10 million characters of new Chinese texts are added to it. 2.2.2 English Corpora (1) JDEST (Jiaotong University Corpus for EST). This corpus was compiled at Shanghai Jiaotong University under the guidance of Prof. Huang Renjie and Prof. Yang Huizhong in 1982. It comprises approximately 1 million words, and was updated to 4,082,368 words in 2000. It has 2,000 texts with a length of 500 words each text covering 10 majors in both British English and American English. It is an annotated corpus and word frequency lists have been produced. It is designed to meet the needs of 224 Yang Xiao-jun students of English used in science and technology. Here is the website: www.sjtu.edu.cn (2) GPEC (Guangzhou Petroleum English Corpus). This corpus was compiled by Prof. Zhu Qi-bo at Guangzhou Training College of the Chinese Petroleum University in 1986. It has 700 passages with 500-600 words in each passage sampled over period of 1975-1986. Its size is 411,612 words in both British English and American English. A pack of concordance programmes for the corpus has been worked out. This corpus is designed to enable the study of the lexicon of Petroleum English and provide information for comparative language analysis. (3) Corpus for EST. This corpus was compiled jointly by Guangzhou University of Foreign Studies and the Hong Kong University of Science and Technology in 1998. It comprises 5 million words. (4) Communicative English Corpus for Chinese Students. This corpus was compiled at the School of Foreign Languages and International Studies, Guangzhou University of Foreign Studies. (5) English Literature Corpus. This corpus was compiled at the School of Foreign Languages and International Studies, Beijing Foreign Studies University. It comprises 5 million words. (6) Chinese Learner English Corpus (CLEC). This corpus was compiled in 1999 under the guidance of Prof. Gui Shichun of Guangzhou University of Foreign Studies and Prof. Yang Huizhong of Shanghai Jiaotong University. It comprises 1 million words. The written texts in this corpus are from the compositions and essays of either English majors or nonEnglish majors or middle school students, (40%, 30% and 30% respectively). A subcorpus, College Learner Corpus, was compiled at School of Foreign Studies, Henan University under the guidance of Prof. Li Wenzhong. (7) The HKUST Corpus of Learner English. This corpus was compiled by J. Milton at the Computer Department of the Hong Kong University of Science and Technology. It comprises 25 million words. This corpus was POS tagged and partly error-tagged. All language materials are from Chinese learners. Kennedy (1998: 42) has made some comments on this corpus as follows: Interlanguage studies of the written English of mainly Cantonese learners of English will be facilitated by the completion of the five-million-word Hong Kong University of Science and Technology (HKUST) Corpus (Milton & Tong, 1991). The corpus will be probably the largest machine-readable corpus yet produced of the written English of Chinese learners and also one of the largest corpora of any single group of learners. It is intended that it will be available with grammatical and discourse feature tags. The use of this corpus to describe the rule base of written English for learners of Chinese background is intended to inform the development of English teaching materials. (8) Corpus for Middle School English Education (MSEE). This corpus was compiled by Prof. He An-ping at South China Normal University in Survey and Prospect of China’s Corpus-Based Research 225 1999. It comprises 2.3 million words in three sections: 1.3 million words of the written and spoken English produced by the secondary school students in Guangdong Province; 0.5 million words of the complete new set of English textbooks used in China’s middle schools, and 0.5 million transcribed words of 130 hours of classroom teaching, matched with video and tape recordings. 2.2.3 Bilingual Corpora (including Parallel Corpora) (1) CONULEXID (the Commercial Press and Nanjing University Lexical Database). This is a bilingual (English and Chinese) Database and was compiled jointly by Nanjing University and Commercial Publishing House in 2000. It is for the purposes of bilingual dictionary making and publishing. (2) Chinese/Japanese Parallel Corpus (CJPC). This corpus was compiled at the National Research Centre for Foreign Language Education, Beijing Foreign Studies University, under the guidance of Prof. Xu Yiping. It is the first Chinese/Japanese parallel corpus compiled in China, upon which many significant results have been produced. (see Xu and Cao, 2002) (3) Chinese/English Parallel Corpus (CEPC). This corpus is being compiled at the National Research Centre for Foreign Language Education, Beijing Foreign Studies University, under the guidance of Dr.Wang Kefei. The objective of the present project is to create a Chinese/English Parallel Corpus of 30 million words, representative of modern Chinese and English in the twentieth century so as to establish a research platform for comparative studies of Chinese and English that can meet observational and descriptive accuracy, translation studies and teaching, statistical analysis such as frequency of occurrence, machine translation and compilation of bilingual dictionaries. This corpus comprises four subcorpora: bilingual corpus of aligned sentences and phrases, multidisciplinary corpus, specialised corpus, and translated texts corpus. All the texts are from written sources. Software such as bilingual text alignment software and bilingual concordance software will be developed for effective and efficient use of this corpus. This corpus has these distinct features: it can be separated for specific purposes or combined as a whole for a general purpose; the original texts have different translation, etc. (see Wang kefei, 2002b). (4) English/Chinese Parallel Corpus (ECPC). Under the Statistical Inter-Lingual Conversion (SILC) project, a 60Mb English/Chinese parallel corpus has been compiled by the Computer Department of the HKUST, containing transcripts of speech recordings and their high-quality translations of the Hong Kong parliament. From the corpus, 29 Mb of English texts (about 5 million words) and 15.5 Mb of Chinese equivalents were used for the experiment. They were all automatically aligned at sentence and paragraph level. 226 Yang Xiao-jun (5) Cantonese-English Bilingual Child Language Corpus. This is a corpus of Cantonese-English Bilingual Children’s early language development which is being compiled by Virinia Yip (CUHK) and Stephen Mattews (HKU). Three simultaneous bilingual subjects will be studied longitudinally for one and a half or two years and video-taped bi-weekly. The resulting transcriptions will form a bilingual corpus, containing English and Cantonese in romanised form for each child. This will be the first Cantonese-English Bilingual Child corpus and will be useful for addressing issues such as the question of differentiation, the degree of balance between the two languages and the possibility of delay relative to monolingual development. 2.3 Leading figures and institutions in the research A: The following is the introduction of the leading figures, whose names are spelt in Chinese Pinyin with the surname followed by the middle and last name, and their contributions to China’s corpus-based research. (1) Yu Shiwen is a doctor supervisor and the director of the Institute of Computational Linguistics, Department of Computer Science and Technology, Peking University. He has been carrying on research on computational linguistics and natural corpus processing since 1986. He compiled the Grammatical Knowledge Base of Contemporary Chinese (GKBCC). This database was published in Tsinghua University Press in 1997. Based on the GKBCC, he also compiled the database the Grammatical Knowledge Base of Contemporary Chinese---a Complete Specification, which was published at Tsinghua University Press in 1998. (2) Huang Changning works in The Research Institute for Computational Language of Tsinghua University and he has written 5 papers on corpusbased research, mainly on Chinese corpora and 1 book—Corpus Linguistics (Chinese version) published by Commercial Press in 2002. (3) Zhang Pu works in The Research Institute for Computational Language of Beijing Language and Culture University and he has written 10 papers on corpus-based research, mainly on Chinese corpora. He is one of the leading researchers for having compiled Modern Chinese Corpus. (4) Gui Shichun is a linguist working in Guangzhou University of Foreign Studies and has compiled the Chinese Learner English Corpus (CLEC) and undertaken some corpus-based research. (5) Yang Huizhong is a doctor supervisor of Foreign Studies Department of Shanghai Jiao Tong University. He has compiled two major corpora with other professors: JDEST and CLEC, and wrote in 1985 a paper—The use of computers in English teaching and research in China and in 2002 with his four doctoral candidates majoring in corpus linguistics wrote a book—An Introduction to Corpus Linguistics. He is one of the pioneers of China’s corpus-based research. Survey and Prospect of China’s Corpus-Based Research 227 (6) Gu Yueguo is a doctor supervisor of the Research Institute for Languages and Linguistics, the Chinese Academy of Social Science. He is in charge of the research programme for compiling a corpus—Situated Discourse for Spoken Chinese corpus, which he is about to complete. In 1998 he edited a special issue for corpus linguistics research as the first issue of Contemporary Linguistics and contributed a paper—Corpora and Language Research to this issue and gave some advice on this kind of research. He has had four doctoral candidates majoring in corpus linguistics. (7) Wang Kefei is a doctor supervisor of the National Research Centre for Foreign Language Education, Beijing Foreign Studies University. He is in charge of compiling Chinese/English Parallel Corpus. He has published four papers on parallel corpus-based approach to both language and translation studies. (8) Chen Guohua is a doctor supervisor of the National Research Centre for Foreign Language Education, Beijing Foreign Studies University and one of the leading members for compiling the Chinese/English Parallel Corpus. He has undertaken some research on corpus-based approaches to lexicography and has one doctoral candidate majoring in corpus linguistics and lexicography. (9) Pan Yongliang is a doctor supervisor in PLA Foreign Studies University and wrote two papers on corpus-based research and has several doctoral candidates majoring in corpus linguistics. (10) Wang Jianxin works at Foreign Languages Department of the Beijing University of Post and Telecommunication and takes part in compiling the Chinese/English Parallel Corpus. He has written six papers on corpus-based research. Among these papers, one was published in the International Journal of Corpus Linguistics (2001, vol. 6, No. 2); another was published in ICAME Journal (2000, 4). He studied corpus linguistics in Bergen University as a visiting scholar for one year. (11) He Anpin works at the Foreign Studies Department of South China Normal University and attended two conferences: ICAME 2000 (Sydney) and Corpus Linguistics 2001 (Lancaster). She has published more than five papers on corpus-based research and compiled a corpus - MSEE. (12) Li Wenzhong. Dr. Li has majored in corpus linguistics and is supervised by Prof. Yang Huizhong. He has compiled a sub-corpus of CLEC— College Learner Corpus wrote some papers on corpus-based research and three chapters of Yang Huizhong’s An Introduction to Corpus Linguistics. B: The following are the leading institutions in this research and their addresses, websites and the e-mail addresses of the project researchers. 228 Yang Xiao-jun (1) The State Key Laboratory of Intelligent Technology and Systems, Tsinghua University, Beijing, 100084, P.R. China. E-mail: [email protected]; website: www.tsinghua.edu.cn (2) The Institute of Computational Linguistics, Dept. of Computer Science and Technology, Peking University, Beijing, 100871, P.R. China. E-mail: [email protected]; website: www.pku.edu.cn (3) The Academia Sinica of Taiwang Website: www.academia.sinica.edu.tw. Scripta Sinica. E-mail: [email protected]; website: academia.sinica.edu.tw/info/index.html#db. (4) The National Research Center for Foreign Languages Education, Beijing Foreign Studies University, Beijing, 100089, P.R. China. E-mail: [email protected]; website: www.sinotefl.com (5) Language Information Processing Institute, Beijing Language and Culture University, Beijing , 100083, P.R. China. E-mail: [email protected]; website: www.blcp.edu.cn (6) The Chinese Academy of Social Science. No:5, Jian Guo Men Wai Da Jie, Beijing, 100732, P.R. China. www.cass.net.cn/jxky/yxsz/kx/index.htm. (7) Beijing University of Aeronautics and Astronautics. www.buaa.edu.cn (8) Open System & Chinese Information Processing Center, Institute of Software, the Chinese Academy of Sciences, Beijing, 100083. P.R. China. E-mail: [email protected] (9) Shanghai Jiaotong University, Shanghai, 200030,P.R. China. www.sjtu.edu.cn (10) Guangdong University of Foreign Studies, Guangdong, 510420, P.R. China. www.gdufs.edu.cn (11) South China Normal University, Guangdong, 510631, P.R. China. www.scnu.edu.cn (12) Wuhan University. www.whu.edu.cn (13) Northeast University, Shenyang, 110006, P.R. China. www.neu.edu.cn (14) Shanxi University, Taiyuan, 030006, P.R. China. www.sxu.edu.cn (15) The City University Of Hong Kong. www.cityu.edu.hk (16) Digital Heritage Publishing Ltd. (Hong Kong). www.skqs.com (17) The Chinese University of Hong Kong. www.cuhk.edu.hk (18) University of Petroleum-Guangzhou. www.ccem.uiuc.edu/chen/up/guangzhou.html. 2.4 Classifications of corpus-based research publications and major academic journals 2.4.1 Dictionaries and books (1) The Comprehensive Dictionary of the Chinese Language (in electronic form). The original 12 volumes of this dictionary was made available on one CD-ROM in 1998 by the Chinese Dictionary Press and the Commercial Press (Hong Kong Ltd.). Survey and Prospect of China’s Corpus-Based Research 229 (2) Grammatical Knowledge Base of Contemporary Chinese (GKBCC). This database was compiled by Peking University and already published in Tsinghua University Press in 1997. Based on the GKBCC, the database the Grammatical Knowledge Base of Contemporary Chinese---a Complete Specification has been compiled by Yu Shiwen, et al. at the ICL of Peking University and was published at Tsinghua University Press in 1998. (3) List of Common Characters in Modern Chinese. It was published in the Chinese Language Press in 1997. (4) Frequency Dictionary of the Modern Chinese Language. This database was compiled at Language Information Processing Institute, Beijing Language and Culture University and has frequency statistics for 1, 310,000 characters of Modern Chinese characters and was published in Beijing Language and Culture University Press in 1997. (5) An Introduction to Corpus Linguistics. This book was written by Yang Huizhong et al. and published by Shanghai Foreign Language Education Press in 2002. (6) An Introduction to Chinese Learner English Corpus (CLEC). This book was written by Yang Huizhong, et al and published by Shanghai Foreign Language Education Press in 2003. (7) Corpus Linguistics. This book was written by Huang Changning and Li Juanzi and published by the Commercial Press in 2002. (8) An Introduction to the Corpus for Middle School English Education (MSEE). This book was written by He Anping and published by Guangdong Electronic Press in 1999. 2.4.2 Classifications of corpus-based research papers As mentioned above, up to now there are about 140 corpus-based research papers (it is by no means exhaustive). They may be classified as follows: (1) In terms of the language in the corpora - Papers on English corpora and bilingual corpora take about 70% of the total number, the rest 30% are on Chinese corpora. (2) In terms of the research orientation and scope - Papers on the introduction to English and Chinese corpora and book reviews on corpus linguistics take about 20% and they mainly cover the period 1985-1995, such as Wang Jianxin (1996) and Pan Yongliang (2000). Papers on theoretical research on corpus linguistics take about 30% and they mainly cover the period 1996-2000. Papers on the applications of corpora take 40% and they mainly cover the period 1998-2003. They may be further classified into the following categories: (a) Papers on the application of corpora to translation studies take about 5%, such as Wang Kefei (2002 a, b), Liao Qiyi (2000), Zhang Meifang (2002), etc., (b) Papers on the application of corpora to language teaching take about 20%., (c) Papers on the application of corpora to contrastive linguistic analysis take about 5%, (d) Papers on the application of corpora to lexicography take about 5%, (e) Papers on the application of corpora to lexical studies take 5%. Papers on the techniques for corpora annotation and concordance take about the rest 10%. 230 Yang Xiao-jun 2.4.3 Classification of these academic journals These academic journals that publish papers on corpus-based research in China may be classified into three types: journals of foreign languages and linguistics, journals of Chinese language and linguistics, and journals of information and computer technology. Here the major ones are listed: (a) Contemporary Linguistics (Beijing); (b) Foreign Language Teaching and Research (Beijing); (c) Journal of Foreign Languages (Shanghai); (d) Modern Foreign Languages (Guangzhou); (e) Foreign Languages and Their Teaching (Dalian); (f) Journal of PLA Foreign Studies University (Luoyang); (g) Foreign Language Teaching (Xi’an); (h) Journal of Languages and Writings (Beijing); (i) Journal of Chinese Information Processing (Beijing); (j) Applied Linguistics (Beijing); and (k) Journal of Computer Science (Beijing). 3. Prospect of China’s corpus-based research In this section, we focus on the developing trend of China’s corpus-based research, corpus annotation problems, corpus processing tools and how to apply the corpora to language teaching and translation studies. 3.1 The developing trend of China’s corpus-based research The present situation in China’s corpus-building development is that the development of spoken Chinese corpora is far behind that of the written Chinese corpora. What is even worse is that there is no spoken English corpora already built or even being built in China, so we should speed up the building of spoken corpora to maintain the balance with that of written corpora. Meanwhile it has been suggested that more small corpora be built for specific research purposes and more bilingual and parallel corpora be built for language research, translation studies, contrastive analysis and dictionary-making. 3.2 Corpus annotation problems and corpus processing tools Technology for tagging and parsing should be standardised and further developed. Annotation on lexical, phonetic, syntactic and phonological levels goes far beyond that of semantic and pragmatic levels, so more attention should be paid to this aspect. The supply of more and more automatic taggers and parsers cannot meet the demand for them and most of these corpus processing tools are only suitable for specific corpora. Technology for alignment is not currently satisfactory. Up to now, we have not developed our own tools to process bilingual corpora. The challenge to corpus analysis tools is to systematise the design of corpora and concordancers so that any concordancer can work on any corpus. Survey and Prospect of China’s Corpus-Based Research 3.3 231 How to apply the corpora to language teaching and translation studies There is great potential for these corpora to enrich China’s corpus-based research. Here we just focus on how to apply the corpora to language teaching and translation studies. 3.3.1 Corpora and language teaching Corpus linguistics has a double role in language teaching; entailing a methodological innovation and a theoretical one, because together they will account for a new way of teaching (Tognini-Bonell, 2001: 14). The development of corpora has the potential for two major effects on the professional life of the language teacher. First, corpora lead to new descriptions of a language, so that the content of what the language teacher is teaching is perceived to change in radical ways (Sinclair, 1991: 100; Stubbs, 1996: 231-232). Secondly, corpora themselves can be exploited to produce language teaching materials, and can form the basis for new approaches to syllabus design and to methodology (Huston 2002: 137). In fact, we can apply corpora to teaching nearly all the language major courses, especially comparative and contrastive studies, vocabulary lessons and writing lessons, etc. 3.3.2 Corpora and translation studies Translation is an increasingly important application of corpora. Research into corpora and translation tends to focus on two areas: practical and theoretical. In practical terms, the question is: What software can be developed that will enable a translator to exploit corpora as an aid in the day-to-day business of translation? In theoretical terms, the question is: What does a corpus consisting of translated texts indicate about the process of translation itself? Because corpora can be used to raise awareness about language in general, they are extremely useful in training translators and in pointing up potential problems for translation. Not only can corpora provide evidence for how a given word or phrase are possible, they also provide an insight into the process and the nature of translation itself (Huston 2002: 123-128). So our bilingual parallel corpora can be a resource and a practical tool to analyse translation, translators’ style and help translation training, teaching and the writing of translation textbooks, etc. 4. Conclusion In this chapter, a general survey of China’s corpus-based research has been conducted from four viewpoints. Prediction of the prospect of the research has also been made from three viewpoints. It is obvious that corpus linguistics is playing a more and more important role in China’s corpus-based research. In the future, a national association of corpus linguistics studies will be created and a national academic journal of corpus linguistics will come into being. In the end, 232 Yang Xiao-jun our motto should never be forgotten: “We are corpus-based, but not corpusbound.” Notes 1 This paper has been polished by my supervisor, Dr. Wang Kefei, Professor of the National Research Centre for Foreign Language Education, Beijing Foreign Studies University. I am very grateful to him and to Prof. Wang Jianxin who offered me some information and data on corpora compilation. 2 Here “Hua” refers to Tsinghua University; “Yu” refers to Beijing Language and Culture University. References Biber, D., S. Conrad and R. Reppen (1998), Corpus Linguistics. Cambridge: Cambridge University Press. He, Anping. (1999), Introduction to the corpus for middle school English Education (MSEE). Guangzhou: Guangdong Electronic Press. Huang, Changning and Li Juanzi. (2002), Corpus Linguistics. Beijing: Commercial Press. Hunston, S. (2002), Corpora in Applied Linguistics. Cambridge: Cambridge University Press. Kennedy, G. (1998), An Introduction to Corpus Linguistics, London: Longman. Liao, Qiyi. (2000), ‘Corpora and translation studies’, Foreign Language Teaching and Research, 5: 380-384. Pan, Yongliang. (2000), ‘Introduction to Corpus Linguistics (Biber, et al 1998)’, Foreign Language Teaching and Research, 5: 389-392. Sinclair, J.M. (1991), Corpus Concordance Collocation,. Oxford: OUP. Stubbs, M. (1996), Text and Corpus Analysis, Oxford: Blackwell. Thomas, J. and M. Short. (1996), Using Corpora for Language Research, London: Longman. Tognini-Bonelli, E. (2001), Corpus Linguistics at Work. Amsterdam/ Philadelphia: John Benjamins Publishing Company. Wang, Jianxin. (1996), ‘Introduction to three contemporary English corpora’, Foreign Language Teaching and Research 3:37-40. Wang, Jianxin. (1998), ‘Important stages in the development of corpus linguistics’, Foreign Language Teaching and Research 4:52-57. Wang, Jianxin. (1999), ‘Some development of researches on corpus linguistics in China’, Foreign Languages and Their Teaching 3: 18-20. Wang, Jianxin. (2001), ‘Recent progress in corpus linguistics in China’, International Journal of Corpus Linguistics, Volume 6: Survey and Prospect of China’s Corpus-Based Research 233 Wang, Kefei. (2002a), ‘Parallel corpus-based approach to translation studies’, Foreign Languages and Their Teaching, 9: 35-39. Wang, Kefei. (2002b), ‘Corpus and network for translator and interpreter training’, Foreign Language Teaching and Research, 3:231-232. Wang, Lidi. and W. Jianxin. (2001), ‘Implementation plan for a 30-million-word Chinese/English parallel corpus’, Presentation at the 3rd International Symposium on EFL in China. May, Beijing. Xu, Yiping. and Cao. Dafeng, (eds.). (2002), The development and application of a Chinese/Japanese parallel corpus, Beijing: Foreign Language Teaching and Research Press. Yang, Huizhong. (2002), An Introduction to Corpus Linguistics, Shanghai: Shanghai Foreign Language Education Press. Yang, Xiaojun. and Li. Saihong. (2003), ‘Advantages of Corpora in Lexicography----Review on OALD (6th edition)’, Foreign Languages and Their Teaching, 4: 47-51. Zhang, Meifang. (2002), ‘Exploiting corpora to analyze the stylistic features of the translators’, Journal of PLA Foreign Studies University 3: 10-14.