Verb Sense Disambiguation Tiago Manuel Paulo Travanca
Transcription
Verb Sense Disambiguation Tiago Manuel Paulo Travanca
Verb Sense Disambiguation Tiago Manuel Paulo Travanca Advisors: Prof. Nuno Mamede and Prof. Jorge Baptista Instituto Superior T´ecnico Spoken Language Systems Laboratory – INESC-ID Lisboa Rua Alves Redol 9, 1000-029 Lisboa, Portugal [email protected] Abstract. This work addresses the problem of Verb Sense Disambiguation (VSD) for the European Portuguese. VSD is a sub-problem of the Word Sense Desambiguation (WSD) problem, which tries to identify the sense of a polysemic word used in a given sentence. Two approaches to VSD were considered. The first, rule-based, approach makes use of the lexical, syntactic and semantic descriptions of the verb senses present in a lexical resource (ViPEr). The second approach uses machine learning with a set of features commonly used in WSD. Both approaches were tested in several scenarios to determine the impact of different features and different combinations of methods. The baseline accuracy of 84%, resulting from the Most Frequent Sense (MFS) for each verb lemma, was both surprisingly high and hard to surpass. Still, both approaches provided some improvement over this value. The best combination of the three techniques yielded an accuracy of 87.2%, a gain of 3.2% above the baseline. 1 Introduction Nowadays, Natural Language Processing (NLP) has many applications: search engines (semantic/topic search rather than word matching), automated speech translation, automatic summarization, among others. Still, one problem that all NLP systems have to deal with is ambiguity, i.e. the fact that a certain expression can be interpreted in more than one way. Ambiguity appears at different stages in the processing of a text or a sentence, such as, tokenization, sentence-splitting, part-of-speech (POS) tagging, syntactic parsing and semantic processing. This work addresses semantic ambiguity of verbs. Considering that the sentence has already been parsed and, even if its syntactic analysis (parse tree) is unique and correct, some words may feature more than one meaning for the grammatical category they were tagged with. For example, sentences (1a) and (1b) show how the verb to count, used in the same position on both sentences, means enumerate on the first sentence while it stands for rely on the second: (1a) I am counting the number of students (1b) I am counting on you to help me 2 In these examples, the meaning of count in (1a) results from its directtransitive construction and plural object, while in (1b) the preposition introducing the second argument and the human trait of that same argument of count determine the verb’s sense. To tackle this problem, two modules were developed: one, rule-based, using a set of linguistic resources, specifically built for Portuguese (ViPEr); and another, based on machine learning, which makes use of a set of features, commonly adopted for the WSD problem, to correctly guess the verb sense. Both modules were integrated in a fully-fledged NLP system, STRING [1]. 2 Related Work 2.1 Lexical Resources Many works on WSD [2,3,4] rely on existing linguistic resources as catalogues of word senses (and their relations). For English, one can cite WordNet [5], FrameNet [6,7,8] and VerbNet [9]. For (European) Portuguese, though other resources have been built they are not yet freely available to the general public. We thus considered using ViPEr [10], which features a classification based on the syntactic-semantic properties of the most frequently occurring verbs. In the particular case of VSD, the most important features to be considered are formal features, such as the syntactic patterns of the sentence forms, and the distributional constraints resulting from the selectional restrictions of the verb on its arguments. Besides the resource here mentioned, there are other known sources with linguistic information on verbal constructions of Portuguese. Among others, the most relevant (and recent) are described in [11] and [12], each developed in a different theoretical framework. However, none of these works have a machine readable form, thus, they cannot be automatically processed. 2.2 Word Sense Disambiguation Approaches An overview of the most used algorithms and features for WSD was conducted, based on the systems evaluated at the SensEval31 [13] All-Words and Lexical Sample tasks (2004). The most commonly used techniques [14] were the Naive Bayes algorithm, Decision Lists, Vector Space Model (with cosine similarity function) and Support Vector Machines; other methods, such as Maximum Entropy models, Adaptative boosting and Latent Semantic Analysis, were also used by some systems. The majority of systems proposed at SensEval3 divided the features used to describe the instances into four different groups, as follows: – Collocations: n-grams (usually bigrams or trigrams) around the target word, composed by the lemma, word-form and part-of-speech tag of each word. – Syntactic dependencies: dependencies connecting the surrounding words to the target word. Typically, the relations of subject, object and modifier are used. 1 http://www.senseval.org/ 3 – Surrounding context: words in a defined window size are collected and used in a bag-of-words approach. Usually, the word form and lemma are extracted; – Knowledge-Based information: some systems make use of information such as WordNet domains, or FrameNet syntactic patterns, for example. The results obtained by the best systems at the SensEval3 conference for the English All-Words and the Lexical Sample tasks were 65.1% and 72.9% precision, respectively. 3 Rule-based Disambiguation Mappings Verb Lexicon Rules Rule Sorting Differences Rule Generation Verb List Difference Finder ViPEr Pre Processing The large number of verbs involved and the complexity of the rule generation task required the construction a specific module for rule generation, whose architecture is illustrated in Figure 1. Disamb. Rules Feature Priority Fig. 1. Rule Generation Module Architecture The first step in the rule generation process, the Pre-Processing, takes as its input the lexical resource information (ViPEr) and converts the linguistic information to an internal structure that represents each verb as a set of meanings. Each meaning, in turn, is represented as a collection of features and their possible values. This module is also responsible for producing the lexicon file (the verb lemmas and their meanings) used by the grammar of XIP [15], the parsing module of STRING. After the parsing is complete, the system passes the processed information to the next module: the Difference Finder. This module compares the different meanings of the same lemma and generates a difference for each argument found different between each pair of meanings. The rule generation module is the last transformational step in the system. The module transforms differences into rules, with each difference usually generating two rules, one for each verb meaning contained in that difference. This step requires an additional parametrization file, mappings, to make the bridge between the features used in ViPEr and those provided by XIP. 4 Finally, the system sorts the rules according to the feature/trait being tested in each of them, since XIP executes the rules in the order they are present in the file. This order favours, firstly, the properties related to structural patterns in sentences, followed by several semantic/selectional restrictions, according to their strength. The base features considered by the system are the selectional restrictions on the verb’s arguments (N0 to N3 ) and their respective prepositions, the property vse, denoting intrinsically reflexive constructions (e.g. queixar-se ‘complain’) and transformational properties regarding the most common two types of passives (with auxiliary verbs ser and estar ‘be’). Additionally, two other properties were added to ViPEr during this work and are used by the rule generation system. The first, vdic property, indicates which verb meanings (usually only one per lemma) allows the verbum dicendi construction pattern, with subject inversion. Then, the aNi=lhe feature, indicating the possibility of a dative pronominalization for some of the verb arguments. 4 Machine Learning Disambiguation The architecture of the supervised classification approach, presented in Figure 2, uses MegaM [16] as the machine learning module, which is based on maximum entropy models [17]. The typical feature extraction and classification modules of supervised classification systems were both implemented inside XIP (that is, inside STRING), to allow XIP rules to be built afterwards to solve other NLP tasks, such as semantic role labeling. In the training phase, STRING takes annotated data (sentences) and extracts the features describing each verb instance and its respective manually annotated verb class (label). Then, MegaM uses that information to build a classification model for each verb lemma. In the prediction phase, STRING assigns a verb class (label) to each verb instance in the raw data, using the models built by MegaM. Label STRING Annotated Data MegaM Classification Model Features STRING Raw Data Labelled Data Fig. 2. Supervised Machine Learning Architecture for VSD using STRING 5 4.1 Training Corpus Due to the time required for manual annotation of the training data, only some verbs were addressed by this approach during this project. The verbs chosen were: explicar, falar, ler, pensar, resolver, saber and ver. This choice was based on the higher number of instances left to disambiguate that these lemmas exhibited after the rule-based testing. Further analysis revealed that these verbs also formed an heterogeneous group, with different MFS percentage values, different number of senses and different ambiguity classes. 4.2 Features The features used by the machine learning module follow those presented in section 2.2, and can be organized into three groups, as follows: – Local features, extracted by the system in a window of size 3 around the target verb, resulting in a total of 6 tokens, with their respective indexes (-3 to +3). Information regarding the POS tag and lemma of each token was considered. – Syntactic features, regarding constituents with a direct relation with the verb. The dependencies used in the system are those of SUBJ (subject), CDIR (direct complement) and MOD (modifier). The POS tag and lemma of the head word of each node in these relations is extracted, together with the dependency name. – Semantic features are extracted for the head words of the nodes that have a direct relation with the verb, as they act as the selectional restrictions on the verb arguments. Each semantic trait found in the head word of the node is also accompanied by the dependency name. The semantic features considered by the system are those defined in ViPEr. Additional features that do no fit exactly into any of those categories were also used. These features concern the subject inversion pattern of verbum dicendi constructions and information regarding the case of clitic personal pronouns. 5 Evaluation To evaluate the two approaches presented in the previous sections, a manually annotated corpus had to be built and validated. The corpus chosen for this evaluation contains around 250 thousand words with, 38,827 verbs manually disambiguated, from which 21,368 (about 55%) are full verbs. From the full verb constructions, 13,030 correspond to ambiguous verbs. The most frequent sense (MFS) for each verb lemma was adopted as the baseline against which the developed modules were compared and evaluated. Since the corpus was the only available annotated data, it had to be used both for counting (to determine the MFS) and for evaluation. To avoid biased results, the common 10-fold cross-validation method was used. An accuracy of 84% was achieved by MFS. This is related to the conservative approach underlying the ViPEr classification, leading to an average lower number of senses per verb when compared to other lexical resources. 6 5.1 Rules+MFS This experiment tested three different combinations of the rule-based approach and the MFS (baseline). In this case, the MFS classifier was applied after the rule-based module to the remaining non-fully disambiguated instances. In the first experiment, the disambiguation rules were applied to all the verbs, regardless of their number of senses per lemma, obtaining an accuracy of 79.15%, below the baseline of 84%. For the second scenario, a finer grained combination was performed, with rule-based disambiguation being applied only to verbs with 8 or more meanings. This resulted in 84.41% accuracy, which was 0.41 percentage points over the baseline. The last experiment, considered the results of the previous one, but also included other verb lemmas that had shown positive results with rules, but had fewer meanings. The final result obtained in the rule-based disambiguation experiments reached the 86.46% accuracy (2.46 percentage points above the baseline). 5.2 Machine Learning Disambiguation Accuracy The first machine learning disambiguation experiment tested the impact of using bias, a special feature calculated by the system during the training step, which indicates the deviation of the model towards each class. Comparing the two scenarios (with and without bias) against the previously presented methods showed that, whenever a verb has a high MFS, it is difficult for another approach to surpass it, and only ML-Bias can equal the MFS accuracy. All verbs with low MFS exhibited some improvement in the results (using ML-NoBias) when compared to the MFS baseline. Still, for the verb pensar, ML-NoBias is outperformed by the combination of rules and MFS. 100 90 80 70 60 50 40 30 20 10 0 explicar falar MFS ler Rules+MFS pensar ML-NoBias resolver ML-Bias saber ver Rules+ML Fig. 3. Machine Learning Results Comparison Looking at the set of classes associated with each lemma, both resolver and ver share the same ambiguity set (06, 32C), while pensar has more senses (06, 08, 7 16, 32C, 35R, 36R). This suggests that ML might not be able to deal with verbs that have many senses or that a specific set of classes is better disambiguated by this technique, though further testing encompassing more verbs should be performed to verify this finding. Afterwards, another scenario was considered, to test the performance of ML as a complementary technique to the rule-based disambiguation, similarly to the scenario that combined rules and MFS, presented in section 5.1. Globally, adding ML as a complementary technique to rules proved to be worse than just using ML for the majority of the verbs here studied. This suggests that the instances being correctly classified by the rules are a subset of the instances correctly disambiguated by ML, and that rules are incorrectly classifying the remaining instances, which ML would otherwise guess correctly. This experiment showed that rules and ML were not very complementary, and that choosing ML alone provides the best improvement to the system. For the cases where Rules+ML surpasses ML alone, other approaches provide better results. Further testing of the ML scenarios should be performed to confirm these findings, namely, using annotated data regarding other verb lemmas and trying out other combination of methods and other scenarios. 5.3 Rules+MFS+ML With all the previous information regarding each verb lemma, a final combination using all the techniques was done, choosing the best method for each lemma. This last experiment obtained the best overall results, settling a new maximum accuracy of 87.2% for the system, improving 3.2 percentage points above the baseline. Finally, a performance test was conducted to assess the impact of the newly added modules to the global performance of STRING. This test revealed a minimal increase in the global time required for processing (around 0.05%, for 5,000 sentences). 6 Conclusions This work addressed the problem of verb sense disambiguation for the European Portuguese. Two approaches were considered when tackling this problem: a ruled-based, and a machine learning disambiguation. The results obtained from the baseline (MFS) were surprisingly high (about 84% accuracy) and proved to be hard to surpass. The integration of both modules provided better results to the system, achieving a final score of 87.2% accuracy, an improvement of 3.2% above baseline. Certainly, there is still much work to be done in VSD, and the system here presented can be improved. Nevertheless, we would like to think that some valid contribution has been made to the domain. The integral work here presented is available at http://www.inesc-id.pt/pt/ indicadores/Ficheiros/7279.pdf 8 References 1. Mamede, N.J., Baptista, J., Diniz, C., Cabarr˜ ao, V.: STRING: An Hybrid Statistical and Rule-Based Natural Language Processing Chain for Portuguese. In: Propor 2012, Demo Session (2012) Paper available at http://www.propor2012. org/demos/DemoSTRING.pdf 2. Buscaldi, D., Rosso, P., Masulli, F.: The upv-unige-CIAOSENSO WSD system. In Mihalcea, R., Edmonds, P., eds.: Senseval-3: 3rd International Workshop on the Evaluation of Systems for the Semantic Analysis of Text, Barcelona, Spain, ACL (2004) 77–82 3. Shi, L., Mihalcea, R.: Open Text Semantic Parsing using FrameNet and WordNet. In: Demonstration Papers at HLT-NAACL 2004. HLT-NAACL–Demonstrations ’04, Stroudsburg, PA, USA, ACL (2004) 19–22 4. Brown, S.W., Dligach, D., Palmer, M.: VerbNet class assignment as a WSD task. In: Proceedings of the 9th International Conference on Computational Semantics. IWCS ’11, Stroudsburg, PA, USA, ACL (2011) 85–94 5. Miller, G.A., Beckwith, R., Fellbaum, C., Gross, D., Miller, K.: WordNet: An On-line Lexical Database. International Journal of Lexicography 3 (1990) 235–244 6. Baker, C.F., Fillmore, C.J., Lowe, J.B.: The Berkeley FrameNet Project. In: Proceedings of the 36th Annual Meeting of the ACL and 17th COLING - Volume 1. ACL ’98, Stroudsburg, PA, USA, ACL (1998) 86–90 7. Johnson, C.R., Fillmore, C.J., Wood, E.J., Ruppenhofer, J., Urban, M., Petruck, M.R., Baker, C.F.: The FrameNet Project: Tools for Lexicon Building (2001) 8. Ruppenhofer, J., Ellsworth, M., Petruck, M.R., Johnson, C.R., Scheffczyk, J.: FrameNet II: Extended Theory and Practice. International Computer Science Institute, Berkeley, California (2006) Distributed with the FrameNet data. 9. Kipper, K., Dang, H.T., Schuler, W., Palmer, M.: Building a Class-Based Verb Lexicon using TAGs. In: TAG+5 5th Workshop on Tree Adjoining Grammars and Related Formalisms, Paris, France (2000) 10. Baptista, J.: ViPEr: A Lexicon-Grammar of European Portuguese Verbs. In: 31st International Conference on Lexis and Grammar, Nov´e Hrady, Czech Republic (2012) 10–16 11. Borba, F.: Dicion´ ario Gramatical de Verbos do Portuguˆes Contemporˆ aneo do Brasil. UNESP, S˜ ao Paulo (1990) 12. Busse, W.: Dicion´ ario Sint´ actico de Verbos Portugueses. Almedina, Coimbra (1994) 13. Mihalcea, R., Edmonds, P., eds.: Senseval-3: 3rd International Workshop on the Evaluation of Systems for the Semantic Analysis of Text, Barcelona, Spain, ACL (2004) 14. Witten, I.H., Frank, E., Hall, M.A.: Data Mining: Practical Machine Learning Tools and Techniques. 3rd edn. Morgan Kaufmann, Amsterdam (2011) 15. Ait-Mokhtar, S., Chanod, J., Roux, C.: Robustness beyond shallowness: incremental dependency parsing. Natural Language Engineering 8(2/3) (2002) 121–144 16. Daum´e, H.: Notes on CG and LM-BFGS Optimization of Logistic Regression. Paper available at http://pub.hal3.name#daume04cg-bfgs, implementation available at http://hal3.name/megam/ (2004) 17. Berger, A.L., Pietra, S.A.D., Pietra, V.J.D.: A Maximum Entropy approach to Natural Language Processing. Computational Linguistics 22 (1996) 39–71