Supervised Domain Adaption for WSD Eneko Agirre and Oier Lopez de Lacalle IXA NLP Group University of the Basque Country Donostia, Basque Contry {e.agirre,oier.lopezdelacalle}@ehu.es Abstract The lack of positive results on supervised domain adaptation for WSD have cast some doubts on the utility of handtagging general corpora and thus developing generic supervised WSD systems. In this paper we show for the first time that our WSD system trained on a general source corpus (B NC) and the target corpus, obtains up to 22% error reduction when compared to a system trained on the target corpus alone. In addition, we show that as little as 40% of the target corpus (when supplemented with the source corpus) is sufficient to obtain the same results as training on the full target data. The key for success is the use of unlabeled data with SVD, a combination of kernels and SVM . 1 Introduction In many Natural Language Processing (NLP) tasks we find that a large collection of manuallyannotated text is used to train and test supervised machine learning models. While these models have been shown to perform very well when tested on the text collection related to the training data (what we call the source domain), the performance drops considerably when testing on text from other domains (called target domains). In order to build models that perform well in new (target) domains we usually find two settings (Daumī III, 2007). In the semi-supervised setting, e the training hand-annotated text from the source domain is supplemented with unlabeled data from the target domain. In the supervised setting, we use training data from both the source and target domains to test on the target domain. In (Agirre and Lopez de Lacalle, 2008) we studied semi-supervised Word Sense Disambigua- tion (WSD) adaptation, and in this paper we focus on supervised WSD adaptation. We compare the performance of similar supervised WSD systems on three different scenarios. In the source to target scenario the WSD system is trained on the source domain and tested on the target domain. In the target scenario the WSD system is trained and tested on the target domain (using cross-validation). In the adaptation scenario the WSD system is trained on both source and target domain and tested in the target domain (also using cross-validation over the target data). The source to target scenario represents a weak baseline for domain adaptation, as it does not use any examples from the target domain. The target scenario represents the hard baseline, and in fact, if the domain adaptation scenario does not yield better results, the adaptation would have failed, as it would mean that the source examples are not useful when we do have hand-labeled target examples. Previous work shows that current state-of-theart WSD systems are not able to obtain better results on the adaptation scenario compared to the target scenario (Escudero et al., 2000; Agirre and Martīnez, 2004; Chan and Ng, 2007). This would i mean that if a user of a generic WSD system (i.e. based on hand-annotated examples from a generic corpus) would need to adapt it to a specific domain, he would be better off throwing away the generic examples and hand-tagging domain examples directly. This paper will show that domain adaptation is feasible, even for difficult domainrelated words, in the sense that generic corpora can be reused when deploying WSD systems in specific domains. We will also show that, given the source corpus, our technique can save up to 60% of effort when tagging domain-related occurrences. We performed on a publicly available corpus which was designed to study the effect of domains in WSD (Koeling et al., 2005). It comprises 41 Proceedings of the 12th Conference of the European Chapter of the ACL, pages 42­50, Athens, Greece, 30 March ­ 3 April 2009. c 2009 Association for Computational Linguistics 42 nouns which are highly relevant in the S PORTS and F INANCES domains, with 300 examples for each. The use of two target domains strengthens the conclusions of this paper. Our system uses Singular Value Decomposition (SVD) in order to find correlations between terms, which are helpful to overcome the scarcity of training data in WSD (Gliozzo et al., 2005). This work explores how this ability of SVD and a combination of the resulting feature spaces improves domain adaptation. We present two ways to combine the reduced spaces: kernel combination with Support Vector Machines (SVM), and k Nearest-Neighbors (k- NN) combination. The paper is structured as follows. Section 2 reviews prior work in the area. Section 3 presents the data sets used. In Section 4 we describe the learning features, including the application of SVD , and in Section 5 the learning methods and the combination. The experimental results are presented in Section 6. Section 7 presents the discussion and some analysis of this paper and finally Section 8 draws the conclusions. not help when tagging the target corpus, showing that tagged corpora from each domain would suffice, and concluding that hand tagging a large general corpus would not guarantee robust broadcoverage WSD. Agirre and Martīnez (2000) used i the DSO corpus in the supervised scenario to show that training on a subset of the source corpora that is topically related to the target corpus does allow for some domain adaptation. More recently, Chan and Ng (2007) performed supervised domain adaptation on a manually selected subset of 21 nouns from the DSO corpus. They used active learning, count-merging, and predominant sense estimation in order to save target annotation effort. They showed that adding just 30% of the target data to the source examples the same precision as the full combination of target and source data could be achieved. They also showed that using the source corpus allowed to significantly improve results when only 10%30% of the target corpus was used for training. Unfortunately, no data was given about the target corpus results, thus failing to show that domainadaptation succeeded. In followup work (Zhong et al., 2008), the feature augmentation approach was combined with active learning and tested on the OntoNotes corpus, on a large domain-adaptation experiment. They reduced significantly the effort of hand-tagging, but only obtained domainadaptation for smaller fractions of the source and target corpus. Similarly to these works we show that we can save annotation effort on the target corpus, but, in contrast, we do get domain adaptation when using the full dataset. In a way our approach is complementary, and we could also apply active learning to further reduce the number of target examples to be tagged. Though not addressing domain adaptation, other works on WSD also used SVD and are closely related to the present paper. Ando (2006) used Alternative Structured Optimization. She first trained one linear predictor for each target word, and then performed SVD on 7 carefully selected submatrices of the feature-to-predictor matrix of weights. The system attained small but consistent improvements (no significance data was given) on the Senseval-3 lexical sample datasets using SVD and unlabeled data. Gliozzo et al. (2005) used SVD to reduce the space of the term-to-document matrix, and then computed the similarity between train and test 2 Prior work Domain adaptation is a practical problem attracting more and more attention. In the supervised setting, a recent paper by Daumī III (2007) shows e that a simple feature augmentation method for SVM is able to effectively use both labeled target and source data to provide the best domainadaptation results in a number of NLP tasks. His method improves or equals over previously explored more sophisticated methods (Daumī III e and Marcu, 2006; Chelba and Acero, 2004). The feature augmentation consists in making three version of the original features: a general, a sourcespecific and a target-specific versions. That way the augmented source contains the general and source-specific version and the augmented target data general and specific versions. The idea behind this is that target domain data has twice the influence as the source when making predictions about test target data. We reimplemented this method and show that our results are better. Regarding WSD, some initial works made a basic analysis of domain adaptation issues. Escudero et al. (2000) tested the supervised adaptation scenario on the DSO corpus, which had examples from the Brown corpus and Wall Street Journal corpus. They found that the source corpus did 43 instances using a mapping to the reduced space (similar to our SMA method in Section 4.2). They combined other knowledge sources into a complex kernel using SVM. They report improved performance on a number of languages in the Senseval3 lexical sample dataset. Our present paper differs from theirs in that we propose an additional method to use SVD (the OMT method), and that we focus on domain adaptation. In the semi-supervised setting, Blitzer et al. (2006) used Structural Correspondence Learning and unlabeled data to adapt a Part-of-Speech tagger. They carefully select so-called `pivot features' to learn linear predictors, perform SVD on the weights learned by the predictor, and thus learn correspondences among features in both source and target domains. Our technique also uses SVD, but we directly apply it to all features, and thus avoid the need to define pivot features. In preliminary work we unsuccessfully tried to carry along the idea of pivot features to WSD. On the contrary, in (Agirre and Lopez de Lacalle, 2008) we show that methods closely related to those presented in this paper produce positive semi-supervised domain adaptation results for WSD. The methods used in this paper originated in (Agirre et al., 2005; Agirre and Lopez de Lacalle, 2007), where SVD over a feature-to-documents matrix improved WSD performance with and without unlabeled data. The use of several kNN classifiers trained on a number of reduced and original spaces was shown to get the best results in the Senseval-3 dataset and ranked second in the SemEval 2007 competition. The present paper extends this work and applies it to domain adaptation. ments the B NC examples play the role of general source corpora, and the F INANCES and S PORTS examples the role of two specific domain target corpora. Compared to the DSO corpus used in prior work (cf. Section 2) this corpus has been explicitly created for domain adaptation studies. DSO contains texts coming from the Brown corpus and the Wall Street Journal, but the texts are not classified according to specific domains (e.g. Sports, Finances), which make DSO less suitable to study domain adaptation. The fact that the selected nouns are related to the target domain makes the (Koeling et al., 2005) corpus more demanding than the DSO corpus, because one would expect the performance of a generic WSD system to drop when moving to the domain corpus for domainrelated words (cf. Table 1), while the performance would be similar for generic words. In addition to the labeled data, we also use unlabeled data coming from the three sources used in the labeled corpus: the 'written' part of the B NC (89.7M words), the F INANCES part of Reuters (32.5M words), and the S PORTS part (9.1M words). 4 Original and SVD features In this section, we review the features and two methods to apply SVD over the features. 4.1 Features 3 Data sets The dataset we use was designed for domainrelated WSD experiments by Koeling et al. (2005), and is publicly available. The examples come from the B NC (Leech, 1992) and the S PORTS and F INANCES sections of the Reuters corpus (Rose et al., 2002), comprising around 300 examples (roughly 100 from each of those corpora) for each of the 41 nouns. The nouns were selected because they were salient in either the S PORTS or F INANCES domains, or because they had senses linked to those domains. The occurrences were hand-tagged with the senses from WordNet (W N) version 1.7.1 (Fellbaum, 1998). In our experi- We relied on the usual features used in previous WSD work, grouped in three main sets. Local collocations comprise the bigrams and trigrams formed around the target word (using either lemmas, word-forms, or PoS tags) , those formed with the previous/posterior lemma/word-form in the sentence, and the content words in a ą4-word window around the target. Syntactic dependencies use the object, subject, noun-modifier, preposition, and sibling lemmas, when available. Finally, Bag-of-words features are the lemmas of the content words in the whole context, plus the salient bigrams in the context (Pedersen, 2001). We refer to these features as original features. 4.2 SVD features Apart from the original space of features, we have used the so called SVD features, obtained from the projection of the feature vectors into the reduced space (Deerwester et al., 1990). Basically, 44 we set a term-by-document or feature-by-example matrix M from the corpus (see section below for more details). SVD decomposes M into three matrices, M = U V T . If the desired number of dimensions in the reduced space is p, we select p rows from and V , yielding p and Vp respectively. We can map any feature vector t (which represents either a train or test example) into the p-dimensional space as follows: tp = tT Vp -1 . p Those mapped vectors have p dimensions, and each of the dimensions is what we call a SVD feature. We have explored two different variants in order to build the reduced matrix and obtain the SVD features, as follows. Single Matrix for All target words (SVD SMA ). The method comprises the following steps: (i) extract bag-of-word features (terms in this case) from unlabeled corpora, (ii) build the term-bydocument matrix, (iii) decompose it with SVD, and (iv) map the labeled data (train/test). This technique is very similar to previous work on SVD (Gliozzo et al., 2005; Zelikovitz and Hirsh, 2001). The dimensionality reduction is performed once, over the whole unlabeled corpus, and it is then applied to the labeled data of each word. The reduced space is constructed only with terms, which correspond to bag-of-words features, and thus discards the rest of the features. Given that the WSD literature shows that all features are necessary for optimal performance (Pradhan et al., 2007), we propose the following alternative to construct the matrix. One Matrix per Target word (SVD - OMT). For each word: (i) construct a corpus with its occurrences in the labeled and, if desired, unlabeled corpora, (ii) extract all features, (iii) build the featureby-example matrix, (iv) decompose it with SVD, and (v) map all the labeled training and test data for the word. Note that this variant performs one SVD process for each target word separately, hence its name. When building the SVD - OMT matrices we can use only the training data (TRAIN) or both the train and unlabeled data (+ UNLAB). When building the SVD - SMA matrices, given the small size of the individual word matrices, we always use both the train and unlabeled data (+ UNLAB). Regarding the amount of data, based also on previous work, we used 50% of the available data for OMT, and the whole corpora for SMA. An important parameter when doing SVD is the number of dimensions in the reduced space (p). We tried two different values for p (25 and 200) in the B NC domain, and set a dimension for each classifier/matrix combination. 4.3 Motivation The motivation behind our method is that although the train and test feature vectors overlap sufficiently in the usual WSD task, the domain difference makes such overlap more scarce. SVD implicitly finds correlations among features, as it maps related features into nearby regions in the reduced space. In the case of SMA, SVD is applied over the joint term-by-document matrix of labeled (and possibly unlabeled corpora), and it thus can find correlations among closely related words (e.g. cat and dog). These correlations can help reduce the gap among bag-of-words features from the source and target examples. In the case of OMT, SVD over the joint feature-by-example matrix of labeled and unlabeled examples of a word allows to find correlations among features that show similar occurrence patterns in the source and target corpora for the target word. 5 Learning methods k- NN is a memory based learning method, where the neighbors are the k most similar labeled examples to the test example. The similarity among instances is measured by the cosine of their vectors. The test instance is labeled with the sense obtaining the maximum sum of the weighted vote of the k most similar contexts. We set k to 5 based on previous results published in (Agirre and Lopez de Lacalle, 2007). Regarding SVM, we used linear kernels, but also purpose-built kernels for the reduced spaces and the combinations (cf. Section 5.2). We used the default soft margin (C=0). In previous experiments we learnt that C is very dependent on the feature set and training data used. As we will experiment with different features and training datasets, it did not make sense to optimize it across all settings. We will now detail how we combined the original and SVD features in each of the machine learning methods. 5.1 k-NN combinations Our k- NN combination method (Agirre et al., 2005; Agirre and Lopez de Lacalle, 2007) takes 45 advantage of the properties of k- NN classifiers and exploit the fact that a classifier can be seen as k points (number of nearest neighbor) each casting one vote. This makes easy to combine several classifiers, one for each feature space. For instance, taking two k- NN classifiers of k = 5, C1 and C2 , we can combine them into a single k = 10 classifier, where five votes come from C1 and five from C2 . This allows to smoothly combine classifiers from different feature spaces. In this work we built three single k- NN classifiers trained on OMT, SMA and the original features, respectively. In order to combine them we weight each vote by the inverse ratio of its position in the rank of the single classifier, (k - ri + 1)/k, where ri is the rank. 5.2 Kernel combination B NC X MFS k- NN SVM S PORTS 39.0 51.7 53.9 F INANCES 51.2 60.4 62.9 Table 1: Source to target results: Train on B NC, test on S PORTS and F INANCES. Finally, we define the kernel combination: n KComb (xi , xj ) = l=1 Kl (xi , xj ) Kl (xi , xi )Kl (xj , xj ) where n is the number of single kernels explained above, and l the index for the kernel type. 6 Domain adaptation experiments The basic idea of kernel methods is to find a suitable mapping function () in order to get a better generalization. Instead of doing this mapping explicitly, kernels give the chance to do it inside the algorithm. We will formalize it as follows. First, we define the mapping function : X F. Once the function is defined, we can use it in the kernel function in order to become an implicit function K(x, z) = (x) · (z) , where · denotes a inner product between vectors in the feature space. This way, we can very easily define mappings representing different information sources and use this mappings in several machine learning algorithm. In our work we use SVM. We defined three individual kernels (OMT, SMA and original features) and the combined kernel. The original feature kernel (KOrig ) is given by the identity function over the features : X X , defining the following kernel: KOrig (xi , xj ) = xi · xj xi · xi xj · xj In this section we present the results in our two reference scenarios (source to target, target) and our reference scenario (domain adaptation). Note that all methods presented here have full coverage, i.e. they return a sense for all test examples, and therefore precision equals recall, and suffices to compare among systems. 6.1 Source to target scenario: B NC X where the denominator is used to normalize and avoid any kind of bias in the combination. The OMT kernel (KOmt ) and SMA kernel (KSma ) are defined using OMT and SMA projection matrices, respectively (cf. Section 4.2). Given the OMT function mapping omt : Rm Rp , where m is the number of the original features and p the reduced dimensionality, then we define KOmt (xi , xj ) as follows (KSma is defined similarly): omt (xi ) · omt (xj ) omt (xi ) · omt (xi ) omt (xj ) · omt (xj ) In this scenario our supervised WSD systems are trained on the general source corpus (B NC) and tested on the specific target domains separately (S PORTS and F INANCES). We do not perform any kind of adaptation, and therefore the results are those expected for a generic WSD system when applied to domain-specific texts. Table 1 shows the results for k- NN and SVM trained with the original features on the B NC. In addition, we also show the results for the Most Frequent Sense baseline (MFS) taken from the B NC. The second column denotes the accuracies obtained when testing on S PORTS, and the third column the accuracies for F INANCES. The low accuracy obtained with MFS, e.g. 39.0 of precision in S PORTS, shows the difficulty of this task. Both classifiers improve over MFS. These classifiers are weak baselines for the domain adaptation system. 6.2 Target scenario X X In this scenario we lay the harder baseline which the domain adaptation experiments should improve on (cf. next section). The WSD systems are trained and tested on each of the target corpora (S PORTS and F INANCES) using 3-fold crossvalidation. 46 X X MFS TRAIN k- NN SVM k- NN - OMT SVM - OMT k- NN - SMA SVM - SMA k- NN - COMB SVM - COMB S PORTS + UNLAB 77.8 84.5 85.1 85.0 86.1 82.9 85.1 81.1 81.3 86. 0 86.7 86.5 F INANCES TRAIN + UNLAB 82.3 87.1 87.0 87.3 85.3 87.9 87.6 86.4 83.2 84.1 88.6 88.5 B NC + X X B NC X X X MFS TRAIN k- NN SVM Table 2: Target results: train and test on S PORTS, train and test on F INANCES, using 3-fold crossvalidation. Table 2 summarizes the results for this scenario. TRAIN denotes that only tagged data was used to train, + UNLAB denotes that we added unlabeled data related to the source corpus when computing SVD . The rows denote the classifier and the feature spaces used, which are organized in four sections. On the top rows we show the three baseline classifiers on the original features. The two sections below show the results of those classifiers on the reduced dimensions, OMT and SMA (cf. Section 4.2). Finally, the last rows show the results of the combination strategies (cf. Sections 5.1 and 5.2). Note that some of the cells have no result, because that combination is not applicable (e.g. using the train and unlabeled data in the original space). First of all note that the results for the baselines (MFS, SVM, k- NN) are much larger than those in Table 1, showing that this dataset is specially demanding for supervised WSD, and particularly difficult for domain adaptation experiments. These results seem to indicate that the examples from the source general corpus could be of little use when tagging the target corpora. Note specially the difference in MFS performance. The priors of the senses are very different in the source and target corpora, which is a well-known shortcoming for supervised systems. Note the high results of the baseline classifiers, which leave small room for improvement. The results for the more sophisticated methods show that SVD and unlabeled data helps slightly, except for k- NN - OMT on S PORTS. SMA decreases the performance compared to the classifiers trained on original features. The best improvements come when the three strategies are combined in one, as both the kernel and k- NN combinations obtain improvements over the respective single classifiers. Note that both the k- NN k- NN - OMT SVM - OMT k- NN - SMA SVM - SMA k- NN - COMB SVM - COMB SVM - AUG S PORTS + UNLAB 53.9 86.0 86.7 68.2 81.3 84.7 84.0 84.7 85.1 84.7 77.1 78.1 84.5 87.2 88.4 85.9 - F INANCES TRAIN + UNLAB 62.9 87.9 73.1 86.0 87.5 87.5 84.2 88.1 88.1 88.5 86.0 85.5 81.6 80.7 88.7 89.7 - Table 3: Domain adaptation results: Train on B NC and S PORTS, test on S PORTS (same for F I NANCES). and SVM combinations perform similarly. In the combination strategy we show that unlabeled data helps slightly, because instead of only combining OMT and original features we have the opportunity to introduce SMA. Note that it was not our aim to improve the results of the basic classifiers on this scenario, but given the fact that we are going to apply all these techniques in the domain adaptation scenario, we need to show these results as baselines. That is, in the next section we will try to obtain results which improve significantly over the best results in this section. 6.3 Domain adaptation scenario B NC + X X In this last scenario we try to show that our WSD system trained on both source (B NC) and target (S PORTS and F INANCES) data performs better than the one trained on the target data alone. We also use 3-fold cross-validation for the target data, but the entire source data is used in each turn. The unlabeled data here refers to the combination of unlabeled source and target data. The results are presented in table 3. Again, the columns denote if unlabeled data has been used in the learning process. The rows correspond to classifiers and the feature spaces involved. The first rows report the best results in the previous scenarios: B NC X for the source to target scenario, and X X for the target scenario. The rest of the table corresponds to the domain adaptation scenario. The rows below correspond to MFS and the baseline classifiers, followed by the OMT and SMA results, and the combination results. The last row shows the results for the feature augmentation algorithm (Daumī III, 2007). e 47 S PORTS B NC X MFS SVM F INANCES 51.2 62.9 accuracy (%) 88 39.0 53.9 77.8 85.1 86.7 68.2 84.7 85.9 88.4 X X MFS SVM k- NN - COMB (+ UNLAB ) B NC +X X MFS SVM SVM - AUG SVM - COMB (+ UNLAB ) 86 82.3 87.0 88.6 73.1 87.5 88.1 89.7 84 82 80 SVM-COMB (+UNLAB, BNC + SPORTS -> SPORTS) SVM-AUG (BNC + SPORTS -> SPORTS) SVM-ORIG (SPORTS -> SPORTS) y=85.1 %25 %32 %50 %62 %75 %82 %100 Table 4: The most important results in each scenario. Focusing on the results, the table shows that decreases with respect to the target scenario (cf. Table 2) when the source data is added, probably caused by the different sense distributions in B NC and the target corpora. The baseline classifiers (k- NN and SVM) are not able to improve over the baseline classifiers on the target data alone, which is coherent with past research, and shows that straightforward domain adaptation does not work. The following rows show that our reduction methods on themselves (OMT, SMA used by kNN and SVM ) also fail to perform better than in the target scenario, but the combinations using unlabeled data (k- NN - COMB and specially SVM COMB ) do manage to improve the best results for the target scenario, showing that we were able to attain domain adaptation. The feature augmentation approach (SVM - AUG) does improve slightly over SVM in the target scenario, but not over the best results in the target scenario, showing the difficulty of domain adaptation for WSD, at least on this dataset. MFS sports (%) Figure 1: Learning curves for S PORTS. The X axis denotes the amount of S PORTS data and the Y axis corresponds to accuracy. 90 88 accuracy (%) 86 84 SVM-COMB (+UNLAB, BNC + FIN. -> FIN.) SVM-AUG (BNC + FIN. -> FIN.) SVM-ORIG (FIN. -> FIN.) y=87.0 %25 %32 %50 %62 %75 %82 %100 finances (%) Figure 2: Learning curves for F INANCES. The X axis denotes the amount of F INANCES data and Y axis corresponds to the accuracy. p < 0.01. In addition, we carried extra experiments to examine the learning curves, and to check, given the source examples, how many additional examples from the target corpus are needed to obtain the same results as in the target scenario using all available examples. We fixed the source data and used increasing amounts of target data. We show the original SVM on the target scenario, and SVM - COMB (+ UNLAB ) and SVM - AUG as the domain adaptation approaches. The results are shown in figure 1 for S PORTS and figure 2 for F I NANCES. The horizontal line corresponds to the performance of SVM on the target domain. The point where the learning curves cross the horizontal line show that our domain adaptation method needs only around 40% of the target data in order to get the same performance as the baseline SVM on the target data. The learning curves also shows 7 Discussion and analysis Table 4 summarizes the most important results. The kernel combination method with unlabeled data on the adaptation scenario reduces the error on 22.1% and 17.6% over the baseline SVM on the target scenario (S PORTS and F INANCES respectively), and 12.7% and 9.0% over the k- NN combination method on the target scenario. These gains are remarkable given the already high baseline, specially taking into consideration that the 41 nouns are closely related to the domains. The differences, including SVM - AUG, are statistically significant according to the Wilcoxon test with 48 that the domain adaptation kernel combination approach, no matter the amount of target data, is always above the rest of the classifiers, showing the robustness of our approach. OntoNotes), to test whether the positive results are confirmed. We would also like to study word-byword behaviour, in order to assess whether target examples are really necessary for words which are less related to the domain. 8 Conclusion and future work Acknowledgments This work has been partially funded by the EU Commission (project KYOTO ICT-2007-211423) and Spanish Research Department (project KNOW TIN2006-15049-C03-01). Oier Lopez de Lacalle has a PhD grant from the Basque Government. In this paper we explore supervised domain adaptation for WSD with positive results, that is, whether hand-labeling general domain (source) text is worth the effort when training WSD systems that are to be applied to specific domains (targets). We performed several experiments in three scenarios. In the first scenario (source to target scenario), the classifiers were trained on source domain data (the B NC) and tested on the target domains, composed by the S PORTS and F INANCES sections of Reuters. In the second scenario (target scenario) we set the main baseline for our domain adaptation experiment, training and testing our classifiers on the target domain data. In the last scenario (domain adaptation scenario), we combine both source and target data for training, and test on the target data. We report results in each scenario for k- NN and SVM classifiers, for reduced features obtained using SVD over the training data, for the use of unlabeled data, and for k- NN and SVM combinations of all. Our results show that our best domain adaptation strategy (using kernel combination of SVD features and unlabeled data related to the training data) yields statistically significant improvements: up to 22% error reduction compared to SVM on the target domain data alone. We also show that our domain adaptation method only needs 40% of the target data (in addition to the source data) in order to get the same results as SVM on the target alone. We obtain coherent results in two target scenarios, and consistent improvement at all levels of the learning curves, showing the robustness or our findings. We think that our dataset, which comprises examples for 41 nouns that are closely related to the target domains, is specially demanding, as one would expect the performance of a generic WSD system to drop when moving to the domain corpus, specially on domain-related words, while we could expect the performance to be similar for generic or unrelated words. In the future we would like to evaluate our method on other datasets (e.g. DSO or References Eneko Agirre and Oier Lopez de Lacalle. 2007. Ubcalm: Combining k-nn with svd for wsd. In Proceedings of the Fourth International Workshop on Semantic Evaluations (SemEval-2007), pages 342­ 345, Prague, Czech Republic, June. Association for Computational Linguistics. Eneko Agirre and Oier Lopez de Lacalle. 2008. On robustness and domain adaptation using SVD for word sense disambiguation. In Proceedings of the 22nd International Conference on Computational Linguistics (Coling 2008), pages 17­24, Manchester, UK, August. Coling 2008 Organizing Committee. Eneko Agirre and David Martīnez. 2004. The effect i of bias on an automatically-built word sense corpus. Proceedings of the 4rd International Conference on Languages Resources and Evaluations (LREC). E. Agirre, O.Lopez de Lacalle, and David Martīnez. i 2005. Exploring feature spaces with svd and unlabeled data for Word Sense Disambiguation. In Proceedings of the Conference on Recent Advances on Natural Language Processing (RANLP'05), Borovets, Bulgaria. Rie Kubota Ando. 2006. Applying alternating structure optimization to word sense disambiguation. In Proceedings of the 10th Conference on Computational Natural Language Learning (CoNLL), pages 77­84, New York City. John Blitzer, Ryan McDonald, and Fernando Pereira. 2006. Domain adaptation with structural correspondence learning. In Proceedings of the 2006 Conference on Empirical Methods in Natural Language Processing, pages 120­128, Sydney, Australia, July. Association for Computational Linguistics. Yee Seng Chan and Hwee Tou Ng. 2007. Domain adaptation with active learning for word sense disambiguation. In Proceedings of the 45th Annual Meeting of the Association of Computational Linguistics, pages 49­56, Prague, Czech Republic, June. Association for Computational Linguistics. 49 Ciprian Chelba and Alex Acero. 2004. Adaptation of maximum entropy classifier: Little data can help a lot. In Proceedings of of th Conference on Empirical Methods in Natural Language Processing (EMNLP), Barcelona, Spain. Hal Daumī III and Daniel Marcu. 2006. Domain adape tation for statistical classifiers. Journal of Artificial Intelligence Research, 26:101­126. e Hal Daumī III. 2007. Frustratingly easy domain adaptation. In Proceedings of the 45th Annual Meeting of the Association of Computational Linguistics, pages 256­263, Prague, Czech Republic, June. Association for Computational Linguistics. Scott Deerwester, Susan Dumais, Goerge Furnas, Thomas Landauer, and Richard Harshman. 1990. Indexing by Latent Semantic Analysis. Journal of the American Society for Information Science, 41(6):391­407. a Gerard Escudero, Lluiz Mī rquez, and German Rigau. 2000. An Empirical Study of the Domain Dependence of Supervised Word Sense Didanbiguation Systems. Proceedings of the joint SIGDAT Conference on Empirical Methods in Natural Language Processing and Very Large Corpora, EMNLP/VLC. C. Fellbaum. 1998. WordNet: An Electronic Lexical Database. MIT Press. Alfio Massimiliano Gliozzo, Claudio Giuliano, and Carlo Strapparava. 2005. Domain Kernels for Word Sense Disambiguation. 43nd Annual Meeting of the Association for Computational Linguistics. (ACL05). R. Koeling, D. McCarthy, and J. Carroll. 2005. Domain-specific sense distributions and predominant sense acquisition. In Proceedings of the Human Language Technology Conference and Conference on Empirical Methods in Natural Language Processing. HLT/EMNLP, pages 419­426, Ann Arbor, Michigan. G. Leech. 1992. 100 million words of English: the British National Corpus. Language Research, 28(1):1­13. i David Martīnez and Eneko Agirre. 2000. One Sense per Collocation and Genre/Topic Variations. Conference on Empirical Method in Natural Language. T. Pedersen. 2001. A Decision Tree of Bigrams is an Accurate Predictor of Word Sense. In Proceedings of the Second Meeting of the North American Chapter of the Association for Computational Linguistics (NAACL-01), Pittsburgh, PA. Sameer Pradhan, Edward Loper, Dmitriy Dligach, and Martha Palmer. 2007. Semeval-2007 task-17: English lexical sample, srl and all words. In Proceedings of the Fourth International Workshop on Semantic Evaluations (SemEval-2007), pages 87­92, Prague, Czech Republic. Tony G. Rose, Mark Stevenson, and Miles Whitehead. 2002. The reuters corpus volumen 1 from yesterday's news to tomorrow's language resources. In Proceedings of the Third International Conference on Language Resources and Evaluation (LREC2002), pages 827­832, Las Palmas, Canary Islands. Sarah Zelikovitz and Haym Hirsh. 2001. Using LSI for text classification in the presence of background text. In Henrique Paques, Ling Liu, and David Grossman, editors, Proceedings of CIKM-01, 10th ACM International Conference on Information and Knowledge Management, pages 113­118, Atlanta, US. ACM Press, New York, US. Zhi Zhong, Hwee Tou Ng, and Yee Seng Chan. 2008. Word sense disambiguation using OntoNotes: An empirical study. In Proceedings of the 2008 Conference on Empirical Methods in Natural Language Processing, pages 1002­1010, Honolulu, Hawaii, October. Association for Computational Linguistics. 50