Machine-Learning-Based Transformation of Passive Japanese Sentences into Active by Separating Training Data into Each Input Particle Masaki Murata National Institute of Information and Communications Technology 3-5 Hikaridai, Seika-cho, Soraku-gun, Kyoto 619-0289, Japan murata@nict.go.jp Tamotsu Shirado National Institute of Information and Communications Technology 3-5 Hikaridai, Seika-cho, Soraku-gun, Kyoto 619-0289, Japan shirado@nict.go.jp Abstract We developed a new method of transforming Japanese case particles when transforming Japanese passive sentences into active sentences. It separates training data into each input particle and uses machine learning for each particle. We also used numerous rich features for learning. Our method obtained a high rate of accuracy (94.30%). In contrast, a method that did not separate training data for any input particles obtained a lower rate of accuracy (92.00%). In addition, a method that did not have many rich features for learning used in a previous study (Murata and Isahara, 2003) obtained a much lower accuracy rate (89.77%). We confirmed that these improvements were significant through a statistical test. We also conducted experiments utilizing traditional methods using verb dictionaries and manually prepared heuristic rules and confirmed that our method obtained much higher accuracy rates than traditional methods. Toshiyuki Kanamaru National Institute of Information and Communications Technology 3-5 Hikaridai, Seika-cho, Soraku-gun, Kyoto 619-0289, Japan kanamaru@nict.go.jp Hitoshi Isahara National Institute of Information and Communications Technology 3-5 Hikaridai, Seika-cho, Soraku-gun, Kyoto 619-0289, Japan isahara@nict.go.jp the meaning of "by", is changed into ga, which is a case postpositional particle indicating the subjective case, and (ii) ga (subject), which is a case postpositional particle indicating the subjective case, is changed into wo (object), which is a case postpositional particle indicating the objective case. In this paper, we discuss the transformation of Japanese case particles (i.e., ni ga) through machine learning.1 The transformation of passive sentences into active is useful in many research areas including generation, knowledge extraction from databases written in natural languages, information extraction, and answering questions. For example, when the answer is in the passive voice and the question is in the active voice, a question-answering system cannot match the answer with the question because the sentence structures are different and it is thus difficult to find the answer to the question. Methods of transforming passive sentences into active are important in natural language processing. The transformation of case particles in transforming passive sentences into active is not easy because particles depend on verbs and their use. We developed a new method of transforming Japanese case particles when transforming passive Japanese sentences into active in this study. Our method separates training data into each input particle and uses machine learning for each input particle. We also used numerous rich features for learning. Our experiments confirmed that our method was effective. In this study, we did not handle the transformation of auxiliary verbs and the inflection change of verbs because these can be transformed based on Japanese grammar. 1 1 Introduction This paper describes how passive Japanese sentences can be automatically transformed into active. There is an example of a passive Japanese sentence in Figure 1. The Japanese suffix reta functions as an auxiliary verb indicating the passive voice. There is a corresponding active-voice sentence in Figure 2. When the sentence in Figure 1 is transformed into an active sentence, (i) ni (by), which is a case postpositional particle with 587 Proceedings of the COLING/ACL 2006 Main Conference Poster Sessions, pages 587­594, Sydney, July 2006. c 2006 Association for Computational Linguistics inu ni watashi ga kama(dog) (by) (I) subjective-case postpositional particle (bite) (I was bitten by a dog.) Figure 1: Passive sentence inu ni watashi ga kamawo ga (dog) (by) (I) subjective-case postpositional particle (bite) (I was bitten by a dog.) Figure 3: Example in corpus inu ga (dog) subject (Dog bit me.) watashi (I) wo object kanda. (bite) reta. passive voice reta. passive voice Figure 2: Active sentence 2 Tagged corpus as supervised data We used the Kyoto University corpus (Kurohashi and Nagao, 1997) to construct a corpus tagged for the transformation of case particles. It has approximately 20,000 sentences (16 editions of the Mainichi Newspaper, from January 1st to 17th, 1995). We extracted case particles in passivevoice sentences from the Kyoto University corpus. There were 3,576 particles. We assigned a corresponding case particle for the active voice to each case particle. There is an example in Figure 3. The two underlined particles, "ga" and "wo" that are given for "ni" and "ga" are tags for case particles in the active voice. We called the given case particles for the active voice target case particles, and the original case particles in passivevoice sentences source case particles. We created tags for target case particles in the corpus. If we can determine the target case particles in a given sentence, we can transform the case particles in passive-voice sentences into case particles for the active voice. Therefore, our goal was to determine the target case particles. Small Margin Large Margin Figure 4: Maximizing margin Murata et al., 2002). Data consisting of two categories were classified by using a hyperplane to divide a space with the support vector machine. When these two categories were, positive and negative, for example, enlarging the margin between them in the training data (see Figure 42 ), reduced the possibility of incorrectly choosing categories in blind data (test data). A hyperplane that maximized the margin was thus determined, and classification was done using that hyperplane. Although the basics of this method are as described above, the region between the margins through the training data can include a small number of examples in extended versions, and the linearity of the hyperplane can be changed to non-linear by using kernel functions. Classification in these extended versions is equivalent to classification using the following discernment function, and the two categories can be classified on the basis of whether the value output by the function is positive or negative (Cristianini and Shawe-Taylor, 2000; Kudoh, 2000): 2 The open circles in the figure indicate positive examples and the black circles indicate negative. The solid line indicates the hyperplane dividing the space, and the broken lines indicate the planes depicting margins. 3 Machine learning method (support vector machine) We used a support vector machine as the basis of our machine-learning method. This is because support vector machines are comparatively better than other methods in many research areas (Kudoh and Matsumoto, 2000; Taira and Haruno, 2001; 588 f (x) = sgn b= i l ( Table 1: Features 1) F1 F2 F3 F4 F5 F6 F7 F8 part of speech (POS) of P main word of P word of P first 1, 2, 3, 4, 5, and 7 digits of category number of P5 auxiliary verb attached to P word of N first 1, 2, 3, 4, 5, and 7 digits of category number of N case particles and words of nominals that have dependency relationship with P and are other than N first 1, 2, 3, 4, 5, and 7 digits of category number of nominals that have dependency relationship with P and are other than N i yi K (xi , x) + b maxi,yi =-1 bi + mini,yi =1 bi 2 jl =1 =1 bi = - j yj K (xj , xi ), where x is the context (a set of features) of an input example, xi indicates the context of a training datum, and yi (i = 1, ..., l, yi {1, -1}) indicates its category. Function sgn is: sgn(x) = 1 (x 0), (2) -1 (other wise). Each i (i = 1, 2...) is fixed as a value of i that maximizes the value of L() in Eq. (3) under the conditions set by Eqs. (4) and (5). L() = F9 F10 case particles of nominals that have dependency relationship with P and are other than N F11 the words appearing in the same sentence F12 first 3 and 5 digits of category number of words appearing in same sentence F13 case particle taken by N (source case particle) F14 target case particle output by KNP (Kurohashi, 1998) F15 target case particle output with Kondo's method (Kondo et al., 2001) F16 case patterns defined in IPAL dictionary (IPAL) (IPA, 1987) F17 combination of predicate semantic primitives defined in IPAL F18 predicate semantic primitives defined in IPAL F19 combination of semantic primitives of N defined in IPAL F20 semantic primitives of N defined in IPAL F21 whether P is defined in IPAL or not F22 whether P can be in passive form defined in VDIC6 F23 case particles of P defined in VDIC F24 type of P defined in VDIC F25 transformation rule used for P and N in Kondo's method F26 whether P is defined in VDIC or not F27 pattern of case particles of nominals that have dependency relationship with P F28 pair of case particles of nominals that have dependency relationship with P F29 case particles of nominals that have dependency relationship with P and appear before N F30 case particles of nominals that have dependency relationship with P and appear after N F31 case particles of nominals that have dependency relationship with P and appear just before N F32 case particles of nominals that have dependency relationship with P and appear just after N il =1 i - 1i 2 l i j yi yj K (xi , xj ) (3) (4) ,j = 1 0 i C (i = 1, ..., l) i l i yi = 0 (5) =1 Although function K is called a kernel function and various functions are used as kernel functions, we have exclusively used the following polynomial function: K (x, y) = (x · y + 1)d (6) C and d are constants set by experimentation. For all experiments reported in this paper, C was fixed as 1 and d was fixed as 2. A set of xi that satisfies i > 0 is called a support vector, (S Vs )3 , and the summation portion of Eq. (1) is only calculated using examples that are support vectors. Equation 1 is expressed as follows by using support vectors. i :xi S Vs f (x) = sgn i yi K (xi , x) + b(7) bi:yi =-1,xi S Vs + bi:yi =1,xi S Vs b= 2 i j yj K (xj , xi ), bi = - :xi S Vs 3 The circles on the broken lines in Figure 4 indicate support vectors. 589 Table 2: Frequently occurring target case particles in source case particles Source case particle ni (indirect object) ga (subject) de (with) to (with) wo (direct object) kara (from) made (to) he (to) no (subject) Frequent target case particles in source case particles 27.57% (493/1788) ni (indirect object) ga (subject) 26.96% (482/1788) wo (direct object) 17.17% (307/1788) ga (subject) de (with) 16.11% (288/1788) to (with) 6.77% (121/1788) wo (direct object) 4.53% ( 81/1788) ga (subject) kara (from) 0.78% ( 14/1788) made (to) 0.06% ( 1/1788) ga (subject) 0.06% ( 1/1788) wo (direct object) Occurrence rate Occurrence rate in source case particles 70.79% (349/493) 27.38% (135/493) 96.47% (465/482) 79.15% (243/307) 13.36% (41/307) 99.31% (286/288) 99.17% (120/121) 49.38% ( 40/ 81) 44.44% ( 36/ 81) 100.00% ( 14/ 14) 100.00% ( 1/ 1) 100.00% ( 1/ 1) Support vector machines are capable of handling data consisting of two categories. Data consisting of more than two categories is generally handled using the pair-wise method (Kudoh and Matsumoto, 2000). Pairs of two different categories (N(N-1)/2 pairs) are constructed for data consisting of N categories with this method. The best category is determined by using a two-category classifier (in this paper, a support vector machine4 is used as the two-category classifier), and the correct category is finally determined on the basis of "voting" on the N(N-1)/2 pairs that result from analysis with the two-category classifier. The method discussed in this paper is in fact a combination of the support vector machine and the pair-wise method described above. case particle being analyzed, and P is the phrase's predicate. We used the Japanese syntactic parser, KNP (Kurohashi, 1998), for identifying N, P, parts of speech and syntactic relations. In the experiments conducted in this study, we selected features. We used the following procedure to select them. · Feature selection We first used all the features for learning. We next deleted only one feature from all the features for learning. We did this for every feature. We decided to delete features that would make the most improvement. We repeated this until we could not improve the rate of accuracy. 4 Features (information used in classification) The features we used in our study are listed in Table 1, where N is a noun phrase connected to the We used Kudoh's TinySVM software (Kudoh, 2000) as the support vector machine. 5 The category number indicates a semantic class of words. A Japanese thesaurus, the Bunrui Goi Hyou (NLRI, 1964), was used to determine the category number of each word. This thesaurus is `is-a' hierarchical, in which each word has a category number. This is a 10-digit number that indicates seven levels of `is-a' hierarchy. The top five levels are expressed by the first five digits, the sixth level is expressed by the next two digits, and the seventh level is expressed by the last three digits. 6 Kondo et al. constructed a rich dictionary for Japanese verbs (Kondo et al., 2001). It defined types and characteristics of verbs. We will refer to it as VDIC. 4 5 Method of separating training data into each input particle We developed a new method of separating training data into each input (source) particle that uses machine learning for each particle. For example, when we identify a target particle where the source particle is ni, we use only the training data where the source particle is ni. When we identify a target particle where the source particle is ga, we use only the training data where the source particle is ga. Frequently occurring target case particles are very different in source case particles. Frequently occurring target case particles in all source case particles are listed in Table 2. For example, when ni is a source case particle, frequently occurring 590 Table 3: Occurrence rates for target case particles Target case particle wo (direct object) ni (indirect object) to (with) de (with) ga (subject) ga or de kara (from) Other Occurrence rate Closed Open 33.05% 29.92% 19.69% 17.79% 16.00% 18.90% 13.65% 15.27% 11.07% 10.01% 2.40% 2.46% 2.13% 3.47% 2.01% 1.79% target case particles are ni or ga. In contrast, when ga is a source case particle, a frequently occurring target case particle is wo. In this case, it is better to separate training data into each source particle and use machine learning for each particle. We therefore developed this method and confirmed that it was effective through experiments (Section 6). 6 Experiments 6.1 Basic experiments We used the corpus we constructed described in Section 2 as supervised data. We divided the supervised data into closed and open data (Both the closed data and open data had 1788 items each.). The distribution of target case particles in the data are listed in Table 3. We used the closed data to determine features that were deleted in feature selection and used the open data as test data (data for evaluation). We used 10-fold cross validation for the experiments on closed data and we used closed data as the training data for the experiments on open data. The target case particles were determined by using the machine-learning method explained in Section 3. When multiple target particles could have been answers in the training data, we used pairs of them as answers for machine learning. The experimental results are listed in Tables 4 and 5. Baseline 1 outputs a source case particle as the target case particle. Baseline 2 outputs the most frequent target case particle (wo (direct object)) in the closed data as the target case particle in every case. Baseline 3 outputs the most frequent target case particle for each source target case particle in the closed data as the target case particle. For example, ni (indirect object) is the 591 most frequent target case particle when the source case particle is ni, as listed in Table 2. Baseline 3 outputs ni when the source case particle is ni. KNP indicates the results that the Japanese syntactic parser, KNP (Kurohashi, 1998), output. Kondo indicates the results that Kondo's method, (Kondo et al., 2001), output. KNP and Kondo can only work when a target predicate is defined in the IPAL dictionary or the VDIC dictionary. Otherwise, KNP and Kondo output nothing. "KNP/Kondo + Baseline X" indicates the use of outputs by Baseline X when KNP/Kondo have output nothing. KNP and Kondo are traditional methods using verb dictionaries and manually prepared heuristic rules. These traditional methods were used in this study to compare them with ours. "Murata 2003" indicates results using a method they developed in a previous study (Murata and Isahara, 2003). This method uses F1, F2, F5, F6, F7, F10, and F13 as features and does not have training data for any source case particles. "Division" indicates separating training data into each source particle. "Nodivision" indicates not separating training data for any source particles. "All features" indicates the use of all features with no features being selected. "Feature selection" indicates features are selected. We did two kinds of evaluations: "Eval. A" and "Eval. B". There are some cases where multiple target case particles can be answers. For example, ga and de can be answers. We judged the result to be correct in "Eval. A" when ga and de could be answers and the system output the pair of ga and de as answers. We judged the result to be correct in "Eval. B" when ga and de could be answers and the system output ga, de, or the pair of ga and de as answers. Table 4 lists the results using all data. Table 5 lists the results where a target predicate is defined in the IPAL and VDIC dictionaries. There were 551 items in the closed data and 539 in the open. We found the following from the results. Although selection of features obtained higher rates of accuracy than use of all features in the closed data, it did not obtain higher rates of accuracy in the open data. This indicates that feature selection was not effective and we should have used all features in this study. Our method using all features in the open data and separating training data into each source particle obtained the highest rate of accuracy (94.30% in Eval. B). This indicates that our method is ef- Table 4: Experimental results Closed data Eval. A Eval. B Baseline 1 58.67% 61.41% Baseline 2 33.05% 33.56% Baseline 3 84.17% 88.20% KNP 27.35% 28.69% KNP + Baseline 1 64.32% 67.06% KNP + Baseline 2 48.10% 48.99% KNP + Baseline 3 81.21% 84.84% Kondo 39.21% 40.88% Kondo + Baseline 1 65.27% 68.57% Kondo + Baseline 2 54.87% 56.54% Kondo + Baseline 3 78.08% 81.71% Murata 2003 86.86% 89.09% Our method, no-division + all features 89.99% 92.39% Our method, no-division + feature selection 91.28% 93.40% Our method, division + all features 91.22% 93.79% Our method, division + feature selection 92.06% 94.41% Method Open data Eval. A Eval. B 62.02% 64.60% 29.92% 30.37% 84.17% 88.20% 27.91% 29.14% 67.79% 70.36% 45.97% 46.48% 80.82% 84.45% 39.32% 41.00% 67.34% 70.41% 53.52% 55.26% 78.30% 81.88% 87.86% 89.77% 90.04% 92.00% 90.10% 92.00% 92.28% 94.30% 91.89% 93.85% Table 5: Experimental results on data that can use IPAL and VDIC dictionaries Closed data Open data Eval. A Eval. B Eval. A Eval. B Baseline 1 57.71% 58.98% 58.63% 58.81% Baseline 2 37.39% 37.39% 37.29% 37.29% Baseline 3 84.03% 86.57% 86.83% 88.31% KNP 74.59% 75.86% 75.88% 76.07% Kondo 76.04% 77.50% 78.66% 78.85% Our method, no-division + all features 94.19% 95.46% 94.81% 94.81% Our method, division + all features 95.83% 96.91% 97.03% 97.03% Method fective. Our method that used all the features and did not separate training data for any source particles obtained an accuracy rate of 92.00% in Eval. B. The technique of separating training data into each source particles made an improvement of 2.30%. We confirmed that this improvement has a significance level of 0.01 by using a two-sided binomial test (two-sided sign test). This indicates that the technique of separating training data for all source particles is effective. Murata 2003 who used only seven features and did not separate training data for any source particles obtained an accuracy rate of 89.77% with Eval. B. The method (92.00%) of using all features (32) made an improvement of 2.23% against theirs. We confirmed that this improvement had 592 a significance level of 0.01 by using a two-sided binomial test (two-sided sign test). This indicates that our increased features are effective. KNP and Kondo obtained low accuracy rates (29.14% and 41.00% in Eval. B for the open data). We did the evaluation using data and proved that these methods could work well. A target predicate in the data is defined in the IPAL and VDIC dictionaries. The results are listed in Table 5. KNP and Kondo obtained relatively higher accuracy rates (76.07% and 78.85% in Eval. B for the open data). However, they were lower than that for Baseline 3. Baseline 3 obtained a relatively high accuracy rate (84.17% and 88.20% in Eval. B for the open data). Baseline 3 is similar to our method in terms of separating the training data into source particles. Baseline 3 separates the training data into Deleted features Not deleted F1 F2 F3 F4 F5 F6 F7 F8 F9 F10 F11 F12 F13 F14 F15 F16 F17 F18 F19 F20 F21 F22 F23 F24 F25 F26 F27 F28 F29 F30 F31 F32 Table 6: Deletion of features Closed data Open data Eval. A Eval. B Eval. A Eval. B Acc. Diff. Acc. Diff. Acc. Diff. Acc. Diff. 91.22% -- 93.79% -- 92.28% -- 94.30% -- 91.16% -0.06% 93.74% -0.05% 92.23% -0.05% 94.24% -0.06% 91.11% -0.11% 93.68% -0.11% 92.23% -0.05% 94.18% -0.12% 91.11% -0.11% 93.68% -0.11% 92.23% -0.05% 94.18% -0.12% 91.50% 0.28% 94.13% 0.34% 91.72% -0.56% 93.68% -0.62% 91.22% 0.00% 93.62% -0.17% 91.95% -0.33% 93.96% -0.34% 91.00% -0.22% 93.51% -0.28% 92.23% -0.05% 94.24% -0.06% 90.66% -0.56% 93.18% -0.61% 91.78% -0.50% 93.90% -0.40% 91.22% 0.00% 93.79% 0.00% 92.39% 0.11% 94.24% -0.06% 91.28% 0.06% 93.62% -0.17% 92.45% 0.17% 94.07% -0.23% 91.33% 0.11% 93.85% 0.06% 92.00% -0.28% 94.07% -0.23% 91.50% 0.28% 93.74% -0.05% 92.06% -0.22% 93.79% -0.51% 91.28% 0.06% 93.62% -0.17% 92.56% 0.28% 94.35% 0.05% 91.22% 0.00% 93.79% 0.00% 92.28% 0.00% 94.30% 0.00% 91.16% -0.06% 93.74% -0.05% 92.39% 0.11% 94.41% 0.11% 91.22% 0.00% 93.79% 0.00% 92.23% -0.05% 94.24% -0.06% 91.39% 0.17% 93.90% 0.11% 92.34% 0.06% 94.30% 0.00% 91.22% 0.00% 93.79% 0.00% 92.23% -0.05% 94.24% -0.06% 91.16% -0.06% 93.74% -0.05% 92.39% 0.11% 94.46% 0.16% 91.33% 0.11% 93.90% 0.11% 92.28% 0.00% 94.30% 0.00% 91.11% -0.11% 93.68% -0.11% 92.34% 0.06% 94.35% 0.05% 91.22% 0.00% 93.79% 0.00% 92.28% 0.00% 94.30% 0.00% 91.16% -0.06% 93.74% -0.05% 92.23% -0.05% 94.24% -0.06% 91.28% 0.06% 93.79% 0.00% 92.28% 0.00% 94.24% -0.06% 91.22% 0.00% 93.74% -0.05% 92.23% -0.05% 94.24% -0.06% 89.54% -1.68% 92.11% -1.68% 90.04% -2.24% 92.39% -1.91% 91.16% -0.06% 93.74% -0.05% 92.28% 0.00% 94.30% 0.00% 91.22% 0.00% 93.68% -0.11% 92.23% -0.05% 94.18% -0.12% 90.94% -0.28% 93.51% -0.28% 92.11% -0.17% 94.13% -0.17% 91.28% 0.06% 93.85% 0.06% 92.28% 0.00% 94.30% 0.00% 91.16% -0.06% 93.74% -0.05% 92.23% -0.05% 94.24% -0.06% 91.28% 0.06% 93.85% 0.06% 92.28% 0.00% 94.24% -0.06% 91.22% 0.00% 93.79% 0.00% 92.28% 0.00% 94.30% 0.00% source particles and uses the most frequent target case particle. Our method involves separating the training data into source particles and using machine learning for each particle. The fact that Baseline 3 obtained a relatively high accuracy rate supports the effectiveness of our method separating the training data into source particles. 6.2 Experiments confirming importance of features This indicates that F25 is particularly effective. F25 is the transformation rule Kondo used for P and N in his method. The transformation rules in Kondo's method were made precisely for ni (indirect object), which is particularly difficult to handle. F25 is thus effective. We could also see not using F7 decreased the accuracy rate (about 0.5%). F7 has the semantic features for N. We found that the semantic features for N were also effective. 6.3 Experiments changing number of training data We next conducted experiments where we confirmed which features were effective. The results are listed in Table 6. We can see the accuracy rate for deleting features and the accuracy rate for using all features. We can see that not using F25 greatly decreased the accuracy rate (about 2%). 593 We finally did experiments changing the number of training data. The results are plotted in Figure 5. We used our two methods of all features "Division" and "Non-division". We only plotted the Figure 5: Changing number of training data accuracy rates for Eval. B in the open data in the figure. We plotted accuracy rates when 1, 1/2, 1/4, 1/8, and 1/16 of the training data were used. "Division", which separates training data for all source particles, obtained a high accuracy rate (88.36%) even when the number of training data was small. In contrast, "Non-division", which does not separate training data for any source particles, obtained a low accuracy rate (75.57%), when the number of training data was small. This indicates that our method of separating training data for all source particles is effective. found that semantic features for nominal targets were effective. We finally did experiments on changing the number of training data. We found that our method of separating training data for all source particles could obtain high accuracy rates even when there were few training data. This indicates that our method of separating training data for all source particles is effective. The transformation of passive sentences into active sentences is useful in many research areas including generation, knowledge extraction from databases written in natural languages, information extraction, and answering questions. In the future, we intend to use the results of our study for these kinds of research projects. References Nello Cristianini and John Shawe-Taylor. 2000. An Introduction to Support Vector Machines and Other Kernel-based Learning Methods. Cambridge University Press. IPA. 1987. (Information­Technology Promotion Agency, Japan). IPA Lexicon of the Japanese Language for Computers IPAL (Basic Verbs). (in Japanese). Keiko Kondo, Satoshi Sato, and Manabu Okumura. 2001. Paraphrasing by case alternation. Transactions of Information Processing Society of Japan, 42(3):465­477. (in Japanese). Taku Kudoh and Yuji Matsumoto. 2000. Use of support vector learning for chunk identification. CoNLL-2000, pages 142­144. Taku Kudoh. 2000. TinySVM: Support Vector Machines. http://cl.aist-nara.ac.jp/~aku-ku//software/TinySVM/ t index.html. Sadao Kurohashi and Makoto Nagao. 1997. Kyoto University text corpus project. 3rd Annual Meeting of the Association for Natural Language Processing, pages 115­118. (in Japanese). Sadao Kurohashi, 1998. Japanese Dependency/Case Structure Analyzer KNP version 2.0b6. Department of Informatics, Kyoto University. (in Japanese). Masaki Murata and Hitoshi Isahara, 2003. Conversion of Japanese Passive/Causative Sentences into Active Sentences Using Machine Learning, pages 115­125. Springer Publisher. Masaki Murata, Qing Ma, and Hitoshi Isahara. 2002. Comparison of three machine-learning methods for Thai partof-speech tagging. ACM Transactions on Asian Language Information Processing, 1(2):145­158. NLRI. 1964. Bunrui Goi Hyou. Shuuei Publishing. Hirotoshi Taira and Masahiko Haruno. 2001. Feature selection in svm text categorization. In Proceedings of AAAI2001, pages 480­486. 7 Conclusion We developed a new method of transforming Japanese case particles when transforming Japanese passive sentences into active sentences. Our method separates training data for all input (source) particles and uses machine learning for each particle. We also used numerous rich features for learning. Our method obtained a high rate of accuracy (94.30%). In contrast, a method that did not separate training data for all source particles obtained a lower rate of accuracy (92.00%). In addition, a method that did not have many rich features for learning used in a previous study obtained a much lower accuracy rate (89.77%). We confirmed that these improvements were significant through a statistical test. We also undertook experiments utilizing traditional methods using verb dictionaries and manually prepared heuristic rules and confirmed that our method obtained much higher accuracy rates than traditional methods. We also conducted experiments on which features were the most effective. We found that Kondo's transformation rule used as a feature in our system was particularly effective. We also 594