ABSTRACT
Spoken monologues feature greater sentence length and structural complexity than do spoken dialogues. To achieve high parsing performance for spoken monologues, it could prove effective to simplify the structure by dividing a sentence into suitable language units. This paper proposes a method for dependency parsing of Japanese monologues based on sentence segmentation. In this method, the dependency parsing is executed in two stages: at the clause level and the sentence level. First, the dependencies within a clause are identified by dividing a sentence into clauses and executing stochastic dependency parsing for each clause. Next, the dependencies over clause boundaries are identified stochastically, and the dependency structure of the entire sentence is thus completed. An experiment using a spoken monologue corpus shows this method to be effective for efficient dependency parsing of Japanese monologue sentences.
- R. Agarwal and L. Boggess. 1992. A simple but useful approach to conjunct indentification. In Proc. of 30th ACL, pages 15--21. Google ScholarDigital Library
- E. Charniak. 2000. A maximum-entropy-inspired parser. In Proc. of 1st NAACL, pages 132--139. Google ScholarDigital Library
- M. Collins. 1996. A new statistical parser based on bigram lexical dependencies. In Proc. of 34th ACL, pages 184--191. Google ScholarDigital Library
- Mark G. Core and Lenhart K. Schubert. 1999. A syntactic framework for speech repairs and other disruptions. In Proc. of 37th ACL, pages 413--420. Google ScholarDigital Library
- R. Delmonte. 2003. Parsing spontaneous speech. In Proc. of 8th EUROSPEECH, pages 1999--2004.Google Scholar
- M. Fujio and Y. Matsumoto. 1998. Japanese dependency structure analysis based on lexicalized statistics. In Proc. of 3rd EMNLP, pages 87--96.Google Scholar
- J. Huang and G. Zweig. 2002. Maximum entropy model for punctuation annotation from speech. In Proc. of 7th ICSLP, pages 917--920.Google Scholar
- H. Kashioka and T. Maruyama. 2004. Segmentation of semantic unit in Japanese monologue. In Proc. of ICSLT-O-COCOSDA 2004, pages 87--92.Google Scholar
- M. Kim and J. Lee. 2004. Syntactic analysis of long sentences based on s-clauses. In Proc. of 1st IJC-NLP, pages 420--427.Google Scholar
- T. Kudo and Y. Matsumoto. 2002. Japanese dependency analyisis using cascaded chunking. In Proc. of 6th CoNLL, pages 63--69. Google ScholarDigital Library
- S. Kurohashi and M. Nagao. 1994. A syntactic analysis method of long Japanese sentences based on the detection of conjunctive structures. Computational Linguistics, 20(4):507--534. Google ScholarDigital Library
- S. Kurohashi and M. Nagao. 1997. Building a Japanese parsed corpus while improving the parsing system. In Proc. of 4th NLPRS, pages 451--456.Google Scholar
- K. Maekawa, H. Koiso, S. Furui, and H. Isahara. 2000. Spontaneous speech corpus of Japanese. In Proc. of 2nd LREC, pages 947--952.Google Scholar
- T. Maruyama, H. Kashioka, T. Kumano, and H. Tanaka. 2004. Development and evaluation of Japanese clause boundaries annotation program. Journal of Natural Language Processing, 11(3):39--68. (In Japanese).Google ScholarCross Ref
- Y. Matsumoto, A. Kitauchi, T. Yamashita, and Y. Hirano, 1999. Japanese Morphological Analysis System ChaSen version 2.0 Manual. NAIST Technical Report, NAIST-IS-TR99009.Google Scholar
- T. Ohno, S. Matsubara, H. Kashioka, N. Kato, and Y. Inagaki. 2005a. Incremental dependency parsing of Japanese spoken monologue based on clause boundaries. In Proc. of 9th EUROSPEECH, pages 3449--3452.Google Scholar
- T. Ohno, S. Matsubara, N. Kawaguchi, and Y. Inagaki. 2005b. Robust dependency parsing of spontaneous Japanese spoken language. IEICE Transactions on Information and Systems, E88-D(3):545--552. Google ScholarDigital Library
- A. Ratnaparkhi. 1997. A liner observed time statistical parser based on maximum entropy models. In Proc. of 2nd EMNLP, pages 1--10.Google Scholar
- S. Shirai, S. Ikehara, A. Yokoo, and J. Kimura. 1995. A new dependency analysis method based on semantically embedded sentence structures and its performance on Japanese subordinate clause. Journal of Information Processing Society of Japan, 36(10):2353--2361. (In Japanese).Google Scholar
- K. Shitaoka, K. Uchimoto, T. Kawahara, and H. Isahara. 2004. Dependency structure analysis and sentence boundary detection in spontaneous Japanese. In Proc. of 20th COLING, pages 1107--1113. Google ScholarDigital Library
- K. Uchimoto, S. Sekine, and K. Isahara. 1999. Japanese dependency structure analysis based on maximum entropy models. In Proc. of 9th EACL, pages 196--203. Google ScholarDigital Library
- T. Utsuro, S. Nishiokayama, M. Fujio, and Y. Matsumoto. 2000. Analyzing dependencies of Japanese subordinate clauses based on statistics of scope embedding preference. In Proc. of 6th ANLP, pages 110--117. Google ScholarDigital Library
- Dependency parsing of Japanese spoken monologue based on clause boundaries
Recommendations
Robust Dependency Parsing of Spontaneous Japanese Spoken Language
Spontaneously spoken Japanese includes a lot of grammatically ill-formed linguistic phenomena such as fillers, hesitations, inversions, and so on, which do not appear in written language. This paper proposes a novel method of robust dependency parsing ...
Simultaneous English-Japanese spoken language translation based on incremental dependency parsing and transfer
COLING-ACL '06: Proceedings of the COLING/ACL on Main conference poster sessionsThis paper proposes a method for incrementally translating English spoken language into Japanese. To realize simultaneous translation between languages with different word order, such as English and Japanese, our method utilizes the feature that the ...
Improve Chinese Semantic Dependency Parsing via Syntactic Dependency Parsing
IALP '12: Proceedings of the 2012 International Conference on Asian Language ProcessingWe address the problem of Chinese semantic dependency parsing. Dependency parsing is traditionally oriented to syntax analysis, which we denote by syntactic dependency parsing to distinguish it from semantic dependency parsing. In this paper, firstly we ...
Comments