用户名: 密码: 验证码:
基于统计学习的词义识别方法研究
详细信息    本馆镜像全文|  推荐本文 |  |   获取CNKI官网全文
摘要
如何解决语言的歧义问题一直困扰着自然语言处理技术的研究人员。语言的歧义最重要的一个表现就是一词多义现象。如何根据歧义词出现的上下文语言环境识别出正确词义是词义识别要解决问题。在自然语言理解领域,词义识别是应用基础研究课题,也是自然语言理解的重点和难点之一。
     早期的词义识别研究以基于规则的方法为主,近些年来随着计算技术和存储技术的改进和提高,统计学习方法越来越受到广泛的欢迎,迅速成为了主流的研究方法。有监督的学习方法应用于词义识别可以获得较高的识别精度,但是该类方法需要有规模足够大的训练样本,这样的样本不是容易获得。无监督的方法不需要人工标注训练样本,但是词义识别相对效果不是十分理想。
     本文分析了统计词义识别研究需要解决的几个关键性问题。从词典资源和语料库资源的建设到词义识别问题的建模方法,以及词义分类问题的特征选择,对所涉及的问题都一一进行了讨论。在这些基础上,本文最后给出了词义类扩展思想,并探讨了如何将其应用于统计词义识别的研究上。
     本文研究成果和创新如下:
     1.从词义的可计算性角度探讨词义刻画与词义识别的关系,探讨如何通过科学控制词义刻画粒度来重新整合现有的词典资源,建设新的机读词典,更好地为词义识别的应用服务。实验分析说明词义刻画粒度的大小直接影响了词义识别的精度,适当控制词义粒度再不产生二义性的前提下可以提高词义识别的精度。本文提出在词义再刻画的基础上整合现有词典资源,建设面向词义计算的新的分类词典;
     2.在探讨面向词义识别的特征选择方法的基础上提出以信息增益改进贝叶斯模型的词义识别新方法。实验中以朴素贝叶斯模型、最大熵方法和支持向量机建立的词义分类器作为参照模型,讨论信息增益改进贝叶斯模型的效果。实验结果显示参照系中使用最大熵和支持向量机构建的词义分类器都比朴素贝叶斯模型强,其中基于支持向量机的词义分类器最好,而经过信息增益改进的贝叶斯模型在词义识别上表现更突出,实验结果比SVM还要高出1.4个百分点,获得了对比实验中最优的识别结果;
     3.从语料库资源难于大规模建设的角度出发,实验分析和探讨了人造歧义词技术的使用问题,并在此基础上提出替换词的概念,以及基于替换词技术的词义识别新方法。实验结果表明,人造歧义词技术可以帮助研究者缓解训练语料短缺的压力,源于此的替换词技术可以让研究者避开人工标注训练样本,实现一种无监督的词义识别方法。实验结果表明基于替换词技术的词义识别方法具有较高的识别精度;
     4.针对词义识别训练语料规模不足够大的问题,提出了词义类扩展思想和基于词义类扩展的词义识别新方法。该方法通过词义类扩展,可以在有限训练语料中获得更多的词义信息,提高训练的效率,改善词义识别的效果,此外词义类扩展技术可以从无词义标注(无词义标记等先验知识)的生语料中统计相关词语信息,以此来为小规模的训练样本提供补充。实验结果表明基于词义类扩展思想的方法提高了训练语料的使用效率,改善了有监督词义识别的效果,这为增强小规模训练样本的统计学习效果提供一个崭新的思路。
     综上所述,本文在资源建设、词义识别的建模、特征选择,以及如何突破训练语料规模的限制实现无监督词义识别和改善有监督的词义识别方法上都作了一些有益的尝试,取得了一些初步成果。随着词义识别研究的不断深入,会涌现出更多更好的新的解决思路和方案。
How to solve the language ambiguity problem is the issue always plaguing the language processing technology researchers. Polysemous situation is thought to be the most important performance of language ambiguity. The problem solved by word sense discrimination (WSD) is how to identify the right meaning of the word by the context of language environment of the ambiguous words. In the field of natural language understanding, word recognition is the application of basic research issue, and also, one of the most important and difficult things of natural language understanding.
     Traditional word recognition is mostly studied by rule-based method, but in recent years, with the improvement and enhancement of computing and storage technology, statistical learning method is becoming more and more popular, and quickly became the mainstream research methods of word recognition. We can get a high identification accuracy when we apply the study method with supervision to word recognition, but such methods require a large enough scale of training sample which is not easy to gain. Although unsupervised method does not need training samples with manual tag, the relative effect of word sense discrimination is not very good.
     Some key issues which should be solved by the statistical word sense discrimination were analyzed in this paper. The questions involved were discussed one by one, not only the construction of dictionary and corpus resources, the modeling method of the problem of word recognition, but also the feature selection of the semantic classification. Basing on these issues, we got the thought of word-sense category extending at the end of this paper, and also, we discussed how to apply it in the research of the statistical word sense discrimination.
     This research and innovation are as follows:
     1. The relationship between word sense discrimination and word sense characterization were discussed from the perspective of semantic computability, and it was also studied how to make re-integration of existing dictionary resources and to construct a new machine readable dictionary by the scientific control of the semantic granularity, providing better service for word sense discrimination. Experimental results showed that the size of semantic granularity in the semantic characterization directly affected the accuracy of word sense discrimination. The accuracy of WSD could be increased by proper control the semantic granularity without ambiguity. It was proposed to integrate the existing dictionary resources, to build a new category dictionary for word sense discrimination;
     2. The improved Bayesian model by information gain for the feature selection of word sense discrimination was proposed. The word sense classifiers established by Naive Bayes model, Maximum Entropy method and Support Vector Machine were used as reference models in experiments which the effectiveness of Bayesian model improved by information gain. The results showed that the classifiers constructed by Maximum Entropy and Support Vector Machine are stronger than the Naive Bayes model, in which Support Vector Machine is the best in several of reference models. But the Bayesian model improved by information gain in word sense discrimination is more prominent comparing with the reference models, and its experimental results were much higher 1.4 percentage points than the SVM classifier, the Bayesian model improved by information gain obtained the best results in comparative experiments;
     3. The use of artificial ambiguous word technology was analyzed and discussed by experiments form the perspective of the difficulty in the construction of large-scale corpus resources, and the concept of vicarious words and the new method of word sense discrimination based on vicarious words were proposed. The results showed artificial ambiguous words technology could help researchers to relieve the pressure of the shortage of training data. The vicarious words technology from artificial ambiguous word allows researchers to achieve an unsupervised method for WSD avoiding the use of the training samples with manual tagging. Experimental results showed the WSD method based on vicarious words had high discrimination accuracy;
     4. The ideas of Word-sense Class Extending (WSE) and a new word sense discrimination method based on WSE were proposed for the problem which the size of training corpus was not sufficiently large. This new method could obtain more semantic information in limited training corpus to enhance training efficiency and improve the effect of word sense discrimination by WSE, in addition, the WSE technology could statistics related word s information in raw corpus (no prior knowledge of semantic tags, etc.) in order to provide additional training samples. Experimental results showed that the word sense discrimination method based on WSE improved the efficiency of the training corpus, made better the effectiveness of WSD. The WSE technology provided a new idea to enhance the effectiveness of statistical learning in small-scale training corpus.
     To sum up, this article had give some useful attempts in resource-building, word sense discrimination modeling, feature Selection, as well as on how to improve word sense discrimination with supervision, and we had achieved some initial results. With the further research of word sense discrimination, more and more new ideas and solutions will be emerged.
引文
[1]冯志伟.当前自然语言处理发展的四个特点.www.lingviko.net/feng/4feature.pdf.2009
    [2]王惠,苏新春.XHK:基于语法知识的汉语词义描述.www.courses.nus.edu.sg/course/chswh/wh_website/papers/XHK%20to%20CLSW6.doc.2010
    [3]卢志茂.无指导的汉语词义消歧方法研究.哈尔滨工业大学博士学位论文.2006.
    [4]杨宇娜.基于统计的中文词义消歧技术研究.哈尔滨工业大学硕士学位论文.2006.
    [5]孙广庆.基于语义范畴扩展的词义消歧的研究.哈尔滨工业大学硕士学位论文.2008.
    [6]Ekaterina Shutova. Models of Metaphor in NLP. In Proceedings of the 48th Annual Meeting of the Association for Computational Linguistics,11-16 July 2010, Uppsala, Sweden.2010:688-697.
    [7]Susan Windisch Brown. Choosing Sense Distinctions for WSD:Psycholinguistic Evidence. In Proceedings of ACL-08:HLT, Short Papers (Companion Volume), Columbus, Ohio, USA, June 2008.2008:249-252.
    [8]Zhi Zhong and Hwee Tou Ng. It Makes Sense:A Wide-Coverage Word Sense Disambiguation System for Free Text.. In Proceedings of the 48th Annual Meeting of the Association for Computational Linguistics,11-16 July 2010, Uppsala, Sweden.2010:78-83.
    [9]Yee Seng Chan and Hwee Tou Ng. Domain Adaptation with Active Learning forWord Sense Disambiguation. In Proceedings of the 45th Annual Meeting of the Association of Computational Linguistics, Prague, Czech Republic, June 2007.2007:49-56.
    [10]Jingbo Zhu and Eduard Hovy. Active Learning for Word Sense Disambiguation with Methods for Addressing the Class Imbalance Problem. In Proceedings of the 2007 Joint Conference on Empirical Methods in Natural Language Processing and Computational Natural Language Learning, Prague, June 2007.2007:783-790.
    [11]Linlin Li, Benjamin Roth, and Caroline Sporleder. Topic Models for Word Sense Disambiguation and Token-based Idiom Detection. In Proceedings of the 48th Annual Meeting of the Association for Computational Linguistics,11-16 July 2010, Uppsala, Sweden.2010:1138-1147.
    [12]Siva Reddy, Abhilash Inumella. WSD as a Distributed Constraint Optimization Problem. In Proceedings of the ACL 2010 Student Research Workshop,13 July 2010, Uppsala, Sweden.2010: 13-18.
    [13]Tim Van de Cruys. Using Three Way Data for Word Sense Discrimination. In Proceedings of the 22nd International Conference on Computational Linguistics (Coling 2008), Manchester, August 2008. 2008:929-936.
    [14]俞十汶,曲维光,王治敏.机器学习与自然语言处理.应用语言学.http://www.docin.com/p-222169.html.2009
    [15]Lucia Specia, Mark Stevenson and Maria das Gracas V. Nunes. Learning Expressive Models for Word Sense Disambiguation. Proceedings of the 45th Annual Meeting of the Association of Computational Linguistics, Prague, Czech Republic, June 2007.2007:41-48.
    [16]A. Kilgarriff and J. Rosenzweig. Framework and Results for English Senseval. Computers and the Humanities. April,2000,34(1-2):15-48.
    [17]Philip Edmonds and Scott Cotton. SENSEVAL-2:Overview. In Proceedings of the Second International Workshop on Evaluating Word Sense Disambiguation Systems (SENSEVAL-2), 2001:1-5.
    [18]Christian Scheible. Sentiment Translation through Lexicon Induction. In Proceedings of the ACL 2010 Student Research Workshop,13 July 2010, Uppsala, Sweden.2010:25-30.
    [19]Scott Cotton, Phil Edmonds, Adam Kilgarriff, et al. SENSEVAL-2:Second International Workshop on Evaluating Word Sense Disambiguation Systems. ACL SIGLEX, Toulouse, France.2001. http://www.sle.sharp.co.uk/senseval2.2008.
    [20]刘小虎.英汉机器翻译中词义消歧方法研究.哈尔滨工业大学博士学位论文.1998.
    [21]全吕勤,何婷婷,姬东鸿等.基于多分类器决策的词义消歧方法.计算机研究与发展.2006,43(5):933-939
    [22]Mitesh M. Khapra, Anup Kulkarni, Saurabh Sohoney. All Words Domain Adapted WSD:Finding a Middle Ground between Supervision and Unsupervision. In Proceedings of the 48th Annual Meeting of the Association for Computational Linguistics,11-16 July 2010, Uppsala, Sweden.2010: 1532-1541.
    [23]Katrin Erk, Sebastian Pad'o. Exemplar-Based Models for Word Meaning In Context. In Proceedings of the ACL 2010 Conference Short Papers,11-16 July 2010, Uppsala, Sweden.2010:92-97.
    [24]Boxing Chen, George Foster and Roland Kuhn. Bilingual Sense Similarity for Statistical Machine Translation. In Proceedings of the 48th Annual Meeting of the Association for Computational Linguistics,11-16 July 2010, Uppsala, Sweden.2010:834-843.
    [25]刘宇鹏,李生,赵铁军.基于WordNet词义消歧的系统融合.自动化学报.2010,36(11):1575-1580
    [26]H. Schutze. Automatic word sense discrimination. Computational Linguistics,1998,24(1):97-123.
    [27]梅家驹,竺一鸣等.同义词词林.上海辞书出版社,1996.
    [28]董振东. HowNet. http://www.keenage.com.2009.
    [29]Diana McCarthy& Rob Koeling. Finding Predominant Word Senses in Untagged Text. In Proceedings of the 42th Annual Meeting of the Association for Computational Linguistics,2004: 279-286.
    [30]张俐,李晶皎等.中文WordNet的研究及实现.东北大学学报(自然科学版).2003,24(4):327-329.
    [31]刘扬,于江生等.CCD构造模型及VACOL辅助软件的设计与实现.语言文字应用.2003,1:83-88.
    [32]Manabu Sassano, Sadao Kurohashi. Using Smaller Constituents Rather Than Sentences in Active Learning for Japanese Dependency Parsing. In Proceedings of the 48th Annual Meeting of the Association for Computational Linguistics,11-16 July 2010, Uppsala, Sweden.2010:356-365.
    [33]Brown Corpus. http://khnt.hit.hib.no/icame/manuals/brown/index.htm.2008.
    [34]王瑞琴,孔繁胜.无监督词义消歧研究.软件学报2009,20(8):2138.2152
    [35]Penn Treebank. http://www.cis.upenn.edu/-treebank.2009.
    [36]Peter F. Brown, Stephen A. Della Pietra, et al. Word-Sense Disambiguation Using Statistical Methods. In Proceedings of the 29th Annual Meeting of the Association for Computational Linguistics,1991: 264-270.
    [37]徐志明,王晓龙等N-gram语言模型的数据平滑技术.计算机应用研究.1999,7:37-44.
    [38]D. Yarowsky. Word-Sense Disambiguation Using Statistical Models of Roget's Categories Trained on Large Corpora. In COLING 14,1992,545-460
    [39]W. Gale, K. Church, and D. Yarowsky. A Method for Disambiguating Word Senses in a Large Corpus. Computers and the Humanities.1992,26:415-439.
    [40]D. Yarowsky. Hierarchical Decision Lists for Word Sense Disambiguation. Computer and the Humanities,2000,34:179-186.
    [41]T. Pederson. A Simple Approach to Building Ensembles of Naive Bayesian Classifiers for Word Sense Disambiguation. In Proceedings of the First Annual Meeting of the North American Chapter of the Association for Computational Linguistics, Seattle, WA, May,2000:63-69.
    [42]Ted Pedersen. Evaluating the Effectiveness of Ensembles of Decision Trees in Disambi guating Senseval Lexical Samples. In the Proceedings of the Workshop on Word Sense Disambiguation: Recent Successes and Future Directions, Philadelphia, July 11,2002.
    [43]Ted Pedersen. A Decision Tree of Bigrams is an Accurate Predictor of Word Sense. In Proceedings of the 2nd Meeting of the North American Chapter of the Association for Computational Linguistics, 2001:79-86.
    [44]G. Chao and M. G. Dyer. Maximum Entropy Models for Word Sense Disambiguation. In Proceedings of COLING-2002, Taipei,2002
    [45]Chen Qingcai, Wang Xiaolong, et al. Word Sense Disambiguation Based on Rough set. Journal of Harbin Institute of Technology (New Series).2002,9(2):201-204.
    [46]Tanja Gaustad. A Lemma-Based Approach to a Maximum Entropy Word Sense Disambiguation System for Dutch. In Proceedings of COLING-2004, Taipei,2002.
    [47]U. S. Kohomban and W. S. Lee. Learning Semantic Classes for Word Sense Disambiguation. In Proceedings of the 43th Annual Meeting of the Association for Computational Linguistics (ACL), July 2005.2005:34-41
    [48]B. Magnini, C. Strapparava, et al. The role of domain information in word sense disambiguation. Natural Language Engineering,2002:8(4):359-373.
    [49]A. Gliozzo, C. Giuliano and C. Strapparava. Domain Kernels for Word Sense Disambiguation. In Proceedings of the 43th Annual Meeting of the Association for Computational Linguistics (ACL), July 2005.
    [50]Cheng Niu, Wei Li, et al. Context Clustering for Word Sense Disambiguation Based on Modeling Pairwise Context Similarities. In Proceedings of the 42th ACL, the Third International Workshop on the Evaluation of Systems for the Semantic Analysis of Text, Barcelona, Spain, July 2004.
    [51]A. Purandare and T. Pedersen. Word Sense Discrimination by Clustering Contexts in Vector and Similarity Spaces. In Proceedings of the Conference on Computational Natural Language Learning, Boston, MA,2004:41-48.
    [52]A. Purandarel and T. Pedersen. Improving Word Sense Discrimination with Gloss Augmented Feature Vectors. In the Proceedings of the Workshop on Lexical Resources for the Web and Word Sense Disambiguation, Puebla Mexico, November 22,2004.
    [53]Joseph Turian, Lev Ratinov, Yoshua Bengio. Word representations:A simple and general method for semi-supervised learning. In Proceedings of the 48th Annual Meeting of the Association for Computational Linguistics,11-16 July 2010, Uppsala, Sweden.2010:384-394,
    [54]Daphna Shezaf, Ari Rappoport. Bilingual Lexicon Generation Using Non-Aligned Signatures. In Proceedings of the 48th Annual Meeting of the Association for Computational Linguistics,11-16 July 2010, Uppsala, Sweden.2010:98-107.
    [55]Hiroyuki Shindo, Akinori Fujino, and Masaaki Nagata. Word Alignment with Synonym Regularization. In Proceedings of the 48th Annual Meeting of the Association for Computational Linguistics,11-16 July 2010, Uppsala, Sweden.2010:137-141.
    [56]李涓子.汉语词义消歧方法研究.北京清华大学博士论文.1999.
    [57]朱靖波,李绗等.基于对数模型的词义自动消歧.软件学报.2001,12(9):1405-1412.
    [58]鲁松,白硕等.基于向量空间模型的有导词义消歧.计算机研究与发展.2001,38(6):662-667.
    [59]车超,金博,滕弘飞等.基于义原关系的多策略汉语词义消歧方法.大连理工大学学报.2010,50(4):603-608
    [60]吴云芳,金澎,郭涛.基于词典属性特征的粗粒度词义消歧.中文信息学报.2007,21(2):3-8
    [61]李旭,刘国华,张东明.一种改进的汉语全文无指导词义消歧方法.自动化学报.2010,36(1):184-187
    [62]黄昌宁,李涓子.词义排歧的一种语言模型.语言文字应用.2000,3:85-90.
    [63]鲁松,白硕等.基于向量空间模型中义项词语的无导词义消歧.软件学报.2002,13(6):1082-1089.
    [64]吴云芳,俞士汶.信息处理用词语义项区分的原则和方法.语言文字应用.2006:2,126-133
    [65]蒋严.“都”的语义刻画和相关语用因素.www3.ctl.cityu.edu.hk:8080/scss/www/2007/resource/jiang_yan.pdf.2007
    [66]吴云芳,段慧明,俞士汶.动词对宾语的语义选择限制.语言文字应用.2005,2:121-128
    [67]Daria Bogdanova. A Framework for Figurative Language Detection Based on Sense Differentiation. In Proceedings of the ACL 2010 Student Research Workshop,13 July 2010, Uppsala, Sweden.2010: 67-72.
    [68]Jun'ichi Kazama, Stijn De Saeger, Kow Kuroda. A Bayesian Method for Robust Estimation of Distributional Similarities. In Proceedings of the 48th Annual Meeting of the Association for Computational Linguistics,11-16 July 2010, Uppsala, Sweden.2010:247-256.
    [69]秦颖,王小捷.基于轨道的汉语词义消歧.全国计算语言联合会议,南京,2005.
    [70]Senseval-3. http://www.senseval.org/senseval3/organization.html.2009.
    [71]刘群,李素建.基于《知网》的词汇语义相似度计算.第三届汉语词汇语义学研讨会,台北,2002.
    [72]车超,滕弘飞.伪实例与人工标注实例相结合的词义消歧方法.中文信息学报.2009,23(6):31-38
    [73]陈彬,洪家荣等.最优特征子集选择问题.计算机学报.1997,20(2):133-138.
    [74]J. R. Firth. A synopsis of linguistic theory. In Studies in Linguistic Analysis Oxford:Blackwell. 1930-1955.
    [75]W. Gale, K. Church, and D. Yarowsky. One Sense Per Discourse. In Proceedings of DARPA Speech and Natural Language Workshop,1992:233-237
    [76]David Yarowsky. One Sense per Collocation. In ARPA Human Language Technology Workshop, Princeton, NJ,1993:266-271.
    [77]Dan Klein and Christopher Manning. Maxent models, conditional estimation, and optimization without the magic. Tutorial presented at NAACL-03 and ACL-03. Sapporo.2003.
    [78]P. Resnik. Selectional Preference and Sense Disambiguation. In Proceedings of ACL Siglex Workshop on Tagging Text with Lexical Semantics, Why, What and How? Washington DC, April 1997.1997:52-58.
    [79]鲁松,白硕.自然语言处理中词语上下文有效范围的定量描述.计算机学报.2001,24(7):742-747.
    [80]W. Martin, et al. On the Processing of Text Corpus. In:Hartmann Red. Lexicography:Principles and Practice. New York,1983:56-64.
    [81]John Hughes. Automatically Acquiring a Classification of Words. Paris:University of Leeds. PhD dissertation.1994.
    [82]K. Church, P. Hanks. Word Association Norms, Mutual Information, and Lexicography. Computational Linguistics.1990,16(1):22-29.
    [83]边肇祺.模式识别.北京:清华大学出版社,2008:104-120.
    [84]J. R. Quinlan. Induction of Decision Trees. Machine Learning.1986,1(1):81-106.
    [85]王亚东,郭茂祖等.一种基于信息增益与费用评价函数的特征选抒准则.计算机研究与发展.1999,36(7):788-793.
    [86]陈清才,王晓龙.一种基于词矢量的汉语语义量化模型.计算机研究与发展.2001,38(2):208-212.
    [87]宋余庆,罗永刚等.应用主分量分析与粗糙集处理的特征提取.计算机工程与应用.2004,22:48-50.
    [88]Dekai Wu, Weifeng Su, Marine Carpuat. A Kernel PCA Method for Superior Word Sense Disambiguation. In Proceedings of the 42th Annual Meeting of the Association for Computational Linguistics,2004.
    [89]杨震,邓贵仕.个性化信息获取中特征提取的遗传算法.系统工程理论与实践.2004,8:86-90.
    [90]Zhongjun He, Yao Meng, Hao Yu. Maximum Entropy Based Phrase Reordering for Hierarchical Phrase-based Translation. In Proceedings of the 2010 Conference on Empirical Methods in Natural Language Processing,9-11 October 2010, MIT, Massachusetts, USA.2010:555-563.
    [91]Wong Pingwai, Yang Yonghsheng. A Maximum Entropy Approach to HowNet Based Chinese Word Sense Disambiguation. http://www.cs.ust.hk/~hltc/semanet02/pdf/wong.pdf.2009
    [92]Ting Liu, Zhimao Lu, Sheng Li. Chinese Word sense disambiguation based on Neural Networks. Journal of Harbin Institute of Technology (New Series),2005,12(4):408-417.
    [93]Tanel Alumae, Mikko Kurimo. Domain Adaptation of Maximum Entropy Language Models. In Proceedings of the ACL 2010 Conference Short Papers,11-16 July 2010, Uppsala, Sweden,2010: 301-306.
    [94]Rosenfeld Ronnald. A Maximum Entropy to Adaptive Statistical Language Learning. Computer Speech and Language.1996,10(3):187-228.
    [95]A. L. Berger, Pietra S. Della, et al. A Maximum Entropy Approach to Natural Language Processing. Computational Linguistics.1996,22(1):40-72.
    [96]Adwait Ratnaparkhi. A Maximum Entropy Approach for Part-of-Speech Tagging. In Proceedings of the First Empirical Methods in Natural Language Processing Conference, Philadelphia, USA,1996, 133-141.
    [97]陈笑蓉,秦进.基于最大熵原理的汉语词义消歧.计算机科学.2005,32(5):174-176.
    [98]张仰森.基于最大熵模型的汉语词义消歧与标注方法.计算机工程.2009,35(18):15-18.
    [99]孙超,张仰森.利用浅层句法分析提取特征的词义消歧.计算机工程与设计.2010,31(21):4704-4707.
    [100]Aditya Bhargava, Grzegorz Kondrak. Language identification of names with SVMs. In Proceedings of the 2010 Annual Conference of the North American Chapter of the ACL.June 2010, Los Angeles, California,2010:693-696.
    [101]James Mayfield, Paul McNamee, and Christine Piatko. Named Entity Recognition Using Hundreds of Thousands of Features. In Walter Daelemans and Miles Osborne, editors, Proceedings of CoNLL-2003, Edmonton, Canada,2003:184-187.
    [102]D. Yarowsky. Unsupervised Word Sense Disambiguation Rivaling Supervised Methods. In Proceedings of the 33rd Annual Meeting of the Association for Computational Linguistics,1995: 189-196.
    [103]William A. Gale, Kenneth W. Church and David Yarowsky. Using Bilingual Materials to Develop Word Sense Disambiguation Methods. In Proceedings of the Fourth International Conference on Theoretical and Methodological Issues in Machine Translation. Montreal, Canada,1992.
    [104]R. Mihalcea and D. I. Moldovan. An Iterative Approach to Word Sense Disambiguation. In Proceedings of FLAIRS-2000, Orlando, FL, May,2000:219-223.
    [105]R. Mihalcea. Bootstrapping Large Sense Tagged Corpora. Proceedings of the 3rd International Conference on Languages Resources and Evaluations (LREC). Las Palmas, Canary Islands, Spain, 2002.
    [106]Zheng-Yu Niu, Dong-Hong Ji and Chew Lim Tan. Serguei Pakhomov. Word Sense Disambiguation Using Label Propagation Based Semi-Supervised Learning. In Proceedings of the 43th Annual Meeting of the Association for Computational Linguistics (ACL), July 2005.
    [107]Seong-Bae Pare, Byoung-Tak Zhang, et al. Word Sense Disambiguation by Learning Decision Trees from Unlabeled Data. Kluwer Academic Publishers. Manufactured in The Netherlands. Applied Intelligence,2003,19:27-38.
    [108]I. Dagan, A. Itai and U. Schwall. Two Languages are More Informative than one. Proceedings of the 29th Annual Meeting of the ACL,18-21 Berkeley, California,1991:130-137.
    [109]I. Dagan and Alon Itai. Word Sense Disambiguation Using a Second Language Monolingual Corpus. Computational Linguistics,1994,20(4):56-596.
    [110]P. Resnik and D. Yarowsky. A Perspective on Word Sense Disambiguation Methods and Their Evaluation. In Proceedings of the ACL SIGLEX Workshop on Tagging Text with Lexical Semantics: Why, What, and How? 1997:79-86.
    [111]Nancy Ide. Cross-lingual Sense Determination:Can it work? In Computers and the Humanities: Special Issue on Senseval,2000,34:147-152.
    [112]H. Kaji and Y. Morimoto. Unsupervised Word Sense Disambiguation Using Bilingual Comparable Corpora. In Proceedings of COLING-2002, Taipei,2002.
    [113]Hwee Tou Ng, Bin Wang and Yee Seng Chan. Exploiting Parallel Texts for Word Sense Disambiguation:An Empirical Study. In Proceedings of the 41th Annual Meeting of the Association for Computational Linguistics,2003.
    [114]Mona Diab. Word Sense Disambiguation Within a Multilingual Framework. University of Maryland College Park. PhD Thesis, USA,2003.
    [115]Mona Diab. An Unsupervised Approach for Boot-strapping Arabic Word Sense Tagging. Proceedings of Arabic Based Script Languages, COLING 2004. Geneva, Switzerland,2004.
    [116]Philip Resnik. Mining the Web for Bilingual Text. In Proceedings of the 37th Annual Meeting of the Association for Computational Linguistics,1999:527-534
    [117]Cong Li and Hang Li. Word translation disambiguation using bilingual bootstrapping. In Proceedings of the 40th Annual Meeting of the Association for Computational Linguistics,2002, 343-351.
    [118]Mona Diab and Philip Resnik. An Unsupervised Method for Word Sense Tagging Using Parallel Corpora. In Proceedings of the 40th Annual Meeting of the Association for Computational Linguistics,2002:255-262.
    [119]Y. Bengio and C. Kermorvant. Extracting Hidden Sense Probabilities from Bi-texts. Technical Report, TR 1231, Departement d'Informatique et Recherche Operationnelle, Universite de Montreal, 2003.
    [120]I. Bhattacharya, L. Getoor, Y. Bengio. Unsupervised Sense Disambiguation Using Bilingual Probabilistic Models. In Proceedings of the 42th Annual Meeting of the Association for Computational Linguistics,2004.
    [121]Iman Saleh, Kareem Darwish, Aly Fahmy. Classifying Wikipedia Articles into NE's Using SVM's with Threshold Adjustment. In Proceedings of the 2010 Named Entities Workshop, ACL 2010,16 July 2010, Uppsala, Sweden.2010:85-92
    [122]William A. Gale, K. W. Church, and D. Yarowsky. Word on Statictical Methods for Word Sense Disambiguation. In Robert Goldman, Peter Norvig. Eugene Charniak, and Bill Gale (eds.). Working Notes of the AAAI Fall Symposium on Probabilistic Approaches to Natural Language, Menlo Park, CA. AAAI Press,1992:54-60.
    [123]H. Schutze. Context Space. In Robert Goldman, Peter Norvig. Eugene Charniak, and Bill Gale (eds.). Working Notes of the AAAI Fall Symposium on Probabilistic Approaches to Natural Language, Menlo Park, CA. AAAI Press,1992:113-112.
    [124]Nathanael Chambers and Dan Jurafsky. Improving the Use of Pseudo-Words for Evaluating Selectional Preferences. Proceedings of the 48th Annual Meeting of the Association for Computational Linguistics,11-16 July 2010, Uppsala, Sweden.2010:445-453.
    [125]陈维振,吴世雄.范畴与模糊语义研究.福州:福建人民出版社,2002.
    [126]Tayor J.R. Linguistic Categorization:Prototypes in Linguistic Theory. Peking:Foreign Language Teaching and Research Press& Oxford UP,2001.
    [127]Clifton J. McFate. Expanding Verb Coverage in Cyc with VerbNet. In Proceedings of the ACL 2010 Student Research Workshop,13 July 2010, Uppsala, Sweden.2010:61-66.

© 2004-2018 中国地质图书馆版权所有 京ICP备05064691号 京公网安备11010802017129号

地址:北京市海淀区学院路29号 邮编:100083

电话:办公室:(+86 10)66554848;文献借阅、咨询服务、科技查新:66554700