自动化技术 |
|
|
|
|
统计机器翻译中大规模特征的深度融合 |
刘宇鹏, 乔秀明, 赵石磊, 马春光 |
1. 哈尔滨工程大学 计算机科学与技术学院, 黑龙江 哈尔滨 150001
2. 哈尔滨理工大学 软件学院, 黑龙江 哈尔滨 150001
3. 哈尔滨工业大学 计算机学院, 黑龙江 哈尔滨 150001 |
|
Deep combination of large-scale features in statistical machine translation |
LIU Yu peng, QIAO Xiu ming, ZHAO Shi lei, MA Chun guang |
1. School of Computer Science and Technology, Harbin Engineering University, Harbin 150001, China;
2. Software School, Harbin University of Science and Technology, Harbin 150001, China;
3. School of Computer Science and Technology, Harbin Institute of Technology, Harbin 150001, China |
引用本文:
刘宇鹏, 乔秀明, 赵石磊, 马春光. 统计机器翻译中大规模特征的深度融合[J]. 浙江大学学报(工学版), 10.3785/j.issn.1008-973X.2017.01.006.
LIU Yu peng, QIAO Xiu ming, ZHAO Shi lei, MA Chun guang. Deep combination of large-scale features in statistical machine translation. JOURNAL OF ZHEJIANG UNIVERSITY (ENGINEERING SCIENCE), 10.3785/j.issn.1008-973X.2017.01.006.
|
[1] ONISHI K H, MURPHY G L, BOCK K. Prototypicality in sentence production [J]. Cognitive Psychology, 2008, 56(2): 103-141.
[2] YOO H J. Deep convolution neural networks in computer vision: a review [J]. IEEE Transactions on Smart Processing and Computing, 2015, 4(1): 35-43.
[3] SAINATH T N, KINGSBUR B, SAON G. Deep convolutional neural networks for largescale speech tasks [J]. Neural Networks, 2015, 64: 39-48.
[4] LE Q, MIKOLOV T. Distributed representations of sentences and documents [C]∥Proceedings of the 31st International Conference on Machine Learning. Beijing: [s. n.], 2014: 1188-1196.
[5] MIKOLOV T, KARAFIAT M, BURGET L. Recurrent neural network based language model [C]∥ Proceedings of the 11th Annual Conference of the International Speech Communication Association (INTERSPEECH). Makuhari: [s. n.], 2010: 1045-1048.
[6] SOCHER R, LIN C C, MANNING C. Parsing natural scenes and natural language with recursive neural networks [C]∥Proceedings of ICML. Bellevue: [s. n.], 2011: 129136.
[7] ROSTI A I, AYAN N F, XIANG B. Combing outputs from multiple machine translation systems [C]∥HLTNAACL. Rochester: [s. n.], 2007: 228-235.
[8] DENERO J, KUAMR S, CHELBA C. Model combination for machine translation [C]∥ Proceedings of NAACL. Los Angeles: [s. n.], 2010: 975-983.
[9] LIANG P,BOUCHART A, KLEIN D. An endtoend discriminative approach to machine translation [C]∥Proceedings of ACL. Sydney: ACL, 2006: 277-285.
[10] TILLMANN C, ZHANG T. A discriminative global training algorithm for statistical MT [C]∥ Proceedings of the 44th Annual Meeting of the Association for Computational Linguistics. Sydney: [s. n.], 2006:721-728.
[11] TARO W, JUN S, HAJIME T. Online largemargin training for statistical machine translation [C]∥Proceedings of EMNLPCoNLL. Prague: [s. n.], 2007: 764-773.
[12] CHIANG D. Hope and fear for discriminative training of statistical translation models [J]. Journal of Machine Learning Research, 2012(13): 1159-1187.
[13] SPENCE G, WANG S, CER D. Fast and adaptive online training of featurerich translation models [C]∥Proceedings of ACL. Sofia: ACL, 2013: 311-321.
[14] HOPKINS M, MAY J. Tuning as ranking [C]∥ Proceedings of the Conference on Empirical Methods in Natural Language Processing. Association for Computational Linguistics. Edinburgh: [s. n.], 2011: 1352-1362.
[15] CHO K, VAN B, GULCEHRE C. Learning phrase representations using RNN encoderdecoder for statistical machine translation [C]∥Proceedings of the Empirical Methods in Natural Language Processing. Doha: [s. n.], 2014.
[16] BAHDANAU D, CHO K, BENGIO Y. Neural machine translation by jointly learning to align and translate [C]∥Proceedings of ICLR. San Diego: [s. n.], 2015: 115.
[17] AULI M, GALLEY M, QUIRK C. Joint language and translation modeling with recurrent neural networks [C]∥Proceedings of the 2013 Conference on Empirical Methods in Natural Language Processing, Association for Computational Linguistics. Seattle: [s. n.], 2013: 1044-1054.
[18] ZHAI F, ZHANG J, ZHOU Y. RNNbased derivation structure prediction for SMT [C]∥ Proceedings of ACL. Baltimore: ACL, 2014: 779-784.
[19] ZHANG J, LIU S, LI M. Bilinguallyconstrained phrase embeddings for machine translation [C]∥Proceedings of ACL. Baltimore: ACL, 2014: 111-121.
[20] LI P, LIU Y, SUN M. Recursive autoencoders for ITGbased translation [C]∥Proceedings of EMNLP. Seattle: [s. n.], 2013: 567-577.
[21] ZHANG J, LIU S, LI M, et al. Bilinguallyconstrained phrase embeddings for machine translation [C]∥Proceedings of ACL. Baltimore: ACL, 2014: 111-121.
[22] SU J, XIONG D, ZHANG B. Bilingual correspondence recursive autoencoder for statistical machine translation [C]∥Proceedings of EMNLP. Lisbon: [s. n.], 2015: 1248-1258.
[23] LIU S, YANG N, LI M. A recursive recurrent neural network for statistical machine translation [C]∥ Proceedings of ACL. Baltimore: ACL, 2014: 1491-1500.
[24] ZHANG J, ZHANG D, HAO J. Local translation prediction with global sentence representation [C]∥ Proceedings of IJCAI. Buenos Aires: [s. n.], 2015:1398-1404.
[25] MENG F, LU Z, WANG M. Encoding source language with convolutional neural network for machine translation [C]∥Proceedings of the 53th Annual Meeting of Association for Computational Linguistics and the 7th International Joint Conference on Natural Language Processing (ACLIJCNLP′15). Beijing: [s. n.], 2015: 2030.
[26] HU B, TU Z, LU Z. Contextdependent translation selection using convolutional neural network [C]∥Proceedings of the 53th Annual Meeting of Association for Computational Linguistics and the 7th International Joint Conference on Natural Language Processing (ACLIJCNLP′15). Beijing: [s. n.], 2015: 536-541.
[27] MIKOLOV T, CHEN K, CORRADO G. Efficient estimation of word representations in vector space [C]∥ICLR 2013. Scottsdale: [s. n.], 2013:112.
[28] CHIANG D. Hierarchical phrasebased translation [J]. Computational Linguistics, 2007, 33(2):201-228.
[29] SOCHER R, PENNINGTON J, MANNING E. Semisupervised recursive autoencoders for predicting sentiment distributions [C]∥Conference on Empirical Methods in Natural Language Processing. Edinburgh: [s. n.], 2011: 151-161.
[30] BODEN M. A guide to recurrent neural networks and backpropagation [R]. Holst: Application of Data Analysis with Learning Systems, 2001: 110.
[31] ZHANG H, HUANG L, GILDEA D. Synchronous binarization for machine translation [C]∥ Proceedings of the 2006 Meeting of the North American Chapter of the Association for Computational Linguistics (NAACL06). New York: [s. n.], 2006: 256-263.
[32] JOERN W, ARNE M, HERMANN N. Training phrase translation models with leaving-one-out [C]∥Proceeding of the 48th Annual Meeting of the Association for Computational Linguistics. Uppsala: [s. n.], 2010: 475-484.
[33] ZHANG M, XIAO X, XIONG D. Topicbased dissimilarity and sensitivity models for translation rule selection [J]. Journal of Artificial Intelligence Research, 2014(50): 130.
[34] XIONG D, LIU Q, LIN S. Maximum entropy based phrase reordering model for statistical machine translation [C]∥Proceeding of the 43rd Annual Meeting of the Association for Computational Linguistics. Sydney: [s. n.], 2006: 521-528.
[35] STOCKLE A. SRILM: an extensible language modeling toolkit [C]∥Proceeding of International Conference on Spoken Language Processing. Denver: [s. n.], 2002: 901-904.
[36] PAPINENI K, ROUKOS S, WARD T. Bleu: a method for automatic evaluation of machine translation[C]∥Proceeding of the 40th Annual Meeting on Association for Computational Linguistics. Barcelona: [s. n.], 2002: 311-318.
[37] KOEHN P. Statistical significance tests for machine translation evaluation [C]∥Proceedings of EMNLP. Doha: [s. n.], 2004: 201-208.
[38] BICICI E, YURET D. Instance selection for machine translation using feature decay algorithm [C]∥Proceedings of the 6th Workshop on Statistical Machine Translation of Association for Computational Linguistics. Edinburgh: [s. n.], 2011: 272-283. |
|
Viewed |
|
|
|
Full text
|
|
|
|
|
Abstract
|
|
|
|
|
Cited |
|
|
|
|
|
Shared |
|
|
|
|
|
Discussed |
|
|
|
|