TY - GEN
T1 - Smoothing for bracketing induction
AU - Duan, Xiangyu
AU - Zhang, Min
AU - Chen, Wenliang
PY - 2013
Y1 - 2013
N2 - Bracketing induction is the unsupervised learning of hierarchical constituents without labeling their syntactic categories such as verb phrase (VP) from natural raw sentences. Constituent Context Model (CCM) is an effective generative model for the bracketing induction, but the CCM computes probability of a constituent in a very straightforward way no matter how long this constituent is. Such method causes severe data sparse problem because long constituents are more unlikely to appear in test set. To overcome the data sparse problem, this paper proposes to define a non-parametric Bayesian prior distribution, namely the Pitman-Yor Process (PYP) prior, over constituents for constituent smoothing. The PYP prior functions as a back-off smoothing method through using a hierarchical smoothing scheme (HSS). Various kinds of HSS are proposed in this paper. We find that two kinds of HSS are effective, attaining or significantly improving the state-ofthe- art performance of the bracketing induction evaluated on standard treebanks of various languages, while another kind of HSS, which is commonly used for smoothing sequences by ngram Markovization, is not effective for improving the performance of the CCM.
AB - Bracketing induction is the unsupervised learning of hierarchical constituents without labeling their syntactic categories such as verb phrase (VP) from natural raw sentences. Constituent Context Model (CCM) is an effective generative model for the bracketing induction, but the CCM computes probability of a constituent in a very straightforward way no matter how long this constituent is. Such method causes severe data sparse problem because long constituents are more unlikely to appear in test set. To overcome the data sparse problem, this paper proposes to define a non-parametric Bayesian prior distribution, namely the Pitman-Yor Process (PYP) prior, over constituents for constituent smoothing. The PYP prior functions as a back-off smoothing method through using a hierarchical smoothing scheme (HSS). Various kinds of HSS are proposed in this paper. We find that two kinds of HSS are effective, attaining or significantly improving the state-ofthe- art performance of the bracketing induction evaluated on standard treebanks of various languages, while another kind of HSS, which is commonly used for smoothing sequences by ngram Markovization, is not effective for improving the performance of the CCM.
UR - https://www.scopus.com/pages/publications/84896061874
M3 - 会议稿件
AN - SCOPUS:84896061874
SN - 9781577356332
T3 - IJCAI International Joint Conference on Artificial Intelligence
SP - 2085
EP - 2091
BT - IJCAI 2013 - Proceedings of the 23rd International Joint Conference on Artificial Intelligence
T2 - 23rd International Joint Conference on Artificial Intelligence, IJCAI 2013
Y2 - 3 August 2013 through 9 August 2013
ER -