学习成分结构的中心词 / Learning Constituent Headedness
1️⃣ 一句话总结
这篇论文提出了一种通过监督学习来预测句子成分结构中心词的新方法,它比传统规则方法更准确,并能有效提升句法分析任务的效果。
Headedness is widely used as an organizing device in syntactic analysis, yet constituency treebanks rarely encode it explicitly and most processing pipelines recover it procedurally via percolation rules. We treat this notion of constituent headedness as an explicit representational layer and learn it as a supervised prediction task over aligned constituency and dependency annotations, inducing supervision by defining each constituent head as the dependency span head. On aligned English and Chinese data, the resulting models achieve near-ceiling intrinsic accuracy and substantially outperform Collins-style rule-based percolation. Predicted heads yield comparable parsing accuracy under head-driven binarization, consistent with the induced binary training targets being largely equivalent across head choices, while increasing the fidelity of deterministic constituency-to-dependency conversion and transferring across resources and languages under simple label-mapping interfaces.
学习成分结构的中心词 / Learning Constituent Headedness
这篇论文提出了一种通过监督学习来预测句子成分结构中心词的新方法,它比传统规则方法更准确,并能有效提升句法分析任务的效果。
源自 arXiv: 2603.14755