Computational Linguistics
About

Dependency Parsing

Dependency parsing analyzes syntactic structure by establishing directed, labeled binary relations (dependencies) between individual words, with one word designated as the root of the sentence.

head(w_i) = w_j, label(w_i) = l (e.g., nsubj, dobj, amod)

Dependency parsing represents the syntactic structure of a sentence as a set of directed binary relations between words. Each relation, called a dependency, connects a head (governor) word to a dependent (modifier) word and is labeled with a grammatical relation such as nsubj (nominal subject), obj (direct object), or amod (adjectival modifier). The resulting structure is a dependency tree: a rooted, directed tree in which every word except the root has exactly one head, and there are no cycles.

Dependency Trees

Dependency Tree Properties A dependency tree for sentence w1 ... wn is a directed graph (V, A):
V = {0, 1, ..., n} (0 = artificial root node)
A ⊆ V × V (set of arcs/dependencies)

Constraints:
1. Single head: each wi (i ≥ 1) has exactly one head
2. Acyclicity: no directed cycles
3. Connectedness: every word is reachable from root
Together ⇒ A forms a directed tree rooted at node 0

Dependency representations have several advantages over constituency representations. They directly encode predicate-argument structure and head-modifier relations, making them closer to the semantic interpretation. They are more suitable for languages with free word order (such as Czech, Turkish, or Hindi), where constituency trees require many discontinuous or crossing branches. They also produce more compact representations — every tree has exactly n arcs for a sentence of n words.

Approaches to Dependency Parsing

The two dominant algorithmic paradigms are transition-based parsing, which constructs the tree incrementally through a sequence of shift-reduce actions, and graph-based parsing, which scores all possible arcs and finds the highest-scoring tree using maximum spanning tree algorithms. Transition-based parsers are faster (typically linear time) but may propagate errors; graph-based parsers are globally optimal but slower (typically quadratic or cubic time).

Dependency vs. Constituency
Dependency and constituency representations are inter-convertible under certain assumptions: given head-finding rules, a constituency tree can be converted to a dependency tree by identifying the head of each phrase and establishing dependencies from each head to its siblings' heads. This conversion is how dependency treebanks are often bootstrapped from constituency treebanks.

Modern dependency parsers achieve labeled attachment scores (LAS) above 95% for English and above 90% for many other languages. The Universal Dependencies project has standardized annotation across more than 100 languages, enabling large-scale multilingual dependency parsing research. Neural network models, particularly those based on BiLSTMs and Transformers, have achieved the highest accuracy across most languages.

Related Topics

References

  1. Kübler, S., McDonald, R., & Nivre, J. (2009). Dependency parsing. Morgan & Claypool. https://doi.org/10.2200/S00169ED1V01Y200901HLT002
  2. Nivre, J. (2008). Algorithms for deterministic incremental dependency parsing. Computational Linguistics, 34(4), 513–553. https://doi.org/10.1162/coli.07-056-R1-07-027
  3. McDonald, R., Pereira, F., Ribarov, K., & Hajiç, J. (2005). Non-projective dependency parsing using spanning tree algorithms. Proceedings of HLT-EMNLP 2005, 523–530. https://doi.org/10.3115/1220575.1220641
  4. de Marneffe, M.-C., & Manning, C. D. (2008). The Stanford typed dependencies representation. Proceedings of the COLING Workshop on Cross-Framework and Cross-Domain Parser Evaluation, 1–8.

External Links