2
2

My go-to reference to nlp, Jurafsky and Martin's book, fails to mention what exactly is a projective dependency/phrase structure grammar, or what is a projective parser, and why is it different (worse?) than a non-projective one. Where can I see this clearly explained?

asked Oct 09 '10 at 18:59

Alexandre%20Passos's gravatar image

Alexandre Passos ♦
2554154278421

edited Jun 17 '11 at 21:22

Oscar%20T%C3%A4ckstr%C3%B6m's gravatar image

Oscar Täckström
2039133450


2 Answers:

Non-projective dependency trees are ones with crossing edges. They allow for more constructions to represented in your parse than the projectivity constraint, and is prevalent in languages with free word order.

Look at Figure 2 of McDonald et al (2005) "Non-projective dependency parsing using spanning tree algorithms". (Actually, this is a seminal work, and it started the whole MST-based approach to dependency parsing. I encourage you to read it.)

From their work:

In English, projective trees are sufficient to analyze most sentence types. ... However, there are certain examples in which a non-projective tree is preferable. Consider the sentence John saw a dog yesterday which was a Yorkshire Terrier. Here the relative clause which was a Yorkshire Terrier and the object it modifies (the dog) are separated by an adverb. There is no way to draw the dependency tree for this sentence in the plane with no crossing edges, as illustrated in Figure 2. In languages with more flexible word order than English, such as German, Dutch and Czech, non-projective dependencies are more frequent.

answered Oct 10 '10 at 11:45

Joseph%20Turian's gravatar image

Joseph Turian ♦♦
579051125146

1

Just a quick historic note: the non-projective dependency parsing as graph search was described already at 1998 for parsing Czech. The parameter estimation was poor (count-and-divide, like most work at that time period), and the decoding was a greedy-heuristic instead of the exact CLE MST search (no good excuse for that one..), but the principle components are there.

(Oct 10 '10 at 21:05) yoavg

A small, loosely related question: what sort of feature is commonly used in a dependency parser? I didn't find even a rough list in the paper, and it doesn't seem obvious to me which way to go.

(Oct 11 '10 at 14:26) Alexandre Passos ♦

For a graph-based parser, you can find a good list in Ryan's thesis (page 51). People are doing more than that these days, but this is the core feature set and it is quite hard to beat.

For a transition based parser, there is a good list in Huang et.al. 2009.

You can also check out the feature list of my easy-first parser.

(Oct 11 '10 at 16:42) yoavg

To add to the discussion, non-projectivity is dependent mostly on the way it is represented. As of now, the present theory of dependency parsing might poke you towards several cases of non-projectivity - partially non-projective, strictly non-projective, etc.. If you are looking into the concept of non-projectivity, then the above definition is sufficient. But if you are deeply interested in it, I would suggest you go through an indepth analysis here: "Marco Kuhlmann, Giorgio Satta: Treebank Grammar Techniques for Non-Projective Dependency Parsing. EACL 2009: 478-486"

answered Jun 11 '11 at 20:46

kakashi_'s gravatar image

kakashi_
1224

Your answer
toggle preview

powered by OSQA

User submitted content is under Creative Commons: Attribution - Share Alike; Other things copyright (C) 2010, MetaOptimize LLC.