Jump to content

Cross-serial dependencies

From Wikipedia, the free encyclopedia
A diagram showing cross-serial dependency using lines and colors to represent the dependent pairs.
A schematic showing cross-serial dependencies. Notice that the w's and v's, which represent words, each form respective series. Notice also that the lines representing the dependency relations mutually overlap.

In linguistics, cross-serial dependencies (also called crossing dependencies by some authors[1]) occur when the lines representing the dependency relations between two series of words cross over each other.[2] They are of particular interest to linguists who wish to determine the syntactic structure of natural language; languages containing an arbitrary number of them are non-context-free. By this fact, Dutch[3] and Swiss-German[4] have been proven to be non-context-free.

Example

[edit]
A Swiss-German sentence containing cross-serial dependencies (shown as lines between the verbs and their objects). The English translation with its dependencies, which do not cross, is shown for comparison.
A more complicated example.

As Swiss-German allows verbs and their arguments to be ordered cross-serially, we have the following example, taken from Shieber:[4]

...mer em Hans es huus hälfed aastriiche.
...we Hans (dat) the house (acc) help paint.

That is, "we help Hans paint the house."

Notice that the sequential noun phrases em Hans (Hans) and es huus (the house), and the sequential verbs hälfed (help) and aastriiche (paint) both form two separate series of constituents. Notice also that the dative verb hälfed and the accusative verb aastriiche take the dative em Hans and accusative es huus as their arguments, respectively.

Non-context-freeness

[edit]

Let to be the set of all Swiss-German sentences. We will prove mathematically that is not context-free.

In Swiss-German sentences, the number of verbs of a grammatical case (dative or accusative) must match the number of objects of that case. Additionally, a sentence containing an arbitrary number of such objects is admissible (in principle). Hence, we can define the following formal language, a subset of :Thus, we have , where is the regular language defined by where the superscript plus symbol means "one or more copies". Since the set of context-free languages is closed under intersection with regular languages, we need only prove that is not context-free (,[5] pp 130–135).

After a word substitution, is of the form . Since can be mapped to by the following map: , and since the context-free languages are closed under mappings from terminal symbols to terminal strings (that is, a homomorphism) (,[5] pp 130–135), we need only prove that is not context-free.

is a standard example of non-context-free language (,[5] p. 128). This can be shown by Ogden's lemma.

Suppose the language is generated by a context-free grammar, then let be the length required in Ogden's lemma, then consider the word in the language, and mark the letters . Then the three conditions implied by Ogden's lemma cannot all be satisfied.

All known spoken languages which contain cross-serial dependencies can be similarly proved to be not context-free.[2] This led to the abandonment of Generalized Phrase Structure Grammar once cross-serial dependencies were identified in natural languages in the 1980s.[6]

Treatment

[edit]

Research in mildly context-sensitive language has attempted to identify a narrower and more computationally tractable subclass of context-sensitive languages that can capture context sensitivity as found in natural languages. For example, cross-serial dependencies can be expressed in linear context-free rewriting systems (LCFRS); one can write a LCFRS grammar for {anbncndn | n ≥ 1} for example.[7][8][9]

References

[edit]
  1. ^ Stabler, Edward (2004), "Varieties of crossing dependencies: structure dependence and mild context sensitivity" (PDF), Cognitive Science, 28 (5): 699–720, doi:10.1016/j.cogsci.2004.05.002.
  2. ^ a b Jurafsky, Daniel; Martin, James H. (2000). Speech and Language Processing (1st ed.). Prentice Hall. pp. 473–495. ISBN 978-0-13-095069-7..
  3. ^ Bresnan, Joan; M. Kaplan, Ronald (1982), "Cross-serial dependencies in Dutch", Linguistic Inquiry, 13 (4): 613–635.
  4. ^ a b Shieber, Stuart (1985), "Evidence against the context-freeness of natural language" (PDF), Linguistics and Philosophy, 8 (3): 333–343, doi:10.1007/BF00630917, S2CID 222277837.
  5. ^ a b c John E. Hopcroft, Jeffrey D. Ullman (1979). Introduction to Automata Theory, Languages, and Computation (1st ed.). Pearson Education. ISBN 978-0-201-44124-6..
  6. ^ Gazdar, Gerald (1988). "Applicability of Indexed Grammars to Natural Languages". Natural Language Parsing and Linguistic Theories. Studies in Linguistics and Philosophy. Vol. 35. pp. 69–94. doi:10.1007/978-94-009-1337-0_3. ISBN 978-1-55608-056-2.
  7. ^ http://user.phil-fak.uni-duesseldorf.de/~kallmeyer/GrammarFormalisms/4nl-cfg.pdf [bare URL PDF]
  8. ^ http://user.phil-fak.uni-duesseldorf.de/~kallmeyer/GrammarFormalisms/4lcfrs-intro.pdf [bare URL PDF]
  9. ^ Laura Kallmeyer (2010). Parsing Beyond Context-Free Grammars. Springer Science & Business Media. pp. 1–5. ISBN 978-3-642-14846-0.