Skip to main content

Recursion in Language: Is It Indirectly Constrained?

  • Chapter
  • First Online:
Recursion: Complexity in Cognition

Part of the book series: Studies in Theoretical Psycholinguistics ((SITP,volume 43))

Abstract

Recursion is a key aspect of formal theories of language. In this chapter we have presented a new perspective on recursion in language, claiming that recursion in language is indirectly constrained. After discussing some aspects of the relation between linguistic theory and mathematical linguistics, we show that this perspective arises from the properties of grammars that are beyond context-free grammars with context-sensitivity that is mild in a sense and that the derivation proceeds from structures to structures. Viewing the generation process inductively, we then show that for each index of recursion, there is a form (we call this a canonical form) that can be generated. However, corresponding to the canonical form, there are variants of the form, called here as non-canonical forms that exhibit gaps in the sense that for a given index of recursion, beyond two levels of embedding, not all non-canonical forms can be generated. This property emerges from the formal architecture of the system and not from any particular linguistic or processing constraint. We call this property Indirectly Constrained Recursion (ICR), suggesting a new perspective on recursion in language. ICR is a formal structural constraint and not a linguistic or processing constraint. In linguistic theories constraints on recursion are usually expressed either in terms of some linguistic constraints or in terms of processing constraints. ICR arises from the properties of the kinds of formal systems considered here. Linguistic constraints and processing constraints are then to be considered as additional constraints. The linguistic phenomena referred to in this chapter are well-known. Hence, the main contribution of this chapter is the presentation of a new perspective on recursion as described above.

In this chapter, we have presented ICR with respect to a particular framework. However, it is expected that similar results will carry over to some of the other systems, for example, in the class of Mildly Context-Sensitive Languages and perhaps, stronger versions of ICR can be found.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 84.99
Price excludes VAT (USA)
  • Available as EPUB and PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 109.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info
Hardcover Book
USD 109.99
Price excludes VAT (USA)
  • Durable hardcover edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

Notes

  1. 1.

    Processing constraints associated with MCSL have been studied in the context of automaton models for MCSL; see, for example, Joshi (1990).

  2. 2.

    This book has been reissued in 2004, with a preface and an Introduction by Naoki Fukui and Mihoko Zushi, Mouton de Gruyter, Berlin, New York. See also “Formal Grammars in Linguistics and Psycholinguistics” William J. M. Levelt, John Benjamin Publishing Company, Amsterdam/Philadelphia, 2008 (reissued as a new edition of the original 1974 edition with an added Postscript summarizing some developments since 1974). This Postscript also refers to the quotation from Chomsky above.

  3. 3.

    A derivation in a TAG proceeds from a sentence form to another sentence form via the operation of adjunction. This is reminiscent of the idea of a transformation as a relation between sentence forms as suggested by Harris (1957). However, in our case, the relation is between structured sentence forms and is formulated as a recursive system.

  4. 4.

    An extension of the pushdown automaton, called an embedded pushdown automaton, EPDA, equivalent to the language of a LTAG (in the same way as a pushdown automaton is equivalent to a CFG) exists and has been studied in the context of processing crossed and nested dependencies (Joshi 1990).

  5. 5.

    In Appendix A, a rather simple example of an LTAG derivation is given for readers who may not be familiar with mildly context sensitive grammars and languages. For more details, including multicomponent LTAGs, see Joshi and Schabes (1997), Joshi (2004), and Vijay-Shanker (1988).

  6. 6.

    In the formal treatment of LTAG recursion enters only via adjoining. In principle, recursion can also enter via substitution. In fact, one way of looking at adjunction is to describe it as a pair of substitutions in the following sense. If a node X in a tree t is targeted for adjoining by an auxiliary tree, say t’, then the adjoining operation can be thought of as a pair of substitutions such that the tree t is pulled apart at the node X, creating two X nodes (top and bottom), and then t’ is spliced into the tree t by making a pair of substitutions. This way of looking at standard adjoining makes the standard TAG as a special case of Tree-Local Multi-Component TAG (TL-MCTAG), which we have discussed later,.

  7. 7.

    We use the term canonical just as a convenient way of referring to the unbounded set of sentence forms, without gaps, that can be derived in a CFG. Nothing further should be read into this terminology.

  8. 8.

    We can also have

    *(3”) Mary, Harry believes John thinks, will win.

    *(3”’) Harry, John thinks, believes Mary will win.

    Judgments for (3’) and (3”) were provided by Tony Kroch and Mark Johnson, respectively.

  9. 9.

    However, consider the following:

    The gardener I told you about showed up again who had done such a bad job.

    Here the extraposed relative clause is meant to modify the subject NP and it sounds much better than (2”). Judgment provided by Tony Kroch.

    It is however possible to extrapose modifiers at the second level in certain cases, as in

    1. (3)

      Supply troubles were on the minds of Treasury investors yesterday who worried about the flood of new government securities. (From WSJ, Penn Treebank Corpus).

    Example 3 was provided by Tonia Bleam and Fei Xia. Stephan Muller has also pointed out similar examples in German, see http://www.cl.uni-bremen/~stefan/Pub/subjacency,html. The relative clause in (3) is non-restrictive. The conditions under which the relative clause can be discontinuous with its head are different for restrictives and non-restrictives.

    Thanks to Tony Kroch for helping me sort out some of these issues.

  10. 10.

    As pointed out by Mark Johnson.

  11. 11.

    German is verb-final only in the subordinate clauses; so the ‘that’ (‘dass’) may not be left out. In our subsequent discussion we will ignore ‘dass’.

    We should really use full NP’s in these examples. We have used proper names only for convenience.

  12. 12.

    There are, of course, many other patterns of scrambling due to the verb sequence itself being scrambled, possibly along with the NP sequence, for example, as in NP1 NP2 NP3 V1 V3 V2. Further, a verb may have more than one NP argument, etc. These can be generated but we do not consider these sequences here.

  13. 13.

    Actually, we use the so-called Tree-Local Multi-Component TAGs (Tree Local MCTAG), which is a generalization of TAGs, allowing more flexibility in composition without increasing the (weak) generative capacity (see Appendix A). We will continue to use the term TAG to include both TAGs and Tree Local MCTAGs.

  14. 14.

    It is possible to add flexible composition (delayed composition) and multiple adjoining under certain circumstances without increasing the weak generative capacity of TL-MCTAG (see Chiang and Scheffler 2008 and Chen-Main and Joshi 2008). The particular derivable and non-derivable sequences mentioned in this paper have been discussed in detail in Chen-Main and Joshi (2008).

  15. 15.

    Our concern here is about the existence of gaps and not about the acceptability judgments necessarily. The acceptability judgments for these sentences are hard in any case. Some are judged as clearly unacceptable and others as marginally acceptable, at best.

    Further, there is no claim here about the gaps being directly related to processing constraints. Processing involves incrementality and that is not clearly an issue here.

  16. 16.

    These three sequences are (1) NP4 NP2 NP3 NP5 NP1 V5 V4 V3 V2 V1 (2) NP4 NP2 NP5 NP3 NP1 V5 V$ V3 V2 V1 and (3) NP4 NP3 NP2 NP5 NP1 V5 V4 V3 V2 V.

    For further details see Chen-Main and Joshi (2008).

  17. 17.

    There is also no claim here that all the sentences outside the gap are equally acceptable.

  18. 18.

    It is interesting to note that even if we consider discourse variants of a canonical sentence form with recursion we find that non-canonical discourse variants break down beyond a certain level. Thus for example, consider the canonical sentence form with tail recursion Harry believes John thinks Mary will win the race and the discourse variant Mary will win the race. John thinks so. Harry believes so. The discourse variant breaks down as the believes clause appears to be taking scope over the first sentence rather than over the think clause. See also Hollebrandse and Roeper (2007) for a discussion about recursion in discourse.

  19. 19.

    There is an automaton model, called an embedded pushdown automaton (EPDA) that corresponds to a TAG in the same way as a pushdown automaton (PDA) is equivalent to a CFG. EPDA have been used to model the processing strategies for crossed and nested dependencies (see Joshi 1990) based on the so-called Principle of Incremental Interpretation as proposed in (Bach et al. 1986). Whether ICR as described above can be related to processing via EPDA is an open question at this time.

  20. 20.

    We have not shown feature structures associated with the nodes of the elementary trees. These feature structures participate in the derivation via unification. We have left out these feature structures for simplicity.

  21. 21.

    By reverse adjoining we mean reversing the direction of composition. The weak generative capacity of Tree-Local MC-TAGs is not affected by such reverse compositions. They do provide more flexibility which is very crucial as we will see in Appendix B.

  22. 22.

    See Joshi (2004) for some early discussion of these issues.

  23. 23.

    In MC-TAG there is the possibility of changing the direction of the derivation under certain conditions, similar in a sense to the change of the direction of the derivation in a categorial grammar by the use of type raising. This sort of flexibility of composition has to maintain the semantic coherence of the composition, i.e., for example, a Vi tree can only compose with a tree of the Vi + 1 type. We omit these details here; see also Chiang and Scheffler (2008) for Delayed Composition.

  24. 24.

    Multiple adjunctions in LTAG are discussed in Schabes and Shieber (1993). These notions can be extended to MC-TAG.

References

  • Bach, E., Brown, C., & Marsen-Wilson, W. (1986). Crossed and nested dependencies in German and Dutch: A psycholinguistic study. Language and Cognitive Processes, 1(4), 249–262.

    Article  Google Scholar 

  • Chiang, D., & Scheffler, T. (2008). Flexible composition and delayed tree-locality. In Proceedings of the TAG + 9 Workshop, Tübingen.

    Google Scholar 

  • Chen-Main, J., & Joshi, A. K. (2008). Flexible composition, multiple adjoining, and word order variation. In Proceedings of the TAG + 9 Workshop, Tübingen.

    Google Scholar 

  • Chomsky, N. (1957). Syntactic structures. The Hague: Mouton.

    Google Scholar 

  • Chomsky, N. (2007). Approaching UG from below. In U. Sauerland & H. M. Gärtner (Eds.), Approaching UG from below (Interfaces + Recursion = Language? Chomsky’s minimalism and the view from syntax-semantics). Berlin: Mouton de Gruyter.

    Google Scholar 

  • Frank, R. (2002). Phrase structure composition and syntactic dependencies. Cambridge: MIT Press.

    Google Scholar 

  • Gärtner, H.-M., & Michaelis, J. (2007). Some remarks on locality conditions and minimalist grammars. In U. Sauerland & H. Gartner (Eds.), Interfaces + Recursion = Language? (pp. 161–196). Berlin: Mouton de Gruyter.

    Google Scholar 

  • Harris, Z. S. (1957). Co-occurrence and transformation in linguistic structure. Language, 33(3), 283–340.

    Article  Google Scholar 

  • Hollebrandse, B., & Roeper, T. (2007, March). Recursion and propositional exclusivity. In Proceedings of the workshop on recursion, University of Illinois, Normal, IL.

    Google Scholar 

  • Joshi, A. K. (1985). Tree adjoining grammars: How much context sensitivity is required to provide reasonable structural descriptions? In D. Dowty, L. Karttunen, & A. Zwicky (Eds.), Natural language parsing (pp. 206–250). Cambridge: Cambridge University Press.

    Google Scholar 

  • Joshi, A. K. (1990). Processing crossed and nested dependencies: An automaton perspective on the psycholinguistic results. Language and Cognitive Processes, 5, 1–27.

    Article  Google Scholar 

  • Joshi, A. K. (2004). Starting with complex primitives pays off: Complicate locally, simplify globally. Cognitive Science, 28(2004), 637–668.

    Article  Google Scholar 

  • Joshi, A. K., & Schabes, Y. (1997). Tree adjoining grammars. In G. Rosenberg & A. Salomma (Eds.), Handbook of formal languages (pp. 99–123). Berlin: Springer.

    Google Scholar 

  • Kallmeyer, L., & Joshi, A. K. (2003). Factoring predicate argument and scope semantics: Underspecified semantics with LTAG. Research on Language and Computation, 2003(1–2), 3–58.

    Article  Google Scholar 

  • Kroch, A. S., & Joshi, A. K. (1986). Analyzing extraposition in a tree adjoining grammar. In G. Huck & A. Ojeda (Eds.), Discontinuous constituents, syntax and semantics. New York: Academic.

    Google Scholar 

  • Peters, S., & Ritchie, R. W. (1973). On the generative power of transformational grammars. Information Sciences, 6(49–83), 1973.

    Google Scholar 

  • Schabes, Y., & Shieber, S. (1993). An alternative conception of tree adjoining derivation. Computational Linguistics, 20–1, 91–124.

    Google Scholar 

  • Vijay-Shanker, K. (1988). A study of tree-adjoining grammars. Ph.D. dissertation, University of Pennsylvania, Philadelphia, PA, USA.

    Google Scholar 

Download references

Acknowledgement

I want to thank Rajesh Bhatt, Joan Chen-Main, Robert Frank, Julia Hockenmaier, Steve Isard, Mark Johnson, Anthony Kroch, Uwe Mönich, and Mark Steedman for very valuable discussions concerning some of the issues discussed in this chapter.

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Aravind K. Joshi .

Editor information

Editors and Affiliations

Appendices

Appendices

1.1 Appendix A

1.1.1 LTAGs and MC-LTAGs

We will give a brief introduction to Lexicalized Tree Adjoining Grammars (LTAGs) and tree-local Multi-Component LTAGs (MC-LTAGs), which will be adequate to follow the derivations in Sects. 1, 2, and Appendix B.

An LTAG consists of a finite set of elementary structures (trees, directed acyclic graphs in general) where each elementary structure is associated with a lexical anchor. Each elementary structure encapsulates the syntactic/semantic arguments and dependencies (corresponding to the lexical anchor associated with the structure) such as agreement, subcategorization, filler-gap dependencies, among others. These structures are large enough to localize all these dependencies and therefore can be larger than one level trees. This property, called the extended domain of locality, is in contrast to the case of CFG’s where the dependencies are distributed over a set of rules.

The elementary trees are either initial trees or auxiliary trees. There is no recursion inside an elementary structure. Of course, an elementary structure may have the same label for the root node and the foot node in an auxiliary tree and can thus participate in a recursive structure. Since there is no recursion internal to an elementary structure, in a sense, in LTAGs recursion is factored away from the domain of dependencies. This is in contrast to CFG’s where recursion is intertwined with the domain of locality, as both may be spread over several rules.

There are two composition operations, substitution and adjoining: These operations are universal in the sense that they are the same for all LTAGs. There are no rules in the sense of the rewriting rules in a CFG.

Since all dependencies of a lexical anchor are localized we may have several elementary structures corresponding to a lexical item, depending on the structural positions each argument can occupy in different constructions associated with the lexical item. There can be many such trees but the total number is finite. One can regard the different trees associated with a lexical item as extended projections associated with that item.

Figures A.1, A.2, and A.3 show an LTAG derivation. Figure A.1 shows the collection of treesFootnote 20 participating in the derivation. Figure A.2 shows the composition of the elementary trees in Fig. A.1 by means of the operations of substitution (shown by a solid line) and adjoining (shown by a dotted line). Figure A.3 shows the derivation tree. The nodes are labeled by the names of trees as well as by the lexical anchors associated with the elementary trees. Solid lines represent substitution and the dotted lines represent adjoining. The lines should also be labeled by the addresses of the nodes in the host trees where substitutions and adjoinings take place. These have been omitted in order to simplify the diagram. Finally, Fig. A.4 shows the derived tree, which is the result of carrying out the compositions shown in Fig. A.3.

Fig. A.1
figure 1

An LTAG derivation

Fig. A.2
figure 2

An LTAG derivation

Fig. A.3
figure 3

A derivation tree

Fig. A.4
figure 4

A derived tree

Note that the dependency between who and e has become long distance. However, it was local in the elementary tree associated with likes in Figs. A.1 and A.2.

The derivation tree is a record of the history of composition of the elementary structures by substitution or adjoining. The derived tree is the result of carrying out all the composition operations as indicated in the derivation tree. In a sense, the derivation tree has all the information and, in fact, compositional semantics can be defined on the derivation tree (Kallmeyer and Joshi 2003).

The order of derivation in the derivation tree in Fig. A.3 is shown bottom up in a sense. This is the normal convention that is followed in a TAG derivation. However, in principle the derivation could begin with any node with the directions of the arrows reversed appropriately. The order of derivation is flexible in a sense; see also Sect. 1 and (Chomsky 2007).

1.1.2 Multi-component TAG’s (MC-TAG’s)

MC-TAG’s are like LTAG’s except that in an MC-TAG the elementary trees can be a set of trees (usually a pair of trees in the linguistic context) and at least one of the trees in the multi-component set is required to have a lexical anchor. In this sense, MC-LTAGs are lexicalized. In a tree-local MC-TAG when a multi-component tree (say with two components) composes with another tree, say t, the two components compose with the same elementary tree, t, by targeting two nodes of that elementary tree. This is why this composition is called tree-local. The two components target two nodes in the local domain of an elementary tree, preserving the notion of locality that is inherent in LTAG. Tree-local MC-LTAGs are known to be weakly equivalent to LTAGs, i.e., they generate the same set of strings but not necessarily the same set of structural descriptions. Instead of formally describing the composition operation in MC-LTAG, we will illustrate it by means of a simple example. In Fig. A.5 we have shown a simple tree-local MC-LTAG derivation.

Fig. A.5
figure 5

A derivation in a tree-local MC-TAG

The derivation proceeds as follows; a3 is substituted in a2 as shown and then it is (reverse) adjoined to the b21Footnote 21 component of b2. The multi-component tree b2 is composed with a1 as follows. b21 is substituted at the NP node of a1 and the b22 is adjoined to a1 at the root node S of a1.

In a multi-component composition in the linguistic context there is usually a constraint between the two components (besides any co-indexing across the components). For example, in our example above there is a constraint (not shown in the Fig. A.5) that the foot node of b22 immediately dominates the NP root node of b21. This constraint is obviously satisfied in the derivation in Fig. A.5.

Schabes and Shieber (1993) introduced an alternate definition of derivation for LTAG in which they allowed multiple adjoining of certain kinds of elementary trees at the same node during the derivation. They showed that multiple adjunctions of the kind described above do not affect the generative capacity of LTAGs, i.e., they are weakly equivalent to LTAGs. This notion of multiple adjunctions can be extended to the derivations in tree-local MC-LTAG where we can allow multiple adjunctions of the NP auxiliary trees that correspond to the scrambled NP’s (see Appendix B). The generative power of tree-local MC-LTAG is not affected. We use such derivations for certain scrambling patterns. In Appendix B, we will show some derivations for sentences with scrambling using tree-local MCTAG.

1.2 Appendix B

Finally, we will describe some derivations for the scrambling (permutation) patterns using LTAGs or MC-LTAGs.Footnote 22 First consider

(1)

…dass

Hans

Peter

Marie

schwimmen

lassen

sah

  

Hans

Peter

Marie

swim

let

saw

  

NP1

NP2

NP3

V3

V2

V1

This canonical sequence and (2) below can be easily derived as shown in Figs. B.1 and B.2 above. Trees for V1, V2, and V3 are all LTAG trees (all nodes on the spine (except the lowest node) are labeled as VP nodes for convenience).

Fig. B.1
figure 6

Derivation for NP1 NP2 NP3 V3 V2 V1

Fig. B.2
figure 7

Derivation for NP3 NP2 NP1 V3 V2 V1

(2)

NP3

NP2

NP1

V3

V2

V1

In Fig. B.1, the V1 tree is adjoined at the root node of the V2 tree and then the derived tree is adjoined to the root node of the V3 tree. In Fig. B.2, the V1 tree is adjoined to the interior VP node of the V2 tree and the resulting tree is then adjoined into the interior VP node of the V3 tree. Finally in Fig. B.3, the V1 tree is adjoined to the interior VP node of the V2 tree and then the resulting tree is adjoined to the root node of the V3 tree.

Fig. B.3
figure 8

Derivation for NP2 NP1 NP3 V3 V2 V1

Finally, in Figs. B.4 and B.5 we show some derivations for sequences with four NP’s and four VP’s. Figure B.4 shows a derivation where the V1 tree is adjoined to the interior VP node of the V2 tree and the resulting tree is adjoined to the interior VP node of the V3 tree. Finally, the resulting tree is adjoined to the interior VP node of the V4 tree.

Fig. B.4
figure 9

Derivation for NP4 NP3 NP2 NP1 V4 V3 V2 V1

Fig. B.5
figure 10

Derivations for two sequences NP1 NP2 NP4 NP3 V4 V3 V2 V1 and NP4 NP1 NP2 NP3 V4 V3 V2 V1

In Fig. B.5 the V1 tree is adjoined to the root node of the V2 tree and the resulting tree is adjoined to the root node of the V3 tree. The V4 tree is a multi-component tree. It has two components. The lower component of the tree has V4 and its subject argument, NP4 which has been scrambled and hence shown with a trace. The upper component has the NP4 argument which is scrambled. There is also a constraint between the two components (not shown explicitly in Fig. B.5). The VP foot node of the upper component must dominate the VP root node of the lower component when the multicomponent is composed with an elementary host tree. The two components of the multi-component tree for V4 must combine with the same elementary tree, i.e., each component must target an elementary tree in order to compose with it, either by substitution or by adjoining. This assures the locality of the composition, as required by the definition of a Tree-Local Multi-Component TAG (TL-MCTAG), which is weakly equivalent to the standard TAG. The lower component of the V4 tree is “attached” (reverse adjoined in a senseFootnote 23) to the VP foot node of the V3 tree (as VP4 is the complement argument of V3) and the upper component of the V4 tree is adjoined to the root node of the V3 tree. Note that we have multiple adjunctions at the root node of the V3 tree.Footnote 24 Figure B.5 thus represents two possible sequences as indicated. It can be thought of as an underspecified representation of these two sequences.

Note that in Fig. B.5, the elementary ‘tree’ for V4 is in two parts (components), the lower and the upper part. The lower part has the trace for NP4 and the upper tree has the scrambled NP4 argument. There is an implied constraint (not shown here) that after these two components of the tree for V4 compose with the tree for V3 at the nodes VP (root node) and the frontier node labeled VP* respectively, these adjoining sites are in a dominating relation. Such constraints are a part of the Tree Local MCTAG composition and further these constraints are specified for each elementary object in a TL-MCTAG.

With 4 verbs and 4 NP’s there are 24 possible sequences, permuting the NP’s and keeping the VP’s in a fixed order. It can be shown that 22 of these sequences can be generated by Tree-local MCTAG and two sequences cannot be generated. These two sequences are

(1)

NP3

NP1

NP4

NP2

V4

V3

V2

V1

(2)

NP2

NP4

NP1

NP3

V4

V3

V2

V1

Thus there is a gap of two sequences in the set of the non-canonical transforms of the canonical sentence forms with three levels of recursion. For every depth of recursion beyond three or more embeddings gaps will continue to exist, for example, with four levels of embedding there will be six gaps, etc.

In a TAG or MC-TAG the composition is always binary. However, the elementary trees are not required to be binary. In the figures above the lower components are non-binary. If one imposes the binary form restriction on the elementary trees of TAG and MC-TAG then it can be shown that the first time the gaps appear is when we have four levels of embedding.

Rights and permissions

Reprints and permissions

Copyright information

© 2014 Springer International Publishing Switzerland

About this chapter

Cite this chapter

Joshi, A.K. (2014). Recursion in Language: Is It Indirectly Constrained?. In: Roeper, T., Speas, M. (eds) Recursion: Complexity in Cognition. Studies in Theoretical Psycholinguistics, vol 43. Springer, Cham. https://doi.org/10.1007/978-3-319-05086-7_7

Download citation

Publish with us

Policies and ethics