Reduplication, the copying operation employed in natural language morphophonology (e.g., Ilokano pluralization, [kaldíŋ] ‘goat’; [kal-kaldíŋ] ‘goats’; Hayes and Abad, 1989, p. 357), creates repetition structures within surface word forms. Though reduplication and surface repetitions have been extensively studied, two questions remain unresolved. First, what are the possible natural language word forms with reduplication? Secondly, how can reduplication be characterized and learned in a unified way with other (morpho)phonological regularities? This dissertation approaches these questions through three studies that combine experimental and computational methods with previous typological studies and phonological theory.
Chapter 2 offers experimental evidence on how human learners tend to generalize reduplicative patterns. Two series of artificial grammar learning experiments using the poverty of the stimulus paradigm (Wilson, 2006) yield the following results. First, human learners rapidly extrapolate and generalize reduplicative hypotheses to novel forms after being exposed to only a small number of familiarized forms. Their generalizations align with coarse-grained phonological abstractions characterizable by the vocabulary of prosody (e.g., syllables, feet, prosodic words), supporting the core claims of Prosodic Morphology (McCarthy and Prince, 1986, et seq.). Moreover, there are strong correlations between participants’ spontaneous responses and naturally occurring reduplicative patterns. The universally preferred patterns followed typological trends, while variations in individually learned grammar reflected the variations attested in natural languages. Lastly, patterns whose empirical status has been controversial appear in participants’ spontaneous responses, offering novel learning-based evidence to support Base-Reduplicant Correspondence Theory (McCarthy and Prince, 1995) as a possible characterization of human learners’ hypothesis space.
Chapter 3 examines the abstract properties of surface repetition structures from a formal language-theoretic view. We revisit the Chomsky Hierarchy, which offers highly abstract characterizations of linguistic processes. Reduplicative patterns with unbounded copying impose a challenge: a model within the classical Chomsky Hierarchy that adequately captures unbounded copying is expected to generate unattested palindrome patterns (e.g., pseudo-Ilokano with reversal, [ŋidlak-kaldíŋ]), which does not match empirical observations. Therefore, we advocate for another language class that cross-cuts the well-known classes in the classical Chomsky Hierarchy. We introduce Finite-state Buffered Machines, an augmentation to the regular class with a primitive copying operation. This is achieved by adding compact memory allocation machinery and an unbounded memory buffer with queue-like storage. We survey the properties of the resulting language class and find this refinement better matches the language typology without sacrificing mathematical rigor.
Chapter 4 proposes a morphophonological learner that extends an expectation-driven maximum entropy lexicon learner proposed by Wang and Hayes (resubmitted) with a component that deals with reduplication learning. Given that the empirical results in Chapter 2 support Base-Reduplicant Correspondence Theory (McCarthy and Prince, 1995), the learner adopts constraints proposed by this theory and learns “hidden structures” (Tesar and Smolensky, 1998), i.e. the prosodic templates. We demonstrate that in this way, the learner can learn different types of reduplication-phonology interactions and capture the population-level results observed in the learning experiments.