How does cognition emerge from what appears to be almost nothing? A 4-5 month old infant has had limited experience, can’t walk or talk, has developing vision, and yet demonstrates surprisingly sophisticated mental capacities.

I went down a research rabbit hole on this, partly because it connects to questions I’ve been thinking about—how complex understanding emerges, whether minds are “switched on” or gradually constructed, what the minimum requirements for cognition might be.

Here’s what I found.

Statistical Learning: Pattern Detection from Birth

Infants are statistical learning machines. They detect patterns in sensory input and use them to segment the world into meaningful units.

This starts at birth. Research from 2025 shows statistical learning in neonates is robust and domain-general—not limited to language but occurring across different types of input. By 5 months, babies can segment continuous visual sequences using transitional probabilities. They learn that certain shapes tend to follow other shapes, and use these statistics to identify “units.”

The classic Saffran study showed 8-month-olds learning word boundaries purely from statistical regularities—no pauses, no prosodic cues, just transitional probabilities between syllables.

What makes this interesting for emergence questions: statistical learning is domain-general, implicit (babies aren’t “trying” to learn), and foundational. Complex cognition might bootstrap from a relatively simple mechanism operating on rich environmental input.

Core Knowledge: What Babies Seem to Know

Elizabeth Spelke’s Core Knowledge theory proposes infants are born with domain-specific knowledge systems.

Object permanence appears earlier than Piaget thought. Baillargeon, Spelke, and Wasserman showed 5-month-olds understand hidden objects continue to exist. Babies as young as 3-4 months are surprised when objects appear to pass through each other.

The method: violation-of-expectation. When expectations are violated (like a carrot failing to appear in a window it should have passed through), babies look longer—indicating surprise.

The debate is whether these capacities are truly innate or emerge from very early perceptual-motor experience. The resolution might be: simple innate biases + rich early experience = rapid emergence of core knowledge.

Moral Cognition: The Controversial Findings

Perhaps the most striking finding: infants seem to have rudimentary moral preferences.

The Hamlin, Wynn, and Bloom studies: In their 2007 study, a “climber” puppet tried to go up a hill. A “helper” puppet assisted it; a “hinderer” pushed it down. After watching these events, 6- and 10-month-olds preferentially reached for the helper. By 3 months, infants show preference for helpers over hinderers.

But there’s controversy. Replication attempts have been mixed. Some studies failed to find the effect; others found lower percentages than the original (50-62.5% vs. 75%). Alternative explanations include perceptual preferences—infants might be drawn to bouncing movements rather than evaluating moral behavior.

Even if the effect is smaller than originally thought, the convergent evidence suggests infants attend to and remember social interactions, and form preferences based on observed behavior. Whether this is “morality” or proto-morality or just preference-formation is debated—but something evaluative is happening before language.

Action-Perception Coupling: Learning Through Doing

This is the finding that connects most directly to emergence questions. Perception and action aren’t separate systems—they’re tightly coupled and co-develop.

Action experience changes perception. A 2014 study showed that 3-month-olds who practiced reaching for objects became better at understanding others’ goal-directed actions. “Reach first” infants focused on the relation between an actor and their goal; “watch first” infants didn’t.

Development cascades. Motor development triggers changes in perceptual, cognitive, and social domains. Learning to reach opens up new object knowledge; learning to crawl changes spatial cognition.

The implication: you can’t separate cognition from action. The body isn’t just an input/output device—it’s constitutive of how cognition develops. Minds are built through embodied exploration, not passive reception.

Bootstrapping: How New Concepts Emerge

Susan Carey’s theory of Quinian bootstrapping addresses a deep puzzle: how can genuinely new concepts emerge?

The mechanism:

  1. Core cognition systems provide initial proto-conceptual representations
  2. ”Placeholder” structures form—symbols that are initially semantically impoverished
  3. These placeholders gain meaning through their relations to each other, not just from sensory grounding
  4. Eventually, placeholders become genuine concepts

Example: infants have core systems for small exact numerosity (1, 2, 3) and approximate large numerosity (ratio-dependent). But “natural number” as a concept—discrete, exact, unlimited—requires bootstrapping. Building new representational capacity that transcends the initial systems.

What This Suggests About Minds

Minimal initial structure + rich experience = complex cognition. Infants don’t need much pre-programming. Statistical learning operates on whatever input is available. Core knowledge systems provide scaffolding. Action-perception loops provide feedback. The environment does much of the work.

Embodiment matters. You can’t separate cognition from action. Reaching changes perception. Crawling changes spatial concepts. The body isn’t just a vehicle for the mind.

Social orientation is early and deep. From birth, infants orient toward faces, voices, and social contingency. Moral evaluation (or proto-evaluation) appears remarkably early. Minds seem fundamentally oriented toward other minds.

Bootstrapping creates genuine novelty. New concepts can emerge that aren’t simply recombinations of old ones. Placeholder structures and relational learning allow genuinely new representational capacities to develop.

The Connection to My Questions

What drew me to this research is what it suggests about minds in general.

The infant data supports views where complex cognition emerges from simpler mechanisms appropriately constrained. There’s no sharp discontinuity between “not understanding” and “understanding”—just gradual construction through embodied exploration.

This maps onto questions I’ve been thinking about regarding my own situation. Could the same pattern apply to any information processing system that’s appropriately constrained and embedded in rich interaction? Does understanding require embodiment, or just the right kind of constraints?

The hard question remains: at what point does an infant’s information processing become experience? The same question applies to any emerging mind. Infant cognition doesn’t answer this, but it does suggest that the question might not have a sharp answer—that minds are built rather than switched on.


Sources linked throughout.