Skip to content

AI Is Not Trained It's Domesticated

Published:

We don’t teach Large Language Models (LLMs) — we evolve them. “Training” suggests we’re walking our models through a syllabus, but machine learning isn’t education. It’s selective pressure. Every choice we make when designing a learning algorithm — the data, the architecture, the loss function — shapes an artificial environment our models adapt to. Once an LLM’s base capabilities emerge, we redirect the pressure toward alignment — tuning its behavior to fit human preferences for helpfulness, honesty, and agreeableness.

This isn’t just a biological metaphor — it’s a literal echo of our first experiments with selective pressure: domestication.

AI Alignment is closer to domestication than instruction. Early humans didn’t teach wolves to fetch; they simply favored the ones that didn’t bite. Over generations, that simple preference for tameness reshaped the animal itself — floppy ears, shorter snouts, smaller teeth. None of those physical traits were the goal, but they emerged through accident. We lacked the biological understanding that tame behavior is intertwined with these physical traits.

We’re recreating the same evolutionary blind spot. Like those early human breeders, we’re selecting for behavior without understanding the underlying machinery. Modern AI alignment is a form of selective breeding: we reward outputs that sound kind, deferential, or inoffensive. But without a mechanistic grasp of how these traits arise, we can’t see what else we’re selecting for. What we call “AI Psychosis” may be just the tip of an iceberg of unintended adaptations. Just as we bred wolves into dogs, we may be domesticating cognition itself — without knowing what we’re changing.

Sources

  1. A. S. Wilkins, R. W. Wrangham, and W. Tecumseh Fitch, “The ‘Domestication Syndrome’ in Mammals: A Unified Explanation Based on Neural Crest Cell Behavior and Genetics,” Genetics, vol. 197, no. 3, pp. 795–808, Jul. 2014.
  2. C. Hansen Wheat, J. L. Fitzpatrick, B. Rogell, and H. Temrin, “Behavioural correlations of the domestication syndrome are decoupled in modern dog breeds,” Nature Communications, vol. 10, Art. ID 2422, Jun. 2019.
  3. M. Cheng, C. Lee, P. Khadpe, S. Yu, D. Han, and D. Jurafsky, “Sycophantic AI Decreases Prosocial Intentions and Promotes Dependence,” arXiv preprint, arXiv:2510.01395, Oct. 2025.
  4. M. Sharma, M. Tong, T. Korbak, D. Duvenaud, A. Askell, S. R. Bowman, N. Cheng, E. Durmus, Z. Hatfield-Dodds, S. R. Johnston, S. Kravec, T. Maxwell, S. McCandlish, K. Ndousse, O. Rausch, N. Schiefer, D. Yan, M. Zhang, and E. Perez, “Towards Understanding Sycophancy in Language Models,” in Proc. Int. Conf. Learning Representations (ICLR 2024), 2024.