This AI thinks it’s the 1800s

An interesting thing about contemporary artificial intelligence models, specifically large language models (LLMs): They can only output text based on whatâs in their training dataset. Models, including ChatGPT and Claude, are âtrainedâ on large databases of text. The models, when asked a question, statistically create a response by calculating, one word at a time, what the most likely next word should be. A consequence of this is that LLMs canât output text about scientific breakthroughs that have yet to happen, because thereâs no existing literature about those breakthroughs. The best an AI could do is repeat predictions written by researchers, or synthesize those predictions.
Adam Mastroianni, writing in his newsletter Experimental History, put this elegantly: âIf you booted up a super-smart AI in ancient Greece, fed it all human knowledge, and asked it how to land on the moon, it would respond, âYou canât land on the moon. The moon is a god floating in the sky.’âÂ
Itâs an interesting thought experiment. What if you intentionally limited the training data? Could you create an AI system that responds as though itâs from a period in the past? What could that reveal about the psychology or everyday experiences of the people from that era?
Thatâs exactly what Hayk Grigorian, a student at Muhlenberg College in Allentown, Pennsylvania, had in mind when he created TimeCapsuleLLM. This experimental AI system was trained entirely on texts from 19th century London. The current release is based on 90 gigabytes of text files originally published in the city of London between 1800 and 1875.Â
This is, to be clear, very much a hobby project. The sample-generated text on GitHub isnât consistently coherent, though Ars Technica did report that it has correctly surfaced names and events from the 1800s. When prompted to continue the sentence âIt was the year of our Lord 1834,â the model recounted a protest: âthe streets of London were filled with protest and petition,â going on to mention the policies of Lord Palmerston, who was the foreign secretary at the time.Â
Itâs an interesting experiment, but could such a thing actually be useful? Potentially.
An opinion piece published by the Proceedings of the National Academy of Sciences of the United States of America (PNAS) by collaborators including Michael E. W. Varnum, a professor of psychology from the Department of Psychology at Arizona State University, is an interesting read. It proposes that models like this could be a way to study psychology outside a modern context. The paper refers to such AI models as Historical Large Language Models, or HLLMs for short, and states that psychology researchers could use them to study the thinking of people in past civilizations.Â
âIn principle, responses from these faux individuals can reflect the psychology of past societies, allowing for a more robust and interdisciplinary science of human nature,â the paper says. âResearchers might, for example, compare the cooperative tendencies of Vikings, ancient Romans, and early modern Japanese in economic games. Or they could explore attitudes about gender roles that were typical among ancient Persians or medieval Europeans.â
Itâs an interesting idea, though the paper does acknowledge this could be tricky.Â
âAll LLMs are a product of their training corpora, and HLLMs face challenges in terms of sampling, given that surviving historical texts are likely not representative samples of people who lived in a particular period,â the paper admits, stating that historical texts tend to be written by elites, not everyday people. âAs a result, it could be hard to generalize from these models.âÂ
And there are other things to keep in mind. Research from Ghent University in Belgium shows that the ideology of the people who work on an LLM shows up in the text those models generate. Thereâs every reason to suspect the same problem will apply to LLMs designed to reflect past cultures.Â
So there are difficulties. Only time will tell if such models end up being used in psychological research, or remain the domain of hobbyists.Â



