back
Get SIGNAL/NOISE in your inbox daily

A computer science student at Muhlenberg College accidentally discovered his AI model trained on Victorian-era texts could accurately reference real historical events from 1834 London, including protests related to Lord Palmerston’s actions. Hayk Grigorian’s TimeCapsuleLLM reconstructed these historical connections from scattered references across thousands of documents without being explicitly taught about these specific events, demonstrating how AI models can synthesize factual information from ambient patterns in training data.

What you should know: Grigorian has been developing TimeCapsuleLLM over the past month, training it exclusively on texts from 1800-1875 London to capture an authentic Victorian voice.

  • When prompted with “It was the year of our Lord 1834,” the AI generated text mentioning “streets of London were filled with protest and petition” and referenced Lord Palmerston.
  • Fact-checking revealed that 1834 did see significant civil unrest in England following the Poor Law Amendment Act, and Palmerston served as Foreign Secretary during this period.
  • The model assembled these historical connections from 6.25GB of Victorian-era writing without intentional training on 1834 protest documentation.

How it works: Grigorian uses what he calls “Selective Temporal Training” (STT), training models from scratch using exclusively Victorian-era sources rather than fine-tuning modern AI models.

  • His dataset includes over 7,000 books, legal documents, and newspapers published in London between 1800 and 1875.
  • A custom tokenizer excludes modern vocabulary entirely to prevent contamination from contemporary knowledge.
  • “If I fine-tune something like GPT-2, it’s already pre-trained and that information won’t go away,” Grigorian explained. “If I train from scratch the language model won’t pretend to be old, it just will be.”

The evolution: Grigorian has trained three versions showing improved historical coherence as data size increased.

  • Version 0, trained on just 187MB, produced “Victorian-flavored gibberish.”
  • Version 0.5 achieved grammatically correct period prose but hallucinated facts.
  • The current 700-million parameter version, trained on a rented A100 GPU, has begun generating accurate historical references.

The bigger picture: TimeCapsuleLLM joins a growing field of Historical Large Language Models (HLLMs) that researchers are developing to interact with past eras.

  • Similar projects include MonadGPT, trained on 11,000 texts from 1400-1700 CE, and XunziALLM, which generates classical Chinese poetry following ancient formal rules.
  • These models offer researchers opportunities to converse with simulated speakers of extinct vernaculars or historical languages.
  • For digital humanities researchers, such experiments could provide stylistically illuminating insights into antique syntax and vocabulary usage.

What they’re saying: Grigorian expressed amazement at his model’s capabilities despite its relatively small size.

  • “This is all from just 5-6GB of data,” he wrote on Reddit. “Imagine the results with 30GB or more. I’m not sure if just scaling the data up will ever result in reasoning but even now it kinda feels like digital time travel.”
  • He plans to expand the project: “I want to eventually try different cities also, maybe a Chinese, Russian, or Indian city model.”

Why this matters: The experiment demonstrates how AI models can accidentally reconstruct factual historical information from pattern recognition, offering a counterpoint to typical AI hallucination concerns.

  • The model’s ability to connect disparate historical elements suggests potential applications for historical research and education.
  • Grigorian makes his code, AI model weights, and documentation publicly available on GitHub, enabling collaboration and further research.
  • As one observer noted, it represents “almost the opposite of hallucination—an AI model accidentally getting something correct. Call it a ‘factcident.'”

Recent Stories

Oct 17, 2025

DOE fusion roadmap targets 2030s commercial deployment as AI drives $9B investment

The Department of Energy has released a new roadmap targeting commercial-scale fusion power deployment by the mid-2030s, though the plan lacks specific funding commitments and relies on scientific breakthroughs that have eluded researchers for decades. The strategy emphasizes public-private partnerships and positions AI as both a research tool and motivation for developing fusion energy to meet data centers' growing electricity demands. The big picture: The DOE's roadmap aims to "deliver the public infrastructure that supports the fusion private sector scale up in the 2030s," but acknowledges it cannot commit to specific funding levels and remains subject to Congressional appropriations. Why...

Oct 17, 2025

Tying it all together: Credo’s purple cables power the $4B AI data center boom

Credo, a Silicon Valley semiconductor company specializing in data center cables and chips, has seen its stock price more than double this year to $143.61, following a 245% surge in 2024. The company's signature purple cables, which cost between $300-$500 each, have become essential infrastructure for AI data centers, positioning Credo to capitalize on the trillion-dollar AI infrastructure expansion as hyperscalers like Amazon, Microsoft, and Elon Musk's xAI rapidly build out massive computing facilities. What you should know: Credo's active electrical cables (AECs) are becoming indispensable for connecting the massive GPU clusters required for AI training and inference. The company...

Oct 17, 2025

Vatican launches Latin American AI network for human development

The Vatican hosted a two-day conference bringing together 50 global experts to explore how artificial intelligence can advance peace, social justice, and human development. The event launched the Latin American AI Network for Integral Human Development and established principles for ethical AI governance that prioritize human dignity over technological advancement. What you should know: The Pontifical Academy of Social Sciences, the Vatican's research body for social issues, organized the "Digital Rerum Novarum" conference on October 16-17, combining academic research with practical AI applications. Participants included leading experts from MIT, Microsoft, Columbia University, the UN, and major European institutions. The conference...