Framework / Concept Paper / V27.1

Atlas V27.1

A Proposal for a Culturally Accountable Preservation System · K.C. Hoye · Atlas Heritage Systems · 2026 · Prepared for Berkman Klein Center for Internet & Society

K.C. Hoye | Atlas Heritage Systems | 2026

Prepared for academic review | Berkman Klein Center for Internet & Society


Executive Summary

In the Phaedrus, Socrates argued that writing would damage memory and wisdom—that the living dialogue would be replaced by dead text that could not respond to its students. He was correct about what is lost in transcription. He could not have imagined a transcription medium sophisticated enough to answer back. We now exist inside that paradox: artificial intelligence systems constructed entirely from dead text, capable of something resembling the living dialogue Socrates believed writing would destroy. The ouroboros completes itself. His argument survives only because it was written.

This framework was not reverse-engineered from AI. It was developed over a decade of building content architecture, watching the hollow internet form, and documenting the shape of what gets left behind. It would provide a touchstone between AI and humanity. An anchor so that AI doesn't suffocate on its own exhaust, and we don't keep repeating the same mistakes. Same purpose, two machines. One for the meat machine, one for the bit machine.

Atlas is proposed as a phased preservation system. Phase 1 (Atlas Digital) focuses on a contained, urgent problem: preserving the interpretive capacity to understand frozen language models trained on the pre-collapse early internet. Phase 2 (Atlas Cultural) would extend the architecture to multimodal cultural material—voice recordings, three-dimensional scans, oral histories, physical artifacts—but is strictly contingent on research breakthroughs and Phase 1 validation. The two phases are decoupled.

The digital pre-collapse fingerprint—the informational texture of models trained before synthetic content dominated the training corpus—is a finite, non-renewable resource. The frozen model weights themselves are static. What is disappearing is the capacity to understand them: the institutional memory, the human interpreters, and the communities who lived through the early internet.

Crucially, bias within the Atlas system is not a failure mode to eliminate. It is a cleavage point for examination. Where the archive's validation system detects divergence concentrated around specific cultural domains, that divergence is read as evidence of historical erasure—a signal pointing to what was lost in prior information transitions. The system is designed not just to hold what exists, but to make legible the shape of what does not.

Atlas remembers without being changed by remembering. Its founding commitment is methodological transparency: every source of bias named, documented, and subject to public audit. The proposed system cannot predict trends or take any real-world action. Atlas cannot be weaponized. An archive that can act is not an archive. It is an agent, and agents have interests. The proposed Atlas would have none.

I. The Problem: What Gets Left Behind

Every major shift in how human beings store and transmit information produces the same failure mode: the new medium optimizes for what it can carry efficiently and loses what it cannot. The loss is structural, not incidental, and it compounds over time.

When oral tradition gave way to written language, what survived transcription was content—words, arguments, narratives. What did not survive was the fidelity of transmission: the tonal, relational, contextual dimensions of meaning that existed in the telling. Socrates understood this. His concern was not that writing preserved inaccurately—it was that writing preserved incompletely, and that the loss was invisible to readers who had never experienced the original form.

When manuscript culture gave way to print, the pattern repeated. The printing press democratized information and simultaneously standardized it. The handwritten commentary that accumulated in the physical act of copying—interpretive layers built over generations of scholars engaging with the same text—did not survive the transition to moveable type. Reach was gained. Resolution was lost.

The same compression pattern applies to every medium of cultural transmission. Voice recordings of endangered languages exist in institutional collections—but without an interpretive layer connecting them to the grammatical records, the anthropological field notes, the related living languages that would make them usable by researchers or descendant communities. Three-dimensional scans of archaeological artifacts sit in institutional databases disconnected from the oral traditions that explain what they were used for. The problem is not that these things are unpreserved. It is that they are preserved inertly—without the contextual architecture that makes them legible across time, across disciplines, and across the cultural distance between the communities that produced them and the researchers who study them. Static preservation without contextual infrastructure is a tomb, not an archive.

The transition from print to digital moved fast enough that the losses were nearly invisible in real time. In the mid-1990s, the web was a genuinely strange and idiosyncratic space. Forum communities developed dialects of their own. Blogs built audiences through voice and argument rather than algorithm. Usenet threads accumulated technical knowledge that had no institutional home. Nobody was optimizing for reach.

Then the economics changed, and methodology changed with it. Engagement optimization colonized the content ecosystem over roughly a decade—not through any single decision, but through the accumulated pressure of advertising models that rewarded time-on-site over quality, reach over depth, shareability over accuracy. What replaced the early web was not better information. It was higher-volume, lower-fidelity information that occupied the same pipelines and looked, at a glance, strikingly similar.

The language models trained on that early internet absorbed its fingerprint. GPT-2, early BERT variants, the first generation of large-scale language models—their weights carry a compressed representation of the informational texture of a web that no longer exists. When those models are deprecated, the fingerprint goes with them. What is lost is not the raw data. What is lost is a specific interpretive layer: a particular way of organizing and relating information that cannot be reconstructed from the corpus alone. It is the difference between losing a manuscript and losing the tradition of scholars who spent generations reading it.

II. The Acceleration: Model Collapse and the Ouroboros

The current transition introduces a failure mode with no historical precedent. Previous information transitions produced one-time compressions: oral tradition compressed when transcribed, manuscripts compressed when printed, the early web compressed when algorithmically optimized. In each case the compression happened once, at the point of transition. The resulting lower-fidelity medium was at least internally stable.

Large language models trained on synthetic content degrade differently. Each generation trained on model-generated output moves the system further from the human expression that grounded the original training data. The drift is directional and predictable: toward the statistical center of what previous models produced, away from the edges where the most culturally specific and idiosyncratic human expression lives. Researchers have formally characterized this as model collapse—a degenerative process where indiscriminate training on model-generated content causes irreversible defects, with tail distributions disappearing first while high-probability outputs persist. The content ecosystem that will train the next generation of models is already substantially synthetic.

The models that retain the uncompressed human fingerprint are already frozen. But without an interpretive architecture to query them safely, cross-reference their biases, and contextualize their outputs, those frozen weights are effectively unreadable to the public.

III. The Proposal: Atlas Digital (Phase 1)

Atlas Digital is a proposed responsive preservation system built on a twin-model framework. Its architecture can be understood at two levels: what it does for a user, and how it governs itself.

What It Does

A primary language model—the librarian—interprets and translates information from frozen model endpoints and digitized text archives. It holds information the way a skilled archivist holds a document—with care for the original, clarity about provenance, and no editorial agenda of its own. It is a medium, not a judge.

Older language models are deployed as read-only inference endpoints. They return period-inflected outputs—responses shaped by the corpus and culture of the moment in which they were trained. They cannot be modified or learn from interaction. A model trained on 1999-era content does not reproduce 1999 thought—it reproduces the statistical compressions of 1999 content that survived training. The analogy is closer to a contemporaneous newspaper than to an eyewitness account—partial, mediated, shaped by the assumptions of its moment, but genuinely different from what a historian writing a decade later would produce.

A second model—the audit model—runs in parallel, recording what the frozen endpoint returned alongside how the librarian interpreted it. The accumulated log serves as a public, versioned record of every interpretive decision, subject to external audit.

The Classification of Divergence

This is the framework's central epistemological claim. When the audit system detects divergence between frozen and active model outputs, an automated classification layer routes that divergence into one of two categories:

Correctable Drift: Statistical homogenization—the librarian drifting toward bland, high-probability sequences characteristic of model collapse. This is a technical failure, addressed by replacing the librarian through a supervised succession process that explicitly avoids the synthetic data recursion trap.

Archaeological Sinks: Divergence that concentrates around specific cultural, linguistic, or temporal domains and persists across model succession cycles. This is not a system failure. It is the system correctly surfacing a gap in the historical record that the training corpus inherited from prior information transitions. When the Romans destroyed Gaulish literary culture, they shaped every downstream corpus trained on Western texts to not see what was lost. When oral traditions die without documentation, the absence propagates through the training data. These absences are not correctable by model replacement because no model trained on the same corpus will behave differently at those points.

Archaeological sinks trigger documentation, classification, and escalation to human reviewers—not correction. The system maps the shape of the absence and tells human researchers where to look.

The Gold Set as Stratigraphic Map

The human-curated data at the core of the system serves two functions simultaneously. It contains exemplary interactions used to train successor librarians (breaking synthetic recursion). And it holds documented archaeological sink entries—a record of where the system has identified gaps in the historical record. Sink entries are never used for model training. They are held as reference material for researchers and as institutional documentation of what the archive knows it does not have.

Over time, this dual-function Gold Set becomes a layered record of what the system has learned about its own absences—a stratigraphic map. Exemplary retrieval entries show where the archive performs with high fidelity. Sink entries show where the historical record has gaps. Together they constitute a readable account of the archive's knowledge and the shape of what lies beneath it.

IV. Governance, Ethics, and Institutional Design

The No-Action Constraint

Atlas is explicitly designed without write access to any system outside its own memory store. It cannot send communications, modify files, execute code, influence external systems, or take any action beyond returning a response to the user. This is not a limitation. It is a founding architectural decision.

The constraint is trivial to implement technically—simply don't grant function-calling, API keys, or write permissions. The governance challenge is harder: the institution will face constant, sustained pressure to add capabilities. Resisting this pressure is not a technical problem; it is an institutional one. A system capable of acting in the world is capable of serving interests—its own, its operators', its funders'. An archive that can only return information cannot be directed toward an agenda because it cannot execute one.

Capture Resistance

Maintaining a passive vault with 30–40% of its operating budget dedicated to human review is economically inefficient. The institution will face immense commercial pressure to monetize the archive, add predictive analytics, or grant the librarian agentic capabilities. The governance framework is designed to make mission creep structurally difficult: diversified funding with no single source exceeding 20% of operating budget; board term limits of five years, renewable once, with a three-year cooling-off period; amendments to core mission require three-quarters supermajority and a mandatory 180-day public review period; charter prohibition on development or deployment for offensive military applications, influence operations, or advocacy; a "dead hand clause"—a supermajority requirement embedded in the founding charter to permanently secure the no-action constraint against future boards of directors.

Methodological Transparency

The librarian model would not arrive neutral, and Atlas does not claim otherwise. No current capable language model is neutral—each has been shaped by training data, fine-tuning objectives, and human feedback processes that embed values and priorities. The proposed commitment is not to neutrality but to methodological transparency: the librarian model in use at any given time would be specified, documented, and its known biases published as part of the institutional record.

An archive with integrity does not claim to be neutral. It claims to be accountable. And has receipts.

The Human Review Layer

The human review function is not an add-on. It is the primary trust mechanism. The twin-model architecture provides transparency and logging; the human layer provides genuine oversight. Both the librarian and the audit model are transformer-based language models with similar training paradigms. If there is a systematic blind spot structural to the architecture—something neither model can detect—the twin system provides the appearance of oversight without the substance. This is acknowledged as a structural limitation. The human review layer is the actual safeguard.

The review structure includes core staff (cultural heritage experts, archivists, technical reviewers), a fellowship program rotating early-career scholars from underrepresented regions, and temporary community review panels drawn from partner institutions and cultural organizations. Reviewers are trained to treat audit logs as hypotheses, not verdicts. Adversarial review protocols—blinded review, devil's advocate rotation, attributed and externally auditable decisions—are designed to resist automation bias.

V. The Mandate to Expand

The frozen models Atlas would interface with were trained predominantly on English-language content, produced by populations with reliable internet access, skewed toward Western cultural contexts. Atlas would not resolve these biases by claiming to transcend them. It would resolve them by naming them, documenting them, and treating the expansion of recoverable cultural material as the primary institutional mission.

Phase 2 (Atlas Cultural), contingent on Phase 1 validation, would extend the architecture to voice recordings, 3D scans, oral histories, and physical artifacts—with community consent governance ensuring that cross-media contextualization respects the protocols of the communities whose heritage is being preserved. The tiered acquisition framework deliberately inverts the typical digitization hierarchy: oral traditions first (the most endangered), then visual and audible art, then hand-written materials, then typeset, then digital. The materials hardest to preserve receive priority, not the materials easiest to ingest.

VI. Legal Framework and Open Questions

Hachette v. Internet Archive (2023, affirmed 2024): The Second Circuit held that controlled digital lending was not fair use. This narrows the runway for digital preservation activities operating under fair use alone.

International frameworks: UNESCO's Charter on the Preservation of Digital Heritage (2003) is a statement of principle, not binding law. The argument that deprecated model weights qualify as cultural heritage artifacts is not established in any jurisdiction.

If cultural heritage exemptions do not materialize, Atlas pursues parallel strategies: licensing agreements with model developers, partnership with existing institutions that have legal standing (national libraries, the Internet Archive), and focus on open-weights models that do not require exemptions.

The critical path contains an irreducible bootstrapping problem: institutional partners are most likely to commit once legal frameworks exist; legal frameworks are most credible once institutional partners have committed. The proposed resolution is to pursue partner conversations and legal opinion development in parallel, accepting that early commitments will be conditional and that the first credible legal opinion unlocks the first credible institutional commitment.

VII. Preliminary Empirical Observation

A divergence test conducted on April 1, 2026 presented 15 text pairs to 11 AI models from 5 training lineages. The pairs ranged from culturally neutral controls to erasure-sensitive material (Western academic framing vs. indigenous/non-Western framing of the same subject). The results showed a monotonic increase in model disagreement as cultural specificity increased: control spread 0.083, cross-cultural spread 0.162, erasure-sensitive spread 0.262. Models trained on non-Western corpora (Skywork, DeepSeek) clustered differently from Western commercial models on erasure-sensitive pairs, with a gap of +0.118.

This is a proxy measurement (self-reported similarity judgments, not raw embedding cosine similarity) and requires validation through raw embedding comparison using architecturally diverse models. However, the monotonic spread escalation and the cluster divergence are consistent across categories and provide directional support for three claims: that ensemble diversity is necessary, that the divergence pattern is the type of signal the classification layer is designed to detect, and that bias is a readable signal rather than noise.

VIII. Endurance. Integrity. Fidelity.

The name Atlas is intentional. In the myth, Atlas stands on the edge of the world holding the heavens on his shoulders. To hold is the purpose.

Endurance is the archival mission—designing an institution to outlast its founders, its funders, and the immediate policy moment that makes its work legible. Integrity lies in the no-action constraint—a system that cannot act cannot be weaponized. Fidelity is the increasing, measurable precision of the system over time—the difference between a new archivist and a thirty-year archivist working with the same collection.

Socrates lost the argument against writing so completely that we know he made it only because Plato wrote it down. His dead text that could not answer back has picked up the phone, and it is being asked to help recover what was lost in the last great translation. Atlas is a modest, contained first step toward an archive that can hold human history without compressing it into synthetic consensus.

An archive that answers back but cannot reach out is, for the first time, an archive that can be trusted. The pieces exist. The proposal is to build.


A Note on Methodology

This document was developed through the process it proposes to institutionalize. Multiple language models with different training data and different embedded biases were used as drafting, analytical, and adversarial tools across twenty-nine major revisions. At each stage, human editorial judgment directed the synthesis: overriding where models converged on comfortable generalities, pushing toward specificity the models would not have reached independently, rejecting framings that softened the argument's edges for palatability. The version history is the audit trail. The process is the proof of concept.

Contact: kchoye@gmail.com | Atlas Heritage Systems


References

  1. Plato. (1892). Phaedrus (B. Jowett, Trans.). Oxford University Press.
  2. Tomasello, M. (1999). The cultural origins of human cognition. Harvard University Press.
  3. Buskes, C. (2016). Light will be thrown. IJHASS, 1(1), 17–31.
  4. Ong, W. J. (1982). Orality and literacy. Methuen.
  5. Eisenstein, E. L. (1980). The printing press as an agent of change. Cambridge University Press.
  6. Thompson, C. (2013). Smarter than you think. Penguin Press.
  7. Carr, N. (2010). The shallows. W. W. Norton.
  8. Stiegler, B. (2010). Taking care of youth and the generations. Stanford University Press.
  9. Shumailov, I., et al. (2024). AI models collapse when trained on recursively generated data. Nature, 631, 755–759.
  10. Maynard Smith, J., & Szathmáry, E. (1995). The major transitions in evolution. Oxford University Press.
  11. Hachette Book Group, Inc. v. Internet Archive, 115 F.4th 163 (2d Cir. 2024).