Plain text wins again: A story of the web, RDF, and LLMs

View profile for Tony Seale

The Knowledge Graph Guy

Plain text has outlived every hype cycle. It has survived the GUI revolution, the database boom, and the reign of the API. Now, in the age of large language models, it’s quietly winning again. When I write or code these days, I keep my notes, rules, and small knowledge models in text. It isn’t nostalgia. It’s because both humans and machines can read them. No translation layers. No black boxes. Just meaning, expressed directly. This practice traces back to the original web. Tim Berners-Lee’s idea was simple: publish information as plain text that anyone - or anything - could inspect, link, and process. Hyperlinks stitched those fragments together, forming a web of knowledge. Today’s foundational models are trained on that corpus. In a sense, they are statistical compressions of its structure and regularities. Back in the day the web’s text wasn’t quite enough for machines to understand. So came RDF which reduced meaning to a minimal grammar: subject, predicate, object. It was a linguistic insight turned into data infrastructure. With that triple pattern you could describe the world in a format that was both legible and formal. A web of meaning, written in text. The next evolution came with JSON-LD and schema.org. Sites began including small application/ld+json blocks beside their prose. JSON remained human-readable; the @context mapped words to globally defined IRIs with unambiguous semantics. Search engines could suddenly read both the story and its structure - the words and the graph beneath them - and index the web not just by text, but by meaning. Now, enter large language models. Trained on that same web, they generate fluent prose that feels like understanding. Yet they highlight an old truth: natural language is ambiguous. A model can speak confidently about “Paris” without knowing whether it’s in France or Texas. RDF solved that decades ago - one plain-text IRI like https://lnkd.in/eBZjiTGt can anchor the word to a specific thing. So the challenge isn’t to replace structure with language; it’s to coordinate the two. We want to feed models text they can reason about, but with hooks that tether words to verifiable meaning. That’s why the plain-text pattern is resurfacing: prose for humans, identifiers and relations for machines, both living side by side. This, more than any breakthrough in model size or token length, may define the next phase of AI. The web taught us that linked plain text could scale to the world. The semantic web taught us how to describe that world formally. LLMs bring the narrative power. The future lies in fusing them - a web where every sentence can be both read and reasoned with. Triples were the first step on that journey. They began as a sparse language for meaning, evolved to sit quietly next to our words, and now return as the hidden scaffolding beneath machine intelligence. After thirty years, the hero of plain text has come home.

  • diagram, text, whiteboard
Jack Jansonius

Developer Decision Intelligence\SQL\Python

2w

No translation layers. No black boxes. Just 𝐦𝐞𝐚𝐧𝐢𝐧𝐠, expressed directly. And then you're back to hermeneutics again.  https://www.linkedin.com/pulse/hermeneutics-age-ai-jack-jansonius-4r5we/

Louis Hendriks

Founder & Initiator of Global Value Web

2w

Tony Seale , again I applaud you for making it so transparent and easy to understand for the masses. Quote “The web taught us that linked plain text could scale to the world. The semantic web taught us how to describe that world formally. LLMs bring the narrative power. The future lies in fusing them - a web where every sentence can be both read and reasoned with.”

I do the same thing! :-) However I'm still searching for robust and easy to use way to embed RDF "mapping" and CSS simple declarations in markdown, without flooding it with too much "instructions" (that would complicate something that should be easy). The best would be a static site generator introducing their own "markdown+rdf" :-)

Kingsley Uyi Idehen

Founder & CEO at OpenLink Software | Driving GenAI-Based AI Agents | Harmonizing Disparate Data Spaces (Databases, Knowledge Bases/Graphs, and File System Documents)

2w

Yep! Which is why RDF-Turtle exists: reduce sentences to their core essence and turbocharge them using standardized identifiers: 1. References — hyperlinks 2. Typed Literals — dates, decimals, floats, booleans, etc. 3. Untyped Literals — optionally language-tagged RDF elegantly unveils natural language as a system of signs, syntax, and semantics for encoding and decoding information. It solves a complex problem in a “deceptively simple” way—its tortured journey was the only path to escape velocity and eventual mass adoption. TimBL realized early that freeing the world from application silos required a pathway back to natural language text and the file create, save, and share pattern. Without this work, a generation of computer users would have been severely hampered by lacking any knowledge of files and folders—an unimaginable disaster for literacy as a whole. See also: https://www.linkedin.com/pulse/file-create-save-share-paradigm-revisited-kingsley-uyi-idehen-phxze — The File Create, Save, and Share Paradigm — Revisited

  • No alternative text description for this image
Rareș I.

Product Expert, focused on AI, Software, Analytics, Engineering. Bottleneck Remover.

2w

Plain text is the cockroach of computing - every shiny new format tries to kill it, yet it keeps crawling back, still perfectly parseable with cat. 🤓

Johan Wilhelm Klüwer

Principal Specialist at Det Norske Veritas

1w

Let me grab the opportunity to plug my plaintext (org-mode) ontology authoring tool! Super convenient if you are an Emacs user. GPL license. Works well with AI assistant for text editing. https://github.com/johanwk/elot

Dan Brickley

Data Standards Engineering

1w

I don’t think we quite solved the Paris problem with RDF, although it is an important foundation. Sometimes we want to be ultra precise eg factchecking media claims about knife crime or air quality. Other times a more expansive merging of levels of detail might make sense, eg capturing “Roman London” as a notion of London. Or in bibliography being able to switch fluidly between a FRBR view of a situation (entities for works, expressions, manifestations, items…) with a simpler and flatter version that may be closer to everyday terminology but awkward when trying to be precise. My hope for the new technologies is the help us navigate between these various legitimate levels of detail in ways pur old “everything has a URI” rhetoric tended to optimistically gloss over.

Dan Brickley

Data Standards Engineering

1w

Tim Bray used to have "Intelligence is a text-based application." as a blog tagline, which I always quite liked

Andrew Noble

Founder & Systems Builder | Creating tensegrity for your practice: a resilient, efficient, and interconnected tech structure.

1w
Niklas Emegård

AI & Knowledge Architect, Länsförsäkringar AB

1w

Poetry.

See more comments

To view or add a comment, sign in

Explore content categories