Plain text has outlived every hype cycle. It has survived the GUI revolution, the database boom, and the reign of the API. Now, in the age of large language models, it’s quietly winning again. When I write or code these days, I keep my notes, rules, and small knowledge models in text. It isn’t nostalgia. It’s because both humans and machines can read them. No translation layers. No black boxes. Just meaning, expressed directly. This practice traces back to the original web. Tim Berners-Lee’s idea was simple: publish information as plain text that anyone - or anything - could inspect, link, and process. Hyperlinks stitched those fragments together, forming a web of knowledge. Today’s foundational models are trained on that corpus. In a sense, they are statistical compressions of its structure and regularities. Back in the day the web’s text wasn’t quite enough for machines to understand. So came RDF which reduced meaning to a minimal grammar: subject, predicate, object. It was a linguistic insight turned into data infrastructure. With that triple pattern you could describe the world in a format that was both legible and formal. A web of meaning, written in text. The next evolution came with JSON-LD and schema.org. Sites began including small application/ld+json blocks beside their prose. JSON remained human-readable; the @context mapped words to globally defined IRIs with unambiguous semantics. Search engines could suddenly read both the story and its structure - the words and the graph beneath them - and index the web not just by text, but by meaning. Now, enter large language models. Trained on that same web, they generate fluent prose that feels like understanding. Yet they highlight an old truth: natural language is ambiguous. A model can speak confidently about “Paris” without knowing whether it’s in France or Texas. RDF solved that decades ago - one plain-text IRI like https://lnkd.in/eBZjiTGt can anchor the word to a specific thing. So the challenge isn’t to replace structure with language; it’s to coordinate the two. We want to feed models text they can reason about, but with hooks that tether words to verifiable meaning. That’s why the plain-text pattern is resurfacing: prose for humans, identifiers and relations for machines, both living side by side. This, more than any breakthrough in model size or token length, may define the next phase of AI. The web taught us that linked plain text could scale to the world. The semantic web taught us how to describe that world formally. LLMs bring the narrative power. The future lies in fusing them - a web where every sentence can be both read and reasoned with. Triples were the first step on that journey. They began as a sparse language for meaning, evolved to sit quietly next to our words, and now return as the hidden scaffolding beneath machine intelligence. After thirty years, the hero of plain text has come home.
Tony Seale , again I applaud you for making it so transparent and easy to understand for the masses. Quote “The web taught us that linked plain text could scale to the world. The semantic web taught us how to describe that world formally. LLMs bring the narrative power. The future lies in fusing them - a web where every sentence can be both read and reasoned with.”
I do the same thing! :-) However I'm still searching for robust and easy to use way to embed RDF "mapping" and CSS simple declarations in markdown, without flooding it with too much "instructions" (that would complicate something that should be easy). The best would be a static site generator introducing their own "markdown+rdf" :-)
Yep! Which is why RDF-Turtle exists: reduce sentences to their core essence and turbocharge them using standardized identifiers: 1. References — hyperlinks 2. Typed Literals — dates, decimals, floats, booleans, etc. 3. Untyped Literals — optionally language-tagged RDF elegantly unveils natural language as a system of signs, syntax, and semantics for encoding and decoding information. It solves a complex problem in a “deceptively simple” way—its tortured journey was the only path to escape velocity and eventual mass adoption. TimBL realized early that freeing the world from application silos required a pathway back to natural language text and the file create, save, and share pattern. Without this work, a generation of computer users would have been severely hampered by lacking any knowledge of files and folders—an unimaginable disaster for literacy as a whole. See also: https://www.linkedin.com/pulse/file-create-save-share-paradigm-revisited-kingsley-uyi-idehen-phxze — The File Create, Save, and Share Paradigm — Revisited
Plain text is the cockroach of computing - every shiny new format tries to kill it, yet it keeps crawling back, still perfectly parseable with cat. 🤓
Let me grab the opportunity to plug my plaintext (org-mode) ontology authoring tool! Super convenient if you are an Emacs user. GPL license. Works well with AI assistant for text editing. https://github.com/johanwk/elot
I don’t think we quite solved the Paris problem with RDF, although it is an important foundation. Sometimes we want to be ultra precise eg factchecking media claims about knife crime or air quality. Other times a more expansive merging of levels of detail might make sense, eg capturing “Roman London” as a notion of London. Or in bibliography being able to switch fluidly between a FRBR view of a situation (entities for works, expressions, manifestations, items…) with a simpler and flatter version that may be closer to everyday terminology but awkward when trying to be precise. My hope for the new technologies is the help us navigate between these various legitimate levels of detail in ways pur old “everything has a URI” rhetoric tended to optimistically gloss over.
Tim Bray used to have "Intelligence is a text-based application." as a blog tagline, which I always quite liked
A method for writing deterministic logic directly in English https://github.com/LogicalContracts/LogicalEnglish
Poetry.
Developer Decision Intelligence\SQL\Python
2wNo translation layers. No black boxes. Just 𝐦𝐞𝐚𝐧𝐢𝐧𝐠, expressed directly. And then you're back to hermeneutics again. https://www.linkedin.com/pulse/hermeneutics-age-ai-jack-jansonius-4r5we/