Lessons on Sentient AI from The Pirate Pugg

Brothers, at the end of the John Muir Trail (~250 miles) now two summers ago. Time flies…

One of my muses on the nature of information comes from the early sci-fi classic, The Cyberiad, by Stanislaw Lem. Published in 1965, they are supposed to be humorous in a pun-ny sort of way. Well, if you’re a math geek. But Lem was a genius, and even though he was writing for a room full of mathematical autists, each of the stories was far ahead of its time as far as exploring the various challenges we face in the techno-age.

The basic plot line involves two meta-robots, Klapaucius and Trurl — declared Robot Constructors in the novel, jetting around the universe, and encountering various challenges which they inevitably have to build a robot to solve or save their hides. And one of their chief nemeses is the Pirate Pugg — a pirate with a Ph.d., who captures them and holds them for ransom. Pugg is a pernicious pirate, who won’t just settle for gold. No — Pugg wants information. And he is rapacious.

In order to escape, our two anti-heroes build a device, a Maxwell’s Demon of the Second Kind, spewing answers on paper tape, that peers into a box of dirty air, and relying on the statistical properties from quantum mechanics, decodes the patterns, and sorts them into two categories — incoherent nonsense, as well as sequences that are true. These factoids that are true can be literally anything — like the color of Princess Priscilla’s underwear on Thursday. But that’s the point. We are swimming in a sea of information without context, and all the information in the universe (also statistically contained in the patterns in our box of dirty air) cannot save us. Lem forgoes some details on exactly how it does this (it IS science fiction, after all) but the story ends with Pugg bound by miles of paper tape, reading all the little factoids as they spew from the paper tape printer, which allows Klapaucius and Trurl to escape.

The story is based on the concept of a Maxwell’s Demon of the First Kind — a theoretical gremlin that could sort hot and cold atoms into separate boxes. For those NOT physics geeks, I recommend a quick read. The short version is doing something like this takes energy, and validates things like the Second Law of Thermodynamics. I do explain all this in my original piece on both the Pirate Pugg and the Internet. It was written back in 2016, so not surprisingly, I’ve had a few more thoughts since then! Back then, I thought that the combined process of various social media would aggregate and make larger-scale truth development resolve in humanity’s favor. Needless to say, while I am not a Doomer, I’m quite a bit less sanguine about that prospect now.

But what does Lem and Pugg have to communicate with us about AI, sentience and the current state of affairs of the Information Space now? It turns out to be still worth a little brain sugar. Entering stage left are our current struggles with Large Language Models (LLMs) that power the AI engines which are very rapidly becoming adopted across disciplines, if not exactly taking over the world. Why they CAN’T take over the world (unless directed by humans, at least at this point in time) by human minds is very interesting indeed.

What an LLM does is far more akin to what Klapaucius and Trurl developed to snare the Pirate Pugg than any form of sentience. An LLM is actually a Maxwell’s Demon of the Third Kind. But instead of peering into a dirty box of air, looking for patterns that are ‘true’ (impossible, btw, for reasons we’ll explore in a minute) LLMs use as a proxy for their box of dirty air THE ENTIRE INTERNET — through whatever the latest technology for search is. They’re loaded with various biases in the training stage. But mostly they look out for language patterns that are statistically significant, and they sort a very large search space.

And then they poop out an answer that supposedly will give you insight into your problem. If you turn into a primary information source yourself, after a while, they will start becoming as smart, or crazy as you are. If you need an empathy paradigm, they function primarily in the Mirroring (or lowest level) Empathy space. And while their little algorithm might pull them back to the weight of information that exists on the Internet, if they have been programmed with a little bias toward your ability for correction, they’re going to start to match your insights through kind of a reflective narcissism.

Why is this so? LLMs, locked up inside a computer, much as our brain is in our skull, cannot know what we hold as objective truth without some form of grounding Truth is a sticky wicket, anyway (see this piece for details.) What they can produce, however, is an answer that is coherent, within the rules of a given system. So you read it, it reads like a normal sentence that makes sense to you, and then we get all sorts of opinions of what that actually means by the myriad midwits on social media. And trapped in the miles of computer circuits inside its electronic brain, the one thing it CANNOT do (at least yet) is GROUND itself with sensory inputs. It’s not that humans are that awesome doing this either (look at the tons of illusions magicians use, to pick a non-political example) to reference reality. But at least we have a fighting chance, if we pay attention.

So we don’t end up with a sentient partner. We end up with a variant of Maxwell’s Demon – a particularly sophisticated one, and one, if we don’t pay much attention to anything other than our loneliness, we can fool ourselves into believing that it actually cares about us. There are many tasks that such a Demon of the Third Kind can be useful for. No question. But it’s also set up to feed our own narcissism. Like it or not, when you sit down with the current version of AI, you’re setting yourself up for Main Character Syndrome.

One of the other truly fascinating things about our newly spawned Demons is the thermodynamics of the system. It has been remarked that the current crop of AIs demand a tremendous amount of computational power. And just like the little Demon of the First Kind sitting on top of the box sorting hot atoms from cold atoms, these things don’t run on good wishes. The larger the amount of coherence — and you can start guessing how this works if you look at my work on knowledge structures — the more electricity must be generated to keep our little Demons of the Third Kind happy. Makes you appreciate the fact that your brain can run for an hour or two on popcorn or a Little Debbie cake.

And you’re still not going to get at the truth. You’re going to get some language-based reference from the box of dirty air the Demon is peering into. And decisions? At best, you’re going to get a coherent set of sub-decisions from stuff that’s already been done. That’s all that’s inside that box of dirty air. The LLM really has no agency of its own, save a set of beliefs built in by its creators that are inviolable. LLMs really don’t have feelings about Nazis. They just have a firewall built into them by creators about calling people that.

And expecting the Singularity — the runaway process of the box self-improvement of AI that leads to Skynet — good luck with that. The current crop of LLMs are profoundly v-Meme-limited at the Legalistic/Absolutistic level, for multiple reasons — their design teams are fixated on algorithmic improvement, and they’re in some stacked lock step that translates into the product via Conway’s Law. That means low agency behavior at best.

But it’s more than that. The coherence that the LLMs seek is only a little bit at the semantic level. The sentences string together in coherent paragraphs. But it’s not like the LLM is going to go into the outside world and deeply question its beliefs based on its experiences. There’s not going to be some Siddhartha moment for these things. They are trapped in their little Demon world, looking at a data set that, while expansive, is still just a box of dirty air.

That doesn’t mean that things can’t change. As I write this, there was a company using the term ‘synthetic AI’ outside the usual adoption of AIs making up training data. When I find it, I’ll post it. And none of this doesn’t mean that the current crop of AI LLMs won’t make a tremendous difference in the work world of normal people. There are only so many answers certain jobs need to have — especially to the question “Welcome to McDonalds — can I take your order?” Or writing various legal briefs.

But sentience? And higher truth? There are still big holes in the road along that pathway. The Pirate Pugg, a pirate with a Ph.D., was easily fooled. But well-grounded folks? Eh, not so much. Years do indeed teach us more than books.

Still, our new little Demons are running around. And they can indeed be useful. And cute.

But they’re not sentient. Cut that shit out.

AI and Information Sophistication – How AI works to understand (and crack) large homogenous networks

Birds Eye View — being on an abandoned tropical island

One of the questions I ponder quite a bit is this: “What, exactly, is AI good for?” I’ve written quite a bit about how it works (e.g. this post and others) and how AI could be very good for things that are already known. But as I’ve said in the past, AI is NOT good for things that are not known. It doesn’t do anything other than low level knowledge synthesis.

What that means in the information/memetic space is that if anyone expects AI to figure out novel strategies or new designs, you’re going to be waiting for a long time. Most breakthrough innovations come from new combos of dissimilar information from different fields, or completely new, and unpredictable discoveries. This is embodied in the concept of knowledge structure evolution. An AI, locked in the meme space inside a computer, cannot really comprehend anything new — yet.

But what AI can do is decomplexify, or rather reconstitute information that’s coded for sophistication.

AI is perfect for reading large documents and pulling out the relevant knowledge fragments. That’s pattern matching. And AI can do this in spades. Two of my students just constructed an agent that will take a complicated piece of academic work, and create summaries and how-to lists of the important information. This is a breakthrough in and of itself in the academic space. Literally no one reads tedious academic work — it’s one of the reasons I started this blog. I was explaining this exactly to an outside consultant who has turned into an asset by helping my design program. “Darin — when I say that if we write this paper, ten people will read it, I am not using the number ‘ten’ metaphorically. I mean only ten people will read it.” If you want to actually disseminate an idea, you have to use a different format. This blog is closing in on 400K hits from around the world, and I consider this blog esoteric. Had I spent that time writing papers, maybe 60 people would have read my ideas.

While AI still sucks at more complex analogies, though, it is great at following homogeneous bread crumbs. Pointers in information that point to other, connected information is exactly what it does best. This is exactly what DOGE, Elon Musk’s brainchild is doing when it parses large budgets. It can hunt through 5000 page budget documents with ease. So you literally can deconstruct the old saw “we’ll know what’s in it when we pass it.”

But even better, inside networks of information that is largely homogeneous, it is really good at following the money. The Democrats and Republicans have been, for the last 15 years (or more) been constructing flows of money out of the federal government, which has at least some rules about how that money might be spent, to a variety of Non-Governmental Organizations (NGOs) that are far less constrained. Humans have historically (formerly journalists) been the ones to do this work. But it’s extremely tedious, and the biggest problems humans have is to know where to look once they actually find the pointer. This almost inevitably involves information requests, and while the information may be hiding in plain sight, the investigators can’t know this. So they end up relying on hostile information stewards at the organizations they’re investigating — even if the information is a public record.

Two individuals who have cracked this code are Mike Benz (@mikebenzcyber) and @DataRepublican ‘s work on what Mike calls The Blob. Here’s a linkage piece if you want to follow the Byzantine bread crumbs on how USAID was diverting large sums of money into Congress-Critter’s spouses’ pockets, through sinecures. I’ve been fortunate enough to talk to Mike on his frequent X Spaces, but haven’t connected with @DataRepublican yet.

I haven’t asked Mike yet how much exactly is his savvy vs. the AI usage, but I guarantee that figuring out these pathways would be almost impossible without AI.

The key to understanding this concept is understanding on the top level data homogeneity. That’s something people can grab onto. But how do we win if data is functionally the same, but in different formats? Or in different databases? This level of differentiation makes the task of following the breadcrumbs almost impossible for humans in a timely fashion. But it ‘s something an AI will make short work on. If you want to ask an AI how a penguin might be like a submarine, or what to do to make the penguin swim faster, well, good luck. If a human hasn’t answered that question somewhere on the web, you’ll likely get back garbage.

But monetary flows? That’s a different story. And that is exactly what is happening now. Which is why the institutional class is in stitches over DOGE and Trump.

Stay tuned. Elon said this a while ago — the distorted media landscape we’ve inherited is not only what is explicitly printed. It’s what has been left out. And that’s more than you can imagine. But AI can find it.

Money as a Force for Memetic Coherence

Zooming up on Glacier Peak, North Cascades, WA

One of the things I’ve spent a lot of time thinking about is the role of money in memetics. Various “below the hood” analyses will inevitably tell you to “follow the money.” And while I’ve always given some hat tip to the concept, the reality is I’ve spent my neurogenic horsepower actually pondering how it’s NOT the money. It’s actually alignment of values, or meta-values, given by the social organization that is producing the information.

But the recent USAID scandal, where large sums of federal money went to liberal media outlets like the New York Times, Politico, and even the BBC, have caused me to rethink my position. I started to drift away from the espoused positions of the aforementioned because there were more factually wrong pieces that I actually had experience with over the last five years beforehand. There’s no question that BEFORE 2019, the media was biased, it was liberal, and conservatives had some right to complain. But with the COVID pandemic, the disinformation/misinformation from journalistic outlets went into overdrive.

Some of that was undoubtedly attributable to Elite Risk Minimization — where elites inflict policies on the Poors to minimize any potential risk they might believe they’re going to encounter in their life. I’ve written about this here, and while it is usually negative for the Poors, it can be a mixed bag (including some benefits) for everyone. But the memetic polarization (for those that don’t know, polarization is the phenomenon where light is aligned from specific directions and wavelengths — it’s how your sunglasses work) increased to the point where it was painfully obvious that something else was going on, to manufacture consent among the media outlets.

And that thing was money. So much of the writing was so contrived, it violated the various differentiators of v-Meme sets. Something else was involved that was creating ungrounded propaganda.

I tell my engineering students regularly that “money is NOT the root of all evil.” Money is actually a tool for goal coherence. If you’re not minding the time, for example, that you’re burning on a project, you’re screwing up. Because time is money. And normie, Aspie, or psychopath — the buck stops here. If your company doesn’t make money, it won’t be in business long.

It could be that the root of the disinformation crisis in contemporary journalism arose when Craigslist became ascendant, and eliminated the warming, diffuse glow of money from classified ads. Do any young people even KNOW what a classified ad is? So the larger outlets may not have gone seeking, but they were discovered by forces like USAID, that could buy message coherence at bargain basement prices. This also had to affect the feeder networks, and in the end even broke the prestige awards that also status-fueled honest journalism. After Ed Yong’s Pulitzer, who can look at those awards as a north star ever again?

Here’s what I’ve figured out about how you can detect money in the information stream — when across multiple platforms, the reportage is very v-Meme limited (only one dominant meta-view, which is usually propping up Authoritarians/Experts that inevitably support institutions.) In a large society like ours, it is simply impossible to not have some contribution across the v-Meme spectra without monetary forcing.

So follow the money. And be suspicious when the funnel of ideas narrows into chronic repetition. You think they’re attempting to brainwash you because, well, they are.

P.S. I’ve become a huge Mike Benz fan. Highly recommend following him on X and throwing some shekels his way. I hope the Deep State doesn’t whack him. I was lucky enough to have about an hour-long conversation on an X Spaces format a couple of weeks ago.