Just renewed my membership for The Kid Should See This, a genuine internet treasure.
Just renewed my membership for The Kid Should See This, a genuine internet treasure.
“I developed a simulation model of a random process that’s highly dependent on input variables and I ran it once with no sensitivity analysis”
If you’re thinking this isn’t a great approach, and certainly not good enough for decision-grade analysis, you’d be right.
And yet this is exactly what many papers and benchmarks do when it comes to large language models. People increasingly use LLMs to perform a coding task, or classify some data, and report the one-off result as if it’s a reliable and meaningful measurement.
A single run. No checking sensitivity to inputs. No reproducibility. No robustness. At best, the person reading the analysis can only hope they’ll get a similar output next time.
Even with a low temperature setting, LLMs can be extremely sensitive to data ordering and prompt structure. Suppose you use an LLM to analyse narrative reports from different regions. It concludes region A is overall 40% positive about an issue, and region B is 30% positive.
That might seem like a meaningful difference, but how confident would you be of getting the same numbers if the LLM was run again with some tweaks to the inputs? How much of that 10% difference is genuine - and how much is down to the quirks of a particular LLM run?
Take a recent report from Anthropic, which surveyed 81,000 Claude users about AI perceptions. It’s been getting a lot of attention, and it does include some nicely visualised interactive comparisons, such as this one looking at sentiment across different regions:
But how real are these differences? For example, are Brazilian users genuinely more positive about AI than UK ones?
It’s worth breaking down the different sorts of uncertainty lurking behind the headline numbers. Some of these are well-known, and some perhaps less so:
Sampling uncertainty: If around 3000 users are included from two countries, then each estimate comes with around ±2% margin of error (because more or less positive people may have been included by chance).
Representativeness uncertainty: More fundamentally, the sample consists of opt-in Claude users rather than a representative cross-section. The study looks at representativeness relative to user base (i.e. usage patterns, region, tier) rather than national populations. So if user composition differs between countries, differences may reflect who uses Claude rather than underlying attitudes.
Measurement uncertainty: Claude won’t necessarily give an identical output each time. When I ran the sentiment classification prompt provided in the methods on a couple of hundred of the accompanying quotes, there was a 79% match between scores on two separate runs (i.e. two agents each running Claude Opus 4.6). The rest of the scores were slightly higher or lower the second time. This was just for short clear quotes, so the inconsistency may be different for full interview transcripts.
Definition uncertainty: Beyond measurement uncertainty, there’s also the issue of what exactly counts as “positive” sentiment. For example, I found Claude classifies this survey comment as being positive about AI: “I don’t want AI to do my work so I can wash dishes, I want AI to wash dishes so I can do intellectual work” – despite it paraphrasing an oft-cited AI criticism. Because transcripts were analysed one-by-one, the LLM has no consistent dataset-level frame of reference for ‘positive’ or ‘negative’.
Translation uncertainty: LLMs can judge different languages in different ways. For example, when I translated those couple of hundred quotes into Portuguese and re-ran the classification prompt, the scores were more tightly centred around the mean. And re-running produced lower agreement than before, with a ~70% match for individual scores.
Binary uncertainty: The map above shows % of people with net positive sentiment (i.e. 5 or more out of 7 on the sentiment scale). But converting continuous scales to yes/no categorisations removes information. If one country has scores that cluster just above the threshold of 5 and another just below, they could have very different net positivity even though individual scores are similar.
It’s important to understand global perceptions of AI. But given the above, we should remember that headline numbers might not be as certain as they seem – and that there are well-established statistical tools out there to help us quantify this uncertainty.
Contention: Whether you find our current “AI moment” at least somewhat intellectually stimulating depends largely (if not entirely) on your comfort with metaphors.
I know this will bother some of my more rabid anti-AI readers, but for all many misgivings about the role of AI in society at present, I am firmly and unapologetically curious about this new technology. More than curious, really. I am intellectually excited by the profound questions that AI poses regarding what we mean by intelligence, cognition, learning, abstraction, understanding—the list is long. What’s more, AI demands that we—and by we, I very much mean we humans—really grapple with the metaphors we use to make sense of ourselves, and those we use to help us understand these strange new tools we’ve created.
Recently, I wrote a mini-defense of the metaphor of “stochastic parrots” as an accurate and useful metaphorical description of the core function of generative AI in the form of large-language models. That’s a claim I’m making about generative AI as it exists today. But what about where things are headed? What metaphor should we use as our “north star”—hey look, another metaphor—to guide where this technology should be going? That’s the question I want to explore this week, using an interesting pre-print article published just a few weeks ago as my intellectual foil.
The article: AI Must Embrace Specialization via Superhuman Adaptable Intelligence, lead authored by Judah Goldfelder (currently a PhD student in the “Creative Machines Lab” at Columbia University), with Philip Wyder, Yann LeCun, and Ravid Schwartz-Ziv listed as co-authors—longtime Cognitive Resonance readers may remember my “AI insider spills secrets!” essay that featured LeCun’s many criticisms of LLMs as currently constructed. In any event, this new article is very readable, not overly technical, and offers up a very useful landscape graphic of the many metaphors and definitions hovering around the aspirations of the AI research field at present. To wit (and please pinch to zoom on mobile, sorry it’s so small):
Stefon voice: Adaptive generalists, cognitive mirrors, economic engines, this landscape HAS IT ALL! Jokes aside, I think it’s very helpful summary of the competing “metaphorical visions” within the AI research world about what exactly the goal of the field should be oriented around. On the Y axis, you’ve got the split between those who want to see AI models cash out in the real world on specific tasks (performance) versus those who want to develop AI models that can adapt to new, unforeseen circumstances (“learn” or adaptability).1 Then on the X axis, you’ve got the scope of tasks that AI models might reasonably address, ranging from literally anything to the somewhat more narrowly defined set of things humans are capable of doing.
Reader, I debated spending the next 1,200 words unpacking each of these “north star” definitions of AI, but I’ll forego that for now—if you’re insatiably curious about any of them, hit me up in the comments, and I’ll do my best to provide a pithy summary. Instead, I want to focus here just on “superhuman adaptable intelligence,” the new term offered by Goldfelder et al as to where the field of AI ought to be headed. More specifically, I want to unpack their definition of “SAI” because I both agree with the general thrust of their argument and think it rests on a somewhat myopic understanding of human behaviors and activities.
So what is “superhuman adaptable intelligence”? It’s “intelligence that can learn to exceed humans at anything important we can do, and that can fill in the skill gaps where humans are incapable.” Note my emphasis on learning: At heart, the argument here is that we should give up on trying to create a singular and generally superintelligent (godlike?) AI model, and instead focus on developing more specialized AI systems that can adapt and improve on specific tasks over time. I won’t go into all the technical details here, but to summarize the argument in a sentence, “the AI that folds our proteins should not be the AI that folds our laundry!”
I’m nodding, I’m nodding. And I particularly appreciate the brief burn of large-language models nestled late in the paper: “Homogeneity kills research. Autoregressive large-language models have become the dominant architecture in the state-of-the-art ‘general’ AI space…[but] they have many flaws; their errors diverge exponentially with prediction length, [making] long-horizon interaction brittle.” Here, Goldfelder et al are echoing something Gary Marcus has been arguing for several years now, namely, that for all their impressive linguistic capabilities and the billions (trillions?) being spent on them, LLMs are largely a detour on the road to developing more advanced AI models and systems.
So on the whole, I’m a fan of the “direction of travel” for future AI research that’s offered here. At the same time, and me being me, I am compelled to highlight one of the more amusing passages contained in this paper that, to my reading, is indicative of a significant blind spot that plagues the AI research community, if not all of cognitive science. It arises in the context of their argument that human beings are not actually “generally intelligent” or all that cognitively capable, and it (yet again) involves chess:
Magnus Carlsen is widely regarded as the greatest chess player of all time, and as such represents the pinnacle of human adaptation when it comes to playing chess. But this begs the question: Is Magnus actually any good at chess?
When compared with the best computers, the answer is clearly no. Even more damning is that with modern day computers, creating a program that plays chess at a much higher level than Magnus is not particularly difficult.
Our perception of his ability is colored by the limitations of humanity. Humans in general are bad at chess; Magnus is much better than most humans. The conclusion, that Magnus is good at chess, is a perfect illustration of our own human centric biases. Magnus Carlsen is not objectively good at chess, he is good at chess with respect to human performance levels.
By any objective metric, playing chess at a much higher level is not difficult from a computational perspective, but it is something that humans are incapable of.
It is undeniably true that chess engines today are more powerful than any humans at calculating chess moves, and thus can generate “solutions” that humans are incapable of. But invoking the “objective metric” of computational prowess as so-called proof that Magnus Carlsen is “not objectively good at chess” is a category error—humans play chess because it’s fun and interesting, it’s the sport of human thinking. As is true with all human games, nothing of any real importance happens as a consequence of winning, the satisfaction instead arises from the experience itself.
What’s more, and I say this as someone who’s spent way too many hours watching Daniel King analyze games on his Power Play Chess channel (he’s great), the capabilities of chess engines are often treated as a semi-interesting sideshow to the far more absorbing endeavor of exploring human-versus-human competition. When commenting on a game, King will often observe, “the computer thinks this is the best move but good luck to any human at seeing that”—and then he’ll go right back to his thoughtful and very human analysis of the game at hand. That’s what we humans like to do, be curious about our fellow humans.2
In my view, AI as a research field, to say nothing of all of cognitive science, suffers from obsessive commitment to what my late-in-life mentor Jim March described as the “consequentialist paradigm,” the view that “our beliefs require that action be justified in terms of ambition for its consequences.” Sometimes that perspective is warranted, no doubt—but not always. Sometimes, we pursue things not because of an outcome we hope to achieve but because the pursuit itself generates meaning, in ways that are entirely subjective, often ephemeral, and we might even say irrational. Why does Don Quixote tilt at windmills, March asks? Because “he embraces the foolishness of obligatory action. Justification for knight-errantry lies not in anticipation of effectiveness but in an enthusiasm for the pointless heroics of a good life. The celebration of life lies in the pleasures of pursuing the demands of duty.”3
Artists, poets, musicians, those who spends time dwelling in the depths of the humanities—and please do note the root of that word—they all naturally intuit this, I think. There is a reason that so many in the artistic community are horrified by AI, and it’s not solely because of the misappropriation of human labor and creativity (though this must not be overlooked). It’s because they see AI as part of a continued technological erosion of the celebration of life that arises from poetic pursuits. The pursuit of artificial intelligence, whether superhuman or adaptable or specialized or whatever else, could be part of the humanities, broadly construed, but only if researchers can resist slapping metrics on everything and declaring that all that matters is what’s measured. Life is not a data set!
I recur to March once more to close us out, and please do take a few minutes to ruminate on this:
We have grown to accept the modern idea that human actions, institutions, and traditions are justified by expectations of their consequences; and we have come to view optimism without hope as an unfortunate surrender to Pollyanna. But education unconditionally celebrates life. It is an arbitrary assertion of optimism. It echoes an ancient conception by which I do what I do because that is what is appropriate to my nature; and my nature is not simply a tautological summary of what I do but an understanding of the essence of my destiny, an interpretation of my history, and an assertion of my humanity.
Speaking of art and assertions of humanity, several months ago Tomáš Baránek, a publisher in the Czech Republic, reached out to me in hopes of translating my Large Language Mistake essay into Czech—there are few things more flattering to a writer than someone wanting to share it with an audience in another language (unfortunately, The Verge’s legal team put the kibosh on that plan).
Sadly, Baránek’s mother Vlasta Baránková recently passed away, but through our correspondence he shared with me her book of gorgeous and haunting artwork—she was an accomplished artist and illustrator of many childrens’ books, including some I’m sure I read as a child. Tomáš also shared with me that she had to maneuver around communist-era restrictions in order to publish her work. As he observes, “this is how real artists and creative people live and die. They cannot help themselves, they enjoy the creativity and its painfulness.”
Indeed. And you will see that creative beauty and pain reflected in the work of Vlasta Baránková, who’s work I share here with permission. She may be gone, but her exploration of the human spirit remains with us. I grieve with you, Tomáš, but I am so grateful to know of your mother and the contributions she made to our shared humanity.
The performance-versus-learning question is one that educators grapple with regularly, of course, and it helps explain why so many get irritated by certain forms of standardized testing.
Just to nerd out on chess a bit more, there’s multiple variants of the game that further test human cognitive flexibility (and often proved harder for AI-driven engines to dominate). Here’s one fun example of such a game involving, yes, Magnus Carlsen.
I have always enjoyed the act of typing words and seeing them come up on screen. While my favorite word processor of all time might be WordPerfect (here), I've used almost all of them. These programs were what sold me on the entire value proposition of computers. They were like typewriters, which I had used in school, except easier in every single way. You could delete things. You could move paragraphs around. It felt like cheating, and I loved it.
As time has gone up what makes up a "document" in word processing has increased in complexity. This grew as word processors moved on from being proxies for typewriters and into something closer to a publishing suite. In the beginning programs like WordPerfect, WordStar, MultiMate, etc had flat binary files with proprietary formatting codes embedded in there.
When word processors were just proxies for typewriters, this made a lot of sense. But as Microsoft Word took off in popularity and quickly established itself as the dominant word processor, we saw the rise of the .doc file format. This was an exponential increase in complexity from what came before, which made sense because suddenly word processors were becoming "everything tools" — not just typing, but layout, images, revision tracking, embedded objects, and whatever else Microsoft could cram in there.
At its base the .doc is a Compound File Binary Format, which is effectively just a FAT file system with the file broken into sectors that are chained together with a File Allocation Table.

It's an interesting design. A normal file system would end up with sort of a mess of files to try and contain everything that the .doc has, but if you store all of that inside of a simplified file system contained within one file then you could optimize for performance and reduced the overhead that comes with storing separate objects in a flat file. It also optimizes writes, because you don't need to rewrite the entire file when you add an object and it keeps it simple to keep revision history. But from a user perspective, they're "just" dealing with a single file. (Reference)
The .doc exploded and quickly became the default file format for humanity's written output. School papers, office memos, résumés, the Great American Novel your uncle was definitely going to finish — all .doc files. But there was a problem with these files.
They would become corrupted all of the goddamn time.
Remember, these were critical documents traveling from spinning rust drives on machines that crashed constantly compared to modern computers, often copied to floppy disks or later to cheap thumb drives you got from random vendor giveaways at conferences, and then carried to other computers in backpacks and coat pockets. The entire workflow had the structural integrity of a sandwich bag full of soup.
Your hard drive filesysystem -> your .doc file (which can get corrupted as a file) -> containers a mini filesystem (which can ALSO get corrupted internally) -> manages your actual document contentSo when Word was saving your critical file, it was actually doing a bunch of different operations. It was:
These weren't atomic operations so it was super easy in an era when computers constantly crashed or had problems to end up in a situation where some structures were updated and others weren't. Compared to like a .txt file where you would either get the old version or a truncated new version. You might lose content, but you almost never ended up with an unreadable file. With .doc as someone doing like helpdesk IT, you constantly ended up with people that had just corrupted unreadable files.
And here's the part that really twisted the knife: the longer you worked on the same file, the more important that file likely was. But Word didn't clean up after itself. As a .doc accumulated images, tracked changes, and revision history, the internal structure grew more complex and the file got larger. But even when you deleted content from the document, the data wasn't actually removed from the file. It was marked as free space internally but left sitting there, like furniture you moved to the curb that nobody ever picked up.
The file bloated. The internal fragmentation worsened. And the probability of corruption increased in direct proportion to how much you cared about the contents.
Users had to be trained both to save the file often (as AutoRecover wasn't reliable enough) and to periodically "Save As" a new file to force Word to write a clean version from scratch. This was the digital equivalent of being told that your car works fine, you just need to rebuild the engine every 500 miles as routine maintenance.
The end result was that Microsoft Word quickly developed a reputation among technical people as horrible to work with. Not because it was a bad word processor — it was actually quite good at the word processing part — but because when a user showed up at the Help Desk with tears in their eyes, the tools I had to help them were mostly useless.
I could scan the raw file for text patterns, which often pulled out the content, but without formatting it wasn't really a recovered file — it was more like finding your belongings scattered across a field after a tornado. Technically your stuff, but not in any useful arrangement. Sometimes you could rebuild the FAT or try alternative directory entries to recover slightly older versions. But in general, if the .doc encountered a structural error, the thing was toast and your work was gone forever.
This led to a never-ending series of helpdesk sessions where I had to explain to people that yes, I understood they had worked on this file for months, but it was gone and nobody could help them. I became a grief counselor who happened to know about filesystems. Thankfully, people quickly learned to obsessively copy their files to multiple locations with different names — thesis_final.doc, thesis_final_v2.doc, thesis_FINAL_FINAL_REAL.doc — but this required getting burned at least once, which is sort of like saying you learned your car's brakes didn't work by driving into a bus.
So around 2007 we see the shift from .doc to .docx, which introduces a lot of hard lessons from the problems of .doc. First, it's just a bundle, specifically a ZIP archive.
my_document.docx (renamed to .zip)
├── [Content_Types].xml
├── _rels/
│ └── .rels
├── word/
│ ├── document.xml ← the actual text content
│ ├── styles.xml ← formatting/styles
│ ├── fontTable.xml
│ ├── settings.xml
│ └── media/
│ ├── image1.png ← embedded images
│ └── image2.jpg
└── docProps/
├── app.xml
└── core.xml ← metadataNow in theory, this is great. Your content is human-readable XML. Your images are just image files. If something goes wrong, you can rename the file to .zip, extract it, and at least recover your text by opening document.xml in Notepad. The days of staring at an opaque binary blob and praying were supposed to be over.
However, in practice, something terrible happened. Microsoft somehow managed to produce the worst XML to ever exist in human history.
Let me lay down the scope of this complexity, because I have never seen anything like it in my life.
Here is the standards website for ECMA-376. Now you know you are in trouble when you see a 4 part download that looks like the following:
If you download Part 1, you are given the following:

Now if you open that PDF, get ready for it. It's a 5039 page PDF.

I have never conceived of something this complicated. It's also functionally unreadable, and I say this as someone who has, on multiple occasions in his life, read a car repair manual cover to cover because I didn't have anything else to do. I once read the Haynes manual for a 1994 Honda Civic like it was a beach novel. This is not that. This is what happens when a standards committee gets a catering budget and no deadline.

There was an accusation at the time that Microsoft was making OOXML deliberately more complicated than it needed to be — that the goal was to claim it was an "open standard" while making the standard so incomprehensibly vast that it would take a heroic effort for anyone else to implement it. I think this is unquestionably true. LibreOffice has a great blog post on it that includes this striking comparison:
The difference in complexity between the document.xml and content.xml files is striking when you compare their lengths: the content.xml file has 6,802 lines, while the document.xml file has 60,245 lines, compared to a text document of 5,566 lines.
So the difference between ODF format and the OOXML format results in a exponentially less complicated XML file. Either you could do the incredible amount of work to become compatible with this nightmarish specification or you could effectively find yourself cut out of the entire word processing ecosystem.
Now without question this was done by Microsoft in order to have their cake and eat it too. They would be able to tell regulators and customers that this wasn't a proprietary format and that nobody was locked into the Microsoft Office ecosystem for the production of documents, which had started to become a concern among non-US countries that now all of their government documents and records were effectively locked into using Microsoft. However the somewhat ironic thing is it ended up not mattering that much because soon the only desktop application that would matter is the browser.
The file formats of word processors were their own problems, but more fundamentally the nature of how people consumed content was changing. Desktop based applications became less and less important post 2010 and users got increasingly more frustrated with the incredibly clunky way of working with Microsoft Word and all traditional files with emailing them back and forth endlessly or working with file shares.
So while .docx was a superior format from the perspective of "opening the file and it becoming corrupted", it also was fundamentally incompatible with the smartphone era. Even though you could open these files, soon the expectation was that whatever content you wanted people to consume should be viewable through a browser.
As "working for a software company" went from being a niche profession to being something that seemingly everyone you met did, the defacto platform for issues, tracking progress, discussions, etc moved to GitHub. This was where I (and many others) first encountered Markdown and started using it on a regular basis.
John Gruber, co-creator of Markdown, has a great breakdown of "standard" Markdown and then there are specific flavors that have branched off over time. You can see that here. The important part though is: it lets you very quickly generate webpages that work on every browser on the planet with almost no memorization and (for the most part) the same thing works in GitHub, on Slack, in Confluence, etc. You no longer had to ponder whether the person you were sending to had the right license to see the thing you were writing in the correct format.
This combined with the rise of Google Workspace with Google Docs, Slides, etc meant your technical staff were having conversations through Markdown pages and your less technical staff were operating entirely in the cloud. Google was better than Microsoft at the sort of stuff Word had always been used for, which is tracking revisions, handling feedback, sharing securely, etc. It had a small subset of the total features but as we all learned, nobody knew about the more advanced features of Word anyway.
By 2015 the writing was on the wall. Companies stopped giving me an Office license by default, switching them to "you can request a license". This, to anyone who has ever worked for a large company, is the kiss of death. If I cannot be certain that you can successfully open the file I'm working on, there is absolutely no point in writing it inside of that platform. Combine that with the corporate death of email and replacing it with Slack/Teams, the entire workflow died without a lot of fanfare.
Then with the rise of LLMs and their use (perhaps overuse) of Markdown, we've reached peak .md. Markdown is the format of our help docs, many of our websites are generated exclusively from Markdown. It's now the most common format that I write anything in. This was originally written in Markdown inside of Vim.
There's a lot of reasons why I think Markdown ended up winning, in no small part because it solved a real problem in an easy to understand way. Writing HTML is miserable and overkill for most tasks, this removed the need to do that and your output was consumable in a universal and highly performant way that required nothing of your users except access to a web browser.
But I also think it demonstrates an interesting lesson about formats. .doc and .docx along with ODF are pretty highly specialized things designed to handle the complexity of what modern word processing can do. LibreOffice lets you do some pretty incredible things that cover a huge range of possible needs.

Markdown doesn't do most of what those formats do. You can't set margins. You can't do columns. You can't embed a pivot table or track changes or add a watermark that says DRAFT across every page in 45-degree gray Calibri. Markdown doesn't even have a native way to change the font color.
And none of that mattered, because it turns out most writing isn't about any of those things. Most writing is about getting words down in a structure that makes sense, and then getting those words in front of other people. Markdown does that with less friction than anything else ever created. You can learn it in ten minutes, write it in any text editor on any device, read the source file without rendering it, diff it in version control, and convert it to virtually any output format.
The files are plain text. They will outlive every application that currently renders them. They don't belong to any company. They can't become corrupted in any meaningful way — the worst thing that can happen to a Markdown file is you lose some characters, and even then the rest of the file is fine. After decades of nursing .doc files like they were delicate flowers that you had to transport home strapped to your car roof, the idea of a format that simply cannot structurally fail is not just convenient. It's a kind of liberation.
I think about this sometimes when I'm writing in Vim at midnight, just me and a blinking cursor and a plain text file that will still be readable when I'm dead. No filesystem-within-a-filesystem. No sector allocation tables. No 5,039-page specification. Just words, a few hash marks, and never having to think about it again.
I lurk in a Facebook group called "Photos of Evansville, Indiana." In it, photographers of various skill levels share photos taken around our city.
Two days ago, a member shared this photo taken across the street from Reitz Memorial High School:
Image Credit: Ted Byrom, via the "Photos of Evansville, Indiana" Facebook group
It's terrible, and I love it.
I could totally see it as a pop art painting in our local museum. Art teachers would ask the students:
I like how the hand sanitizer sits in the foreground and immediately catches your eye as if it's the subject of the photo. Then you realize that it's out-of-focus, and move up to the stop sign. But that's... also not an interesting subject. From there, the composition draws you to the Road Closed sign, which isn't very interesting either. The photo is almost interesting, but lacks a clear subject. It only leaves you confused.
Lose the hand sanitizer and add a girl walking by with an boldly-colored umbrella, and it would work just fine. Or keep the hand sanitizer, add an N95 mask, and post it during COVID as a commentary. Five years ago, the stop sign and construction barrier could have been meaningful:
"Life has stopped, and all roads to happiness are blocked. I feel stagnant, and all I can do is keep washing my hands and hoping to survive it"
Obviously, this is just someone's snapshot of a view that he liked, but couldn't convey. He accidentally managed a composition that fell into the uncanny valley between "Just someone's snapshot" and "Art that's trying to say something", and it's hilarious.

I recently came across the following joke map online:

Surprisingly, and probably unintentionally, there’s a grain of historical truth to this “Megachusetts”. Massachusetts’ original colonial charter specified northern and southern borders, and that the colony should extend all the way from the Atlantic Ocean to the Pacific Ocean, or rather: “…from the Atlantick and Westerne Sea and Ocean on the East Parte, to the South Sea on the West Parte;” The grant wasn’t quite the line depicted in the meme, but it was a line across the continent nevertheless.
Massachusetts wasn’t the only New England colony to get such a “sea to sea” grant. Connecticut’s colonial charter of 1662 granted the young colony all English territory “…from the said Narrogancett Bay on the East to the South Sea on the West parte, with the Islands thereunto adioyneinge…”
Besides the creative spellings of “Narragansett” and “adjoining” (from before standardized English spelling), this line gave Connecticut a basis to claim to its own narrow strip of entire continent, again hypothetically reaching the Pacific ocean. And like any reasonable government of the era, the colonial governments of Connecticut and Massachusetts kept on claiming this land as long as they feasibly could.
Several colonies had western claims along these lines. New York and Virginia claimed the most western land, but North Carolina, South Carolina and Georgia all had claims, too. Many of these seven state claims overlapped. The southern states in particular actually had access to land past the Appalachians, and several of them tried to actually settle those areas and enforce their claims after American independence.
Connecticut and Massachusetts, however, had no way to enforce most of their claims. The clearest issue was that French Louisiana and Spanish New Spain were in the way of the westernmost section, but also Pennsylvania and New York territory cut off direct access New England the old Northwest. (And of course, there were numerous indigenous nations in the territory as well, but in the Americans’ minds they were going to be conquered eventually either way.)
But even without direct access Connecticut didn’t entirely give up on its long western extension. States of the young USA started ceding their western land claims to the federal government in the decades following independence. This kept the peace between states and gained goodwill from the union, and made it easier for Americans to expand into the area. Connecticut ceded most of their claim, too but held onto a small piece of what would become Northeast Ohio.

Connecticut even started sending settlers to the area in the 1790s, under the auspices of the “Connecticut Land Company”. A few towns were even founded, most notably one on the banks of the Cuyahoga river by a certain Moses Cleaveland. Even today, Cleveland, Ohio1 maintains its original New England-style common grazing land as its central plaza, Public Square.
This area, known as the “Connecticut Western Reserve”, eventually was ceded to the federal government as well. It was just too impractical for Connecticut to manage an exclave on the other side of the Appalachians. But the legacy of this last bit of Long Connecticut (Disconnecticut?) remains, most notably in the name of Case Western Reserve University in Cleveland.
Coming soon: The elementary school in a nuclear bunker