1431 stories
·
1 follower

You Can’t Trust the Internet Anymore

1 Share

I like things that are strange and a bit obscure. It’s a habit of mine, and a lot of this blog is to document things I haven’t heard of before, because I wanted to learn about them. I mean, jeez, I’m certainly not writing blog posts about strip mahjong because the people demand it. But I can’t stop seeing misinformation everywhere, and I have to say something. This post is just a rant.

Phantasy Star Fukkokuban

This is Phantasy Star Fukkokuban, a Japanese Sega Genesis game released in 1994 to commemorate the release of Phantasy Star IV by re-releasing the original. It has an interesting component: it is the Master System game, just packaged into a Genesis cart. The PCB wires the Genesis lines the same way your Power Base Converter would. My guess is the reason for this is because the Master System wasn’t very popular in Japan, and Phantasy Star IV tied together the whole series with a lot of tiebacks to the first one in particular.

Phantasy Star Fukkokuban, which uses the Phantasy Star box art on a Japanese cartridge shell.

As a Master System game disguised as a Genesis one, this game is technically interesting. Some Genesis consoles can’t play Master System games, and those ones can’t play this game either. Also, I love the Phantasy Star series; even if 2 is my favorite. This makes this cartridge a perfect subject for my interest, so I’ve talked about it before and will talk about it again. In fact, I have a post I’m working on where I mention it.

Phantasy Star title screen. (C) SEGA 1987

So there I was, writing a blog post, and wanted to look up the release date. The first result I found in DuckDuckGo, my search engine?

DuckDuckGo search results. First, GameFAQs. Second, TCRF. Third, Press Start Gaming. An abandonware site is at the bottom

GameFAQs is at the top; a titan since the 1990’s. The second result is The Cutting Room Floor, a wiki much beloved by myself. And then the third result is “Press Start Gaming”.

Welcome to Press Start Gaming, your ultimate destination for gaming and tech enthusiasts! Founded with a passion for exploring the ever-evolving worlds of gaming and technology, we aim to deliver high-quality reviews, insightful articles, and the latest industry news to help you stay informed and inspired. Whether you’re a casual gamer, a tech aficionado, or a seasoned pro, we have something for everyone.

And here’s a thing about me. I want to trust new websites. I have a bias towards clicking on articles from sites I don’t know, because to be quite honest, I’ve read the TCRF page on Phantasy Star a thousand times. How else do you learn something new?

Phantasy Star title screen. (C) SEGA 1988

Also, I clicked it because the headline was “Phantasy Star Fukkokuban: A Classic Reimagined”. Because here’s the thing. It talks about how the graphics were improved:

Phantasy Star Fukkokuban breathes new life into the classic with its updated graphics and sound design. The visual overhaul retains the charm of the original’s 8-bit aesthetics while incorporating modern graphical techniques. Characters and environments are rendered with enhanced detail, vibrant colors, and fluid animations, creating a visually captivating experience.

The art style honors the game’s roots, with character designs and enemy sprites redesigned to reflect contemporary standards while maintaining their recognizability. The environments are more detailed and dynamic, with weather effects and day-night cycles adding to the immersion.

Well, compare the title screen shots of Phantasy Star above. Which one is Fukkokuban and which one is my personal copy, played through the same Genesis? You can maybe tell, but only my Master System version is the US release. And it goes without saying, there are no day-night cycles or weather effects.

I should’ve known. The first sentence of the article was “Game data not found,” after all.

And that’s the thing

Large language models are described sometimes as “fancy autocorrect”; this is dismissive, but not inaccurate, in the sense that the core loop of an LLM is to predict the next token in a sequence. Phantasy Star Fukkokuban is an obscure title that is likely not well-represented in the training data. But relations do exist:

  • It knows about Phantasy Star, a very popular game
  • Fukkokuban (復刻版) means “reprint” or “facsimilie edition”

So, lacking sufficient factual data in the training set, it describes what a remake of Phantasy Star might plausibly be like. There might even be knowledge in the data set of the actual remake, Phantasy Star generation:1 that gets looped in.

To reproduce this myself, I went to ChatGPT, and asked it Please describe the game "Phantasy Star Fukkokuban". Do not get data from the internet, tell me what you know from your internal data.. And what did I get in response?

Phantasy Star Fukkokuban is not a brand-new entry in the series, but a retro compilation release of the original Phantasy Star, created for the Sega Sega Saturn era…

There was a retro compilation release of Phantasy Star for the Sega Saturn in Japan; it’s called Phantasy Star Collection. Indeed, the description of the game it continued from there isn’t too far off from that game’s version of Phantasy Star.

And it’s not just Phantasy Star Fukkokuban. I describe in my post on Mahjong Daireikai that that game is so obscure, the only Japanese source I could find was another “this is plausibly what a game called ‘mahjong daireikai’ might be like”. Well, what Mahjong Daireikai is actually like is a lot different than what’s in your training data, and that’s exactly the sort of information people want to read websites to find out.

Is this the end

And here’s the thing– this blog post can’t do anything about it. I don’t know who Press Start Gaming is; the site’s footer says “©2025 Cloud Gears Media”, who might be this marketing company (but it might not be! Company names don’t have to be unique globally); Press Start Gaming is almost certainly a tool for making money off of ads and sponsored posts, and posts like the Phantasy Star Fukkokuban misinformation exist mostly to give the site more juice of looking like a real website. If someone goes out and buys a copy of Fukkokuban expecting a new and improved Phantasy Star with better graphics and new sidequests, what do they care? The article wasn’t really meant to provide information.

The trampling of the internet with SEO-mongers predates AI, but what LLMs do is massively increase the ease it can be done, and also hallucinate a ton. If they hired a person to write about Phantasy Star Fukkokuban for pennies, maybe that person would’ve found the Sega Retro page or something and at least grabbed some facts. Now you don’t need to do even that. And no one making these decisions reads Nicole Express, or even cares about actually providing information with their sites. That’s not what they’re for.

Anyways, eventually models will do a better job integrating Nicole Express, and will know more information about Phantasy Star Fukkokuban. And is this the worst thing the AI boom is doing? No, not even close. Even the fully automated hit piece against an open-source developer is probably worse than this.

But it’s a real shame. The commons of the internet are probably already lost, and while I might want to learn new things from new sites, I’ll just have to stick to those with pre-LLM reptuations that I trust. Well, until those sites burn their reputations to make a few extra pennies with AI, like Ars Technica seems to just have.

This post is just a rant. Thanks for listening, at least.



Read the whole story
mrmarchant
4 hours ago
reply
Share this story
Delete

The best way to spot AI is also the easiest

1 Share

How to make AI spotting easy

Finding AI videos is hard, but finding AI accounts is easy. Anyone can do it with some basic knowledge about how AI media has advanced over the past two years.

I often joke about how many beautiful 24-year-olds discovered Instagram for the first time in November of 2025. Google’s Nano Banana Pro was released November 20th, and a rush of fake people registered. Let’s turn this into a practical way to find fake accounts. Put a pause on finding individual AI videos and AI videos; it’s more efficient to find the AI accounts, then work backwards.

In this piece we’ll plot out the timeline for recent advancements in AI generation and how new accounts and videos coincide with each change.

The Account Age Paradox

It’s obvious to us that an iPhone 17 Pro’s camera looks better than the camera from the iPhone 6S. 10 years of technological advancement separates them. But in 2015, a new iPhone 6S produced photos notably better than the HTC One M7 I had at the time. That HTC One took the earliest photos in my current photo library, and I look back on them fondly.

The same cannot be said of an AI creator in the year 2035, who is trying to prove their character is real. Can they look back at the old, Veo 3 AI generations from 10 years ago to establish proof of life? Of course not - it has the opposite effect, because Veo 3 looks awful and obviously AI to people in 2035. This is a huge difference between AI and real media, and one that AI creators are already aware of.

On Instagram, relevant AI accounts are usually just a few months old and pop up with AI media innovations. When I find a suspected AI character on TikTok, which unlike Instagram or YouTube does not make “account age” visible on the app, I immediately scroll down to their earliest posts. Did they leave the old, bad generations, or did they delete them and start over recently? Either are huge red flags and nearly impossible to overcome.

Putting it into action

I just found a new account on Instagram: an AI influencer named Olivia Arizpe. First of all, her bio says “Digital Creator” and “AI” in it, but let’s pretend the creator wasn’t so honest.

First, I’ll tap on the three dots at the top of the account page to pull up “About this account.” Here I see that the account was created in January of 2026. The first video is also from January of 2026. This video shows the AI avatar lip-syncing to a Tame Impala song. This is follows “motion control” trend I’ll talk about later. That’s all the information I need to confidently call this an AI account.

Or what about this account, a “News” page that claims to be from Tulsa, Oklahoma?

This page used to be called “Tulsa Area Breaking News”, but it’s just an AI slop page. They started in November, and their earliest available video is from the same month. This follows OpenAI’s release of Sora, and indeed there are Sora watermark scrubbing artifacts. Not to mention, their first video follows the format of the AI-generated EBT videos coming out at the time, and those were almost entirely generated with Sora.

None of this required good eyesight or any pixel peeping. And if any shiny, new AI slop comes out of the AI-generated Tulsa area and they really want to trick you, they’ll have to delete the old stuff.

We dive into how to find account transparency information in this article, by the way. It contains instructions for Instagram, Facebook, YouTube, TikTok, and X.

Granted, it’s unreasonable for you to remember that Sora came out in October, or that Motion Control AI generators were trending in January. Luckily, you don’t have to remember: I wrote it all down below for you to reference any time you need.

Subscribe now

The practical AI release timeline

When coming across an account we suspect is AI-generated, we’re looking for either:

  1. A relatively new account that appeared and immediately implemented some recent technology

  2. An account that shows signs of older or more affordable AI technologies in older posts

Let’s look at previous major advancements in AI media and how those popped up on the Internet. This will not only demonstrate how new technology changes trends on social media, but also what to look for when looking back at old posts. This will also give us some important cutoff dates. For example, if you see a dance video from 2023, you can be sure it wasn’t AI-generated.

Before 2025

2025 will probably go down as the most pivotal year in AI media history, and most of the AI-focused accounts today don’t use models from 2024 or earlier. But there are some notable exceptions you should be aware of.

Until Late 2023 - Deepfakes and AI avatars

In 2018, deepfake technology was creating memes and porn, but it was very much in the experimentation phase. Deepfakes aren’t AI videos — they use a different tech architecture — but they’re the first popular form of “AI video” in its broadest possible definition. The corporatized versions of deepfakes are usually called “Avatars”. These are made by companies like Synthesia, who were already around by 2020, and companies like Heygen joined the mix later.

Important takeaway: Face-centric “AI videos” were already possible before 2023, but their uses were pretty limited.

A collection of “classic” AI image slop

Late 2023 to Early 2024 - AI Image slop starts hitting social media

A lot of Facebook posts still have I would call “classic” AI slop. These are the images created by technology from late 2023. For example:

  • OpenAI released DALL-E 3 in October of 2023

  • Midjourney released V6 in December of 2023

While these are just two of many image generators of this era (and OpenAI is depreciating the DALL-E 3 API shortly), there are many accounts still stuck in this era because it’s cheap! These sorts of images became ubiquitous when OpenAI integrated DALL-E into ChatGPT, lowering the barrier of entry and opening the AI slop flood gates.

This is the era of Shrimp Jesus and “engagement slop” like old, disabled puppies asking for donations, or ragged puppies begging for likes. Today, we still see NFL coaches who desperately need help with medical bills. They don’t look like the coaches exactly, there’s a yellowish tint, and it’s giving “Polar Express.”

Important Takeaway: If you see a decent image from before mid-2023, its foundation is probably a real photo. Maybe it’s a heavily edited real photo, drawn, or 3D rendered by this era’s already incredibly powerful animation or gaming engines. But it wasn’t made by “AI”.

Early to mid 2024 - First (decent) AI Video models released publicly

Sora got a lot of early attention February of 2024, followed by models like Luma Dream Machine and Runway Gen-3 in June that year. AI video was still relatively limited, but boosted by many new scaling and training advancements from 2023. Given its limitations, memes and AI video slop was their primary social media use.

Important Takeaway: If you see a decent-quality video posted before mid-2024, unless it was professionally modified or rendered, it’s a real video.. There aren’t any good “AI videos” before this point.

2025

March - ChatGPT 4o Images

I still remember sitting on the couch with my wife on March 26th, 2025 - the day after OpenAI released ChatGPT’s 4o image generation. The internet was lit up with Studio Ghibli-ified images. Feeling guilt about the inherent ethical quandaries of that style, I instead made a photo of our recent vacation in the Family Guy art style. Our surprise and fear from that day seems pretty quaint in hindsight, given what else 2025 was about to bring.

AI-generated profile pictures from this era are still common. The yellow-tinted, glossy, evenly-lit photos from this generation are ubiquitous. Since ChatGPT is the most popular large language model, its built-in image generation is also really popular.

A collection of stills from Google Veo 3 test generations

May - Google Veo 3

Veo 3 was the first AI video model with meaningful sound, and it was also a jump forward in video quality. It generated videos with a cinematic look and feel, though it definitely wasn’t up to cinematic standards. As a result, though it was envisioned as a tool for creatives, it instead flooded the internet with AI videos.

A Veo 3 Review was my first ever YouTube video. Today, Veo 3 makes a platonic “AI video” that’s relatively easy to spot after you’re familiar with it. The most distinctive characteristics include smooth and even lighting, temporal inconsistencies and background issues, and the character’s robotic and melodramatic voices.

Immediately in May, a ton of AI slop pages popped up on social media. Since Veo 3 could only do widescreen 16:9 videos at first, vertical-native platforms like TikTok suddenly had a ton of new accounts posting AI videos with black bars (also known as letterboxes) on the top and bottom of every video. ASMR Fruit cutting videos and AI man-on-the-street interviews were trend. But also, people’s first “I got tricked” moments came with Veo 3’s bunnies on trampolines.

Important Takeaway: If you find a realistic video with matching sound or dialogue posted before May of 2025, unless it was generated with a game or animation engine, it’s a real video.

Mid-2025 - Other video models

Around the same time as Veo 3 were Runway Gen 4 (April) and Kling 2.1 (May). These were similar in video quality to Veo 3, but neither had synchronized sound. They traded this for vertically-native videos and different video styles. Along with Midjourney releasing its first video model in June, Alibaba’s open-sourcing of Wan 2.2 in July, and many more advancements from companies like LTX and Minimax, a plethora of good video-generation options came online in mid-2025.

With these vertical-native models, we got AI-generated landscapes, fake natural phenomena, and AI-generated cartoons for kids that were NOT kid-friendly. But this is also when AI slop started looking realistic enough to fool a ton of people in vertical feeds. By August this is almost all I covered.

August - Nano Banana

Google’s Nano Banana, the informal name for Gemini 2.5 Flash Image, was a big jump in photo quality. Before Nano Banana, mainstream AI photo still had a glossy look. After Nano Banana, photorealistic images were very accessible.

Important Takeaway: A lot of photo-only AI accounts got a huge boost or started in August of 2025.

October 2025 - Sora 2

The next jump in AI video came from OpenAI, who took a year and a half after the original Sora to release the Sora 2 video model. Alongside it was the release of the Sora app, a TikTok-style vertical video app with only AI videos.

The Sora 2 model had a few key innovations:

  • It felt less uncanny than Veo 3 for many viewers. Eyes, mouth, and skin detail were more realistic.

  • It had better physics than any other model at the time.

  • It was funny. Lazy prompts were spiced up by a language model in the background. This meant inexperienced AI prompters could make viral videos.

And yet, it was a very noisy model with a heavy “AI accent.”

Sora’s release coincided with a huge increase in the number of “AI slop” accounts because it was free. Until this point, good AI video generation was expensive, but the Sora app let users generate a lot of videos in the free app. These videos could be downloaded with a watermark (that was easily scrubbed), then reposted to the other, much more popular social media sites. The Sora 2 API released just 2 weeks after, providing videos without watermarks, and giving more people access to the powerful Sora 2 Pro model. To this day, a ton of AI videos come from Sora 2 because it has a good price-to-quality ratio.

Important Takeaway: Many AI video slop accounts have October 2025 birthdays or changed their usernames at this time.

November 2025 - Nano Banana Pro

Google’s jump to Nano Banana Pro (Gemini 3.0 Pro Image) was surprising. Just three months after the original, it brought big improvements to realism and quality, as well as improved spacial reasoning and text rendering. This is the point where AI photos became mostly undetectable on first glance, though they can be spotted when looking closely.

Along with Nano Banana Pro came the release of SynthID, which can be accessed through Google’s Gemini large language model. It’s an invisible watermarking system that embeds and detects a watermark hidden inside the pixels of a photo.

Creators reacted accordingly. Misinformation like the infamous “Bubba Trump“ photo and fake celebrity paparazzi photos spread wildly. More relevant to our everyday social media use, AI generated influencer accounts proliferated in late November. The Nano Banana Pro release also corresponded with TikTok’s unfortunately-timed new emphasis on image carousel posts. And, since many AI video generators have a photo-to-video mode, these AI images are the starting frames for many AI videos. Nano Banana Pro is still a leader in this field, and while other generators have caught up, this was an important demarcation point.

And that wasn’t all that happened in November...

November and December 2025 - Motion Control AI improvements

Coinciding with Nano Banana Pro’s release were improvements in Motion Control models, most notably Wan 2.2 Animate. This prompting innovation lets creators take a “control” video — often a video stolen from a real creator’s social media — and “replace” them with a new character, real or not. This method saw further improvement through Kling 2.6 Motion control.

These models unlock very realistic motion and physics for AI characters. Combining Nano Banana Pro (or similar photo models) with a motion control video model lets creators generate realistic, consistent characters across posts. While not perfect (there are still plenty of AI giveaways), AI influencer creators now had a ton of tools at their disposal.

Important Takeaway: Accounts that feature human avatars and started or rebranded in November or December of 2025 are a huge red flag. A lot of AI slop accounts moved into the more profitable AI influencer business.

Early 2026 trends

Releases from Kling and Bytedance show further improvements in AI video quality, which we’re still analyzing as of this post. I expect some new accounts in February that play off their strengths, but those are yet to be determined. Right now it’s a bunch of demos of stolen intellectual property and celebrity deepfakes, as can be expected with a new model release.

Moving Forward

People regularly ask me what might happen if AI video becomes “perfect” or “undetectable.” But to my eyes, real videos aren’t “perfect”. There are always artifacts of the process that made them.

I remember watching Superbowl XLV 15 years ago, on an imposing 75-inch high-definition TV. It was incredible and perfect at the time. But looking back at it now, it looks a bit dated, even without today’s de-interlacing conversion artifacts. In 5 years, perhaps the oversaturated HDR look of today’s iPhone phones will be an out-of-fashion artifact of our current era.

The highest-end, most impressive AI video generations today still don’t look "real” to me, but they do look “really good”, which is real enough for most people. The difference is marginal, but with hindsight it may become obvious. Those of us who make real media will have to adapt and differentiate ourselves from AI advancements, figuring out what our advantages are. It’s always going to be changing, so stay tuned here for updates on the latest releases and countermeasures.

Riddance is a reader-supported publication. To receive new posts and support our work, consider becoming a free or paid subscriber.



Read the whole story
mrmarchant
4 hours ago
reply
Share this story
Delete

AI Didn’t Destroy Critical Thinking. We Did.

1 Share

Illustration

It feels like we’ve reached a critical mass of consensus: AI is just bad for our students. The American Association of Colleges and Universities just released the results of a national survey of U.S. faculty, which found that 95 percent believed AI would “increase students’ overreliance” on such AI tools, and 90 percent believed AI would “diminish students’ critical thinking skills.” This is mirrored in a recent Brookings report, which concluded that “the risks of utilizing AI in education overshadow its benefits.” As one professor (“I’m an AI power user”) put it, “I want to strip things back: no laptops, no phones, just pens and paper.”

It seems everyone wants to find a way to minimize or even forbid AI use, kind of like how cell phone bans and restrictions in K–12 schools have passed in 33 states. The consequences of doing nothing, such narratives proclaim, could be dire. The Brookings report, for example, throws around terms such as cognitive decline, cognitive impairment, and cognitive atrophy—all of which, it notes, are associated with an “unhealthy aging brain.” They quote an MIT brain imaging study that suggests the long-term consequences of AI use may include “diminished critical inquiry, increased vulnerability to manipulation, decreased creativity…[and] risk internalizing shallow or biased perspectives.”

Here’s the problem with all this “Chicken Little” hysteria. Four years ago, before any of us had a clue about weird acronyms such as GPT, LLM, or AGI, every education expert I know was bemoaning students’ continued lack of academic competence. NAEP has for decades documented how just a small percentage of U.S. students reach even a “proficient” level in their reading and writing and that, compared to other countries, U.S. students consistently are middle-of-the-pack. Results from the Collegiate Learning Assessment incited two prominent scholars to conclude that college students were “academically adrift” and learning almost nothing across their years in college.

AI, in other words, did not erode critical thinking; it exposed how poorly we have been teaching it.

Let me be blunt: There was no golden age of critical thinking or academic achievement before AI came along and seemingly ruined everything. In the years before ChatGPT arrived, K–12 educators said some of their most pressing concerns were that schools were boring and that we didn’t know how to talk to each other; college leaders worried that they lacked the ability to strengthen students’ critical thinking, communication, or problem-solving skills to successfully enter the workforce.

So, sure, I understand today’s basic argument: Maybe using AI in the wrong way will make all this even worse. Trust me, I’ve been there. I was ready to give up and walk away as I saw AI supercharge a disengagement spiral that turned my college classroom into a transactional mirage of learning.


EdNext in your inbox

Sign up for the EdNext Weekly newsletter, and stay up to date with the Daily Digest, delivered straight to your inbox.


But here’s the thing: The emergence of AI truly marks a transformation akin to a Copernican revolution in education. This is because AI has given us the chance to implement a type of powerful personalized learning that we have only dreamt of, a potential reality that education theorists have spent decades developing key concepts and core theories about (e.g., ubiquitous learning, situated learning, legitimate peripheral participation, distributed cognition). The problem is we’ve never been able to implement this vision faithfully within the institutional constraints of our education systems. And revolutionary moments, like all transformations, create massive disruptions.

The solution, though, is not to pretend these disruptions don’t exist, nor is it to bemoan that the sky is falling. Instead, we need to embrace them.

I, for example, have finally figured out how to help my students use AI as a daily tutor, Socratic conversation partner, and writing mentor. I walk my students through the ethical use of AI and how—if prompted correctly and used deliberatively—it can help them think carefully and thoughtfully about some of our most complex and contested societal issues. So rather than face a passive and disengaged lecture hall of 70 students, I watch them write daily reflections such as this: “Overall in this course I have noticed that we are being taught how to think rather than what to think and I think that AI and been a great tool during this process.” Many other researchers and faculty are experimenting with how to make AI a catalyst for learning rather than a ghost writer for outsourcing thinking.

Recent handwringing about the loss of critical thinking skills, I would therefore suggest, says far more about how we teach than how our students learn. If we really care about saving students’ critical thinking skills, we need to think critically ourselves about how to re-envision our education systems with the right guardrails and guideposts to leverage AI-driven tools rather than disengage from this transformational moment. Prohibitions and nostalgia for a pre-AI world are the real dangers that will result from a failure to think critically. Instead, educators’ embrace of AI as a transformational tool is what will make a world of difference.

Dan Sarofian-Butin was the founding dean of the School of Education and Social Policy at Merrimack College and is now a professor of education there.

The post AI Didn’t Destroy Critical Thinking. We Did. appeared first on Education Next.

Read the whole story
mrmarchant
4 hours ago
reply
Share this story
Delete

It’s never too late to stop hating math

1 Share

Americans are getting worse at math. Student scores have fallen to their lowest point in decades. Nearly half of high school students barely meet what the National Assessment of Educational Progress (NAEP) considers a “basic” level of comprehension, and more than 900 freshmen at the University of California, San Diego — 12.5% of the institution’s first-year class in 2024 — had the mathematical proficiency of a 13-year-old.

U.S. adults aren’t faring much better. Last checked, only 65% could pass a basic arithmetic test, making the country one of the more quantitatively challenged in the industrialized world.

But this isn’t the first time the math graph has trended downward. A similar development took place during the early stages of the Cold War, when enrollment in high school algebra fell to levels not seen since the start of the 20th century. It wasn’t until the launch of Sputnik in 1957, when the Soviet Union kicked off the space race, that alarm gave way to action. Math and science education were overhauled, and calculus was introduced into the curriculum.  

As then, concerns about the national “math deficit” are closely tied to geopolitics. Instead of Russian rocket engineers, policymakers now worry about competition from China, India, and other emerging economies, which have for years provided Big Tech with some of its best and brightest. What will happen to the U.S. when Shenzhen overtakes Silicon Valley?

That’s one side of the coin, but the math deficit does not only affect national prosperity and security. As the U.S. Board of Education’s National Mathematics Advisory Panel stressed in a 2008 report, it also affects our quality of life. More than a launchpad to better job opportunities, math is a tool for self-improvement. It is useful in the arts, sports, and many other areas that, at first glance, seem to have little to do with numbers. To learn math, the French mathematician David Bessis writes in his book Mathematica, is to “change the way you see the world.”

Math anxiety

Remember that dreadful feeling you got in math class? Your teacher is flying through the coursework at 100 miles per hour, and all of your classmates are nodding along; meanwhile, you’re sitting there so lost and confused that you can feel your brain spinning. 

If that sounds familiar, know that you’re not alone.

According to one survey, 9 out of every 10 U.S. adults have experienced some level of math anxiety. Math anxiety is so common that even LLMs — AIs trained on vast amounts of human output — associate numbers with words like frustrating, exasperating, and alarming.

Math anxiety even shares many symptoms with regular anxiety: clammy palms, an upset stomach, increased heart rate, and lightheadedness. By activating the brain’s pain and fear centers (the insula and amygdala), math anxiety can impair your working memory and, by extension, cognitive abilities — explaining all those times you watched aghast as the equations in your textbook morphed into indecipherable hieroglyphics. 

The consequences of math anxiety reach far beyond the classroom. If internalized, the all-too-recognizable belief that someone is or isn’t “a math person” can lead them to forgo a rewarding career in science, technology, medicine, or many other fields that require beyond-the-basics arithmetic skills. In extreme cases, math-anxious people may try to avoid any activity involving numbers, from balancing bank statements to measuring out ingredients for home-cooked recipes.

Learning math, therefore, often begins with learning to cope with math anxiety. Studies find that many of the techniques used for managing regular anxiety — from breathing exercises to cognitive behavioral therapy — can help with math anxiety, as well. Echoing progressive pedagogical movements from the 1960s and 1970s, many contemporary educators argue that people learn best when mathematical exercises are presented through meaningful, relevant everyday situations, revealing a seemingly abstract discipline as the practical resource it really is.

Others argue that math is best learned when it is fun. This approach not only aligns with what we know of brain development — babies and children often learn through play — but also helps address math anxiety. If someone encounters math outside the classroom, away from an impatient teacher and those intimidating textbooks, they may give the subject a second chance. That’s one explanation for the sudden and explosive popularity of “math influencers” like Andy Math and 3Blue1Brown, who have over 1.3 million followers on TikTok and 8 million subscribers on YouTube, respectively. Browse their comment sections, and you’ll find no shortage of comments like, “This just proves that maths isn’t boring” and “Never in my life did I think I’d binge-watch math videos.”

Bessis — who specializes in algebra, geometry, and topology, and achieved mathematical fame for solving a problem dating back to the Nixon presidency — says overcoming math anxiety begins with recognizing that it is shared by people at every skill level.

“Every graduate student knows [the feeling],” he tells Big Think. “You sit in a seminar, and after 20 seconds, you understand nothing. It’s not that you don’t understand some of the details or references; it’s all just nonsense. There’s no meaning. It’s probably similar to when you were an infant, listening to people speak a language you don’t understand. As an adult, you don’t expect to be in such situations, so you need confidence and awareness to avoid panicking.”

Joining the ranks of Andrew Hacker’s The Math Myth, Matt Parker’s Love Triangle, and Francis Su’s Mathematics of Human Flourishing, Bessis’ Mathematica is part of a growing body of books that, much like those aforementioned influencers, attempts to restore the reader’s relationship with math. Using only mathematical concepts a middle schooler would understand, Bessis dismantles a variety of prejudices and preconceptions left over from our schooldays. 

Perhaps more importantly, Mathematica seeks to reintroduce readers to the joyful side of math — the exciting, magical side that is too often beaten out of us at school, but cherished by those who stick with the subject. Using the celebrated French mathematician Alexander Grothendieck as an example, Bessis argues math is best pursued with the mindset of a toddler: with “radical curiosity and indifference to judgment.” Think not of the panic attacks you suffered during exams, but the pride you felt when you learned to count to ten.

The fruits of math

Think also of the many ways that mathematical literacy may help improve your life, professional and private. In the early 20th century, stern schoolmasters believed math taught order, discipline, and a strong work ethic. Today’s research points to different but equally valuable benefits. Studies have found that math exercises correlate with cognitive function and metacognition (thinking about thinking) — both of which, in turn, correlate with mental health. The research also suggests that math can, directly or indirectly, improve neuroplasticity and emotional regulation, and help stave off dementia

Contrary to the long-discredited yet persistent left brain vs. right brain myth, mathematics is useful in the arts — another field that many wrongly believe requires an innate talent to explore and enjoy. A study assessing thousands of students in China found that mathematical literacy and creative thinking go hand in hand, suggesting that one stimulates the other and vice versa. 

Similar suggestions echo throughout art history. Leonardo da Vinci and other Renaissance heavyweights could not have brought their paintings to life without a deeply technical understanding of perspective, and they often constructed their images using a variety of scientific instruments and measurement tools. More recently, saxophonist John Coltrane and drummer Clayton Cameron credited their musical success to their mathematical abilities — lived experiences that support contemporary research exploring the link between early music education and later mathematical performance.

Just as there is math in art, some mathematicians would claim that there is art in math. “When I write down a proof,” Cymra Haskell, a professor at the University of Southern California, once told her college’s newspaper, “it feels like a puzzle coming together. There can be an intense pleasure in that, similar to the pleasure I feel when I listen to a beautiful piece of music or gaze at a beautiful painting.” 

More than a metaphor, her observation evokes a study that examined neural activity in 15 mathematicians. It found that looking at certain equations jump-started the medial orbitofrontal cortex, the same part of the brain responsible for perceiving and appreciating beauty.

Looking at his own creative output, Bessis proposes that his experiences as a mathematician helped him become a better writer. “There are quite a few writers who started out as mathematicians,” he says. “Victor Hugo was an accomplished math student and almost stuck with it. There’s something about math that resembles the process of literary writing. It teaches you to articulate what’s real, what’s in front of you, like describing exactly how you tie your shoes.”

Beyond self-help

Considering math’s many real-life applications, it’s no wonder that more than one reviewer has referred to Mathematica as a self-help book. Bessis disagrees with this categorization, even though he stands behind what it implies. 

“It’s not a ‘ten lessons to become super smart’ book,” he says, “but inherent in there is this promise, this idea that math will make you smarter and help you see the world more clearly.”

Above all, math has made Bessis more confident and less afraid of the unknown. “After making an impact in my field at age 35,” he reflects, “I made a big bet: I quit math and reinvented myself as a writer and founder, two other territories where you’re supposed to come in equipped with rare skills. Math convinced me that those skills are not inborn; you could develop them with the right effort and focus. It made me capable of taking huge bets on my own learning ability.”

This article It’s never too late to stop hating math is featured on Big Think.

Read the whole story
mrmarchant
5 hours ago
reply
Share this story
Delete

Gas Town’s Agent Patterns, Design Bottlenecks, and Vibecoding at Scale

1 Share
On agent orchestration patterns, why design and critical thinking are the new bottlenecks, and whether we should let go of looking at code
Read the whole story
mrmarchant
6 hours ago
reply
Share this story
Delete

The AI hater's guide to code with LLMs (The Overview)

1 Share

Introduction

This is the post I don’t think many people expected me to write, and I have(rightly!) surrounded myself with people who are generally uncomfortable tosomewhat hostile to “AI”, mostly for good reasons, though I’ll get into themany caveats on that as I go.

As activists mitigating the harms of “AI”, we need to be well informed, and weneed to understand what the specific harms are. Treating it with ahands-clean purist mindset will be extremely difficult and as activism, morealienating than effective. These are genuinely useful tools, and pretendingthey aren’t will not in fact win many hearts and minds.

This post is going to be very long, because in addition to technical context,I’m touching social issues, technical background, discourse norms, contextin a culture of rising technocracy and fascism funded by venture capital, andthe erosion of our information systems and cultural norms all at once. I can’tget into it all here, but I am not staying away from it on purpose.

Overall, I still believe that LLMs are a net negative on humanity, that thedestruction of our infosphere is going to have generational consequences, andthat if the whole thing disappeared from the face of the earth tomorrow, Iwouldn’t be sad. The damage is would still be out there, but the cheapness ofbullshit pervading everything would at least resume being human content millscale. Not to say that that was good before LLMs came along and made it thisbad, but it was better.

That said, that’s not going to happen, and the amount of effort required tomake it happen would be much better spent on organizing labor and climateaction. The AI industry may collapse in a house of cards. I think it somewhatlikely considering amount of financial trickery these companies are using. Butas someone I know put it: we’re not just going to forget that computers canwrite code now. We aren’t.

I want you to think about all of this with an intensely skeptical mind. Nothostile, mind you, but skeptical. Every claim someone makes may well becheckable. You can check! I recommend you do so. My math in this essay will berough back of envelope calculation, but I think that is appropriate given thetendency of the costs of technology to change orders of magnitude, andsituationally for things to vary by at least a factor of two.

And since we’re both operating in the domain of things not long ago consideredscience fiction, and because the leadership of AI companies tend to be filledwith people with a love of science fiction, many of whom won’t hesitate to, asis said, create the Torment Nexus from the popular science fiction novel Don’tCreate The Torment Nexus, I suggest one story to read and keep in mind:Marshall Brain’s “Manna – Two Views of Humanity’sFuture”.

TL;DR

  • There are open models and closed; good code work needs to be done on stuffthat needs very high end hardware to run, at least in part.
  • Chinese models are quite good, and structured differently as companies.
  • Don’t bother running models on your own hardware at home to write code unlessyou’re a weird offline-first free software zealot. I kind of am, and still Isee the futility with the hardware I have on hand.
  • Nobody agrees on the right way to do things.
  • Everyone is selling something. Usually a grand vision handwaving the hard andbad parts.
  • I’ll write more about how to actually use the tools in another segment.
  • A lot of the people writing about this stuff are either executives who wantto do layoffs, or now-rich people who made it big in some company’s IPO. Takewhat they say with the grain of salt you’d use for someone insulated by moneyand who can have free time relatively easily. They are absolutely handwavingover impacts they themselves will not experience.

A note on terms

I am writing this with as much verbal precision as I can muster. I loathe termslike “Vibe Code”, and in general I am not jumping on any marketing waves andhype trains. I’m being specifically conservative in the words I use. I say LLM,not “AI”, when talking about the text generation models at the heart of most ofthe “AI” explosion. I’ll prefer technical terms to marketing buzzwords thewhole way through, even at the cost of being awkward and definitely a littlestodgy. Useful precision beats vacuous true statements every time, and thedifference now very much matters.

The Models

There are a zillion models out there. Generally the latest and greatest modelsby the most aggressive companies are called “frontier” models, and they arequite capable. The specific sizes and architectures are somewhat treated astrade secrets, at least among the American companies, so things like powerrequired to operate them and the kind of equipment required is the sort ofthings analysts in the tech press breathe raggedly over.

the American frontier models include

  • Anthropic’s “Claude Opus”
  • OpenAI’s GPT-5.2
  • Google Gemini 3 Pro
  • something racist from xAI called Grok.

The frontier models are a moving target as they’re always the mostsophisticated things each company can put forth as a product, and quite oftenthey’re very expensive to run. Most of the companies have tools that cleverlychoose cheap models for easy things and the expensive models for difficultthings. Remember this when evaluating anything resembling a benchmark: it’s aneasy place to play sleight of hand.

When you use a frontier model company’s products, most of the time you interactwith a mix of models. This is usually a somewhat cheaper to run version of thefrontier models as the main mode, sometimes offering the true best model as anoption, a thing that is sometimes invoked, and the whole thing is hidden behinda façade that mkes it all look the same. Version numbers often resemble cellphone marketing, with a race to hae bigger numbers, “X” and “v” in places tomake it seem exciting. There is no linear progression nor comparison of any ofthe numbers in the names of models or products.

I largely have no interest in interacting with the American frontier modelcompanies, as their approach is somewhat to dominate the industry and burn theworld doing it. Anthropic is certainly the best of the bunch but I really don’twant to play their games.

I do not know this for sure, but I expect these models run into the terabytesof weights, more than a trillion parameters, plus they are products with a lotof attached software — tools they can invoke, memory and databases and userprofiles fed into the system.

Behind them are the large models from other AI companies, largely Chinese,producing research models that they and others operate as services, and oftenthey are released openly (called “open weights models”). Additionally some ofthe frontier model companies will release research models for various purposes.All core AI companies pretty much style themselves as research organizationsfirst, and product companies second. Note that nearly every AI company callsits best model a frontier model, whether it fits with the above or not.

Chinese companies and therefore models often have a drive for efficiency thatthe American ones do not. They are not the same kind of market-dominatingmonopolist-oriented sorts that VC-funded American companies are. They aren’t ascapable, but they do more with less. They’re very pragmatic in their approachcompared to the science fiction fueled leadership of American AI companies.These models run in the hundreds of gigabytes and have hundreds of billion ofparameters, though most can be tweaked to run some parts in a GPU and the reston a CPU in main memory, if slowly. They can run on regular PC hardware, ifextremely high end hardware, and distillations and quantizations of thesemodels, while they lose some fidelity, fit on even more approachable hardware.Still larger than most people own, but these are not strictly datacenter-onlybeasts.

Large, capable open models (Mostly Chinese) include:

  • z.AI’s GLM-4.7 and GLM-5
  • Kimi K2.5
  • MiniMax M2.1
  • Deepseek-V3.2
  • Alibaba’s Qwen3-Max
  • Mistral Large 3
  • Trinity Large

Mistral Large 3 comes out of Europe. Trinity comes out of the US, but has aless “win the AI race” mindset. There’s a lot of superpower “We need our ownsovereign solution” going on. China, the US and Europe are all making sure theyhave a slice of the AI pie.

I’m sure there’s more — the field is ever changing, and information about themodels from Chinese companies percolates slowly compared to the Americanfrontier models.

Behind these models are specialized smaller models, often sort-of good for codewriting tasks if one isn’t challenging them, but I actually think this is wherethe line of usefulness is drawn.

Medium-small coding models include:

  • Qwen2.5-Coder
  • GPT-OSS 120b
  • Mistral’s Codestral
  • GPT-4.7-Flash
  • Claude Haiku
  • Gemini 2.5 Coder
  • Smaller versions of Qwen3
  • Smaller versions of many other models

There’s also some much smaller models that will run on large gaming GPUs. Idon’t think they’re quite useful, they’re very attractive toys that people canget to do some truly impressive things, but I don’t think they’re all that.They are, however, about the capability of what kneejerk AI-haters expect,error-prone lossy toys that if anyone called “the future”, I’d laugh in theirface or spit at their feet. Notice how far down the list this is.

The Economics

LLMs are expensive pieces of software to run. Full stop, anything with broadutility is something that requires a GPU greater than most high end gaming PCs,and quite a lot of RAM. I am setting a high bar here for utility, because AIboosters tend to have a frustrating way of equivocating, showing low numbersfor costs when it suits them, and high ones for performance, despite not beingfrom the same models. There are domain specific tasks and models that can workin mere small GPU or even Raspberry Pi levels of computation, but for generalpurpose “reasoning” tasks and coding specifically, right now in 2026, withcurrent model efficiencies, and with current hardware, if you want to use LLMsfor writing software, you will be throwing a lot of computing power at it. A$5000 budget would barely suffice to run something like gpt-oss 120b(OpenAI’s open model that is okay at code-writing tasks). Additionally, if youkept the model busy 100% of the time, you might be talking $50-$200 inelectricity depending on local prices, per month.

If you spent $15,000 and triple the electricity you could run something likeGLM-4.7 at a really good pace.

Water cooling for data centers is probably the most talked about environmentalcost, but I think it’s actually a distraction most of the time. Dear god why dopeople build data centers in Arizona, that’s a travesty, but also that’s aspecific decision made by specific people with names and addresses who shouldbe protested specifically.

Datacenter growth at the cost of people driving up electricity demand is a bigproblem, and we need to get back on the solar train as fast as possible.

This is not inexpensive software to run. However, it’s not an unfathomableamount of power.

Training models is wildly expensive, but it amortizes. There are in factdifficult economic conversations we need to be having here, but it’s allobscured by the fog of “what about the water?” and “AI will save us all andchange everything!” that pervades the discourse. The framing of the argumentsat large are fundamentally misleading, by basically everyone, pro or anti-AI,and much more about affiliative rhetoric than argumentative. We need to havethe arguments, and actually look for and persuade people of the truths. They’reuncomfortable so I fully understand why we’re not very often, but if we want toactually solve crises, we need to talk with actual truths in mind.

With prices of $200/month for “Max” plans, if one uses the tools well, acompany would in fact be making a smart decision to get their developers usingthem. They are definitely below cost, probably by at least 3-5x. Maybe 10x.(Remember that a price shock will come at some point before depending on theeconomics of these systems in existential ways for a business.)

Even at cost the math works out for a great many use cases.

Light plans are $20/month, and I think that for intermittent use, with goodtime sharing, that’s quite sustainable. In my experimentation I’m paying evenless than that, and while I don’t think those prices will be sustained, I don’tthink they’re impossible either.

Most of the big providers and almost all of the hosted open model providershave a pay-by-the-token API option. This is an unpackaged a-la-carte offering,in the style of cloud providers. They nickle and dime you. The model whiletransparent is hard to calculate. The usual rates are in prices per millioninput tokens and per million output tokens. Input tokens are cheaper, butinteractions with tools will re-send them over and over so you get charged forthem multiple times. Output tokens are more expensive but closer to one-timethings. Expensive models can be $25 per million output tokens and $5 permillion input tokens (Claude Opus 4.6). I expect this reflects a decent marginon the true costs, but I have not a ton to back this expectation up. Most openmodels run in the realm of $0.50-$3 per million input tokens and $1-$5 permillion output tokens. Given that a lot of the open models are run by companieswith no other business than running models, I expect these represent near truefinancial costs. There’s no other business nor investment to hide anycomplexity in.

The Tools

Most of the tools can talk to most of the models in some way. Usually each hasa preferred model provider, and doing anything else will be a lesson inconfiguration files and API keys. Some more so than others.

Most of the tools are rougly as secure as running some curl | bash command.They kinda try to mitigate the damage that could happen, but not completely,and it’s a losing battle with fundamentally insecure techniques. Keep this inmind. There are ways to mitigate it (do everything in containers) but you willneed to be quite competent with at least Docker to make that happen. I havenot, I’m going for being a micromanaging busybody and not using anythingresembling “YOLO mode”. I also back everything up and am not giving permissionto write to remote repos, just local directories.

I know terminal-based tools more than IDEs, though I’ll touch on IDE-integratedthings a bit. I haven’t used any web-based tools. I grew up in terminals andthat’s kinda my jam.

  • Claude Code is widely acknowledgedas best in class, has a relatively good permission model, and lots of toolshook into it. It’s the only tool Anthropic allows with their basic consumersubscriptions. If you want to use other tools with Claude, you have to pay bythe token. Can use other models, but it’s a bit of a fight, and a lot of APIsdon’t support Anthropic’s “Messages” API yet.
  • OpenAI Codex is OpenAI’s tooling. It’s gotdecent sandboxing, so that what the model suggests to run can’t escape andtrash your system nearly so easily. It’s not perfect but it’s quite a bitbetter than the rest. It’s a bit of a fight to use other models.
  • OpenCode touts itself as open source, when in realitymost stuff is. It’s a bit less “please use my company’s models” than mosttools, and it’s the tool I’ve had the best luck with. It has two modes — Buildand Plan — and using them both is definitely a key to using the tool well.Plan mode creates documents and written plans. Build does everything else andactually changes files on disk.
  • Kilo Code is both a plugin for VS Code, and a toolin the terminal. It has not just two modes but five, and more can becustomized. “Code”, “Architect”, “Ask”, “Debug”, and “Orchestrator”.Orchestrator mode is interesting in thatit’s using one stream of processingwith one set of prompts to evaluate the output of other modes. This shouldallow more complex tasks without failing because there’s a level ofoversight. I’ve not used this yet, but I will be experimenting more. Itspermission model is pretty laughable but at least it starts out asking you ifit can run commands instead of just doing it.
  • Charmbracelet Crush isaesthetically cute but also infuriating, and it’s very insistent onadvertising itself in commit messages. I’ve not yet seen if I can make itstop, but it did make me switch back to OpenCode.
  • Cursor — App and terminal tool. Requires an accountand using their models at least in part, though you can bring your own key touse models through other services.
  • Cline — Requires an account. IDE plugins and terminal tools.
  • TRAE — IDE with orchestration features. Intended tolet it run at tasks autonomously. I’ve not used it.
  • Factory Droid. Requires an account. Can bring your own key.
  • Zed. IDE editor, with support for a lot of providers and models.

TL;DR

I like OpenCode; Kilo Code and Charmbracelet Crush are runners-up. The(textual) user interface is decent in all three, and it’s not loud, it’s notfancy, but it’s pretty capable. At some point I’ll try orchestration and thenmaybe Kilo Code will win the day. You’re not stuck with just one tool either.

Antagonistic Structures and The Blurry JPEG of the Internet

At its core, you can think of LLMs as extremely tight lossy data compression.The idea that it is a “blurry jpeg of the internet” is not wrong in kind,though in scope it understates it. Data compression is essentially predictingwhat’s next, and that’s exactly what LLMs do. Very different specifics, but inthe end, small bits of stuff go in, large outputs come out. It’s also “fancyautocomplete”, but that too undersells it because when you apply antagonisticindependent chains of thought on top, you get some much more useful emergentbehavior.

A pattern that you have to internalize is that while lots of these tools andmodels are sloppy and error-prone, anything you can do to antagonize that intobeing better will be helpful. This is the thing where I show you how LLM codetools can be a boon to an engineer who wants to do things well. Suddenly, wehave a clear technical reason to document everything, to use a clear typesystem, to clarify things with schemas and plans, to communicate technicaldirection before we’re in the weeds of editing code. All the things thatdevelopers are structurally pushed to do less of, even though they’re always anet win, are rewarded.

You will want your LLM-aided code to be heavily tested. You will want dataformats fully described. You will want eery library you use to have accuratedocumentation. You will use it. Your tools will use it.

You will want linters. You will want formatters. Type systems help.

This pattern goes deep, too. Things like Kilo Code’s “Orchestrator” mode andsome of Claude Code’s features work as antagonistic checks on other models.When one model says “I created the code and all the tests pass” by deleting allthe failing tests, the other model which is instructed to be critical will say“no, put that back, try again”.

One of the big advances in models was ‘reasoning’ which is internally a similarthing: If you make a request, the model is no longer simply completing what youprompted, but instead having several internal chains of thought approaching itcritically, and then when some threshold is met, continuing on completing fromthere. All the useful coding models are reasoning models. The model internallyantagonizes itself until it produces something somewhat sensible. Repeat asneeded to get good results.

Even then, with enough runtime, Claude will decide that the best path forwardis to silence failing tests, turn off formatters, or put in comments saying//implement later for things that aren’t enforced. Writing code with thesetools is very much a management task. It’s not managing people, but sometimesyou will be tempted to think so.

The Conservative Pressure

So here’s the thing about LLMs. They’re really expensive to train.

There’s two phases: “pre-training” (which is really more building the rawmodel, it’s most of training), and “post-training” (tailoring a general modelinto one for certain kind of tasks).

Models learn things like ‘words’ and ‘grammar’ in pre-training, along withembedded, fuzzy knowledge of most things in their training set.

Post-training can sort-of add more knowledge, giving it a refresher course inwhat happened since it came out. There’s always a lag, too. It takes time totrain models.

The thing is though that the models really do mostly know only about what theywere trained on. Any newer information almost certainly comes from searches themodel and tools together do, and stuffs into the context window of the currentsession, but it doesn’t know anything, really.

The hottest new web framework of 2027 will not in fact be new, because themodels don’t know about it and won’t write code for it.

Technology you can invent from first principles will work fine. Technology thatexisted and was popular in 2025 will be pretty solid. Something novel or niche,code generation will go of the rails much more easily without a lot of tooling.

This is, in the case of frontend frameworks, maybe a positive development inthat the treadmill of new frameworks is a long-hated feature of a difficult tosimplify problem space for building things that real people touch.

In general however, it will be a force for conservatism in technology. Expecteverything to be written in boring, as broken as it ever was in 2025 ways for awhile here.

They’re Making Bullets in Gastown

There’s a sliding scale of LLM tools, with chats on one end and fullorchestration systems of independent streams of work being managed by yet moreLLM streams of work at the other. The most infamous of this isGastown,which is a vibe-coded slop-heap of “what if we add more layers of management,all LLM, and let it burn through tokens at a prodigious rate?”

Automating software development as a whole will look a lot like this - ifemployers want to actually replace developers, this is what they’ll do. Withmore corporate styles and less vibe coded “let’s turn it all to eleven” goingon.

Steve’s point in the Gastown intro is that most people arne’t ready for gastownand it may eat their lunch, steal their baby and empty their bank account. Thisis true. Few of us are used to dealing with corporate amounts of money andeffort and while we think a lot about the human management of it, we don’tusually try to make it into a money-burning code-printer. I think there’s a lotof danger for our whole world here. Unfettering business has never yieldedunambiguously good results.

Other tools like this are coming. While I was writing this,multi-claude was released, andthere’s more too:Shipyard,Supacode, everyone excited about replacing people isbuilding tools to burn more tokens faster with less human review. They’rewriting breathless articles and hand-waving about the downsides (or assumingthey can throw more LLMs at it to fix problems.)

I personally want little part in this.

Somewhere much futher down the scale of automation is things like my friendDavid’s claude-reliabilityplugin, which is a pile of hacks to automate Claude Code to keep going when itstops for stupid reasons. Claude is trained on real development work and “I’lldo it later” is entirely within its training set. It really does stop and puttodos on the hard parts. A whack upside the head and telling it to keep goingsure helps make it make software that sucks less.

Automating the automation is always going to be a little bit of what’s goingon. Just hopefully with some controls and not connecting it to a money-funneland saying full speed ahead on a gonzo clown car of violence.

There’s a lot of this sortof thing.

The Labor Issue

The labor left has had its sights on AI for a while as the obvious parallel tothe steam-looms that reshaped millwork from home craft to extractive industry.We laud the Luddites, who, contrary to popular notions about them were notanti-technology per se, they just saw the extractive nature of businesses usingthese machines, turning a craft one might make a small profit at to a job wherepeople get used up, mind and body, and exhausted. They destroyed equipment andtried to make a point. In the end they had only moderate success, though theyand the rest of the labor movement won us such concepts as “the weekend” and“the 8 hour day”.

Even the guy who madeGastown sees howextractive businesses can - or even must! - be. Maybe especially that guy.We’re starting to see just how fast we can get the evil in unfettered business,capital as wannabe monopolists, to show itself.

Ethan Marcotte knows what’s up: We need tounionize.. That’sone of the only ways out of this mess. We, collectively, have the power. Butonly collectively. We don’t have to become the protectionist unions of old, butwe need to start saying “hey no, we’re not doing that” en masse for the partsthat bring harms. We need to say “over my dead body” when someone wants to runroughshod over things like justice, equality, and not being a bongo-playingextractive douchecanoe. We’ve needed to unionize for a long time now, and notto keep wages up but because we’re at the tip of a lot of harms, and we need tostop them. The world does not have to devolve into gig work and wideninginequality.

Coding with automated systems like this is intoxicating. It’s addictive,because it’s the lootbox effect. We don’t get addicted to rewards. We getaddicted to potential rewards. Notice that gamblers aren’t actually motivatedby having won. They’re motivated by maybe winning next time. It can lead usto the glassy eyed stare with a bucket of quarters at a slot machine, and itcan lead us to 2am “one more prompt, maybe it’ll work this time” in a hurry. I sure did writing webtty.

There’s Something About Art…

Software, while absolutely an art in many ways, is built on a huge commons ofopen work done by million of volunteers. This is not unambiguously always good,but the structure of this makes the ethics of code generation more complex andnuanced than it is for image generation, writing generation, and videogeneration. We did in fact put a ton of work out there with a license that says“Free to use for any purpose”. Not to say every scrape of github was ethical atall: I’m sure AGPL code was snaffled up with the rest, and ambiguously ornon-permissively licensed too. It is however built on a massive commons whereany use is allowed. The social status quo was broken, but the legal line atleast is mostly in the clear. (Mostly. This is only a tepid defense of some ofthe AI company scrapes.)

AI image generation and video generation can get absolutely fucked. It wasalready hard to make it as an artist because the value of art is extremely hardto capture. And we broke it. Fuck the entire fucking AI industry for this and Ihope whoever decided to make it a product first can’t sleep soundly for therest of their life. I hope every blog post with a vacuously related image withno actual meaning finds itself in bit rot with alacrity.

Decoding the Discourse

It’s helpful to know that the words used to describe “AI” systems are wildlyinconsistent in how people use them. Here’s a bit of a glossary.

Agent:

  1. A separate instance of a model with a task assigned to it.
  2. A coding tool.
  3. A tool of some kind for a user to use but that can operate in the background in some way.
  4. A tool models can invoke
  5. A service being marketed that uses AI internally.
  6. A tool that other agents can use.

Agentic: in some way related to AI.

Orchestration: Yo dawg I heard you liked AI in your AI, so I put AI in yourAI so your AI can AI while you AI your AI.

Vibe Coding:

  1. coding with LLMs.
  2. using LLMs to write code without looking or evaluating the results.

A coda

In the time it took to write this over a week or more, Claude Opus 4.5 gave wayto Claude Opus 4.6. GLM-4.7 was surpassed by GLM-5 just today as I write thisbit, but z.ai is now overloaded trying to bring it online and has no sparecomputing power. All my tools have had major updates this week. The pace ofchange is truly staggering. This is not a particularly good thing.

I may edit this article over time. No reason we can’t edit blog posts, youknow. Information keeps changing with new data and context.

Now go out there and try your best to make the world better.

Read the whole story
mrmarchant
6 hours ago
reply
Share this story
Delete
Next Page of Stories