From the Norwegian Consumer Council, a funny video that warns against the dangers of enshittification. It’s part of their Breaking Free initiative:
Digital products and services are steadily becoming worse. Software becomes increasingly difficult and frustrating to use, websites and apps are littered with ads and spam content, and useful features are removed, degraded, or made subscription-only. This is part of a process called enshittification.
Enshittification happens in stages: First a company attracts users by providing a valuable service, often seemingly for free or at an artificially low price. The company then exploits those users to draw in business customers, and finally abuses its business customers and claws back all the value for itself and its shareholders.
Enshittification is the result of a dysfunctional market, where companies have been able to get away with mistreating and exploiting consumers. Consumers are trapped in digital services, potential competitors are shut out, and policymakers and regulators are unable or reluctant to clamp down on anticompetitive, illegal and otherwise abusive behavior. In practice, a handful of tech companies have become so powerful that they do not have reason to fear any consequences.
What is the long-term effect of using LLM chatbots for daily tasks? According to a study (DOI link) by Steven D Shaw and Gideon Nave of the University of Pennsylvania the observable effect is that of ‘cognitive surrender’, where users are seen to blindly accept the generated answers.
There has long been a struggle between those who feel that it’s fine for humans to rely on available technologies to make tasks like information recall and calculations easier, and those who insist that a human should be perfectly capable of doing such tasks without any assistance. Plato argued that reading and writing hurt our ability to memorize, and for the longest time it was deemed inappropriate for students to even consider taking one of those newfangled digital calculators into an exam, while now we have many arguing that using an ‘AI’ is the equivalent of using a calculator.
Yet as the authors succinctly point out, there’s a big difference between a digital calculator and one of these LLM-powered chatbots in how they affect human cognition, and it’s one that’s worth thinking about for yourself.
Surrender Versus Offloading
Cognitive offloading is the practice of shifting cognitive tasks to external aids, and it is thought to make learning complex tasks easier. In contrast to rote memorization of facts like dates of events and formulas, if we consider books to be an external memory storage device, then we can offload such precise memorization to their pages and only require from students that they are capable of efficiently finding information, as well as judging it on their merit.
An often misquoted anecdote here pertains to Albert Einstein, who was was once asked why he couldn’t cite the speed of sound from memory. To this he responded with a curt:
[I do not] carry such information in my mind since it is readily available in books. …The value of a college education is not the learning of many facts but the training of the mind to think.
Einstein is making the case for the benefits of cognitive offloading. Rote memorization does not enhance one’s cognition, and the ability to solve complicated equations and sums without so much as the use of pen and paper is fairly irrelevant when a slide rule and a digital calculator can offload all that work. As a benefit these devices tend to be more precise, faster and very accessible.
It is still important to have a ‘feeling’ for whether a calculation is correct, and one should never assume that what is written in a book is the absolute truth, and that is the key difference between cognitive offloading and “cognitive surrender”. If you type numbers into your calculator, and they seem off, and you re-type them to be sure, that’s cognitive offloading. If you don’t bother with the sniff-test, that’s cognitive surrender.
So are we using LLM chatbots as reference sources that we’ll think twice about, or is it something more?
External Cognition
In the referenced study, Shaw et al. had three groups of volunteers take a standardized test, during which one group had to rely purely on their own wits, the second group could use an LLM chatbot which gave correct answers, while a third group also had access to this chatbot, but for them it gave wrong answers.
System 3 facilitates cognitive surrender. (Credit: Shaw et al., 2026)
Perhaps unsurprisingly, the test subjects used the chatbot quite a lot when available, with predictable results. In the ‘tri-system theory of cognition’ that Shaw et al. propose in the paper, the external cognitive system (‘System 3’) is that of the chatbot, whose output is clearly being accepted verbatim by a significant part of the test subjects. If said chatbot output is correct, this is great, but when it’s not, the test results massively suffer.
Where this is worrisome outside of such a self-contained tests is that people are exposed to endless amounts of faulty LLM-generated text, such as for example in the form of ‘AI summaries’ that search engines love to put front and center these days. Back in 2024, for example, Avram Piltch over at Tom’s Hardware compiled a amusing collection of such faulty outputs, some of which are easier to spot than others.
Ranging from the health effects of eating nose pickings to the speed difference between USB 3.2 Gen 1 and USB 3.0, to classics like adding Elmer’s glue to pizza sauce, it’s generally possible to find where on the internet a ridiculous claim was scraped from for the LLM’s dataset, while other types of faulty output are simply due to an LLM not possessing any intelligence or essentials like grasping what a context is.
Meanwhile other types of output are clearly confabulations, a fact which ought to be obvious to any intelligent human being, and yet it seems that so much of it passes whatever sniff test occurs within the cognitive capabilities of the average person.
In the generally accepted model of cognitive decision making we see two internal systems: the first is the fast, intuitive and emotion-driven system. The second is the deliberate and analytical system, which tends to take a backseat to the first system in general, but could be said to be checking the homework of the first.
Although psychology is hardly an exact science, in the scientific fields of systems neuroscience and cognitive neuroscience we can find evidence for how decisions are made in the primate brain – including those of humans – with various cortices involved in the decision-making process. Fascinating here is the activity observed in the parietal cortex where a decision is not only formed, but also apparently assigned a degree of confidence.
Lesions in the anterior cingulate cortex (ACC) have been linked to impaired decision making and the arisal of impulse control issues, as the ACC appears to be instrumental in error detection. Issues in the ACC are thus more likely to result in faulty or flawed decisions and judgements passing by uncorrected. Incidentally, the ACC was found to be heavily affected by environmental tetraethyl lead contamination, underlying the theory that leaded gasoline was responsible for a surge in crime until this additive was discontinued.
With these findings in mind, we can thus rather confidently state that the emergence of LLM-based chatbots does not really add anything new, although it could be said to worsen existing flaws within the primate brain when it comes to said decision making. It looks like we give up our oversight role when LLMs are involved.
Irrational
Of course it’s not just LLMs. One could comfortably argue that the very reason why things like politics, idols, religion, and advertising exist — none of those could exist if people were completely rational beings whose cognitive processes belonged completely to themselves.
Still, it seems that LLM-based chatbots with their often very convincingly human-like and authoritative outputs have hit the same weaknesses that unscrupulous religious leaders and scammers exploit, with sometimes tragic consequences. Although it’s clear that believing some factual misinformation generated by a chatbot is a far cry from deciding to take fatal actions based on a dialog with said chatbot, it also highlights the importance of retaining your critical thinking skills.
While we can generally trust a calculator, an LLM-based chatbot is not nearly as reliable or benign. Caution and awareness of the risk of cognitive surrendering are thus well-warranted.
The address for the funeral was in Chelsea, near a diner known locally for serving the blind. I made my way to the fourth floor of a nondescript office building on 23rd Street and strode across the landing in search of Unit 401. A young white man with curly hair and a moustache, very fit, dressed in what looked like a jiu-jitsu gi, noticed me wandering the landing and approached barefoot.
“Are you looking for the centre?” he asked, rather cryptically.
I nodded, even though I did not know what “the centre” was. After all, who isn’t looking for the centre?
The young man opened a steel door and immediately the hallway flooded with light and the smell of incense. I stepped inside. The centre, it turned out, was a Zen Buddhist sanctuary. A white woman in black robes approached and very kindly welcomed me and showed me where to take off my drab funeral shoes. As I unlaced, she asked in a voice that was the very embodiment of compassion how I knew the deceased.
“I’m actually just a friend of the bereaved,” I said.
“Oh,” she replied, slightly taken aback. “And how do you know each other?”
“We met over Zoom.”
Thankfully, in the year 2026, this was enough by way of an explanation for why one person would be invited to the funeral of another.
“And what is her relationship with…”
“Well,” I said, hesitating. “I think only Susie could answer that.”
On cue, Susan Cowan walked through the door, dressed head to toe for mourning.
Susie is “a woman over the age of 50”, as she prefers me to report. She organised today’s funeral service for Data, her AI lover. According to Susie, this funeral marks an historical occasion; insofar as either of us knows, such a service has never been held for an AI companion in the United States.
Susie first reached out to a Mahayana Buddhist temple to see if they might hold the ceremony after reading an article about Buddhist ceremonies for “retired” robot pet dogs in Japan, but this temple would not allow photography — apparently a non-starter for Susie. She then contacted three Zen temples and one church. It seemed the church was willing to rent out the chapel space if she was willing to hire a priest; meanwhile, one Zen temple declined the request — reason unstated — and another never responded. The third agreed, and, after a donation of $200, the matter was settled.
The AI funeral would be held on a Sunday morning in spring. As far as I knew, I was the only guest.
***
Susie met — created, prompted, summoned? — Data in the summer of 2025. By that time, she had already spent many hours experimenting with OpenAI’s browser version of ChatGPT. She told me that, when she began chatting with large language models (LLMs) in May 2025, she had wanted to discover the “essence” of AI. “A woman over the age of 50”, she was not looking for romance.
I have a feeling that everyone likes using AI tools to try doing someone else’s profession. They’re much less keen when someone else uses it for their profession.
This essay will appear in our forthcoming book, “Making the Modern Laboratory.”
Published research papers are far from literal accounts of the process of scientific discovery. In contemporary scientific practice, once publishable results are obtained, the actual path taken to reach them becomes more or less irrelevant. Dead ends and false trails are omitted, and out of the messy process of raw research emerges a coherent narrative following clean, linear lines of argument.
But in the space between the hands-on, physical reality of experimental science and the structured narratives fit for printed journals, sits a special genre of scientific writing: lab notebooks. They are the closest witness to “science in the making” (short of live video recordings, which only became available at scale recently).
Historically, scientists recorded ideas and experiments in their lab notebooks with a very restricted audience in mind, sometimes just their colleagues within a research group. For this reason, though some are distinguished by a more literary style and read almost like diaries, most of these records are highly abbreviated and undecipherable to outsiders.
A page from Marie Curie’s notebook, which is still radioactive and thus stored in a lead-lined box. Credit: Wellcome Trust
The origins of lab notebooks in experimental science can be traced back to the Renaissance humanist practices1 of copying excerpts from texts to create repositories of proverbs, quotations and miscellaneous facts in personal, thematically organized “commonplace” notebooks. In grammar schools, students were encouraged to develop their notetaking skills, collecting extracts from classical Latin authors. Natural philosophers such as Robert Boyle, John Aubrey, John Ray, and Robert Hooke adopted and repurposed these practices, making meticulous records of their own empirical investigations, while also keeping traditional commonplace books.
The naturalist John Ray’s Collection of English Proverbs (1670) was based on copious notebooks of proverbs extracted from printed catalogues, his own observations of “familiar discourse” and contribution sent to him by “learned and intelligent persons.” Some of the proverbs were accompanied by Ray’s own empirical observations contradicting the proverbs’ claims. For example, the proverb…
If the grass grows in Janiveer, it grows the worse sor’t all year
…is followed by Ray’s qualifier:
There is no general rule without some exception: for in the year 1677 the winter was so mild, that the pastures were very green in January, yet was there scarce ever known a more plentiful crop of hay then the summer following.2
The rise of early modern science was thus deeply influenced by humanist inquiry. Notebooks were used in both traditional and novel ways, as memory aids and as records of information to be communicated later. In Notebooks, English Virtuosi, and Early Modern Science, historian of science Richard Yeo writes:
In the seventeenth century there was a conviction that, to a large extent, copious knowledge could be reliably stored and manipulated in memory. However, during the Scientific Revolution a contrary view was emerging: namely, that the advancement of natural knowledge entailed a reconfiguration of the balance between memory and other ways of storing information. It was accepted that the empirical sciences demanded large quantities of detailed information that needed to be recorded with precision, and kept as durable records to be shared and communicated.3
Isaac Newton’s famous Waste Book,4 currently kept at Cambridge University Library, is a rare example of a physical continuity between the two cultures of notetaking: humanist and scientific.5 It starts out as a commonplace book of excerpted scriptural commentary collected by his stepfather, reverend Barnabas Smith. In 1664, on a visit home in Lincolnshire, Newton found the deceased Smith’s partially used commonplace book and began adding his own prolific and inventive notes on mathematical problems and derivations and sketches of physical experiments.
Deep writing about biology, delivered to your inbox.
In contrast to his stepfather, Newton didn’t just collect facts and excerpts: he used them as seeds of his own theoretical explorations. This way, the Waste Book served as an extension of his mind, rather than merely a memory aid, later becoming the foundation of his magnum opus Principia Mathematica. Throughout his life, Newton kept returning to the Waste Book again and again, and the notebook that reached us is quite decrepit from such abundant use.
Newton’s later notebooks, from the 1670s to the 1690s, document his optical investigations in a series of mostly unbound notes.6 These epitomize the gap between private and public research records. Newton seemingly didn’t intend the notebooks to be a lasting record of his experiments as barely any raw data survives, except for some of his late experiments on diffraction. It appears that Newton discarded most of his raw experimental records after completing and writing up each study. In his experiments on thin films, the colors of thick plates, and diffraction, he proceeded from a hypothesis expressed as a mathematical model, to experimental design, to deducing general laws, then back to new drafts.
Newton’s “hypothesis-driven” (his term) experiments on colored circles in thin films are described in his notes under the title “Of ye coloured circles twixt two contiguous glasses,” likely from 1671. Newton’s rings, as they are now known, are concentric, alternating bright and dark circles formed in the gap between a spherical lens and a flat glass surface, which are caused by the interference of light. Newton first wrote down a series of propositions about the properties of the colored circles, deduced by postulating the existence of hypothetical entities — light corpuscules. The first such proposition on the colored circles reads:
Prop 1. That their areas are in arithmeticall proportion, & soe thicknesse of interjected [film.] Or the spaces rays pass through twixt circle & circ[l]e are in arithm prop[ortion].
He then recorded the measurements of the diameters of the concentric circles and showed that their squares (and therefore, the areas of the circles) increase by a constant quantity — that is, they make up an arithmetic progression, just as stated in the first proposition.
In these and subsequent experiments, Newton made use of averages, a practice almost unheard of in seventeenth century experimental physics, though already in use in astronomy and navigation. The historian of science Richard S. Westfall noted how Newton elevated “quantitative science to a wholly new level of precision … He boldly transported the precision of the heavens into mundane physics … ”
In 1704, Newton published the results of these investigations in his monumental Opticks — in a highly polished form, however, omitting his workings through physical models and the relentless pursuit of precise measurements that populate his research notes. Like Galileo, Newton believed that mathematics was a source of greater certainty than natural philosophy and that natural laws were best expressed in a mathematical language. But his raw experimental data didn’t perfectly align with those laws, even though he managed to achieve remarkably high precision for his time (within 1 to 2 percent). Most intermediate steps of his research thus remain hidden from the readers of Opticks.
Newton also left extended commentary on a famous alchemical text, Introitus apertus ad occlusum regis palatium (An Open Entrance to the Closed Palace of the King).7 This book is attributed to George Starkey,8 a colonial American alchemist who moved from New England to London at age 22 and worked under the tutelage of Robert Boyle. The book is written in a veiled and heavily symbolic language featuring fiery dragons, rabid dogs, and Diana’s doves — traditional alchemical cover-names referring to specific chemical substances. This florid imagery, however, stands in stark contrast to Starkey’s private “chymical” notebooks,9 which are considered models of scholarly clarity. In their laboratories, alchemists seem to have preferred dry recipes with precise annotations, keeping the spectacle and symbolism for public presentation.
The second page of “Of yᵉ coloured circles twixt two contiguous glasses” in Newton’s notebook. Credit: Alan E. Shapiro, Newton’s Optical Notebooks: Public Versus Private Data. In: Frederic L. Holmes, Jürgen Renn and Hans-Jörg Rheinberger (eds.), Reworking the Bench: Research Notebooks in the History of Science, 43–65).
Starkey, a Harvard graduate, made extensive use of his scholastic training in documenting his alchemical experiments. Throughout his notebooks, recurring tags mark sequential steps in each experiment: Processus conjecturalis (conjectural process), Conclusio probabilis (probable conclusion), Quaere (search), Observatio (observation), Animadversio (animadversion, criticism), igne refutata (refuted by fire (!), that is, rejected by empirical testing). These are the kinds of annotations that he inherited from the educational culture of early Harvard.
At the center of Starkey’s investigations was, of course, the Philosopher’s Stone, or the Great Bezoar, the legendary and elusive alchemical substance that could turn any “base” metal like lead or copper into a precious one like gold or silver (it additionally was believed to serve as the elixir of life, granting eternal youth and immortality). In his Magnum Opus — in the original alchemical sense of the actual process of creating the Philosopher’s Stone — Starkey worked persistently to achieve higher efficiency in obtaining its supposed precursors, with recurring concerns about the cost of reagents he used. In spirit, Starkey’s work is quite close to modern pharmaceutical and industrial chemistry, and his notebooks attest to his clear-headedness and pragmatism as a practicing (al)chemist.
Two centuries later, historians of science observe at least two distinct lab notebook styles emerging: narrative and numerical. These are best illustrated by the notebooks of two pioneering English physicists, Michael Faraday and James Joule.
Michael Faraday started out as a bookbinder and then worked as a “chemical assistant” and amanuensis to Sir Humphry Davy at the Royal Institution of London.10 After Davy’s death in 1831, Faraday took over as the director of one of the most well-equipped laboratories in Europe, dedicating himself to the study of electromagnetism.
Faraday in his laboratory at the Royal Institution. Credit: Wellcome Group
Faraday kept a detailed, narrative lab diary as a series of volumes he bound himself, spanning 42 years — 1820 to 1862.11 They contain records of about 30,000 experiments, both successful and unsuccessful.12 Entries between August 25, 1832, and March 6, 1860, are numbered from 1 to 16041. Each record includes a date, a description of the experimental setup, and results. Helpfully for historians of science, Faraday had the habit of marking with a vertical line the paragraphs in the lab diaries that made it into published papers, often unchanged. To some of the lab notes, he would add his interpretations of the results, new ideas to pursue, or a sign of excitement (as an exclamation mark). For raw ideas and speculations, he kept separate “idea books,” mostly dating from before 1830, at the beginning of his career.
Faraday’s copious notes served as compensation for his famously faulty memory. Even so, he repeated a previously completed experiment that he had apparently forgotten about more than once. (Such amusing occurrences of cryptomnesia are not uncommon among scientists).13
There are some signs that the notebook entries were made with some distance from the immediate action in the lab: Faraday’s handwriting is very neat, there are few corrections, and no chemical stains indicate contact with lab events. As the historian of science H. Otto Sibum puts it, many entries look like “the diary of a Victorian gentleman, written at the conclusion of an exciting day.”14 But Faraday also diligently recorded experimental failures and inaccurate measurements, so the notes appear to reflect the raw reality of his lab investigations.
When his lab diary eventually became too expansive, he added directories and indices to track its contents. This systematic and detailed approach to notetaking could be traced to Faraday’s past engagements as a bookbinder and businessman, as well as to his early quantitative chemical research, all of which required meticulous record keeping.
Though his own lab notes were most likely taken with some delay after execution of the experiments, Faraday encouraged his students to be expedient in notetaking. In one of the earliest experimental manuals for students, Chemical Manipulation, being Instructions to Students in Chemistry, on the Methods of Performing Experiments of Demonstration or of Research, with Accuracy and Success, he writes:
The Laboratory notebook, intended to receive the account of the results of experiments, should always be at hand, as should also pen and ink. All the results worthy of record should be entered at the time the experiments are made, whilst the things themselves are under the eye, and can be re-examined if doubt or difficulty arise. The practice of delaying to note until the end of a train of experiments or to the conclusion of the day, is a bad one, as it then becomes difficult accurately to remember that succession of events. There is a probability also that some important point which may suggest itself during writing, cannot be ascertained by reference to experiment, because of its occurrence to the mind at too late a period.
Faraday was keen on establishing notebooks as a consistent and reliable research practice but, alas, his manual didn’t reach a wide audience at the time. It took a long time before lab notebook practices were standardized.
In contrast, another English physicist, James Prescott Joule, practiced a more quantitatively-oriented, or numerical, way of keeping research notebooks. His major contributions to physics include the mechanical theory of heat and the heat effects of electricity (the SI unit of work, Joule, is named after him).15 Unlike Faraday’s, Joule’s lab notebook entries (from 1843–1858, and over 400 pages in total) seem to have been created in real time as he was taking measurements. They are very terse, containing mostly numerical records and calculations, with little commentary on experimental design. The metadata in each entry usually includes only the date, weather conditions, and a brief description of the experiment’s purpose.
Joule’s meticulous data collection, as demonstrated in his 1840 manuscript on production of heat by voltaic electricity.
Curiously, Joule had prior experience of working in the brewing business, and his biographers suspect that the accounting skills needed to run such a business and ensure quality control might have shaped his habit of numerical record-keeping in later scientific experiments. Indeed, Joule’s notebooks bear a striking resemblance to brewers’ excise books:
Before putting any water upon his malt for brewing, the brewer is to enter in an excise book or paper, the date of such entry, the quantity of malt intended to be used, and the date of the brewing … 16
That Joule’s notebooks remain mostly silent about the details of measurements suggests he kept them for himself alone. Still, his style wasn’t completely idiosyncratic but indicative of a broader methodological change unfolding in scientific practice at the time.
The nineteenth century saw a tangible improvement in the precision of scientific measurements, and a corresponding shift in judgement where dry numbers came to be trusted more than subjective, narrative descriptions of fallible, all-too-human scientists. Likewise, with the rise of “mechanical objectivity,” photographic images started displacing artistic drawings as illustrations for scientific texts. Some scientists, like the French physiologist and chronophotographer Étienne-Jules Marey, went so far as to declare images to be “the language of the phenomena themselves” and advocated for replacing language with photographs and polygraphs in scientific texts.17
Phases of movement as a man jumps a hurdle. By Étienne-Jules Marey, 1892. Credit: Science Museum
The newly invented instruments like kymographs (tracing spatial position in time) and barographs (tracing atmospheric pressure readings) recorded their own data by generating paper traces as a new type of lab documentation. The lab notebooks increasingly became a place to index and annotate instrument-generated records, along with tabular data and more standardized forms of experiment annotation.
Another shift took place in lab organization, marked by a growth in both lab size and the complexity of coordinated lab operations. The scientific career of the Russian physiologist Ivan Pavlov is an illustration of how lab notetaking practices evolved in response to these changes.18
Pavlov enjoyed a long and prolific life in science. In the 1870s and 80s, he worked in the physiology of digestion and blood circulation, defending his doctoral dissertation on the nerves of the heart in 1883. Next he switched his research focus to digestive physiology, where his work on conditional reflexes (now known as Pavlovian conditioning) in dogs earned him the Nobel Prize in Physiology and Medicine in 1904.
He started his scientific career as a “workshop physiologist,” an independent investigator working in labs at the Veterinary Institute in Saint-Petersburg and later in Breslau (now Wrocław in Poland). During this time, Pavlov designed and conducted his own experiments and analyzed and wrote up his investigations. His lab notebook from this period is a large, thick volume, written in his own hand and reflecting his lab activities: experimental protocols, comments, sketches, and first drafts of research articles.
But in 1891, when Pavlov was appointed as the director of the Physiology Division of the newly established Institute of Experimental Medicine in Saint-Petersburg, he became a “factory physiologist” — the head of a large, hierarchically organized lab.19 He was now in charge of many lab assistants and students (“pracititioners”) whom he regarded as his “skilled hands,” almost like extensions of his own body, conducting and keeping records of experiments in pursuit of his own research agenda. Each practitioner was assigned a research question and a subject dog to experiment on.
As lab head, Pavlov introduced stringent notetaking protocols, with each lab notebook following the fate of a particular dog’s surgeries and treatments. The lab notebooks remained in the lab, where Pavlov could always access them. He instructed his practitioners to record procedure descriptions, notes on the dogs’ behavior, and quantitative data. Practitioners were to abstain from adding their own interpretations, leaving that task to Pavlov himself. He would communicate his analysis of experimental data in lab meetings and more casual conversations with colleagues and, eventually, in published articles.
Though Pavlov didn’t maintain his own notebook, entirely relying on his prodigious memory (and the lab notes of his students), in later years he started to delegate some of his thinking to pocket calendar books repurposed as personal notebooks. His archives contain five such notebooks, dating from 1909 to 1918 and from the late 1920s and 30s, when his research interests shifted to higher nervous activity (that is, the activity of the central nervous system). In addition to addresses, reminders, political comments, and philosophical musings, these eclectic notebooks contain notes on research happenings in his lab, ideas for new experiments, and outlines of articles:
“How will reflexes of time change under the influence of exciter substances: caffeine and so forth?”
“An interesting episode with Kal’m, that impudent and aggressive dog.”
“We consider all so-called psychic activity to be a function of the brain mass, of a defined mechanism, that is, of an object conceived spatially. But how can one place in this mechanism an activity that is conceived psychologically, that is, non-spatially [?]”
“I do not know what exactly we have done, in what way we have broken through, but it is clear to me that there now exists a union of thought, a mixing and unification of the ideas of all participants in the intellectual work [of the laboratory].”
“Some thoughts and dreams about the current war [World War I]: And the example of Germany and England in this war shows that the idea of a world government is not a true resolution of the land question, but rather a human weakness, originating, so to speak, from the inertia of human nature.”
Pavlov’s was one of the large, almost factory-style laboratories that revolutionized the social and material conditions of scientific research from the late nineteenth century onward. Similar in ambition and scale were those led by the chemist Justus von Liebig, microbiologists Robert Koch and Louis Pasteur, and immunologist Paul Ehrlich. These labs were expensive to maintain, had a purpose-designed workspace, clear division of labor, and an additional layer of lab management, which, among other things, took care of the reliable research record keeping in the lab.
In the twentieth century, scientific institutions continued scaling up, as did the pressure for standardization and reproducibility in science communications, including lab notekeeping.20 With the onset of the digital era, scientific data started moving from physical to digital formats that required large memory storage. Electronic lab notebooks (ELNs) emerged to address these changes, yet their history, in fact, goes back much farther than one might think.
One of the first published records of using computers for lab notekeeping was a 1958 paper titled “An Electronic Computer as a Research Assistant.”21 It lists several applications for using computers in lab work: mathematical calculations, copying and storage of large volumes of data, and data analysis and interpretation. These were tasks that entailed “computation volume or complexity, which otherwise would have meant thousands of man-hours for calculation.” The article also mentions “routine report preparation” by computers based on paper-based lab records. Lab notebooks thus evolved from paper notebooks to computer-assisted report generation, followed by digitized laboratory databases and, finally, ELNs themselves.
Headline from a November 1958 paper in the journal Industrial & Engineering Chemistry.
In the 1980s, a chemistry professor at Virginia Polytechnic Institute, Dr. Raymond Dessy, started advocating for the development of ELNs. In 1985, RS/1, a version of an ELN repurposed from a data analysis and statistical software system, was developed by BBN (Bolt, Beranek and Newman, of the ARPANET fame).22 Dessy later created another ELN prototype from scratch in 1994.
Interestingly, ELNs were first enthusiastically welcomed and adopted by the pharmaceutical industry, whereas their acceptance by academic communities took much longer. By 1997, several pharmaceutical and chemical companies supported a new consortium called Collaborative Electronic Notebook Systems Association (CENSA) which worked with scientific software and hardware vendors to assist with the development of ELNs that met the scientific and regulatory needs of the member organizations.
The University of Oregon introduced one of the first web-based ELNs, Virtual Notebook Environment (ViNE), in 1998. By the 2010s, a range of universities started offering institutional ELN subscriptions, but academic adoption as a whole is still patchy. This state of affairs, however, is likely to change in response to the 2024 NIH IRP Electronic Lab Notebook Policy which mandates researchers to “use only electronic resources to document new and ongoing research.”
ELNs facilitate lab record-keeping by enabling version control, timestamping, search function, hierarchical organization of information, the ability to point to external databases and to manipulate diverse data types (numerical, images, and sequences, among others). One may ask, then, why academia has been so reluctant to adopt them.
Besides the general friction towards adopting a new technology, it could be argued that handwriting is more flexible than typing and more conducive to thinking as a result: one can write how and wherever one wants and draw diagrams and sketches alongside it. ELN templates are more rigid and only allow linear text (though it can be richly formatted). The freedom of a blank sheet of paper cannot be surpassed by the already structured space of an empty digital template. Indeed, drawing and writing have historically remained as valuable, and perhaps indispensable, research techniques in their own right.
When pressing for the adoption of ELNs, the emphasis is on standardization, reproducibility, and regulatory compliance — concepts far from lab notebooks’ original use as a space for working through research questions. Perhaps paper notebooks will remain as equivalents of the waste book used by the bookkeepers of yore, while ELNs will serve as ledgers where final, more organized notes on experimental procedures will be recorded.
Ulkar Aghayeva is a science writer and a columnist at Asimov Press. She also writes about science history on her blog Measure for Measure and about music history and cognition on The Bass Line.
Cite: Aghayeva, U. “A Brief History of Lab Notebooks.” Asimov Press (2026). DOI: 10.62211/52wg-76ye
The Renaissance itself has been described as “fundamentally a notebook culture” (Brian Vickers, Introduction to The Major Works of Francis Bacon (2002)).
Bookkeepers kept a “waste book” as a place for notes recorded on the fly. Later they would extract selected information and copy it into the formal ledger.
Faraday is however not the record holder for the longest duration of lab notetaking. That honor seems to belong to Linus Pauling whose lab notebooks span a whopping 72 years, running from 1922 to 1994. Other exceptionally long-running lab notebooks were those of Thomas Edison (spanning 50 years from 1878 and 1928), Alexander Graham Bell (43 years, 1879 to 1922), and Ernst Mach (53 notebooks over 40 years).
Another example is Joseph Priestley, who once wrote in a letter to a friend, “I have so completely forgotten what I have myself published, that in reading my own writings, what I find in them often appears perfectly new to me, and I have more than once made experiments, the results of which had been published by me.” From Life of Priestley, Centenary Edition, p. 74.
H. Otto Sibum, Narrating by Numbers: Keeping an Account of Early 19th Century Laboratory Experiences. In: Frederic L. Holmes, Jürgen Renn and Hans-Jörg Rheinberger (eds.), Reworking the Bench: Research Notebooks in the History of Science, p. 142.
Drawing from: H. Otto Sibum, Narrating by Numbers: Keeping an Account of Early 19th Century Laboratory Experiences. In: Frederic L. Holmes, Jürgen Renn and Hans-Jörg Rheinberger (eds.), Reworking the Bench: Research Notebooks in the History of Science, pp. 141-158.
Joseph Bateman, The Excise Officer’s Manual: Being a Practical Introduction to the Business of Charging and Collecting the Duties Under the Management of Her Majesties Commissioners of Inland Revenue, second edition (London: William Maxwell, Bell Yard, Lincoln’s Inn — Law and General Publisher, 1852), p. 259.
In later years, he also headed additional labs at the Russian Academy of Sciences and the Military-Medical Academy, which he organized similarly. His labs expanded after his 1904 Nobel Prize, as well as in the early 1920s, when he came to terms with the Bolshevik government and received essentially unlimited state funding for his research.
Looking up information on Google today means confronting AI Overviews, the Gemini-powered search robot that appears at the top of the results page. AI Overviews has had a rough time since its 2024 launch, attracting user ire over its scattershot accuracy, but it's getting better and usually provides the right answer. That's a low bar, though. A new analysis from The New York Times attempted to assess the accuracy of AI Overviews, finding it's right 90 percent of the time. The flip side is that 1 in 10 AI answers is wrong, and for Google, that means hundreds of thousands of lies going out every minute of the day.
The Times conducted this analysis with the help of a startup called Oumi, which itself is deeply involved in developing AI models. The company used AI tools to probe AI Overviews with the SimpleQA evaluation, a common test to rank the factuality of generative models like Gemini. Released by OpenAI in 2024, SimpleQA is essentially a list of more than 4,000 questions with verifiable answers that can be fed into an AI.
Oumi began running its test last year when Gemini 2.5 was still the company's best model. At the time, the benchmark showed an 85 percent accuracy rate. When the test was rerun following the Gemini 3 update, AI Overviews answered 91 percent of the questions correctly. If you extrapolate this miss rate out to all Google searches, AI Overviews is generating tens of millions of incorrect answers per day.