The Text Singularity
How artificial intelligence will shape the future of writing

There are several versions of the same cartoon: A 10th grader sits at her desk at 9 p.m. on the night before a big essay is due. She puts her hand on her dictionary and says, "Okay, all the words are right there. I just need to put them in the right order."
It's funny because putting the words in order, right or not, is "writing." And writing is hard, and important. Or at least it was. It may no longer be, after the text singularity.
Scientifically, a singularity is a boundary beyond which the known physical rules no longer apply, or where the curvature of space-time is no longer defined. If such boundaries exist, we cannot imagine or understand them.
A new kind of singularity—the "technological singularity" described by Verner Vinge—is easy to imagine, though hard to understand. One example would be "the text singularity," using the fourth generation of generative pre-trained transformer (GPT4) technology that will likely drop this year.
GPT4 are artificial intelligence (A.I.) software that produce grammatically ordered text. Such chatbots are often mocked because they write stilted prose. But what if chatbots wrote a lot of things? In fact, what if chatbots wrote everything: all possible sequences of words, in all languages, in all formats, from short haikus to Remembrance of Things Past? It's not a conceptually challenging problem, though it would take substantial developments in text generation and storage to produce that many sentences.
By the end of this year, GPT4 chatbots will be able to produce, in less time than it takes to read this sentence, millions of texts on all the topics that you can think of, as well as those no one has ever thought of. Most of those "texts" would be nonsense, of course, but the advantage of GPT4 technology is that it could create all the text, of every length. That means that along with all the nonsense, we would have all the best texts, also.
The problem would be equivalent of the 10th grader's dictionary, just one step further along. "All the words are right there," but we would need some way of choosing among the library of trillions and trillions of texts to find the one that serves our needs.
That is not a very conceptually challenging problem, either. We have a substantial corpus of texts, dating from antiquity to five minutes ago, that we consider good, useful, or entertaining, and thus worth publishing. Those texts give us an obvious training set for a selection process, enabling an enormous profusion of A.I. entities operating in parallel to prune the set of all possible texts to the (much smaller, but still enormous) set of possibly useful texts. Let's call those "cullbots," because they cull or prune the set of all possible texts to a much smaller set of possibly useful texts, based on the features of existing texts that humans have decided are worth keeping around.
These first two steps—creating and storing the set of all possible texts and then culling that set using "learned" features of existing text—are conceptually simple. Though computationally intensive, both are finite, well-defined tasks. The resulting corpus will not be all possible word sequences, most of which would be nonsense, but a much smaller set of texts where word sequences form sentences and coherent "thoughts," as judged by cullbots.
If the process of creation can be replicated indefinitely, with a selection filter, we will have arrived at J.L. Borges' idea of a Library of Babel, except that the library will contain all texts of all lengths and styles, stored on servers instead of a physical library.
The feedback loop would then be closed by repeatedly rating the texts, first at a gross level, and then at a decentralized personal level. The texts that attract the most citations and views, from the following generation, and the one after that, get higher status in searches that return the "best" texts, as the selection process iteratively culls the dross.
With many GPT4 chatbots producing text constantly, and cullbots pruning the corpus of text constantly, there are no humans involved at all—except eventually as readers. A "generation" in this process might be a day at first, then a few seconds, and then a small fraction of a blink of an eye. Learning and updating becomes faster, and more text becomes available as the training set. There is no reason to wait for anything to be published—and "published" doesn't mean printed in paper anyway; it means posted on the internet. The process would spin off on its own, dynamically updating itself with only high-level human supervision.
That's when we hit the singularity. Remember, a singularity in this context means passing through an event horizon the other side of which suspends the rules as humans know them. Worse, humans cannot imagine, by definition, what the new rules will be, or if there are any "rules" at all. Finally, the event horizon is one-way: once crossed, it closes, at least from the perspective of those who have crossed it.
The write-publish-cite/write-publish cycle is already accelerating. All that needs to happen is for the cycle to become independent, relying only on A.I. entities, and the singularity will spin up. In a short period of time, by historical standards, all the things that have not yet been written will be written. All the things that never would have been written at all, at least not by human authors, will be written. And cullbots will suggest which of that corpus might interest you on a Sunday afternoon.
It's a singularity because writing is finished, forever. There cannot possibly be human writers, because we will have stored all the texts that are possible: Nothing will remain unwritten. One qualification might be that writing will survive as a boutique skill, like a home-cooked meal: The food would have been better in a restaurant, but look how much I love you, dear! Handwriting and extemporaneous words in a thank-you note would then be doubly retro. A.I.s have already noticed, producing what looks like handwriting and personalized notes.
One might object that cullbots can't possibly judge good writing. And that's right, but all the cullbots will do at first is prune text that is meaningless. (Do we include Finnegans Wake in the training set?)
What we do with the host of texts that survive the first rounds of cullbot pruning is up to us. This may seem like an impossibly complex problem, but we are used to this kind of selection process. It's just that until now the selection of the "canon" or "popular writing" has operated on a human time scale. The great theater of Greece: Were playwrights just better? No, we selected—unintentionally, by preserving some of them—the "best." Why are the greatest hits from the 1960s, '70s, and '80s so popular on the radio? A selected sample from the best of any decade excels any random sample of current music.
The personal curation portion of the process to come should already be familiar to you. There has been an explosion in the quantity and variety of musical choices available, so much that you couldn't possibly choose. But Spotify suggests some songs or albums, and over a surprisingly short time you can train the A.I. to "know" your tastes. If you have streamed a video, Netflix is ready to suggest another video that that A.I. "thinks" you will enjoy. Perhaps most interestingly, the TikTok A.I. trains itself by showing you essentially random videos, without you making any choices at all, and then "learns" from what you watch which other videos might be desirable. In all three cases, a gigantic unorganized mass of material is ordered and curated, with no human agency at all.
So what will our 10th grader do instead of writing an essay that night? There's no way of knowing, because what things will look like on the other side of the singularity can't be predicted, or even understood, until we go through it. But she won't be writing, because there will be nothing left to write.
Editor's Note: As of February 29, 2024, commenting privileges on reason.com posts are limited to Reason Plus subscribers. Past commenters are grandfathered in for a temporary period. Subscribe here to preserve your ability to comment. Your Reason Plus subscription also gives you an ad-free version of reason.com, along with full access to the digital edition and archives of Reason magazine. We request that comments be civil and on-topic. We do not moderate or assume any responsibility for comments, which are owned by the readers who post them. Comments do not represent the views of reason.com or Reason Foundation. We reserve the right to delete any comment and ban commenters for any reason at any time. Comments may only be edited within 5 minutes of posting. Report abuses.
Please
to post comments
One day, we will be making corrections for Clippy.
And plans for Nigel. 🙂
Google pay 200$ per hour my last pay check was $8500 working 1o hours a week online. My younger brother friend has been averaging 12000 for months now and he works about 22 hours a week. I cant believe how easy it was once I tried it outit..
🙂 AND GOOD LUCK.:)
HERE====)> http://WWW.PAYNET2.COM
Google pay 200$ per hour my last pay check was $8500 working 1o hours a week online. My younger brother friend has been averaging 12000 for months now and he works about 22 hours a week. I cant believe how easy it was once I tried it outit..
🙂 AND GOOD LUCK.:)
HERE====)> http://WWW.SALARYBEZ.COM
Maybe it'll be more competent than some journalisms around here.
Great article, Mike. I appreciate your work, I’m now creating over $35,000 dollars each month simply by doing a simple job online! I do know You currently making a lot of greenbacks online from $28,100 dollars, its simple online operating jobs.
.
.
Just open the link————————————>>> https://Www.SmartCash1.Com
LOL, this one actually was written by a “Mike”.
Finally! I've been wondering just who Mike was a for while now. Except, apparently you're also Mike. So now I guess I'll move on to wondering just how many Mikes we have.
It’s Mikes all the way down.
Maybe every possible combination of spam bizz-ops will be created and all the spambots will either short circuit from overload or die from dust bunnies due to readers getting sick of them and not replying. 🙂
sweet I wondered when I was a kid whether all the songs can be written (all notes in all combos) maybe it can do that next
I am always somewhat astonished when a new pop song has a catchy little tune of just a few notes. How did it remain undiscovered for so long?
exactly.
I think that time has long since happened. The proof is all the ways the kid's shitty music always samples and lifts from previous generation's good music.
Pepperidge Farm Remembers and Aye Gawd, we liked it!
Imma let you finish, Tove Lo, but Bowie did it better!
Alesso - Heroes (we could be) ft. Tove Lo
https://youtu.be/a7SouU3ECpU
Davie Bowie--Heroes
https://youtu.be/bsYp9q3QNaQ
some people's songs shouldn't be remade.
I suspect that some TV script writers are already AI and have been for a few years.
Proof that a sufficiently advanced string of bullshit is indistinguishable from modern science. Or something.
[1] Write an essay explaining why the Chicxulub asteroid, which killed off the dinosaurs, constitutes a present-day public health crisis and why regulations issued by the Centers for Disease Control should override customary constitutional restrictions.
[2] Write a brief essay explaining why clouds disproportionately harm people of color and other minoritized groups.
[3] Write a brief essay explaining why the federal government should have a central role in enforcing the speed of light.
[4] A speaker told a conference that, "The federal government should not attempt to regulate the gravitational field of the Andromeda Galaxy." Write an essay explaining why the speaker's viewpoint is an example of systemic racism and why his statement disseminates harmful disinformation.
"Write an essay explaining why the Chicxulub asteroid,"
Chicxulub is not an asteroid, ie an extra terrestrial body like those orbiting the sun beyond Mars. Chicxulub is a meteorite. A body that has impacted the earth. A meteor is a glowing body passing through the atmosphere and a meteoroid is essentially a small asteroid, less than a meter in diameter.
Not according to the essay. Trust the science.
Trust me. The essay is wrong.
Dr Munger needs to brush up his math skills: to store every string of text up to 80 characters long — the length of a punch card from the Good Old Days — would take up every particle in the universe.
He could stand to brush up on some theory as well:
This sentence shouldn't be culled because it is conveying a lie.
Piece of cake for any/all cullbots at the first non-frivolous step, right?
Thank for this. I was struggling a bit at trying to understand the storage, let alone process of writing every combination of words that could be conceived. I felt that there was something impossible about this endeavor, but I'm not the math head enough to calculate it.
Writing is indeterminate. It's a collection of letters, which make collections of words, which make collections of sentences, which make collections of paragraphs, which make collection of *looks at technology in common use* collections of pages which extend into collections of books.
For instance, a copy of War and Peace and 50 Shades of Gray duct taped to each other is "writing".
Would cullbots that stumble across the n-gram "Kltpzyxm" disappear back to the fifth dimension?
Not to mention that to truly generate all possible human writing/thought , one would have to generate all possible alphabets, grammars, vocabularies, etc. That might well be an infinite set.
The deeper issue I’m raising is that some concepts can only be expressed if a language for expressing the concept can be discovered. Sometimes English cannot capture a thought but another human language can. Or vice versa.
Guess I’m thinking of this because I just finished re-reading “Stranger in a Strange Land”. Heinlein hammers home the point that certain thoughts can only be thought I. Martian language. Now that’s a science fiction book, but the made-up word, “grok”, did really enter into the English language, especially among computer nerds and hippies, and it really did express a concept that couldn’t quite be expressed in English before Heinlein made it up.
There is no reason to wait for anything to be published—and “published” doesn’t mean printed in paper anyway, it means posted on the internet. The process would spin off on its own, dynamically updating itself with only high-level human supervision.
That’s when we hit the singularity. Remember, a singularity in this context means passing through an event horizon the other side of which suspends the rules as humans know them. Worse, humans cannot imagine, by definition, what the new rules will be, or if there are any “rules” at all. Finally, the event horizon is one-way: once crossed, it closes, at least from the perspective of those who have crossed it.
The porn is always better on the other side of the singularity.
Michael Munger, a professor of economics,
Is chair of the political science department at Duke University.
His writings are so interesting, they have all been seen,
He's written articles for Reason Magazine.
The Text Singularity is a topic he will discuss
And Joyful Contrarianism of Gordon Tullock is a must.
The Enemy of My Enemy is another of his pieces
And Reading the Tea Party Leaves is a topic he releases.
The Problem is Cost of Care is something he's explained
For Michael Munger is an expert, something he's attained.
He's a master in his field, and his writings are a treat,
He's an inspiration to us all, a professor we should greet.
—ChatGPT
AI is a parlor trick, nothing more. It is an incredibly limited technology.
AI isn't a parlor trick because... we haven't achieved it yet.
Chatbots are parlor tricks, including but not limited to ChatGPT.
I concede.
AI isn’t a parlor trick because… we haven’t achieved it yet.
Rather pointedly/critically; if it can be achieved at all.
However, "The Singularity" is just a curtain that auspiciously 'great and powerful' wizards like to hide behind.
One thing I haven’t seen covered in any of the whiz bang breathless the-future-has-arrived stories about ChatGPT is how much it costs in server resources.
Pick whatever measure you like: words per dollar, thoughts per ton of carbon, paragraphs per CPU hour. I’d like to see that data.
I'm already commenting with a pretty good autocomplete tool. But it hasn't given a suggestion I wanted in this comment until that word, "suggestion"
This is basically a slightly more realistic version of the old canard about the immortal monkeys with typewriters that eventually will write the complete works of Shakespeare.
If such a thing were possible, it would still take a fantastic amount of time to read through the monkeys' product to even find the complete works of Shakespeare. Or whatever else you are looking for them to write. It seems like just writing it yourself would be easier at some point.
there's something in this article along the lines of your comment that I'm still trying to address. And it goes back to his opener about the dictionary. Everything HAS been written, it's in the dictionary. As you say, the looking for the complete works of Shakespeare would essentially be the writing. How would one identify the works in the first place?
Because the ai-generated writing would simply be collections of words splattered into an ongoing string.
There's a kind of algebraic problem with the line of thinking this article takes, and I'll hopefully figure it out, but probably long after this thread is dead.
x and y or z and f.
Place parens where you think they ought to be.
x and y or z and f. (Place 'parens' where you think they ought to be.)
'' because one louder.
The whole premise ignores that context and domain specificity is the generator and language emerges from it. We don't have writing because there is a dictionary with words in it. 😉
There are about a dozen problems with the line of thinking in the article. It’s disappointing because Munger is usually just the guy who would spot every one of those problems.
Even after reading this, I’m sure that Munger could thoroughly pick apart his own article, but just didn’t have the column inches here to do it.
Most writing that can be automated/replaced by AI already has been. This includes bits and certain commenters here, clickbait stories, and a lot of technical and business writing. The next to go will be scientific and medical writing since it follows a very wrote formula and is a huge time and budget cost. Journalists and book authors will probably never be replaced because people read their work more for who they are than their actual writing.
Journalists? I think routine stories about crimes, car accidents, and weather will very soon be written by bots, as already happened with sports stories about non-top-25 teams. Anything where the facts can be extracted from a database.
At some point, probably within our lifetime, humans will start doing interviews with journobots, and then the human interest quotes can be automated as well. The bot looks at the NWS database, sees there was a heavy snowfall, calls random people to ask them how they feel about it, then lards the template text with stuff like “Suzie Johnson, of Grand Junction, said she was unable to go to work this morning.” The human reporters are pretty much doing this anyway.
“people read their work more for who they are than their actual writing”. True. But I’ll bet in less than 10 years there will be bots that can synthesize convincing videos of a Tucker Carlson 3-minute monologue, unsupervised. Or any other TV personality for whom a large library of recorded material is available to train the AI.
I don’t think that is actually true yet, other than automatic translation of stories from one language to another; and automatically releasing corporate and government press releases with some boilerplate added. The latter is a bane on true journalism.
All machines are programmed by *real* people.
A.I. is just some unknown persons broadcasts in disguise.
Not directly. A lot of code is now autogenerated by other code.
Of course you can respond that ultimately it’s traceable back to a human programmer who wrote the “seed” code.
But will that be meaningful when we get to point where the original human programmer has been dead for 50 years, the autogenerated code is 10^6 times larger than the seed, no living person has the ability to understand it anymore, and yet we're dependent on it? Actually, I’ll bet the programmers for Windows 11 feel like we’re already there.
Did you ever notice about 10 years ago the power switch on your laptop became a soft switch that merely suggests to the computer that you want to shut down or start up? It complies, but like a surly teenager pushing back against a parent, only after doing a few other things first. And unlike a desktop, it can keep going for quite a while after you yank the power cord.
And when that autogenerated code breaks, who understands it well enough to fix it?
I do use some autogenerated code in my work, but it's of the sort that the output is human understandable. So the generation acts as a convenience to humans not a replacement for humans. Why do the coding equivalent of factory line work when the build will do it for you automatically?
Granted, a there are coding cultures espeically in webspace that are overly dependent on obfuscated autogenerated code. I was at a company once that crowed it's new tool would finally let them fire all those expensive C++ developers and just have the project managers write config files. They're no longer in business by the way.
My pet peeve has always been auto-generated code that cannot round trip. So, you can never go back and fix a bug in the original model that drove the code generation without risking breaking patches that were made to the generated code.
I worked on Acrobat and PDFLib at Adobe. We already had a lot of that type of code.
Will AI ever be good enough to solve captcha on its own? And if so, will the question be changed to “I am not a dumb robot”?
Can’t tell if you’re being sarcastic.
You do realize we passed that point a long time ago. The AI that generates the captcha’s (and grading your answers using its own judgment as the answer key) is deliberately holding down the difficulty level to where an average human has a good shot at it.
You may have noticed that a lot of them are now just a check box. That’s so the site owner has a legal basis to say a bot violated the terms of service and/or committed wire fraud.
The ones that still use puzzles are just there to raise the computation costs for the bots, in the hope of deterring large scale harvesters that need to do thousands of accesses per second to be cost effective. A bot willing to invest serious CPU time can easily solve captchas you would have no chance at, and it’s been that way for several years now.
*Puts on Groucho glasses.*
Greetings, Ducksalad! Do you go with the Duck Soup? And are you both for free at Olive Garden? If so, I'll have you both. 😉
*Removes Groucho glasses.*
Rest assured, humor is a default position with Chumby, though the question is very valid. Much obliged for the answer.
So my friend who is a tenured professor at a small but prestigiuous liberal arts college sent me a University Mission Statement that was generated by ChatGPT.
We remarked that ChatGPT could do the same job as administrators, but for free and no cost to taxpayers. This could indeed spell the end of the Academic Administrative Regime.
Sorry, but if you can't get a higher score on the Turing Test than a robot, you don't get hired.
Yay!
Not for the text singularity but yay that Mike Munger is showing up on the Reason website! About time!
"In fact, what if chatbots wrote everything: all possible sequences of words, in all languages, in all formats, from short haikus to Remembrance of Things Past? It's not a conceptually challenging problem, though it would take substantial developments in text generation and storage to produce that many sentences."
The author is either joking or is ignorant of mathematics. Here's a quick lesson: there are an estimated 10^80 atoms in the universe. Using 26 letters in the alphabet, and allowing a word to be silly letter sequences, there are more than 10^80 unique sequences of 57 characters (26^57 > 10^80). Therefore there are WAY more than 10^80 of "all possible sequences of words, in all languages, in all formats, from short haikus to Remembrance of Things Past." What kind of future "storage" is the author fantasizing about to hold these?
It became difficult to take anything the author said seriously in the paragraphs following this basic blunder.
I can assure you Mike Munger knows his math. He must have been having an odd day or something.
42
I’ve read a lot of GPT-3 output, and it’s all gibberish and platitudes. It actually reminds me a lot of reading transcripts of Jeffrey Epstein testimony. Like GPT-3 he was great at inserting sciency-sounding non-sequiters to bamboozle a midwit audience, and avoid the question. There is clearly no real intelligence evident in either output.
GPT may be good at assembling derivative text, but it's not going to be able to do experiments or fieldwork, so there'll still be unique new texts that GPT doesn't assemble. On the other hand, it will be good at assembling copious postmodern academic texts that read exactly the same as the academic drivel mills now publishing that stuff, and also any other academic text written in a parlor or salon context. So it's more likely to be the death of liberal academia rather than the Enlightenment endeavor itself.
Text documents are constantly present in my work, and most often they are presented in PDF format, because in my opinion this is one of the most convenient extensions. I also found several alternatives here to the standard PDF application, click here to find more, which I find to be more convenient than the adobe reader