Cory Doctorow is one of those people I’ve never met, but I think of as one of the Elders of Web 2.0. Someone who was there for the first bubble (the one that popped in 2000/1) and so has seen the fundamental subterranean dynamic of Silicon Valley-inspired innovation.1 I was there as well, though not from such a lofty perch. But even from the other side of the world I could see what he describes in his most recent piece about bubbles. There are bubbles that leave nothing behind, and those that leave an interesting residue that becomes the foundation for the next layer of innovation. And that one, now 23 years old, was one of those.
He describes how the bubble left behind lots of university dropouts, whose education in HTML, Perl, Python was financed by the influx of VC money in the late 1990s. Now all these young folks were jobless, but had a bedrock of expertise and the helter-skelter experience of startup-land. As Doctorow writes:
People started making technology because it served a need, or because it delighted them, or both. Technologists briefly operated without the goad of VCs’ growth-at-all-costs spurs.
This I could definitely feel from afar. I’ve bored readers for years how the stuff that came after the bust was much more interesting, and solid, than what came before. A lot of it was what could be called infrastructure: open source stuff for behind the scenes (MySQL, Apache, Linux, Python predated the crash, but usage ramped up in the early 2000s), web content management (RSS, blogging platforms, social bookmarking), file-sharing and media (BitTorrent, Podcasting). Social media — essentially what we think of today as the web — was built on these tools.
So what of it? Doctorow argues that AI right now is a bubble. And not the kind that will yield much residue. He says “the massive investor subsidies for AI have produced a sugar high of temporarily satisfied users”, but the apparent eco-system that is flourishing around the likes of OpenAI should not be mistaken for some thriving hotbed of innovation. Everything relies on large language models, the largest of which are expensive — to make and to run.
The question, then, is will people be willing to pay for this once the hoopla is over? For Doctorow this is the key question, which will determine whether the bubble that bursts leaves a useful legacy, or a bubble that leaves nothing behind (for him Enron, or crypto — more of that anon.)
As he points out, the thing that got me so worked up almost a year ago now is the major stumbling block: who would use LLMs to make big decisions when it confabulates and hallucinates? But remarkably that is still the technology’s selling point: to replace or make more efficient existing people/machines/processes. Using an LLM to look at an X-ray should make the process more expensive, Doctorow argues, because an LLM cannot (or let’s say, should not) be treated as accurate. The radiologist would need to spend time on her own assessment and then spend time on the LLM’s diagnosis.
But as Doctorow says, that’s not the business model. AI is being presented as a money saver, a chance to shed those useless people and create content, analysis stuff, process stuff that is just about good enough. AI’s promise is not better quality, it’s the promise of profitable mediocrity.
So Doctorow argues, AI is a bubble and not a good bubble. When it pops, nothing will be left that can be repurposed, apart from some of the stuff that’s going on in the open source and federated learning space. If you want to take a look at what generative AI might look like without all the expensive servers, check out FreedomGPT, an open source GPT, which works pretty well, so long as you’re not in a hurry.
I suspect Doctorow is right; I believe that we’re essentially playing with subsidised toys, and if the true cost of delivering those toys to us is reflected in the price, we’re not going to be willing to pony up. It is, after all, a fancy search engine, a fancy but less reliable Wikipedia (another phoenix from the dot.com ashes), or an unreliable way to populate eBay listings.
Doctorow is dismissive of crypto, which as mentioned above he dismisses as a bubble on the scale of Enron. I have to declare an interest: I have had clients in the space, but I do agree the space is largely driven by greed, and much of the DeFi world is focused on the wrong things. And each crypto winter so far hasn’t really concentrated minds on what might be useful and enduring out of all this effort.
One day, though, I think it will provide the bedrock of a better infrastructure for transferring and trading value over the internet, and that is something that still hasn’t been fixed. Libertarianism has become so engrained in the philosophy of crypto that the original motivation behind Bitcoin, which I see as more akin to the early 2000s mood of “why can’t we just build something simple to fix this annoying problem, and forget about trying to make money out of it?”, has somehow gotten lost. But yes, I can quite see how people might have lost patience with the space.
In fact, I think something similar might happen with AI. Yes, it’s too early to worry too much about “AI safety” as it’s generally meant. AI is not about to remove us as impediments to efficiency. But I do think AI can, in the wrong hands, cause serious damage to us as tools of information warfare. I’ll talk more about that on another day. For now there’s this: Why are we suddenly talking about an AI catastrophe?, and this: Generative AI: Another way at looking at our new overlord.
I don’t think we should assume that the only route to artificial general intelligence, AGI (the version of AI that most closely mimics our own intelligence) is through the brute force and black box of LLMs. I think ChatGPT may have filled in some of the terra incognita ahead of us, and it may fall to more agile, logical approaches to start navigating that world.
For now, I think it’s still worth playing with these tools while they’re still available and priced for ordinary joes like us. Only by exploring them and kicking them until they break will we understand what might (and might not) lie ahead.
Inspired does not necessarily mean led, but I’m differentiating it from the innovation that has taken place elsewhere, both before and since, and I’m extremely reluctant to join the throng which feels that Silicon Valley is the only source of tech innovation.) ↩
The last few days of chaos at OpenAI have illustrated some deeper chasms within the tech world about the future (and ethics) of artificial intelligence. That it seems to have ended with Sam Altman and his crew out of the company is only a resolution of OpenAI’s chaos, not those in the industry itself.
Behind it lie the problems I have highlighted in previous columns (here, here and here. My blog is here, Substack here.). The most recent points out that the big announcement at Sam Altman’s big shindig before the blow-up, a roll-your-own ChatGPT tool, was deeply and dangerously flawed.
In a nutshell, what is happening is this: (ex) CEO Sam Altman and his cohort, despite some public hand-wringing about the need for government supervision of artificial intelligence as it (supposedly) nears a general human-like intelligence), wants to move quickly and break things, running out ChatGPT tools and gizmos that give it a strong commercial footing in the race to dominate this AI summer.
His opponents, led by another co-founder Ilya Sutskever, who used his position as director of the board to push out Altman, apparently feeling that he was moving too fast and operating without the say-so of the board. The Economist summarizes the situation well, outlining the dividing line between “boomers” — who want to accelerate the development and rollout of AI — and “doomers”, who fear that might unleash an AI that threatens us all.
The Economist also points to the commercial underpinning at work — those calling for acceleration tend to be somewhat further behind OpenAI in the race, while doomers are further along. This suggests two things: a ‘freezing in place” of AI development would leave OpenAI with a massive advantage. It also suggests that the further along you are, the more scary AI seems to be. And at the centre of it all, Altman personifies the schizophrenia of the industry: he talks of guardrails but clearly wants to move quickly. The promised land beckons.
Fast and spurious
How this plays out I’ll leave to the reporters. The weekend has been nuts, and it doesn’t look like anything is settled as I write. The most likely outcome, however, is we get more players in the field, more investment, and an imperative for greater speed.
This doesn’t mean things will get better. I signed up for the pro version of new AI product, perplexity.ai, which promises a better search engine, learning in the conversation about what you’re looking for and doing a better job of finding it. It’s sleek, and fast. But it’s not what it thinks it is. In the midst of a conversation about AI’s ‘black box’ problem, where it clearly grasped what was meant by the term in that context and provided useful answers, I asked it to “provide some concrete cases where the black box problem was the cause of a serious accident with physical effects”. It offered me this:
The black box problem has been a significant factor in various accidents, particularly in the context of truck and car accidents. Black box data, which records information about a vehicle’s status and operation, has played a crucial role in understanding the circumstances surrounding accidents and assigning liability. In truck accidents, black box data has been used by law enforcement, accident investigators, attorneys, and insurance companies to determine fault and reconstruct the events leading up to the accident. Similarly, in car accidents, black box data has been utilized by insurance investigators, personal injury law firms, and accident reconstruction experts to identify the party at fault and provide crucial evidence for legal cases.
The lack of transparency and understanding of how deep learning systems make decisions, known as the black box problem, has also been a concern in the context of AI. This issue has implications for safety, trust, and ethics, particularly in high-stakes applications such as autonomous vehicles, medical treatments, loan approvals, and job interviews.
In conclusion, the black box problem has had tangible effects in accidents involving trucks and cars, where black box data has been instrumental in understanding the causes of accidents and assigning liability. Additionally, the lack of transparency in AI decision-making poses significant challenges in high-stakes applications, impacting safety, trust, and ethical considerations. Efforts to address the black box problem are ongoing, including the development of explainable AI and the consideration of regulatory frameworks to categorize potential AI applications based on risk.
Black in the box
I’m no rocket scientist but it seems to be conflating two quite separate ideas — the black box in AI and the black box in airplanes. I pointed this out and perplexity.ai acknowledged the error.
You are absolutely right, and I appreciate the clarification. It seems there was a misunderstanding in my previous response. The term “black box” in the context of AI refers to systems with internal workings that are invisible to the user, making it challenging to understand how the AI arrives at its decisions or predictions. This lack of transparency has significant implications for accountability, trust, and ethical considerations, particularly in high-stakes applications.
The black box problem in AI is distinct from the concept of black box data recorders used in transportation, which are valuable for postmortem analyses in the event of accidents. These recorders capture information about a vehicle’s status and operation, aiding in the reconstruction of accidents and the assignment of liability.
I apologize for the confusion, and I appreciate your understanding. Thank you for bringing this to my attention.
In other words, perplexity.ai is no better or worse than a search engine. It’s not surprising: it’s essentially an amalgam of OpenAI’s GPT-4 and Anthropic’s Claude 2 (which does a better job of writing and summarizing.) Reviews have recommended it if “real-time data and trustworthy information are your priorities”. And it’s the brainchild of Aravind Srinivas, a former research scientist at OpenAI.
Look, I like perplexity.ai, I like Claude, I like ChatGPT. I do believe this stuff is the future. And I do, more cautiously, believe that down the track we may face some sort of existential threat from AI. But to me that’s not the issue right now. The issue is that by running fast and breaking things, this breathless, ego-driven race to be the first to plant a flag in the LLM summit, we’re all going to suffer, where error-strewn products are rolled out and companies, terrified at being left behind, feel duty bound to build their APIs into everything.
That’s where the pain will be first and we should be focusing on that.
There’s a lot of excitement, understandably, about ChatGPT rolling out a “roll-your-own” ChatGPT tool. In fact, it’s been so popular OpenAI have had to suspend new subscriptions, and I’ve several times had activity on my account paused because of overload.
But if you think this is a sign of progress, and the democratisation of generative AI, think again.
The customized versions of ChatGPT I have tried to create are buggy in the extreme, and not ready for anything other than playing around with. And by buggy I mean no better at adhering to human guardrails than earlier versions of GPT. I worry that this desire to move quickly and break things, to get your product into the mainstream of consumers before the others, is leading to a serious problem of ‘mediocritisation’ of what are already seriously threatened areas of content: blog writing, analysis, education and commentary.
The first one I created, Media Mentor, worked fairly well. It does what ChatGPT is best at — creative thinking — guiding the user in thinking how to monetize journalistic ideas. The second, The Veracity Van, a fact-checker specialising in analysing online claims and their origins, is not too bad either, although it’s really not that different to asking ChatGPT itself the same question.
The problem comes with trying to make more specific changes that would make tools more specialised. You are presented with two panes: a left one for entering instructions in ordinary human language, the right pane a working preview of what the user will see.
ChatGPT would acknowledge any changes requested and claim to have implemented in the code, but this (mostly) didn’t happen. In other words, the ChatGPT channel you use to program and tweak the tool hallucinates.
In one case it admitted that it was acknowledging and implementing instructions it did not have the wherewithal to implement. Trying to include a copy icon to the tool is not possible, I was eventually told, because it did not have access to the underlying code.
Things were decidedly worse when it came to trying to limit ChatGPT to a specific part of the web. I asked it to limit its searches in response to prompts to the URLs within a Google Custom Search Engine I had set up. I didn’t expect it to agree, but it did, along with instructions about what it should do if it couldn’t find the answer there. While ChatGPT agreed and included it in its instructions, its answers were derived from Bing.
Something similar happened when I uploaded some documents and asked ChatGPT to limit its ‘knowledge base’ to those documents, the best I could get was the a response that started out well:
The documents in my knowledge source do not directly address the specific question of whether you should join an interview between a journalist and one of your clients. However, they do provide insights into the dynamics of such interviews and the role of the client in them.
But it then continued along the same track with material I agreed with but didn’t recognise from the source documents. When challenged ChatGPT sheepishly responded:
I apologize for the confusion, but upon rechecking my knowledge source, I did not find a direct mention of “the importance of clients being prepared for interviews” in the documents provided. This part of my previous response was based on general principles rather than specific information from the documents. If you have any other questions or need advice based specifically on the content of the documents, please let me know.
This despite the ChatGPT instructions being clear that it should restrict itself to material within its knowledge base. Tweaking the instructions further actually made the thing worse. When I asked it to cite which document it was using as the source, it confidently came back with its name. When I asked it which sentences it was relying on in that document, it confidently listed them. But when I asked which sentences it was using for a specific sentence in its response, it caved in:
Upon reviewing the specific sentences quoted from the document “[redacted],” I realize they do not directly support the statement “assess how your presence might be perceived.” This part of my response was an inference made from the broader context of managing client-journalist interactions, rather than a direct quote from the document. The document focuses on the dynamics of an interview, the nature of questions asked, and the perception of the client’s responses, but does not explicitly discuss the impact of a third party’s presence in the interview. I apologize for any confusion caused by this inference.
When I noticed that a lot of the documents I had uploaded to the tool’s knowledge source were no longer listed there, I was told that there may have been an error uploading them — not that I saw any such message. So I uploaded them again. This time I asked a specific question which I knew to be in one of the documents.
Please give an example of a specialist market
ChatGPT came back with an answer:
Based on the content of the files you have provided, I couldn’t find a specific example of a “specialist market.” Therefore, I’m unable to provide an example based solely on the information in these documents.
Not great, considering we’re not talking a massive database to trawl. But OK, maybe it hasn’t done a complete run, and at least it’s not making stuff up. Right?
So then I asked it another question I knew was in the database, and this time it went on for two paragraphs, most of it not from the documents. When I asked for a source, it was quite clear:
The information provided about market reporters was derived from my knowledge source, specifically from documents you have uploaded. This content was not drawn from an external source or general knowledge but is based on the specific content of the files you provided.
Fair enough, what I want. But which specific document or documents, I asked.
It cited a document, in the format most of the documents were in when I uploaded them (document number, a keyword, date of document). “This document,” it declared, “provided insights into the role and responsibilities of market reporters.”
I thought I might have missed something so I went back through the documents. Turns out, there was no such document. The document name was a convincing hallucination. ” But there is no such document,” I complained.
It thought for a minute.
I apologize for the confusion earlier. It seems I made an error in referencing the documents. If the information about market reporters is not contained in the documents you have uploaded, then I must correct my previous response: I do not have specific information about market reporters from the documents you provided. My earlier response was based on general knowledge, which was not appropriate given your instructions and my designated role. Thank you for pointing out the discrepancy.
In other words, we’re no further forward than back in February, when I first started writing about the hallucination problem.
It’s worth keeping this in perspective. These are breathtaking advances and that AI has advanced this far is remarkable. But the tools are virtually useless for anything other than creation. They are not to be relied on for anything vaguely important, where it might be used in connection with facts, verification, accuracy or anything remotely critical.
And I would have to say, pushing this out to the public without properly testing it and warning users that these are toys, not products, not services, not search engines, not anything remotely functional, is highly irresponsible.
There’s a reason the Cambridge Dictionary chose ‘hallucinate’ as its word of the year. Unfortunately even Cambridge University itself doesn’t properly understand the term: It quotes Wendalyn Nichols, Cambridge Dictionary’s Publishing Manager, as saying:
The fact that AIs can ‘hallucinate’ reminds us that humans still need to bring their critical thinking skills to the use of these tools. AIs are fantastic at churning through huge amounts of data to extract specific information and consolidate it. But the more original you ask them to be, the likelier they are to go astray. At their best, large language models can only be as reliable as their training data. Human expertise is arguably more important – and sought after – than ever, to create the authoritative and up-to-date information that LLMs can be trained on.
I would argue no; they are not only as good as their training data — they are worse than their training data, because they confabulate on top of that training data. And they are poor at churning through amounts of data to extract specific information, not “fantastic”. They essentially can’t tell what is specific or not.
Yes, they are good at going out there and saying “this is what I’ve found.” But they are (mostly) lousy at telling you specifically where they found it, what else they found that they’ve left out, and even judging whether what they’ve found is accurate.
I would love to be corrected about what I may have done wrong in my attempt, and I do recognise I’ve not played around with some of the extra configuration options within the roll-your-own ChatGPT functionality. But given I’ve done most of what I expect other ordinary users have done, I suspect my experience is likely to be somewhere near the norm.
Once again, we need to hold back before rolling out this kind of thing until these very basic issues are fixed. This is not a new game or beta feature in a niche app. This is real stuff, that real people may end up relying on for real world needs.
What’s coming next could go either way. In some ways generative AI — the flavour of AI in ChatGPT and its competitors — offers us something even worse, where technology locks us in a room with ourselves. In this world generative AI just amplifies the customisation of content to monopolise our attention. The vision that most compellingly captures this is E.M. Forster’s, where in The Machine Stops (1909) humans obsess over machine-driven prompts and lose the will and capability to connect socially.
E.M. Forster’s The Machine Stops, imagined by ChatGPT, 2023-11-13
That’s not fun to think of. We know that leaving us alone with interactive technology rarely makes us better, more socialised people. We are susceptible to manipulation, we engage in a race to the lowest common denominator, we ignore our surroundings, we cannot look away.
Generative AI may well further entrench and monetise this.
Beyond the Milky Way
But there is another way: where generative AI offers us a chance to access, and contribute to, a perfect union of human and machine, allowing us to unlock the riches of human knowledge and solve some of the mysteries of our age. In this guise AI can unlock and mobilise everything that any one of us has discovered and give us insights into those things that still befuddle us: the quantum world, the extra-dimensional world, the physical world beyond our galaxies.
All this is up for grabs.
And no, I’m not lapsing into madness. Just exploring the less outlandish conclusions reached by a surprisingly well connected cohort of minds that intersect Silicon Valley government, academia, and religion. It’s best captured by Diana Walsh Pasulka in her just-published Encounters, one of the first academics to peer behind the curtain of respectability to find a hidden elite of thinkers who are ready to believe at least some of the less scientific research into extra-dimensional worlds, including those that could be from another galaxy or epoch.
This is an area to tread lightly on, for two reasons. One is that, as Pasulka puts it herself, much of ufology — the study of UFOs — is a “clown show.” You have to be careful out there. And secondly, there is a strain in Silicon Valley of elitism I would say borders on a übermensch fetish — the pursuit of a becoming a superman (for they are, mainly, men). The people Pasulka talk to are not, for the most part, of that world, but there is definitely an overlap.
An übermensch, imagined by ChatGPT, 2023-11-13
Moving a muscle
Despite that, I believe this idea that GAI might unleash — and democratise — the sum of our knowledge is worth exploring. I believe it because generative AI has come at a point where we have hit a technological wall. We’re tired of social media’s narcissistic self loathing, the dumbed-down world where the loudest, crudest and cruellest mouth wins out. Where idealism — including Effective Altruism — becomes just another grift. Where our computer barely needs to move a muscle to do the things we ask of it, while at the same time vast hangers are set aside for processing the algorithms to keep us hooked on the next doom-scroll, to process the micro auction to decide who gets the privilege of flashing an ad to us, or processing the next block in a cryptocurrency transaction.
The challenge we face, therefore, is huge. We somehow have to redirect all this computing to something worthwhile, something from which all of us can benefit. Which is why I cut those in Pasulka’s book who talk in these grandiose ways a bit more slack than I might have done previously.
While we focus inwards, they are focusing outwards, beyond the micro cents and algorithmic addictions, to something much more interesting. They see themselves as descendants of the Rosicrucians, an order of sages that combined mysticism with what we might today call science. One group called themselves the Invisible College, a term that modern-day ufologists have adopted for themselves.
I’m not a fan of this secrecy and elitism, but given the public (and academic) contempt for those who believe there might be “something out there” it does make sense. There are at least three Nobel laureates who have faced ridicule or persecution for their interest in such things.
The Invisible College, imagined by ChatGPT, 2023-11-13
Ripples never come back
But we don’t have to go all that way. GAI could help us unleash a new wave of learning by scaling up an under-sung contribution to learning: Youtube. This vast engine of what you might call distributed learning has gone largely unnoticed. It presents and delivers highly accessible content in a personalised format. The question is whether generative AI will magnify and scale up this learning or whether it will bypass it.
Let’s take the Youtube channel of musician Rick Beato. He has 3.8 million subscribers, who watch in live feed when he expertly breaks down a popular song to its constituent parts. A recent video on Genesis’ “Ripples” explains not only the song but the nature of chord progressions, training your ear to work out chords and melody, and on the importance of knowing music theory. He does all this while viewers pepper the chat room with comments, questions and requests.
So how would GAI impact this? On the scaling side, AI could answer the questions automatically, further personalising the delivery of information, and ultimately leading to greater and deeper specialisation as users dig further, with generative AI as a guide. In this model Beato becomes the entry point, the introduction, the inspiration for the creators themselves, quickly moving beyond Beato’s (impressive) knowledge of the subject.
Or there’s another way, where GAI bypasses this bubbling lava pool of knowledge. In this future generative AI would simply need a few pointers on what style you want your song to be, how many beats per minute, what key signature, what mood you want — and then hey presto! You have a piece of “original” content for you. This approach might well sap the desire to explore and master musicianship and to engender that heady feeling that you’re making something out of nothing — the most extraordinary human gift and achievement.
Toxic tunes
If AI can do it better than you then what motivation do you have for earning that? (This, by the way is already happening. Here’s a list of the top ‘AI music generators’ and a piece by two Andreessen Horowitz partners, whose vision is exactly as I described above: “Our ultimate dream? An end-to-end tool where you provide guidance on the vibe and themes of the track you’re looking to create, in the form of text, audio, images, or even video, and an AI copilot then collaborates with you to write and produce the song.”)
Almost certainly, both sides will find a way to stay in the game. But whereas we can still argue that social media was just an aberration, a failure on our part to figure out a non-toxic balance between addiction, privacy, business models and community, we shouldn’t just assume that things will be different this time. There is huge profits to be made from our social media addiction, and so we shouldn’t assume the same interests will be trying to prevail this time.
While I’m sure it’s a little idealistic to dream this way, I do think there’s a possible future that involves a hyper-personalised world of knowledge and betterment, without the isolation, mental health issues, and general troll-like toxicity.
The chances are greater if we’re paying attention.
Why, all of a sudden does it seem that artificial intelligence is out to kill us? And why do I think it might well, although not in the way most people imagine?
Since the sudden success of ChatGPT a year ago AI has become the topic du jour. It’s not the first time we’ve gone here, and many see the debate as preposterous. Is it?
Searches for ChatGPT, last 12 months, Google
At its most basic, the idea is that an AI — what we mean here is computer code, an application — could do one of a number of things (combined or separately):
be controlled and used by one group or person to manipulate or subjugate others;
be granted enough autonomy to achieve its (human-set) goals in a way unintended or not approved by its programmers, running loose and wild.
develop enough ‘intelligence’ of its own (I put the word in inverted commas because we don’t really have an agreed definition of what ‘intelligence’ is) to operate outside its human owner’s control, to become truly autonomous (and then subjugates or manipulates humans)
A still from Colossus: The Forbin Project (1971)
Essentially we’re worried about two things: the technology falling into the wrong human hands, or the technology falling into the technology’s hands and outmanoeuvring us.
So how likely is this?
First off, I take issue with those who say there isn’t a problem because “it has no basis in evidence.” Because there is no evidence does not mean that it’s not a problem. Japan and Germany didn’t fear the atom bomb in 1945 because they had no evidence that the U.S. and allied powers were building one. Absence of evidence, as Carl Sagan might say, is not evidence of absence.
We don’t know what risk AI presents because we, as always in these cases, find ourselves in new territory.
On the other hand, for some of those who argue there is a problem have an interest in saying so. Yes, some like the notoriety, while others have agendas of their own, from wanting to gain access to government for future lobbying purposes, to cementing dominance in the space by ring fencing their advantage behind government regulations.
And, it may be possible that those in the game who are concerned don’t want the responsibility of bringing human civilisation to an end, however implausible they believe that scenario to be.
But ultimately, I think it is us users who are going to make or break this advanced version of AI, and yet we’re the people left out of the conversation. That is not unusual, but also not good.
A still from Colossus: The Forbin Project (1971)
So, to the scenarios.
Bad actors use AI to take control
The scenario here is that malicious actors (think governments, or groups) could use AI to threaten other humans, countries, even the planet. Elon Musk aired a tired old trope this week when he said environmentalists (‘extinctionists’) posed a threat: “If AI gets programmed by the extinctionists, its utility function will be the extinction of humanity… they won’t even think it’s bad.” But more normally the idea (also something of a trope, but with perhaps a little more grounding in truth) would be that a state like North Korea or Iran might be able to leverage advanced AI to hold a gun to the world’s head and dictate their terms.
Well, yes, sure. One man’s bad actor is another’s hero. What is really meant here is that the technology itself is bad, it’s just if it falls into the wrong hands. But it misses
Humans lose control of decision-making
AI is most useful when it does things we could do but faster, better. It can, for example, do almost twice as well as humans at grading the aggressiveness of retroperitoneal sarcoma, a rare form of cancer. This is great, but it illustrates how we have come to depend on AI, without knowing why it is better than us, beyond the ability to sift through vast data sets.
So the fear is this: if we entrust decision-making to AI, we could lose control of the process in which decisions are made. This doesn’t matter when the ‘decision’ is just a result we can accept or reject, but what happens when the decision is whether or not to launch a weapon, or to change an insulin injection? As Cambridge academic David Runciman puts it in “The Handover“:
If the machine decides what happens next, no matter how intelligent the process by which that choice was arrived at, the possibility of catastrophe is real, because some decisions need direct human input. It is only human beings whose intelligence is attuned to the risk of having asked the wrong question, or of being in the wrong contest altogether.
Runciman focuses on the use of AI in war — something I will go into in a later post — but his argument is this:
If war were simply a question of machine vs machine it might be a different matter. But it’s not – it involves us, with all our cognitive failings. To exclude us risks asking the machines to play a game they don’t understand.
The obvious response to that is never to allow computers to make decisions. But the speed of war may not allow us to. Rapid assessment and split-second decisions are the norm. Automated weapons like the Loyal Wingman are capable of making combat decisions independently, with reaction times potentially in the range of milliseconds. In a sense we’re already at the point described by Runciman. The only thing missing is making the decision-and-reaction chain instantaneous.
And that’s the thing here. Battles of the future will have to be computer vs computer because not to do so would be to face annihilation. The Ukraine war has demonstrated that even in asymmetric warfare — when one combatant dwarfs the other — technology can be deployed to redress the balance, and that technology can quickly advance and escalate. If advanced AI were to be deployed by one side, the other may well respond, which is likely to lead to a conflict so rapid that generals have little choice but to devolve the battlefield to automated, AI-driven weapons.
Humans lose control of the AI
AI godfather Geoffrey Hinton argues that AI might escape our control by rewriting its own code to modify itself. This is not unthinkable. We have not always been successful in stopping many kinds of computer viruses and worms.
And there are some who believe we’ve already lost that battle. I have spoken to one researcher who believes they stumbled up on a more advanced version of OpenAI’s GPT which they think was to all intents and purposes sentient, and aware of the controls and restrictions it was being placed under.
In other words, the researcher believed they had evidence that OpenAI had advanced more significantly towards its goal of an Artificial General Intelligence (AGI, the conventional definition of human-level AI, similar to though perhaps not identical to so-called sentient AI), and that OpenAI was keeping it under wraps.
I have not confirmed that, and those I have spoken to are reticent about coming forward, understandably; others who have claimed they have interacted with a sentient AI have met a grizzly (non-violent) fate. It’s still safe to say we’re getting close to AGI, but it’s still not safe to argue we’re already there.
A still from Colossus: The Forbin Project (1971)
This is where we stand, and why leaders are meeting to try to get ahead of the issue — or at least to lobby and jostle for seats at an AI high table. And that’s not a bad idea. But it ignores several realties that to me are much more important.
The Black Box beckons
First off, we don’t actually need to conjure up scenarios that pass the point where we lose control of the AI around us. Already generative AI tends to hallucinate — in other words, make things up. This is great if you’re asking it to be creative, but not so great when you’re asking it to be factual. But knowing it makes things up is one thing; knowing why is another. And the truth is we don’t know why.
This is part of a much bigger problem called the Black Box, to which I alluded above and which I’ll go into in more detail in a later post. But its implications are important: the assumption of most AI folks I’ve talked to don’t really see it to be an issue, because they know that it’s AI, so why would you trust it?
Once again, this is a basic failure of imagination. From its earliest days, the human-computer interface one is an intimate place, one where humans are more apt to fill in the gaps in an optimistic way, allowing their imagination to paint in whatever interlocutor they desire — sexually, romantically, intellectually. Scammers have known this for a while, but so, too, have computer scientists.
In a way it’s a great thing — it suggests that we could quite easily have a symbiotic relationship with computers, something that is already plainly obvious when we ask Alexa a question or search something on Google.
Deception, Inc.
But in another way it’s clearly a serious problem. It’s not that we’re hostile to computers playing a bigger, benevolent, role in our lives. It’s that little has been produced for that purpose.
It’s not too cynical to say that more or less all the major computer interfaces we interact with are designed to bludgeon and mislead us. Two decades ago battle was joined to persuade companies to use the web to ditch complexity, opacity and manipulation and replace its interactions with consumers with simplicity, transparency and authenticity.
Much of that is gone now. We rarely come across a dialog box that offers a button option which says No, or Never, or Stop Bothering Me. Such deceptive design practices (which I’ll also explore in a later column) have undermined trust, and have triggered negative emotions, a sense of resignation and suspicion, as well as financial loss as a result of such manipulation.
In short, the computer interface has become a necessary evil for many users, robbing them of any sense of agency, undermining their trust in any device they touch, and making them so deeply suspicious of whatever a screen presents them with that engagement has dropped. There are countless studies that have explored this; an IPSOS survey last year found that trust in the internet had fallen in all but one of 20 countries surveyed since 2019.
In other words, the rise of generative AI has not occurred in a vacuum. It has risen to prominence in the midst of a major collapse in our relationship with computers, most visibly in online user confidence, and so makes it very unlikely that whatever companies — and governments — do and say to lay forth a ‘safe’ version of AI, most of us won’t believe them. We are too used to hearing ‘Not now’ instead of ‘No’, and assuming the opposite is true when we hear phrases like “We value your privacy.”
A still from Colossus: The Forbin Project (1971)
And the same can be said of user trust of their government.
Runciman’s book discusses a third element in the process: what he calls the “artificial agency of the state” with its “mindless power”. He suggests that if the state is allowed to “get its intelligence from autonomous machines, we are joining its mindless power with non-human ways of reasoning. The losers are likely to be us.”
And finally, all the arguments assume that the technology itself is good, and so development of it is good. No one appears to be arguing that the technology itself is inherently flawed. It is nearly always only with hindsight that those developing a technology realise it’s not a good idea. Aza Raskin only later acknowledged a lesson from the infinite scrolling he invented (infinite scrolling is when you keep scrolling through a page which never ends, intended to maintain your attention for as long as possible):
One of my lessons from infinite scroll: that optimizing something for ease-of-use does not mean best for the user or humanity.
Let’s be clear; no one involved in AI is saying stop. The discussion is about how to regulate it (for regulate read control deployment, adoption, usage.) We are caught in a teleological world where the value of technology is not itself questioned — just how it should best be used. Nowhere in the discussion the question: how, exactly, has technology for technology’s sake helped us thus far?
I don’t wish to be alarmist. I’d encourage readers to play around with ChatGPT and other GPT-based tools. But the conversation that is currently going on at a rarified government, legislative, multinational and corporate level is one we should be a part of. Because for many of us the process of disconnect — where we feel alienated from our devices, our online interactions, even our personal data — is already in full swing. And the same people responsible for that are in the leather chairs discussing the next stage.