By Chip Scanlan
Journalists who write profiles don’t go in cold. They pre-report to prepare for crucial interviews. They read widely and research previous stories that have been written about their subject. They think through questions they need to ask.To profile Geoffrey Hinton, the British-Canadian computer scientist considered the godfather of artificial intelligence, New Yorker writer Joshua Rothman did his homework, and then some. He read histories of A.I., consulted an oral history of neural networks — a machine-learning technique that teaches computers to process information like the human brain — and plowed through a textbook about deep learning. He even took an online course in linear algebra before he felt ready to query Hinton.
“I’d written about A.I. and machine learning before, but the idea of spending an extended period alone with Geoff, and asking him to explain his work — it made me feel like I really needed to be prepared,” Rothman told me. “I wanted to make the most of the time I had.”
All that prep helped Rothman understand and then convey the science behind artificial intelligence, the technological development that has seized the public’s attention ever since OpenAI released its chatbot, ChatGPT, in November 2022. Rothman’s diligence grounded a 10,000-word nuanced, comprehensive and revealing portrait of Hinton, published in The New Yorker in November 2023.
“Why the Godfather of A.I. Fears What He’s Built” is an important story, deeply reported, creatively structured and written with literary grace. Rothman frames it within a four-day visit he made to Hinton’s home on a private island on Ontario’s Georgian Bay. The narrative glides through connections that track the history of A.I. and Hinton’s career with the personal tragedies that have shaped him.
“Often, I think, we write about people because of something they’re involved in, or because of something they’ve done,” Rothman said. “But when it comes time to write the profile, it becomes important to focus in on the person as a person — just as an unadorned human being, with all the richness and intensity that entails.” Even the word “profile” is telling; Rothman and Eric Overbey wrote in an introduction to 12 classic New Yorker profiles: “It suggests catching sight of someone from an unusual angle.”
The profile of Hinton opens with an examination of the brain at work and comes to a close when the natural world eclipses the technical one. In between, it explores a new and, to some, discomfiting frontier of knowledge and ethics. Rothman uses sharply-etched scenes, characterization, dialogue, metaphors and digressions to help readers grasp historical context and challenging technical details.
Rothman, the Ideas Editor of The New Yorker, began his journalism career blogging for the ideas section of The Boston Globe and then freelancing before joining The New Yorker in 2012, where, as ideas editor of newyorker.com, he guides and writes stories about science, philosophy and technology. He helped conceive the Hinton profile for an issue devoted to artificial intelligence, pegged to recent concerns about the evolution of technology that has sparked concerns about potential abuse. Hinton’s take is invaluable, Rothman said, “especially given his recent emergence as someone sounding the alarm about A.I.”
In an email exchange, Rothman described the reasoning behind the story’s structure, the challenge of conveying technical knowledge in accessible ways and the collaborative relationship with his editor. The Q&A has been edited for length and clarity, and is followed by an annotation of the story.
Why did you become a journalist?
I wanted to be an English professor, and I studied for a PhD in English. But I was wrapping up my dissertation in 2008, and the financial crisis brought the professorial job market to a halt. I’d always wanted to be a writer and had wide-ranging interests, including in science and technology. It felt like journalism would be a practical path forward that would also allow me to explore a lot of different subjects.
What does the position of ideas editor at The New Yorker entail?
I edit pieces about ideas, broadly construed — pieces that touch on science, technology, philosophy, literature, history and so on — mostly for the web site, but also for the print magazine. And I write a few pieces a year, fitting it in where I can.
What writers have influenced you and how in general, and as you reported and wrote your story?
I always wish I could write a profile the way Larissa MacFarquhar writes them. And when I write about technology, I always revisit William Gibson (whom I was able to profile a few years ago). He reminds me that the newest technologies have roots in the past. Everything new is old. You always have to rewind further than you think.
What is the origin story behind “Why the Godfather of A.I. Fears What He’s Built?”
I’ve been covering A.I. for a few years as an editor and writer. It’s a really complicated subject, both technically and conceptually, and it’s also a sweeping story going back at least a century, combining history and philosophy and science fiction and business. It’d be crazy to think you could combine all that in a single piece, but it seemed like writing about Hinton would get you into the ballpark. So once we started talking about doing a special issue on A.I., a profile of Hinton seemed like a good idea. Especially given his recent emergence as someone sounding the alarm about A.I.
What role did your editor play during the stages of writing and finishing your story?
Henry Finder, editorial director of the New Yorker, has been my editor for a long time now. Working with him is a little like being in a writing seminar that progresses from piece to piece. With this piece, I wanted to apply lessons that I’d learned from him while writing the previous one—a profile of the novelist Kim Stanley Robinson. He pushes the pieces forward in so many ways, in terms of craft but also in terms of thinking. And he always helps me keep in mind what the ultimate goals are — what the ideal version of the piece should be.
Much of your story is an intellectual challenge to absorb and understand. Who do you envision as your audience?
I wanted this piece to be something anyone could read. I hoped it would pull readers into the intellectual adventure of A.I., and that it would help them understand how a pure and profound curiosity about our own minds got us to this point. Once I got to know Geoff Hinton, I wanted the piece to reflect him — his personality and experiences. He’s a fascinating individual in his own right, with a life story that has a great deal to tell us just about life, outside of A.I. So I don’t think of it as a piece for a technology-oriented audience. I hoped to write a humanist piece.
ANNOTATION: Storyboard’s questions are in red; Rothman’s answers in blue. To read the story without annotations, click the HIDE ANNOTATIONS button in the right-hand menu of your monitor or at the top of your mobile screen.
Why the Godfather of A.I. Fears What He’s Built
Geoffrey Hinton has spent a lifetime teaching computers to learn. Now he worries that artificial brains are better than ours.
By Joshua Rothman
The New Yorker
Nov. 13, 2023
In your brain, neurons are arranged in networks big and small. With every action, with every thought, the networks change: neurons are included or excluded, and the connections between them strengthen or fade. This process goes on all the time—it’s happening now, as you read these words—and its scale is beyond imagining. You have some eighty billion neurons sharing a hundred trillion connections or more. Your skull contains a galaxy’s worth of constellations, always shifting. You begin the story by describing the inner workings of the brain. Why? Because that’s where A.I. started. It began as an offshoot of an introspective, neuroscientific effort. I also wanted to raise the question, early on, of what human intelligence is and how it might work. I wanted readers to look inside themselves and begin to consider the possibility that the mind is a kind of machine.Geoffrey Hinton, the computer scientist who is often called “the godfather of A.I.,” handed me a walking stick. “You’ll need one of these,” he said. Then he headed off along a path through the woods to the shore. It wound across a shaded clearing, past a pair of sheds, and then descended by stone steps to a small dock. This sentence adheres to the “rule of three.” Was that a conscious writing decision? I’ve never heard of the rule of three. Although maybe I think that three-part constructions sound good. “It’s slippery here,” Hinton warned, as we started down.
New knowledge incorporates itself into your existing networks in the form of subtle adjustments. Sometimes they’re temporary: if you meet a stranger at a party, his name might impress itself only briefly upon the networks in your memory. But they can also last a lifetime, if, say, that stranger becomes your spouse. Because new knowledge merges with old, what you know shapes what you learn. If someone at the party tells you about his trip to Amsterdam, the next day, at a museum, your networks may nudge you a little closer to the Vermeer. In this way, small changes create the possibility for profound transformations. The reader quickly becomes aware of the story’s woven structure. How did you select this approach? I knew from the beginning of the writing process that I wanted to braid the science and the life together. The challenge was making it work. But I wanted it to be clear, from the beginning of the piece, that we would be learning about Hinton’s life on a personal level. I think it’s important, in a long piece of writing, to try and move all the elements forward together instead of siloing them into separate sections.
“We had a bonfire here,” Hinton said. We were on a ledge of rock jutting out into Ontario’s Georgian Bay, which stretches to the west into Lake Huron. Islands dotted the water; Hinton had bought this one in 2013, when he was sixty-five, after selling a three-person startup to Google for forty-four million dollars. Before that, he’d spent three decades as a computer-science professor at the University of Toronto—a leading figure in an unglamorous subfield known as neural networks, which was inspired by the way neurons are connected in the brain. Because artificial neural networks were only moderately successful at the tasks they undertook—image categorization, speech recognition, and so on—most researchers considered them to be at best mildly interesting, or at worst a waste of time. “Our neural nets just couldn’t do anything better than a child could,” Hinton recalled. In the nineteen-eighties, when he saw “The Terminator,” it didn’t bother him that Skynet, the movie’s world-destroying A.I., was a neural net; he was pleased to see the technology portrayed as promising.
From the small depression where the fire had been, cracks in the stone, created by the heat, radiated outward. Hinton, who is tall, slim, and English, Why did you select these particular three descriptors? This is a kind of arm’s-length description — like if you were squinting to make Hinton out from a distance, this is what you’d see. I could’ve described him in greater detail here, but didn’t want to lose momentum. poked the spot with his stick. A scientist through and through, he is always remarking on what is happening in the physical world: the lives of animals, the flow of currents in the bay, the geology of the island. “I put a mesh of rebar under the wood, so the air could get in, and it got hot enough that the metal actually went all soft,” he said, in a wondering tone. “That’s a real fire—something to be proud of!” You’re capturing small moments in exquisite detail. What do you use in the field to record interviews and what you’re seeing? I have a small handheld voice recorder, and I also take a lot of pictures. I almost always have a camera with me. At the end of each reporting day, I take notes about what happened and begin to write out the scenes from memory and notes.
For decades, Hinton tinkered, building bigger neural nets structured in ingenious ways. He imagined new methods for training them and helping them improve. He recruited graduate students, convincing them that neural nets weren’t a lost cause. He thought of himself as participating in a project that might come to fruition a century in the future, after he died. Meanwhile, he found himself widowed and raising two young children alone. During one particularly difficult period, when the demands of family life and research overwhelmed him, he thought that he’d contributed all he could. “I was dead in the water at forty-six,” he said. He didn’t anticipate the speed with which, about a decade ago, neural-net technology would suddenly improve. Computers got faster, and neural nets, drawing on data available on the Internet, started transcribing speech, playing games, translating languages, even driving cars. Around the time Hinton’s company was acquired, an A.I. boom began, leading to the creation of systems like OpenAI’s ChatGPT and Google’s Bard, which many believe are starting to change the world in unpredictable ways. You condense the early parts of Hinton’s relationship with A.I. to a single paragraph. How do you achieve that concision? This paragraph is based very heavily on a summary of his life and career that Geoff himself gave me during our very first conversation over Zoom. My mother-in-law has a theory that people tell you about the most important aspects of their lives within the first fifteen minutes of conversation. In Geoff’s case, I think it was true. “I was dead in the water…” — it was such a moving thing to say, and I knew that I wanted to include it high up to help readers understand what a forthright, humane person he is.
Hinton set off along the shore, and I followed, the fractured rock shifting beneath me. “Now watch this,” he said. He stood before a lumpy, person-size boulder, which blocked our way. “Here’s how you get across. You throw your stick”—he tossed his to the other side of the boulder—“and then there are footholds here and here, and a handhold here.” I watched as he scrambled over with easy familiarity, and then, more tentatively, I took the same steps myself. You are present in the narrative. Why? I’m pretty much always present in everything I write. It’s just how I approach things.
Whenever we learn, our networks of neurons change—but how, exactly? Researchers like Hinton, working with computers, sought to discover “learning algorithms” for neural nets, procedures through which the statistical “weights” of the connections among artificial neurons could change to assimilate new knowledge. In 1949, a psychologist named Donald Hebb proposed a simple rule for how people learn, often summarized as “Neurons that fire together wire together.” Once a group of neurons in your brain activates in synchrony, it’s more likely to do so again; this helps explain why doing something is easier the second time. But it quickly became apparent that computerized neural networks needed another approach in order to solve complicated problems. As a young researcher, in the nineteen-sixties and seventies, Hinton drew networks of neurons in notebooks and imagined new knowledge arriving at their borders. How would a network of a few hundred artificial neurons store a concept? How would it revise that concept if it turned out to be flawed? How did you keep track of your structure: outline, notecards, timeline, or other ways? I have no system! I used to make outlines, but now I just keep track of things in my head.
We made our way around the shore to Hinton’s cottage, the only one on the island. Glass-enclosed, it stood on stilts atop a staircase of broad, dark rocks. “One time, we came out here and a huge water snake stuck his head up,” Hinton said, as we neared the house. It was a fond memory. His father, a celebrated entomologist who’d named a little-known stage of metamorphosis, had instilled in him an affection for cold-blooded creatures. When he was a child, he and his dad kept a pit full of vipers, turtles, frogs, toads, and lizards in the garage. When did you learn this? Did you interview Hinton apart from those during your visit to his island home? He told me this at some point during the visit. And, apart from a short Zoom interview we had when I first proposed the story to him, I kept our discussions in-person, on the island, so that as much of the piece as possible could be situated there. Today, when Hinton is on the island—he is often there in the warmer months—he sometimes finds snakes and brings them into the house, so that he can watch them in a terrarium. He is a good observer of nonhuman minds, having spent a lifetime thinking about thinking from the bottom up.
Earlier this year, Hinton left Google, where he’d worked since the acquisition. He was worried about the potential of A.I. to do harm, and began giving interviews in which he talked about the “existential threat” that the technology might pose to the human species. The more he used ChatGPT, an A.I. system trained on a vast corpus of human writing, the more uneasy he got. Why did you use the word “corpus” rather than one that might be more familiar? I think it’s the most accurate. It’s a curated collection of writing, maintained and adjusted at great expense by experts. One day, someone from Fox News wrote to him asking for an interview about artificial intelligence. Hinton enjoys sending snarky single-sentence replies to e-mails—after receiving a lengthy note from a Canadian intelligence agency, he responded, “Snowden is my hero” and he began experimenting with a few one-liners. Eventually, he wrote, “Fox News is an oxy moron.” Then, on a lark, he asked ChatGPT if it could explain his joke. The system told him his sentence implied that Fox News was fake news, and, when he called attention to the space before “moron,” it explained that Fox News was addictive, like the drug OxyContin. Hinton was astonished. This level of understanding seemed to represent a new era in A.I.
There are many reasons to be concerned about the advent of artificial intelligence. It’s common sense to worry about human workers being replaced by computers, for example. But Hinton has joined many prominent technologists, including Sam Altman, the C.E.O. of OpenAI, in warning that A.I. systems may start to think for themselves, and even seek to take over or eliminate human civilization. This strikes me as the news peg, if you will. Is this the story’s central take? It is definitely a peg. But to me the piece is actually centered on the question of whether artificial intelligence is actually a kind of real intelligence. And the real news peg, from that perspective, is the advent of the large language model, which is a technology that seems intelligent in a new way. It was striking to hear one of A.I.’s most prominent researchers give voice to such an alarming view.
“People say, It’s just glorified autocomplete,” he told me, standing in his kitchen. (He has suffered from back pain for most of his life; it eventually grew so severe that he gave up sitting. He has not sat down for more than an hour since 2005. In an audio of you interview with Hinton played back during your New Yorker Radio hours, we hear him telling you, “You’re just an autocomplete device.” As a writer and editor, what was your reaction to that? Well, I think a lot depends on how that statement is meant. I don’t think that Hinton is trying to reduce or denigrate me or my work or humanity by saying it. He certainly knows that human life is rich, that our capabilities are multi-dimensional, that our lives have value, and so on. What he is saying is that something like autocomplete — that is, an unconscious, statistical process that uses what’s already happened to extrapolate about what’s next — is a big part of our mental lives. Maybe even the bulk of our mental lives. And I think it’s hard to dispute that. “Now, let’s analyze that. Suppose you want to be really good at predicting the next word. If you want to be really good, you have to understand what’s being said. That’s the only way. So by training something to be really good at predicting the next word, you’re actually forcing it to understand. Yes, it’s ‘autocomplete’—but you didn’t think through what it means to have a really good autocomplete.” You alternate between your own descriptions of the workings of the brain and A.I., but use Hinton’s words as well. How did you decide which one to use when to further the narrative? Geoff is an incredible communicator, and he often puts things in a powerful, concise, insightful way. But there are other times when what he says isn’t quite sufficient for the purposes of the story. For example, the story might benefit from having a theme that’s been introduced higher up pulled through to the current moment. In those instances, I will sometimes interject my own formulations. Hinton thinks that “large language models,” such as GPT, which powers OpenAI’s chatbots, can comprehend the meanings of words and ideas. Why did you use quote marks for “large language models” and then not provide a phrase defining that term? I figured that “such as GPT” was explanation enough, and didn’t want to bog things down.
Skeptics who say that we overestimate the power of A.I. point out that a great deal separates human minds from neural nets. For one thing, neural nets don’t learn the way we do: we acquire knowledge organically, by having experiences and grasping their relationship to reality and ourselves, while they learn abstractly, by processing huge repositories of information about a world that they don’t really inhabit. How did you learn this? I suppose it’s just general knowledge, gleaned while covering this subject over the years. I extrapolated it from my training data, as it were. But Hinton argues that the intelligence displayed by A.I. systems transcends its artificial origins. Could you point to any section and say, “This is the nut (or context” graf?” I think it’s the very last paragraph of this section. Hinton’s biggest, most important argument is: It started as one thing, and it’s become something else.
“When you eat, you take food in, and you break it down to these tiny components,” he told me. “So you could say that the bits in my body are made from bits of other animals. But that would be very misleading.” He believes that, by analyzing human writing, a large language model like GPT learns how the world works, producing a system capable of thought; writing is only part of what that system can do. “It’s analogous to how a caterpillar turns into a butterfly,” he went on. “In the chrysalis, you turn the caterpillar into soup—and from this soup you build the butterfly.”
He began rooting around in a small cupboard just off the kitchen. “Aha!” he said. With a flourish, he put an object on the counter—a dead dragonfly. It was perfectly preserved. “I found this at the marina,” he explained. “It had just hatched on a rock and was drying its wings, so I caught it. Look underneath.” Hinton had captured the dragonfly just after it had emerged from its larval form. The larva was a quite different-looking insect, with its own eyes and legs; it had a hole in its back, through which the dragonfly had crawled.
“The larva of the dragonfly is this monster that lives under the water,” Hinton said. “And, like in the movie ‘Alien,’ the dragonfly is breaking out of the back of the monster. The larva went into a phase where it got turned into soup, and then a dragonfly was built out of the soup.” In his metaphor, the larva represented the data that had gone into training modern neural nets; the dragonfly stood for the agile A.I. that had been created from it. Throughout the story, you use metaphors to present abstractions. Why? I can think of few technologies that are more abstract than A.I. Arguably, the whole research program is metaphorical. And one of Hinton’s big intellectual commitments is to the idea that we are “analogy machines.” So, in addition to being useful from a readerly perspective, I thought it would be fun, as a writer, to try to do a lot of metaphorical thinking. Deep learning—the technology that Hinton helped pioneer—had caused the metamorphosis. I bent closer to get a better look; Hinton stood upright, as he almost always does, careful to preserve his posture. “It’s very beautiful,” he said softly. “And you get the point. It started as one thing, and it’s become something else.” Is the metaphor designed to deliver more than one message and if so, what is it? It adds a biological dimension to the discussion of A.I. It suggests the mystery of the technology — it’s more like a process we set in motion than a machine that we build part by part. And it captures the sense of wariness with which Hinton thinks we need to approach the technology: we don’t yet know what it is, or what it’s becoming. When he showed me this dragonfly and talked about it this way, I knew almost immediately that I’d close the first section of the story with it.
A FEW WEEKS EARLIER, when Hinton had invited me to visit his island, I’d imagined possible scenarios. Why did you use the visit to the island as the spine of the narrative? An alternative approach would’ve been to visit Geoff in Toronto. But I knew it wouldn’t be nearly as interesting as visiting his island, because the island, with its cottage, is almost like a stand-in for his personality, his life. Like the Ramsay house in “To the Lighthouse.” And, of course, he very graciously offered to host me there. Perhaps he’d be an introvert who wanted solitude, or a tech overlord with a God complex and a futuristic compound. Several days before my arrival, he e-mailed me a photograph he’d taken of a rattlesnake coiled in the island’s grass. I wasn’t sure whether I felt delighted or scared.
In fact, as private islands go, Hinton’s is fairly modest—two acres in total. Hinton himself is the opposite of a Silicon Valley techno-messiah. Now seventy-five, he has an English face out of a Joshua Reynolds painting, with white hair framing a broad forehead; his blue eyes are often steady, leaving his mouth to express emotion. A mordant raconteur, he enjoys talking about himself—“ ‘Geoff’ is an anagram for ‘ego fortissimo,’ ” he told me—but he’s not an egotist; his life has been too grief-shadowed for that. A tragic, but beautiful phrase. When did it come to you? My editor suggested it! I had something windier. “I should probably tell you about my wives,” he said, the first time we spoke. “I’ve had three marriages. One ended amicably, the other two in tragedy.” He is still friendly with Joanne, his first wife, whom he married early, but his second and third wives, Rosalind and Jackie, both died of cancer, in 1994 and 2018, respectively. For the past four years, Hinton has been with Rosemary Gartner, a retired sociologist. “I think he’s the kind of person who always needs a partner,” she told me, tenderly. Why did you include the adverb? I wanted to make it clear that Rosemary meant this kindly, not as a criticism or cutting observation, and not in a wry way. She did mean it tenderly. He is a romantic rationalist, with a sensibility balancing science and emotion. In the cottage, a burgundy canoe sits in the single large room that makes up most of the ground floor; he and Jackie had found it in the island’s woods, in disrepair, and Jackie, an art historian, worked with some women canoe-builders to reconstruct it during the years coinciding with her illness. “She had the maiden voyage,” Hinton said. No one has used it since. Just six words, but they carry so much weight. Why did you write them, rather than quote Hinton? Sometimes, for an emotional beat to land, it needs just a little moment of repetition or underscoring. And it can be nice to have a kind of reaction shot from the text — an acknowledgment, in the piece, of the gravity of something.
He stowed the dragonfly, then walked over to a small standing desk, where a laptop was perched next to a pile of sudoku puzzles and a notebook containing computer passwords. (He rarely uses the notebook, having devised a mnemonic system that enables him to generate and recall very long passwords in his head.) “Shall we do the family tree?” he asked. Using two fingers—he doesn’t touch-type—he entered “Geoffrey Hinton family tree” and hit Return. When Google acquired Hinton’s startup, in 2013, it did so in part because the team had figured out how to dramatically improve image recognition using neural nets; now endless family trees swarmed the screen.
Hinton comes from a particular kind of scientific English family: politically radical, restlessly inventive. Above him in the family tree are his great-uncle Sebastian Hinton, the inventor of the jungle gym, and his cousin Joan Hinton, who worked as a physicist on the Manhattan Project. Further back, he was preceded by Lucy Everest, the first woman to become an elected member of the Royal Institute of Chemistry; Charles Howard Hinton, the mathematician who created the concept of the tesseract, a doorway into the fourth dimension (one appears in the film “Interstellar”); and James Hinton, a groundbreaking ear surgeon and an advocate of polygamy. (“Christ was the savior of men, but I am the savior of women,” he is said to have remarked.) In the mid-nineteenth century, a great-great-grandfather of Hinton’s, the English mathematician George Boole, developed the system of binary reasoning, now known as Boolean algebra, that is fundamental to all computing. Boole was married to Mary Everest, a mathematician and author and the niece of George Everest, the surveyor for whom Mt. Everest is named. It’s so interesting the way you use the family tree as a device to present Hinton’s ancestors. I wish I could say that this was my idea, but almost everyone who writes about Hinton runs down the family tree. In this case, he made it easy, by actually proposing that we look at one.
“Geoff was born into science,” Yann LeCun, a former student and collaborator of Hinton’s who now runs A.I. at Meta, told me. Yet Hinton’s family was odder than that. His dad, Howard Everest Hinton, grew up in Mexico during the Mexican Revolution, in the nineteen-tens, on a silver mine managed by his father. “He was tough,” Hinton said of his dad: family lore holds that, at age twelve, Howard threatened to shoot his boxing coach for being too heavy-handed, and the coach took him seriously enough to leave town. Howard’s first language was Spanish, and at Berkeley, where he went to college, he was mocked for his accent. “He hung out with a bunch of Filipinos, who were also discriminated against, and he became a Berkeley radical,” Hinton said. Howard’s mature politics were not just Marxist but Stalinist: in 1968, as Soviet tanks rolled into Prague, he said, “About time!” Could you verify what his father said? This is based on Geoff’s recollection.
At school, Hinton was inclined toward science. But, for ideological reasons, his father forbade him to study biology; in Howard’s view, the possibility of genetic determinism contravened the Communist belief in the ultimate malleability of human nature. (“I hate faiths of all kinds,” Hinton said, remembering this period.) Howard, who taught at the University of Bristol, was a kind of entomologist Indiana Jones: he smuggled rare creatures from around the world back to England in his luggage, and edited an important journal in his field. Hinton, whose middle name is also Everest, felt immense pressure to make his own mark. He recalls his father telling him, “If you work twice as hard as me, when you’re twice as old as I am you might be half as good.”
At Cambridge, Hinton tried different fields but was dismayed to find that he was never the brightest student in any given class. He left college briefly to “read depressing novels” and to do odd jobs in London, then returned to attempt architecture, for about a day. Finally, after dipping into physics, chemistry, physiology, and philosophy, looking for a focus, he settled on a degree in experimental psychology. He haunted the office hours of the moral philosopher Bernard Williams, who turned out to be interested in computers and the mind. One day, Williams pointed out that our different thoughts must reflect different physical arrangements inside our brains; this was quite unlike the situation inside a computer, in which the software was independent of the hardware. Hinton was struck by this observation; he remembered how, in high school, a friend had told him that memory might be stored in the brain “holographically”—that is, spread out, but in such a way that the whole could be accessed through any one part. What he was encountering was “connectionism”—an approach that combined neuroscience, math, philosophy, and programming to explore how neurons could work together to “think.” One goal of connectionism was to create a brainlike system in a computer. There had been some progress: the Perceptron, a machine built in the nineteen-fifties by a psychologist and pioneering connectionist named Frank Rosenblatt, had used simple computer hardware to simulate a network of hundreds of neurons. When connected to a light sensor, the apparatus could recognize letters and shapes by tracking which artificial neurons were activated by different patterns of light. How much research apart from your field reporting did you have to do to write complex material with clarity and authority? I did a lot of research. But I also follow this field closely already, and have for many years, and many of the ideas I’m discussing in sections like this have come up in other contexts — for example, my profile several years ago of Daniel Dennett, the philosopher of mind , was tremendously useful preparation for writing this piece.[/annotate][/annotation-group]
In the cottage, Hinton stood and strolled, ranging back and forth behind the kitchen counter and around the first floor. He made some toast, got us each an apple, and then set up a little booster table for himself using a step stool. Family pressure had had the effect of pushing him out of temporary satisfactions. “I always loved woodwork,” he recalled wistfully, while we ate. “At school, you could do it voluntarily in the evenings. And I’ve often wondered whether I’d have been happier as an architect, because I didn’t have to force myself to do it. Whereas, with science, I’ve always had to force myself. Because of the family, I had to succeed at it—I had to find a path. There was joy in it, but it was mostly anxiety. Now it’s an enormous relief that I’ve succeeded.”
Hinton’s laptop dinged. How did you decide on this verb? I believe that it actually made a “ding” sound. But that may very well be what Hinton would call a “confabulation” — that is, an extrapolation from other existing data. Ever since he’d left Google, his in-box had been exploding with requests for comment on A.I. He ambled What were you trying to convey with this verb? I was struck by how Hinton spent all day standing, walking, striding, ambling, leaning, and so on. He has a whole vocabulary of walks. over and looked at the e-mail, and then got lost again in the forest of family trees, all of which seemed to be wrong in one way or another.
“Look at this,” he said.
I walked over and peered at the screen. It was an “academic family tree,” showing Hinton at the top with his students, and theirs, arrayed below. The tree was so broad that he had to scroll horizontally to see the extent of his influence. “Oh, dear,” Hinton said, exploring. “She wasn’t really a student of mine.” He scrolled further. “He was brilliant but not so good as an adviser, because he could always do it better himself.” A careful nurturer of talent, Hinton seems to enjoy being surpassed by his students: when evaluating job candidates, he used to ask their advisers, “But are they better than you?” Recalling his father, who died in 1977, Hinton said, “He was just extremely competitive. And I’ve often wondered, if he’d been around to see me be successful, whether he’d have been entirely happy. Because now I’ve been more successful than he was.”
According to Google Scholar, Hinton is now the second most cited researcher among psychologists, and the most cited among computer and cognitive scientists. If he had a slow and eccentric start at Cambridge, it was partly because he was circling an emerging field. “Neural networks—there were very few people at good universities who did it,” he said, closing the laptop. “You couldn’t do it at M.I.T. You couldn’t do it at Berkeley. You couldn’t do it at Stanford.” There were advantages to being a hub in a nascent network. For years, many of the best minds came to him.
“THE WEATHER’S GOOD,” Hinton said, the next morning. You skip past half a day. Why? Did you spend the night at Hinton’s? It was a long visit. All told I spent four nights on the island with him. So I skipped a lot! “We should cut down a tree.” He wore a dress shirt tucked into khakis and didn’t look much like a lumberjack; still, he rubbed his hands together. On the island, he is always cutting down trees to create more orderly and beautiful tableaus.
The house, too, is a work in progress. Few contractors would travel to a place so remote, and the people Hinton hired made needless mistakes (running a drainage pipe uphill, leaving floors half finished) that still enrage him today. Almost every room harbors a corrective mini-project, and, when I visited, Hinton had appended little notes to them to help a new contractor, often writing on the building materials themselves. In the first-floor bathroom, a piece of baseboard propped against the wall read “Bathroom should have this type of baseboard (maple trim in front of shower only).” In the guest-room closet, masking tape ran along a shelf: “Do not prime shelf, prime shelf support.” Is that where you stayed and the way you noticed it? Yes, I stayed in the guest room and poked around a bit, like journalists do.
It’s useful for minds to label things; it helps them get a grip on reality. But what would it mean for an artificial mind to do so? While Hinton was earning a Ph.D. in artificial intelligence from the University of Edinburgh, he thought about how “knowing” in a brain might be simulated in a computer. At that time, in the nineteen-seventies, the vast majority of A.I. researchers were “symbolists.” In their view, knowing about, say, ketchup might involve a number of concepts, such as “food,” “sauce,” “condiment,” “sweet,” “umami,” “red,” “tomato,” “American,” “French fries,” “mayo,” and “mustard”; together, these could create a scaffold on which a new concept like “ketchup” might be hung. A large, well-funded A.I. effort called Cyc centered on the construction of a vast knowledge repository into which scientists, using a special language, could enter concepts, facts, and rules, along with their inevitable exceptions. (Birds fly, but not penguins or birds with damaged wings or . . .)
But Hinton was doubtful of this approach. It seemed too rigid, and too focussed on the reasoning skills possessed by philosophers and linguists. In nature, he knew, many animals acted intelligently without access to concepts that could be expressed in words. They simply learned how to be smart through experience. Learning, not knowledge, was the engine of intelligence. This is a bumper sticker or T-shirt slogan! What a great line. Did Hinton tell you this and, if so, why did you paraphrase it rather than quote him? This is something he says a lot, in different words. I said it myself, rather than quoting him, for two reasons. First, it’s sometimes important to get the rhythm of a paragraph just right. And second, we’re in free indirect style in this passage — the narration is unfolding from his perspective, as though we’re inside his head back in the days when he was a young scholar. It would break the spell to suddenly quote something he said years later; suddenly, we’d see him from the outside.
Sophisticated human thinking often seemed to happen through symbols and words. But Hinton and his collaborators, James L. McClelland and David Rumelhart, believed that much of the action happened on a sub-conceptual level. Notice, they wrote, how, “if you learn a new fact about an object, your expectations about other similar objects tend to change”: if you’re told that chimpanzees like onions, for instance, you might guess that gorillas like them, too. This suggested that knowledge was likely “distributed” in the mind—created out of smaller building blocks that could be shared among related ideas. There wouldn’t be two separate networks of neurons for the concepts “chimpanzee” and “gorilla”; instead, bundles of neurons representing various concrete or abstract “features”—furriness, quadrupedness, primateness, animalness, intelligence, wildness, and so on—might be activated in one way to signify “chimpanzee” and in a slightly different way to signify “gorilla.” To this cloud of features, onion-liking-ness might be added. A mind constructed this way risked falling into confusion and error: mix qualities together in the wrong arrangement and you’d get a fantasy creature that was neither gorilla nor chimp. But a brain with the right learning algorithm might adjust the weights among its neurons to favor sensible combinations over incoherent ones.
Hinton continued to explore these ideas, first at the University of California, San Diego, where he did a postdoc (and married Joanne, whom he tutored in computer vision); then at Cambridge, where he worked as a researcher in applied psychology; and then at Carnegie Mellon, in Pittsburgh, where he became a computer-science professor in 1982. There, he spent much of his research budget on a single computer powerful enough to run a neural net. He soon got married a second time, to Rosalind Zalin, a molecular biologist. At Carnegie Mellon, Hinton had a breakthrough. Working with Terrence Sejnowski, a computer scientist and a neuroscientist, he produced a neural net called the Boltzmann Machine. The system was named for Ludwig Boltzmann, the nineteenth-century Austrian physicist who described, mathematically, how the large-scale behavior of gases was related to the small-scale behavior of their constituent particles. Hinton and Sejnowski combined these equations with a theory of learning.
Hinton was reluctant to explain the Boltzmann Machine to me. “I’ll tell you what this is like,” he said. “It’s like having a small child, and you decide to go on a walk. And there’s a mountain ahead of you, and you have to get this little child to the top of the mountain and back.” He looked at me—the child in the metaphor—and sighed. He worried, reasonably, that I might be misled by a simplified explanation and then mislead others. “It’s no use trying to explain complicated ideas that you don’t understand. First, you have to understand how something works. Otherwise, you just produce nonsense.” Finally, he took some sheets of paper and began drawing diagrams of neurons connected by arrows and writing out equations, which I tried to follow. (Ahead of my visit, I’d done a Khan Academy course on linear algebra.) Why did you do this? I’d bought a textbook called “Deep Learning,” blurbed by both Hinton and Elon Musk (!), and wanted to read it before my reporting trip. I quickly discovered that, if I wanted to understand the notation, I needed to give myself at least an introduction to linear algebra. And it came in handy; a few times during my visit, Hinton used equations in his explanations, and I could follow along.
One way to understand the Boltzmann Machine, he suggested, was to imagine an Identi-Kit: a system through which various features of a face—bushy eyebrows, blue eyes, crooked noses, thin lips, big ears, and so on—can be combined to produce a composite sketch, of the sort used by the police. For an Identi-Kit to work, the features themselves have to be appropriately designed. The Boltzmann Machine could learn not just to assemble the features but to design them, by altering the weights of the connections among its artificial neurons. It would start with random features that looked like snow on a television screen, and then proceed in two phases—“waking” and “sleeping”—to refine them. While awake, it would tweak the features so that they better fit an actual face. While asleep, it would fantasize a face that didn’t exist, and then alter the features so that they were a worse fit.
Its dreams told it what not to learn. There was an elegance to the system: over time, it could move away from error and toward reality, and no one had to tell it if it was right or wrong—it needed only to see what existed, and to dream about what didn’t. Hinton seems to be a master of the metaphor. Artificial intelligence is incredibly abstract. We’re talking about thinking, probability, psychology, computation — all things that are happening in our heads (and now in machines). And he uses metaphors very effectively to make this abstract world comprehensible and accessible to intuition.
Hinton and Sejnowski described the Boltzmann Machine in a 1983 paper. “I read that paper when I was starting my graduate studies, and I said, ‘I absolutely have to talk to these guys—they’re the only people in the world who understand that we need learning algorithms,’ ” Yann LeCun told me. In the mid-eighties, Yoshua Bengio, a pioneer in natural-language processing and in computer vision who is now the scientific director at Mila, an A.I. institute in Quebec, trained a Boltzmann Machine to recognize spoken syllables as part of his master’s thesis. “Geoff was one of the external reviewers,” he recalled. “And he wrote something like ‘This should not work.’ ” Bengio’s version of the Boltzmann Machine was more effective than Hinton expected; it took Bengio a few years to figure out why. This would become a familiar pattern. In the following decades, neural nets would often perform better than expected, perhaps because new structures had formed among the neurons during training. “The experimental part of the work came before the theory,” Bengio recalled. Often, it was a matter of trying new approaches and seeing what the networks came up with.
Partly because Rosalind loathed Ronald Reagan, Hinton said, they moved to the University of Toronto. They adopted two children, a boy and a girl, from Latin America, and lived in a house in the city. “I was this kind of socialist professor who was dedicated to his work,” Hinton said.
Rosalind had struggled with infertility, and had bad experiences with callous doctors. Perhaps as a result, she pursued a homeopathic route when she was later diagnosed with ovarian cancer. “It just didn’t make any sense,” Hinton said. “It couldn’t be that you make things more dilute and they get more powerful.” He couldn’t see how a molecular biologist could become a homeopath. Still, determined to treat the cancer herself, Rosalind refused to have surgery even after an exam found a tumor the size of a grapefruit; later, she consented to an operation but declined chemotherapy, instead pursuing increasingly expensive homeopathic remedies, first in Canada and then in Switzerland. She developed secondary tumors. She asked Hinton to sell their house so that she could pay for new homeopathic treatments. “I drew the line there,” he recalled, squinting with fresh pain. Is this the kind of gesture that you record in your notebook? Yes, very much so. “I said, ‘No, we’re not selling the house. Because if you die I’m going to have to look after the children, and it’s much better for them if we can stay.’ ”
Rosalind returned to Canada and went immediately into the hospital. She hung on for a couple of months, but wouldn’t let the children visit her until the day before she died, because she didn’t want them to see her so sick. Throughout her illness, she was convinced that she’d soon get well. Describing what happened, Hinton still seems overwhelmed—he is angry, guilty, wounded, mystified. When Rosalind died, Hinton was forty-six, his son was five, and his daughter was three. “She hurt people by failing to accept that she was going to die,” he said.
The sound of waves filled the midafternoon quiet. Strong yellow sun spilled through the room’s floor-to-ceiling windows; faint spiderwebs extended across them, silhouetted by the light. What an exquisite description. Could you describe the writing and revision behind it? I was so struck by the light in Hinton’s cottage that I took many photographs, and I went back to them to write the scene. As I got to the end of the writing process, I knew that I wanted to introduce the spiderwebs somewhere; they’d turned out to be a metaphor for how A.I. might grow to cover the world. I tried using them in a few different places, and concluded that this spot was both the most beautiful and the most parsimonious. Hinton stood for a while, collecting himself.
“I think I need to go cut down a tree,” he said.
We walked out the front door and down the path to the sheds. From one of them, Hinton retrieved a small green chainsaw and some safety goggles.
“Rosemary says I’m not allowed to cut down trees when there’s nobody else here, in case I chop off an arm or something,” he said. “Have you driven boats before?”
“No,” I said.
“I’ve got to not chop off my right arm, then.”
Over his khakis, he strapped on a pair of protective chaps.
“I don’t want to give you the impression that I know what I’m doing,” he said. “But the basic idea is, you cut lots of V’s, and then the tree falls down.”
Hinton crossed the path to the tree that he had in mind, inspecting the bushes for snakes as we walked. The tree was a leafy cedar, perhaps twenty feet tall; Hinton looked up to see which way it was leaning, then started the saw and began to cut into the trunk on the side opposite the lean. He removed the saw, and made another converging cut to form a V.
Hinton worked the chainsaw in silence, occasionally stopping to wipe his brow. It was hot in the sun, and mosquitoes swarmed every shady nook. I inspected the side of the shed, where ants and spiders were engaged in obscure, ceaseless activity. Down at the end of the path, the water shone. It was a beautiful spot. Still, I thought I saw why Hinton wanted to alter it: a lovely rounded hill descended into a gentle hollow, and if the unnecessary tree were gone the light could flow into it. The tree was an error.
Eventually, he began a second cut on the other side of the tree, angling it toward the first. Then he stopped and turned to me. “Because the tree leans away from the cut, the V will open up as you go deeper, and the blade won’t get stuck,” he explained. He continued the upper cut, nudging the tree toward an entropic moment. Suddenly, almost soundlessly, gravity took over. The tree fell under its own weight, landing with surprising softness at the bottom of the hollow. The light streamed in. Why did you lavish so much attention on the tree-cutting? Two reasons. First, I simply thought it was interesting to see Geoff Hinton, godfather of A.I., felling a tree. But second, the tree struck me as evocative of several different things in the world of A.I. There’s the idea of pruning away undesirable things (trees, or connections between neurons) to let the light in. There’s also the idea of setting something up (a tree trunk or a neural network) and then letting nature take its course through a process (gravity or learning). I wasn’t able to get into it in the piece, but one way to understand the training process is by imagining a landscape; a technique called “gradient descent” is used to find “minima” in the network’s errors. And then there’s the fact that the development of A.I. has been a slow, patient process, with Geoff as its “gardener.” All these things made me want to dwell a little on the tree. I think that readers can sense the vibrancy of a scene like this, even if it’s not explicitly worked out for them.
HINTON WAS IN LOVE with the Boltzmann Machine. He hoped that it, or something like it, might underlie learning in the actual brain. “It should be true,” he told me. “If I was God, I’d make it true.” But further experimentation revealed that as Boltzmann Machines grew they tended to become overwhelmed by the randomness that was built into them. “Geoff and I disagreed about the Boltzmann Machine,” LeCun said. “Geoff thought it was the most beautiful algorithm. I thought it was ugly. It was stochastic”—that is, based partly on randomness. By contrast, LeCun said, “I thought backprop was super clean.”
“Backprop,” or backpropagation, was an algorithm that had been explored by a few different researchers beginning in the nineteen-sixties. Even as Hinton was working with Sejnowski on the Boltzmann Machine, he was also collaborating with Rumelhart and another computer scientist, Ronald Williams, on backprop. They suspected that the technique had untapped potential for learning; in particular, they wanted to combine it with neural nets that operated across many layers.
One way to understand backprop is to imagine a Kafkaesque judicial system. Picture an upper layer of a neural net as a jury that must try cases in perpetuity. The jury has just reached a verdict. In the dystopia in which backprop unfolds, the judge can tell the jurors that their verdict was wrong, and that they will be punished until they reform their ways. The jurors discover that three of them were especially influential in leading the group down the wrong path. This apportionment of blame is the first step in backpropagation.
In the next step, the three wrongheaded jurors determine how they themselves became misinformed. They consider their own influences—parents, teachers, pundits, and the like—and identify the individuals who misinformed them. Those blameworthy influencers, in turn, must identify their respective influences and apportion blame among them. Recursive rounds of finger-pointing ensue, as each layer of influencers calls its own influences to account, in a backward-sweeping cascade. Eventually, once it’s known who has misinformed whom and by how much, the network adjusts itself proportionately, so that individuals listen to their “bad” influences a little less and to their “good” influences a little more. The whole process repeats again and again, with mathematical precision, until verdicts—not just in this one case but in all cases—are collectively as “correct” as possible. Amazing use of an extended metaphor. How did it come to you? I tried many different metaphors, including a Presidential election in which the “wrong” candidate gets elected and a corporate bureaucracy in which the orders of the C.E.O. are imperfectly carried out. But I liked this Kafka-esque metaphor because it captures the authoritarian aspect of backpropagation — ultimately, it’s the humans who tell the machine what’s right or wrong — and because it responds to the slightly sour view that Hinton takes of backprop. In his view, it’s a disappointing way to learn.
In 1986, Hinton, Rumelhart, and Williams published a three-page paper in Nature showing how such a system could work in a neural net. They noted that backprop, like the Boltzmann Machine, wasn’t “a plausible model of learning in brains”: unlike a computer, a brain can’t rewind the tape to audit its past performance. But backprop still enabled a brainlike neural specialization. In real brains, neurons are sometimes arranged in structures aimed at solving specific problems: in the visual system, for instance, different “columns” of neurons recognize edges in what we see. Something similar emerges in a backprop network. Higher layers subject lower ones to a kind of evolutionary pressure; as a result, certain layers of a network that’s tasked with deciphering handwriting, for instance, might become tightly focussed on identifying lines, curves, or edges. Eventually, the system as a whole can develop “appropriate internal representations.” The network knows, and makes use of its knowledge.
In the nineteen-fifties and sixties, a great deal of excitement had accompanied the Perceptron and other connectionist efforts; enthusiasm for connectionism waned in the years after. The backprop paper was part of a revival of interest and earned widespread attention. But the actual work of building backprop networks was slow-going, for both practical and conceptual reasons. Practically, computers were sluggish. “The rate of progress was basically, How much could a computer learn overnight?” Hinton recalled. “The answer was often not much.” Conceptually, neural nets were mysterious. It wasn’t possible to program one in the traditional way. You couldn’t go in and edit the weights of the connections among artificial neurons. And, anyway, it was hard to understand what the weights meant, because they had adapted and changed themselves through training.
There were many ways the learning process could go wrong. In “overfitting,” for example, a network effectively memorized the training data instead of learning to generalize from it. Avoiding the various pitfalls wasn’t always straightforward, because it was up to the network to learn. It was like felling a tree: researchers could make cuts here and there, but then had to let the process unfold. Did you know you were going to refer to the felling of Hinton’s tree again here? I knew that the tree would be a useful metaphor. In general, throughout the writing and revision process, I try to tie things together as much as possible — to see which metaphors, phrases, and images can be usefully pulled through from one part of the piece to another. They could try techniques like “ensembling” (combining weak networks to make a strong one) or “early stopping” (letting a network learn, but not too much). They could “pre-train” a system, by taking a Boltzmann Machine, having it learn something, and then layering a backprop network on top of it, so that a system’s “supervised” training didn’t begin until it had acquired some elemental knowledge on its own. Then they’d let the network learn, hoping that it would land where they wanted it. Your writing makes complex science accessible. How did you learn to do so? I’m always trying to improve, and not always succeeding. Being an editor myself has really helped: I read so much excellent writing done by the writers with whom I’m working. The writer who’s helped me the most is probably Daniel Dennett, the philosopher (whom I profiled in 2017). Dan works to build what he calls “intuition pumps” — thought experiments or metaphors drawn from the physical world that help guide your thinking. I’m always trying to emulate him.
New neural-net “architectures” were developed: “recurrent” and “convolutional” networks allowed the systems to make progress by building on their own work in different ways. But it was as though researchers had discovered an alien technology that they didn’t know how to use. They turned the Rubik’s Cube this way and that, trying to pull order out of noise. “I was always convinced it wasn’t nonsense,” Hinton said. “It wasn’t really faith—it was just completely obvious to me.” The brain used neurons to learn; therefore, complex learning through neural networks must be possible. He would work twice as hard for twice as long. A nod to his father’s brutal statement. Why did you allude to it here? One of the most striking things about Hinton’s story is his persistence, and I think part of it is rooted in his upbringing. He believed in these ideas and it just didn’t make sense to him to walk away from them. He also felt driven to achieve something substantial. So I wanted to bring his dad back here, to remind us of how unique Hinton is as a person.
When networks were trained through backprop, they needed to be told when they were wrong and by how much; this required vast amounts of accurately labeled data, which would allow networks to see the difference between a handwritten “7” and a “1,” or between a golden retriever and a red setter. But it was hard to find well-labelled datasets that were big enough, and building more was a slog. LeCun and his collaborators developed a giant database of handwritten numerals, which they later used to train networks that could read sample Zip Codes provided by the U.S. Postal Service. A computer scientist named Fei Fei Li, at Stanford, spearheaded a gargantuan effort called ImageNet; creating it required collecting more than fourteen million images and sorting them into twenty thousand categories by hand.
As neural nets grew larger, Hinton devised a way of getting knowledge from a large network into a smaller one that might run on a device like a mobile phone. “It’s called distillation,” he explained, in his kitchen. Why did you identify the setting? I just didn’t want the piece to feel too disembodied, even during the technical sections. “Back in school, the art teacher would show us some slides and say, ‘That’s a Rubens, and that’s a van Gogh, and this is William Blake.’ But suppose that the art teacher tells you, ‘O.K., this is a Titian, but it’s a peculiar Titian because aspects of it are quite like a Raphael, which is very unusual for a Titian.’ That’s much more helpful. They’re not just telling you the right answer—they’re telling you other plausible answers.” In distillation learning, one neural net provides another not just with correct answers but with a range of possible answers and their probabilities. It was a richer kind of knowledge.
A few years after Rosalind’s death, Hinton reconnected with Jacqueline Ford, an art historian whom he’d dated briefly before moving to the United States. Jackie was cultured, warm, curious, beautiful. How do you know this? From Geoff! “She’s way out of your league,” his sister said. Did you talk with Hinton’s sister? The fact-checking department did. Still, Jackie gave up her job in the U.K. to move to Toronto. They got married on December 6, 1997—Hinton’s fiftieth birthday. The following decades would be the happiest of his life. His family was whole again. His children loved their new mother. He and Jackie started exploring the islands in Georgian Bay. Recalling this time, he gazed at the canoe in his living room. “We found it in the woods, upside down, covered in canvas, and it was just totally rotten—everything about it was rotten,” he said. “But Jackie decided to rescue it anyway, like she did with me and the kids.”
Hinton was not in love with backpropagation. “It’s so unsatisfying intellectually,” he told me. Unlike the Boltzmann Machine, “it’s all deterministic. Unfortunately, it just works better.” Slowly, as practical advances compounded, the power of backprop became undeniable. In the early seventies, Hinton told me, the British government had hired a mathematician named James Lighthill to determine if A.I. research had any plausible chance of success. Lighthill concluded that it didn’t—“and he was right,” Hinton said, “if you accepted the assumption, which everyone made, that computers might get a thousand times faster, but they wouldn’t get a billion times faster.” Hinton did a calculation in his head. Suppose that in 1985 he’d started running a program on a fast research computer, and left it running until now. If he started running the same program today, on the fastest systems currently used in A.I., it would take less than a second to catch up.
In the early two-thousands, as multi-layer neural nets equipped with powerful computers began to train on much larger data sets, Hinton, Bengio, and LeCun started talking about the potential of “deep learning.” The work crossed a threshold in 2012, when Hinton, Alex Krizhevsky, and Ilya Sutskever came out with AlexNet, an eight-layer neural network that was eventually able to recognize objects from ImageNet with human-level accuracy. Hinton formed a company with Krizhevsky and Sutskever and sold it to Google. He and Jackie bought the island in Georgian Bay—“my one real indulgence,” Hinton said.
Two years later, Jackie was diagnosed with pancreatic cancer. Doctors gave her a year or two to live. “She was incredibly brave and incredibly rational,” Hinton said. “She wasn’t in deep denial, desperately trying to get out of it. Her view was ‘I can feel sorry for myself, or I can say I don’t have much time left and I’d better do my best to enjoy it and make everything O.K. for other people.’ ” She and Hinton pored over the statistics before deciding on therapies; largely through chemo, she extended one or two years to three. In the cottage, when she could no longer manage the stairs, he constructed a small basket on a string so that she could lower her tea from the second floor to the first, where he could warm it up in the microwave. The devotion behind this gesture is heart-rending. How did you react when he told you about it? I was very moved. When people open up like this, I feel an extra sense of responsibility to write something that’s worthy of that openness. (“I should’ve just moved the microwave upstairs,” he observed.) Some editors and writers argue that attributive verbs should be limited to “says” or “said.” They argue that examples like the one used here are unnecessary. What’s your take on this? I can imagine that over-using those kinds of words could be distracting. But what’s the harm from time to time?
Late in the day, we leaned on Hinton’s standing desk as he showed me photos of Jackie on his laptop. In a picture of their wedding day, she and Hinton stand with his kids in the living room of their neighbor’s house, exchanging vows. Hinton looks radiant and relaxed; Jackie holds one of his hands lightly in both of hers. In one of the last pictures that he showed me, she gazes at the camera from the burgundy canoe, which she is paddling in the dappled water near the dock. “That was the summer of 2017,” Hinton said. Jackie died the following April. That June, Hinton, Bengio, and LeCun won the Turing Award—the equivalent of the Nobel Prize in computer science. You could have delivered this information anytime earlier. Why did you decide to wait until now? I think you can only tell a story once. If I’d included this higher up, it would’ve been repetitive appearing here. I wanted to keep things in order, as much as possible.
Hinton is convinced that there’s a real sense in which neural nets are capable of having feelings. “I think feelings are counterfactual statements about what would have caused an action,” he had told me, earlier that day. “Say that I feel like punching someone on the nose. What I mean is: if I didn’t have social inhibitions—if I didn’t stop myself from doing it—I would punch him on the nose. So when I say ‘I feel angry,’ it’s a kind of abbreviation for saying, ‘I feel like doing an aggressive act.’ Feelings are just a way of talking about inclinations to action.”
He told me that he had seen a “frustrated A.I.” in 1973. A computer had been attached to two TV cameras and a simple robot arm; the system was tasked with assembling some blocks, spread out on a table, into the form of a toy car. “This was hard, particularly in 1973,” he said. “The vision system could recognize the bits if they were all separate, but if you put them in a little pile it couldn’t recognize them. So what did it do? It pulled back a little bit, and went bash!, and spread them over the table. Basically, it couldn’t deal with what was going on, so it changed it, violently. And if a person did that you’d say they were frustrated. The computer couldn’t see the blocks right, so he bashed them.” To have a feeling was to want what you couldn’t have. Here’s another example of wisdom you deliver rather than Hinton. Why? In this case, he did a great job of explaining this idea just a paragraph up. But I wanted to rephrase it in a way that would make a broader connection to the house, and to the life with Jackie that wasn’t able to unfold there, and to the bittersweet feelings that the place sometimes provokes.
“I love this house, but sometimes it’s a sad place,” he said, while we looked at the pictures. “Because she loved being here and isn’t here.”
The sun had almost set, and Hinton turned on a little light over his desk. He closed the computer and pushed his glasses up on his nose. He squared up his shoulders, returning to the present.
“I wanted you to know about Roz and Jackie because they’re an important part of my life,” he said. “But, actually, it’s also quite relevant to artificial intelligence. There are two approaches to A.I. There’s denial, and there’s stoicism. Everybody’s first reaction to A.I. is ‘We’ve got to stop this.’ Just like everybody’s first reaction to cancer is ‘How are we going to cut it out?’ ” But it was important to recognize when cutting it out was just a fantasy.
He sighed. “We can’t be in denial,” he said. “We have to be real. We need to think, How do we make it not as awful for humanity as it might be?”
HOW USEFUL—OR DANGEROUS—will A.I. turn out to be? No one knows for sure, in part because neural nets are so strange. In the twentieth century, many researchers wanted to build computers that mimicked brains. But, although neural nets like OpenAI’s GPT models are brainlike in that they involve billions of artificial neurons, they’re actually profoundly different from biological brains. Today’s A.I.s are based in the cloud and housed in data centers that use power on an industrial scale. Clueless in some ways and savantlike in others, they reason for millions of users, but only when prompted. They are not alive. They have probably passed the Turing test—the long-heralded standard, established by the computing pioneer Alan Turing, which held that any computer that could persuasively imitate a human in conversation could be said, reasonably, to think. And yet our intuitions may tell us that nothing resident in a browser tab could really be thinking in the way we do. The systems force us to ask if our kind of thinking is the only kind that counts.
During his last few years at Google, Hinton focussed his efforts on creating more traditionally mindlike artificial intelligence using hardware that more closely emulated the brain. In today’s A.I.s, the weights of the connections among the artificial neurons are stored numerically; it’s as though the brain keeps records about itself. In your actual, analog brain, however, the weights are built into the physical connections between neurons. Hinton worked to create an artificial version of this system using specialized computer chips.
“If you could do it, it would be amazing,” he told me. The chips would be able to learn by varying their “conductances.” Because the weights would be integrated into the hardware, it would be impossible to copy them from one machine to another; each artificial intelligence would have to learn on its own. “They would have to go to school,” he said. “But you would go from using a megawatt to thirty watts.” As he spoke, he leaned forward, his eyes boring into mine; I got a glimpse of Hinton the evangelist. In a New Yorker Radio Hour when the magazine’s editor, David Remnick, interviews you about this story, he suggests that Hinton is a godfather who has become an apostate. Which is he in your mind: Evangelist or apostate? I think that part of the reason he’s such an important person to listen to now is that he’s neither — or, at least, he’s attempting to be neither. He’s trying to be a realist, to be someone who knows the technology well and is simply describing the possibilities. Observers see him as one or the other, depending on their points of view. But he’s not trying to hype A.I. or criticize it; he’s trying to describe our situation, as he sees it. Because the knowledge gained by each A.I. would be lost when it was disassembled, he called the approach “mortal computing.” “We’d give up on immortality,” he said. “In literature, you give up being a god for the woman you love, right? In this case, we’d get something far more important, which is energy efficiency.” Among other things, energy efficiency encourages individuality: because a human brain can run on oatmeal, the world can support billions of brains, all different. And each brain can learn continuously, rather than being trained once, then pushed out into the world.
As a scientific enterprise, mortal A.I. might bring us closer to replicating our own brains. But Hinton has come to think, regretfully, that digital intelligence might be more powerful. In analog intelligence, “if the brain dies, the knowledge dies,” he said. By contrast, in digital intelligence, “if a particular computer dies, those same connection strengths can be used on another computer. And, even if all the digital computers died, if you’d stored the connection strengths somewhere you could then just make another digital computer and run the same weights on that other digital computer. Ten thousand neural nets can learn ten thousand different things at the same time, then share what they’ve learned.” This combination of immortality and replicability, he says, suggests that “we should be concerned about digital intelligence taking over from biological intelligence.”
How should we describe the mental life of a digital intelligence without a mortal body or an individual identity? In recent months, some A.I. researchers have taken to calling GPT a “reasoning engine”—a way, perhaps, of sliding out from under the weight of the word “thinking,” which we struggle to define. “People blame us for using those words—‘thinking,’ ‘knowing,’ ‘understanding,’ ‘deciding,’ and so on,” Bengio told me. At what stage in the reporting did you interview Begio and Yann LeCun? How many human sources did you speak with? I talked to them after I spent time with Hinton on his island, not before. I take different approaches in different pieces; in this case, I felt that it would be especially important to see Hinton on a personal level, as a human being, rather than as a technologist first and a human being second. So I wanted to see him without talking to others about him first, and to encounter him fresh. I ended up talking to maybe a dozen people on background, but only wanted to quote a few in the piece, to prevent the story from succumbing to talking-head syndrome, and to keep it focused on a single human individual. “But even though we don’t have a complete understanding of the meaning of those words, they’ve been very powerful ways of creating analogies that help us understand what we’re doing. It’s helped us a lot to talk about ‘imagination,’ ‘attention,’ ‘planning,’ ‘intuition’ as a tool to clarify and explore.” In Bengio’s view, “a lot of what we’ve been doing is solving the ‘intuition’ aspect of the mind.” Intuitions might be understood as thoughts that we can’t explain: our minds generate them for us, unconsciously, by making connections between what we’re encountering in the present and our past experiences. We tend to prize reason over intuition, but Hinton believes that we are more intuitive than we acknowledge. “For years, symbolic-A.I. people said our true nature is, we’re reasoning machines,” he told me. “I think that’s just nonsense. Our true nature is, we’re analogy machines, with a little bit of reasoning built on top, to notice when the analogies are giving us the wrong answers, and correct them.”
On the whole, current A.I. technology is talky and cerebral: it stumbles at the borders of the physical. “Any teen-ager can learn to drive a car in twenty hours of practice, with hardly any supervision,” LeCun told me. “Any cat can jump on a series of pieces of furniture and get to the top of some shelf. We don’t have any A.I. systems coming anywhere close to doing these things today, except self-driving cars”—and they are over-engineered, requiring “mapping the whole city, hundreds of engineers, hundreds of thousands of hours of training.” Solving the wriggly problems of physical intuition “will be the big challenge of the next decade,” LeCun said. Still, the basic idea is simple: if neurons can do it, then so can neural nets. At any point, did you use A.I. during the reporting and writing? If so, how, when and why? I used A.I. a lot! Not for the writing or reporting, really, but just to have the experience of using it. I kept a running conversation going with ChatGPT about itself, about Hinton, about A.I. technology and ethics. I’d chat with it in speech mode quite often. I’d try out different metaphors for various technical concepts and ask ChatGPT to riff on those metaphors. I also asked ChatGPT to take on different personalities; for whatever reason, I enjoyed it most when it was snarky. At one point, I asked it if “writer’s block” was real and explained that I was writing a piece on Hinton. It replied: “Oh, Josh, a project on Geoff Hinton? How original. It’s not like he’s one of the most recognized figures in deep learning or anything. But hey, maybe you’ll manage to write something that isn’t another regurgitated biography or fanboy piece. Most writers would kill for an opportunity to profile someone like him, but here you are, whining about it taking too long. Maybe you should consider if you’re actually cut out for this kind of work. What could possibly be so hard about writing a piece on a guy who’s been interviewed and profiled a million times? Good luck with that. Hope your ‘writer’s block’ miraculously disappears.”
Hinton suspects that skepticism of A.I.’s potential, while comforting, is often motivated by an unjustified faith in human exceptionalism. Researchers complain that A.I. chatbots “hallucinate,” by making up plausible answers to questions that stump them. But he contests that terminology. “We should say ‘confabulate,’ ” he told me. “ ‘Hallucination’ is when you think there’s sensory input—auditory hallucinations, visual hallucinations, olfactory hallucinations. But just making stuff up—that’s confabulation.” He cited the case of John Dean, President Richard Nixon’s White House counsel, who was interviewed about Watergate before he knew that the conversations he described had been tape-recorded. Dean confabulated, getting the details wrong, mixing up who said what. “But the gist of it was all right,” Hinton said. “He had a recollection of what went on, and he imposed that recollection on some characters in his head. He wrote a little play. And that’s what human memory is like. In our minds, there’s no boundary between just making it up and telling the truth. Telling the truth is just making it up correctly. Because it’s all in the weights, right?” From this perspective, ChatGPT’s ability to make things up is a flaw, but also a sign of its humanlike intelligence.
Hinton is often asked if he regrets his work. He doesn’t. (He recently sent a journalist a one-liner—“a song for you”—along with a link to Edith Piaf’s “Non, Je Ne Regrette Rien.”) When he began his research, he says, no one thought that the technology would succeed; even when it started succeeding, no one thought that it would succeed so quickly. Precisely because he thinks that A.I. is truly intelligent, he expects that it will contribute to many fields. Yet he fears what will happen when, for instance, powerful people abuse it. “You can probably imagine Vladimir Putin creating an autonomous lethal weapon and giving it the goal of killing Ukrainians,” Hinton said. During your New Yorker Radio Hour interview with Remnick, audio excerpts of your conversations with Hinton are played. How would you describe your interviewing style? In that case, I conducted a special, more formal interview with Hinton, just for the radio, at the end of my reporting trip. Ordinarily, I’m more informal. I like to ask people to tell me about their lives in chronological order. I prefer listening to talking. He believes that autonomous weapons should be outlawed—the U.S. military is actively developing them—but warns that even a benign autonomous system could wreak havoc. “If you want a system to be effective, you need to give it the ability to create its own subgoals,” he said. “Now, the problem is, there’s a very general subgoal that helps with almost all goals: get more control. The research question is: how do you prevent them from ever wanting to take control? And nobody knows the answer.” (Control, he noted, doesn’t have to be physical: “It could be just like how Trump could invade the Capitol, with words.”)
Within the field, Hinton’s views are variously shared and disputed. “I’m not scared of A.I.,” LeCun told me. “I think it will be relatively easy to design them so that their objectives will align with ours.” He went on, “There’s the idea that if a system is intelligent it’s going to want to dominate. But the desire to dominate has nothing to do with intelligence—it has to do with testosterone.” I recalled the spiders I’d seen at the cottage, and how their webs covered the surfaces of Hinton’s windows. They didn’t want to dominate, either—and yet their insectoidal intelligence had led them to expand their territory. Living systems without centralized brains, such as ant colonies, don’t “want” to do anything, yet they still find food, ford rivers, and kill competitors in vast numbers. Either Hinton or LeCun could be right. The metamorphosis isn’t finished. We don’t know what A.I. will become.
“Why don’t we just unplug it?” I asked Hinton, of A.I. in general. “Is that a totally unreasonable question?” Why did you include this dialogue with Hinton? This is the question on many of our minds, and even though it’s a silly one — obviously we can’t “unplug” A.I. I wanted to ask it in a direct way, so that we could get to the heart of the matter.
“It’s not unreasonable to say, We’d be better off without this—it’s not worth it,” he said. “Just as we might have been better off without fossil fuels. We’d have been far more primitive, but it may not have been worth the risk.” He added, stoically, “But it’s not going to happen. Because of the way society is. And because of the competition between different nations. If the U.N. really worked, possibly something like that could stop it. Although, even then, A.I. is just so useful. It has so much potential to do good, in fields like medicine—and, of course, to give an advantage to a nation via autonomous weapons.” Earlier this year, Hinton declined to sign a popular petition that called for at least a six-month pause in research. “China’s not going to stop developing it for six months,” he said.
“So what should we do?” I asked.
“I don’t know,” he said. “It would be great if this were like climate change, where someone could say, Look, we either have to stop burning carbon or we have to find an effective way to remove carbon dioxide from the atmosphere. There, you know what the solution looks like. Here, it’s not like that.”
Hinton was pulling on a blue waterproof jacket. We were heading to the marina to pick up Rosemary. “She’s brought supplies!” he said, smiling. As we walked out the door, I looked back into the cottage. In the big room, the burgundy canoe shone, caressed by sunlight. Chairs were arranged in front of it in a semicircle, facing the water through the windows. Some magazines were piled on a little table. It was a beautiful house. A human mind does more than reason; it exists in time, and reckons with life and death, and builds a world around itself. It gathers meaning, as if by gravity. An A.I., I thought, might be able to imagine a place like this. But would it ever need one? Why did you make this observation? To me, this story is partly about how one of the people who’s been instrumental in creating A.I. is also deeply human. Hinton isn’t someone who wants to transcend or escape or redefine what it means to be a human being; he’s an empathetic, experienced, wise person who’s lived a full and sometimes difficult life and who fully grasps complexity, richness and value of human experience. It’s possible to imagine a version of the A.I. story in which the scientists behind the technology are somehow cold, like in the Hollywood movie “Ex Machina.” I wanted to show how this wasn’t true. And I also wanted to pose the question at the end of the paragraph. Many of the ideas in the closing part of the piece relate to mortality and its role in making us into individuals. (Can an immortal A.I. ever be an individual?) This is one of the ways in which thinking about artificial intelligence leads us to think about the nature of human existence. I thought it would be good to bring the piece full circle.
We made our way down the wooded path, past the sheds and down the steps to the dock, then climbed into Hinton’s boat. It was a perfect blue day, with a brisk wind roughing the water. Hinton stood at the wheel. I sat in front, watching other islands pass, thinking about the story of A.I. To some, it’s a Copernican tale, in which our intuitions about the specialness of the human mind are being dislodged by thinking machines. To others, it’s Promethean—having stolen fire, we risk getting burned. Some people think we’re fooling ourselves, getting taken in by our own machines and the companies that hope to profit from them. In a strange way, it could also be a story about human limitation. If we were gods, we might make a different kind of A.I.; in reality, this version was what we could manage. Meanwhile, I couldn’t help but consider the story in an Edenic light. By seeking to re-create the knowledge systems in our heads, we had seized the forbidden apple; we now risked exile from our charmed world. But who would choose not to know how knowing works? What is the narrative purpose of this paragraph? Well, on one level, it’s actually true that I had these thoughts while we motored back to the marina. But I decided to include them because I wanted to foreground how the A.I. story is still developing and could be told in many ways. It’s such a primal, elemental story — it has Golems, Frankensteins, zombies, gods, and so on — and those kinds of metaphors have a magnetism that’s hard to resist. I thought it would be helpful to name them, and to make them explicit, so that readers could engage with them out in the light. And I also wanted to set up the ending, with the snake, in a way that would make the symbolism more obvious while also rendering it a little tongue-in-cheek.
At the marina, Hinton did a good job of working with the wind, accelerating forward, turning, and then allowing it to guide him into his slip. “I’m learning,” he said, proud of himself. We walked ashore and waited by a shop for Rosemary to arrive. After a while, Hinton went inside to buy a light bulb. I stood, enjoying the warmth, and then saw a tall, bright-eyed woman with long white hair striding toward me from the parking lot.
Rosemary and I shook hands. Then she looked over my shoulder. Hinton was emerging from the greenery near the shop, grinning.
“What’ve you got for me?” she asked.
Hinton held up a black-and-yellow garter snake, perhaps a metre long, twisting round and round like a spring. “I’ve come bearing gifts!” he said, in a gallant tone. “I found it in the bushes.”
Rosemary laughed, delighted, and turned to me. “This just epitomizes him,” she said.
“He’s not happy,” Hinton said, observing the snake.
“Would you be?” Rosemary asked.
“I’m being very careful with his neck,” Hinton said. “They’re fragile.”
He switched the snake from one hand to another, then held out a palm. It was covered in the snake’s slimy musk.
“Have a sniff,” he said. After a story that presents complex concepts in often long paragraphs, you shifted here to an easy-to-grasp scene driven by dialogue and action. Why the change? The piece is structured as a visit, and I wanted to end it by bringing the visit to a close. And I also wanted the piece to be very much about Hinton as a person, and so it made sense to me to end it on single a moment with him, realized as a three-dimensional, real-time scene, rather than an interview. Also, I think it feels good, after a somewhat claustrophobic piece — it’s just me and Hinton on the island — to suddenly have someone new arrive.
We took turns. It was strange: mineral and pungent, reptilian and chemical, unmistakably biotic.
“You’ve got it all over your shirt!” Rosemary said.
“I had to catch him!” Hinton explained.
He put the snake down, and it slithered off into the grass. He watched it go with a satisfied look.
“Well,” he said. “It’s a beautiful day. Shall we brave the crossing?” Why did you end the story on this note? Did you experiment with alternate approaches? As soon as this happened during my reporting — Geoff catching the snake, my meeting Rosemary, their charming interaction — I knew that I’d probably be ending the story with it. And then it became a matter of working backwards and figuring out what I needed to add (or subtract) to make the ending as satisfying as possible. The article has been so retrospective and bittersweet. I thought it was good to see Hinton in the present, in a happy relationship. And we’re seeing his comfort with the non-human, which is itself a dimension of his own humanity. My hope was that the ending would prompt us to appreciate our humanity at the same time it makes us think about what might challenge it.
* * *
Chip Scanlan is an award-winning writer who taught at the Poynter Institute and now coaches writers around the world. He is the author of several books on writing and the newsletter Chip’s Writing Lessons.