John McCarthy
| John McCarthy | |
|---|---|
| Institutions | Stanford University |
| Fields | Artificial intelligence; Computer science; Mathematics |
| Occupation | Computer scientist |
| Coined terms | artificial intelligence |
| Known for | Lisp; Artificial intelligence; Time-sharing systems |
| Invented | Lisp programming language |
| Wikidata | Q738352 |
John McCarthy: Inventor of Lisp and Pioneer of Artificial Intelligence
John McCarthy (1927–2011) was an American computer scientist renowned as a founding figure of artificial intelligence (AI). He coined the term “artificial intelligence” and organized the famous 1956 Dartmouth conference that sparked the field. McCarthy invented the Lisp programming language – the second-oldest high-level language after Fortran – and conceived time-sharing systems that presaged modern cloud computing. Throughout a long career at MIT and especially Stanford, he championed logical, mathematical approaches to AI and left a legacy of ideas that still influence computing today.
Early Life and Education
Born in Boston on September 4, 1927, John Patrick McCarthy grew up during the Great Depression and later moved to Los Angeles. He showed unusual academic promise from a young age. McCarthy skipped two years of high school and entered the California Institute of Technology (Caltech) in 1944. He majored in mathematics, graduating with a Bachelor of Science in 1948. He went on to earn a Ph.D. in mathematics from Princeton University in 1951 under the supervision of Solomon Lefschetz, a prominent topologist. McCarthy’s early work in mathematics laid the groundwork for his logical and formal approach to computer science.
After finishing his Ph.D., McCarthy held brief academic positions at Princeton and then Stanford University (1953–1955). He began teaching computer science at Stanford when the field was very new. In 1955, he took a post at Dartmouth College but soon moved to the Massachusetts Institute of Technology (MIT). At MIT in the late 1950s, he and Marvin Minsky (among others) set up the first AI research project. In 1962 McCarthy joined Stanford University as a professor of mathematics, and when Stanford started a new Computer Science Department in 1965, he became one of its founding faculty. He remained at Stanford for nearly four decades, directing research and mentoring students until retiring in 2001.
Major Works and Ideas
Origins of Artificial Intelligence
John McCarthy’s 1955 proposal for a summer research project on “artificial intelligence” at Dartmouth College is widely seen as the birth of AI as a formal field. In that proposal, McCarthy and his colleagues (including Marvin Minsky, Nathaniel Rochester, and Claude Shannon) described AI as “the endeavor to make machines use language, form abstractions and concepts, solve kinds of problems now reserved for humans, and improve themselves.” This summer 1956 Dartmouth workshop was the first time the term “artificial intelligence” appeared in print, and it helped launch the study of machine intelligence.
McCarthy believed that any aspect of learning or intelligence could in principle be precisely described so that a machine could simulate it. He famously wrote that “every aspect of learning or any other feature of intelligence can in principle be so precisely described that a machine can be made to simulate it.” This vision of general intelligence guided much of his work. He proposed the “Advice Taker” in 1958 – a thought experiment about a program that would accept general rules and facts in logical form, effectively taking advice from written instructions. This idea inspired later work in question-answering systems and logic programming.
In 1960 McCarthy authored a landmark paper called “Programs with Common Sense”. In it he outlined the goal of creating a program that would use logical sentences (“common sense” knowledge) to make decisions. He described a hypothetical “suggestor” program that could reason about party invitations and attendance by understanding facts like “if it’s probably going to rain the party might be indoors.” This paper laid out the principle that AI systems should use declarative knowledge: facts and rules stated in a high-level language, rather than hard-coded procedural instructions. His guidance was, in McCarthy’s words, “Sentences can be true in much wider contexts than specific programs can be useful.” This philosophy – separating knowledge (sentences) from the code that uses it – influenced knowledge representation in AI for decades.
Lisp and Programming Languages
One of McCarthy’s most enduring legacies is the creation of Lisp in 1958 (with major publications in 1960). Lisp, which stands for LISt Processing, was designed for symbolic computation – manipulating symbols rather than numbers – which made it well-suited to AI research. Lisp introduced ideas that became fundamental in computer science, such as using linked lists as a basic data structure, treating code as data (so programs could manipulate themselves), and automatic garbage collection (a way to manage unused memory). In Lisp, functions and data share the same structure (both are lists), allowing very flexible programming.
Lisp quickly became the preferred language for AI development. It allowed researchers to write programs that could, for example, manipulate symbolic expressions, search through possibilities, or represent logical rules. Some of McCarthy’s own AI programs, such as those for playing chess or reasoning about physical objects, were written in Lisp or its descendants. Lisp also influenced later programming languages and paradigms. For example, the concept of functional programming (writing programs in terms of mathematical functions) and features like recursion and higher-order functions (functions that take other functions as arguments) were popularized by Lisp. Dialects of Lisp remain in use today (for instance, Scheme, Clojure, and Lisp’s heir used in the Emacs text editor), illustrating Lisp’s lasting impact.
In addition to Lisp, McCarthy contributed to other programming language designs. He was involved in the early definition of ALGOL, an influential language of the 1960s, suggesting features such as recursion and conditional expressions. These features later appeared in the final ALGOL 60 standard. In summary, McCarthy’s work on programming languages was guided by the idea of creating powerful, expressive notations for computation.
Time-Sharing and Utility Computing
In the late 1950s McCarthy turned his attention to how computers should be used efficiently. The computers of that era were large and expensive; he proposed that multiple users should be able to share a single machine “timesharing” style. In a now-famous January 1959 memo, he described a general way to let several people use one computer simultaneously by giving each user the machine for a short turn at a time. This idea of time-sharing allowed interactive use of computers (rather than running one job at a time on punch cards) and greatly increased productivity for researchers. His concept inspired development of early time-sharing systems at MIT by teams that included Fernando Corbató and others.
McCarthy also saw a commercial and societal future for shared computing resources. In 1961 he publicly suggested that computing power might become a utility – a service sold like water or electricity – thanks to time-sharing. Today’s cloud computing, where companies buy computing time over the internet, can trace its roots to this vision. Many colleagues credit McCarthy’s time-sharing ideas with speeding up development of networked computing. In his words relayed by one Stanford colleague, McCarthy “started” the concept that eventually led to the Internet’s core, since the ARPANET began as a network of timesharing computers.
Other Research Contributions
McCarthy made numerous other contributions across computer science. In the 1960s he and his students created one of the first computer vision and robotics systems. This “hand-eye” project used a camera and a robotic arm: the computer could see blocks with the camera and command the robot to stack them. It was an early demonstration of a machine processing visual data and acting on it.
From 1978 to the mid-1980s, McCarthy developed circumscription, a technique for non-monotonic reasoning in logic. (Non-monotonic means that conclusions drawn can change when new information comes in, unlike standard logic where adding facts only increases what follows.) Circumscription was his mathematical approach to giving computers common-sense “defaults”. For example, most birds can fly unless specified otherwise; circumscription attempts to formalize reasoning such as “by default, assume things are normal unless told otherwise.” This work addressed problems of representing everyday knowledge that often has exceptions. It influenced research in knowledge representation and reasoning in AI: techniques like default logic and answer set programming owe something to his ideas.
He also contributed to formal verification, the rigorous proving that computer programs meet their specifications. McCarthy and his colleagues investigated ways to mathematically prove properties of programs, a field still important for ensuring software correctness. Additionally, McCarthy proposed far-reaching ideas outside core AI: in 1982 he described a “space fountain,” a thought experiment for building a tower into space kept aloft by a stream of objects; this was one of many creative concepts he floated about advanced technology.
Research Methodology and Philosophy
McCarthy’s approach to research was methodical and mathematically oriented. He firmly believed that intelligence, at its core, could be captured by formal logic and precise statements. As noted above, he insisted that machines’ knowledge be represented in declarative form – as logical sentences or facts – instead of hidden in procedural code. He urged the AI community to focus on clear reasoning and knowledge representation. One of his Stanford colleagues summed up the idea: rather than Write a program that does something, McCarthy wanted machines that know things in logical form. This separation of knowledge and reasoning algorithms was a hallmark of his style.
He extended this logic-based view in several ways. John McCarthy realized that ordinary first-order logic (the basic logical mathematics) couldn’t by itself capture all aspects of common sense and context. So he worked on ideas like contexts (ways to say “in this situation these rules apply”) and circumscription (mentioned above) to handle exceptions and defaults. He thought deeply about how a logical system could represent changing or incomplete information – for example, formalizing everyday reasoning like “normally assume I have a key, unless I remember I lost it.” This work on common-sense reasoning showed his commitment to grounding AI in mathematical clarity, even for messy human knowledge.
In personality and style, McCarthy was known for intense focus and sharp wit. He expected precision and often quipped about the importance of mathematics. A favorite motto was, “He who refuses to do arithmetic is doomed to talk nonsense.” He meant that clear, quantitative thinking was essential; without it, one could only make unfounded claims. This became part of his personal brand: his email signature and a license-plate frame once read “Do the arithmetic or be doomed to talk nonsense.” His approach to problems was rigorous and uncompromising: he typically concentrated deeply on one issue at a time. Colleagues noted that he kept a sense of humor about everything, even as he worked on abstract or futuristic ideas. He described his own outlook as “radical optimism,” meaning he firmly believed almost any problem could ultimately be solved with enough ingenuity.
McCarthy was also a staunch advocate for open, wide-ranging inquiry. He supported free speech wholeheartedly and believed scientists should apply the same rigorous standards of logic and evidence to all claims, even in politics or social issues. In the AI community, he often played devil’s advocate to test ideas. For instance, he engaged in public debates on whether machines can truly “think” or have beliefs, defending the notion that computers could be seen to have goals and beliefs in a technical sense. This philosophical open-mindedness – along with his humor and mathematical bent – made him an influential but sometimes polarizing intellectual leader.
Influence on AI and Computing
John McCarthy’s influence on computer science and AI is profound and varied. First and foremost, he helped create the field itself. By coining “artificial intelligence” and organizing the Dartmouth conference, he set the agenda and gave researchers a unifying goal. For years afterward, he guided others in that goal. He co-founded two major AI research centers: the MIT Artificial Intelligence Project (in 1959) and the Stanford Artificial Intelligence Laboratory (SAIL) in 1965. As a director and mentor at Stanford, McCarthy built SAIL into a thriving community. During the 1960s and 1970s, SAIL became a pioneer in many technologies beyond AI: its researchers developed early computer graphics, networked printers, speech understanding, computer music, and much more. In this way he indirectly influenced fields like human–computer interaction and multimedia.
Many leading AI figures were McCarthy’s students or collaborators. For example, Edward (Ed) Feigenbaum, a student at Stanford, became a father of “expert systems” (AI programs that use vast rules of knowledge) and won a Turing Award himself. Nils Nilsson, co-author of a classic AI textbook, was another Stanford colleague. The community that McCarthy helped forge propagated his methods: generations of AI researchers adopted logic-based methods, knowledge representation, and the Lisp language in their work. Even today, many computer science courses start by teaching Lisp or its concepts, partly due to McCarthy’s legacy.
Lisp’s influence on computing can be seen even for those who never used it directly. Concepts like “garbage collection” (automatic memory management) became standard in later languages such as Java or Python. The idea of treating code as data paved the way for modern techniques (for example, many machine learning frameworks embed models as data). Functional programming, which is now common in languages like Haskell or functional libraries for JavaScript, traces back to Lisp’s emphasis on functions as first-class objects. In short, the principles McCarthy introduced broadened beyond AI into programming at large.
McCarthy’s early work on timesharing also shaped the computing era. Because he advocated making computing powerful and interactive, the eventual spread of networked computers and the internet owes something to his vision. As mentioned, the connection from time-sharing to ARPANET to today’s cloud means that many take-for-granted resources ultimately emerged from the ideas he first sketched.
Finally, McCarthy’s influence extends to how people think about AI’s potential. He was one of the first to articulate the aspiration that machines could eventually replicate human-level intelligence in full. This ideal – passing the Turing test, or handling any intellectual task a person can – has been a guiding light for some researchers (though not all share this uncompromising aim). By setting such a high bar, McCarthy helped provoke deep questions about cognition, reasoning, and common sense that continue in AI research. Even in modern debates over narrow vs. general AI, his legacy looms large: his work reminds us that AI is not only about useful applications but about understanding intelligence itself.
Critique and Debate
While John McCarthy’s contributions were pioneering, his ideas were also subject to debate and critique as the field evolved. Early AI was dominated by the logic-based, “symbolic” approach he championed. Some later researchers argued that this approach was too brittle or narrow. For example, Marvin Minsky and others developed neural network models (connectionist approaches) to mimic aspects of the human brain, something McCarthy was skeptical of. Critics say purely symbolic AI struggled with real-world uncertainty and perception in ways that neural networks handle naturally. In the late 1980s and 1990s, “connectionism” (neural methods) resurged and some of McCarthy’s colleagues felt his logic-heavy orientation was out of step with data-driven machine learning. McCarthy himself quipped about the limitations of focusing on domains like chess. After IBM’s Deep Blue beat Garry Kasparov in 1997, McCarthy joked that computer chess was the “genetics of artificial intelligence” – useful for research but not the same as understanding true intelligence. In other words, he was critical of narrow AI successes, just as others later critiqued symbolic AI for not handling noisy data or learning well.
Some have also noted that McCarthy’s vision of rapidly achieving general AI has proven too optimistic. In 1956 at Dartmouth he and colleagues guessed that a machine with commonsense reasoning might be created “within a generation,” but by the 21st century human-level AI remained unrealized. Skeptics say this shows that human cognition is even more complex than early pioneers believed. Nonetheless, McCarthy acknowledged these difficulties. He continued working on problems like common-sense reasoning for decades, and often lamented that no one had fully solved them. In a sense his critics were proven right on timing, but his focus on those challenges remained influential.
Philosophy of mind issues also put McCarthy’s ideas in conversation with critics. For instance, in 1979 McCarthy wrote an essay arguing that even simple machines (e.g., thermostats or calculators) could be said to “have beliefs” or “goals” in a technical sense. He viewed mentalistic language as a useful abstraction: if a computer acts to achieve an outcome, we can metaphorically say it “believes” that outcome is desirable. Philosopher John Searle famously responded with the “Chinese Room” thought experiment in 1980, contending that syntactic manipulation of symbols (like computers do) could never amount to true understanding or consciousness. The Searle-McCarthy exchange highlights a fundamental critique: does McCarthy’s logical model of mind really capture what consciousness or genuine understanding is? McCarthy would respond that clear-cut logical models at least give a starting point for analysis, but the debate reveals that his optimistic view of machine “beliefs” was not universally accepted in philosophy. It remains an open question.
Lastly, some found McCarthy’s personal style challenging. He was known to be direct, even brusque at times, when discussing theory. Younger students sometimes found him intimidating, though his emphasis on clarity arguably served a purpose: he wanted people to have tough, honest conversations about difficult problems. Over time, the AI community as a whole has balanced his high-level vision with more incremental, empirical approaches. It’s fair to say that while McCarthy was revered as a leader, AI matured by drawing on both his insights and those of others who focused on different methods (statistical learning, brain-inspired models, etc.).
Legacy
John McCarthy’s legacy in computing and AI is vast and enduring. He is often called the “father of AI” because he not only helped start the field but shaped its questions and methods for decades. Technologies and ideas he introduced remain in use. Lisp and its descendants are still taught and used in areas of AI research and academia, reflecting his influence on programming. The phrase “artificial intelligence” is now global; McCarthy put that label on the map. The concept of writing computer programs that reason, with knowledge organized in logic, continues in many AI systems today – for example, rule-based expert systems or planning programs.
His vision of time-sharing lives on every time we run a program in the cloud or click on a web app. His optimistically extreme goal – passing the Turing test – set a high bar that still inspires AI researchers to consider the big picture of intelligence. Several modern AI labs trace their roots to McCarthy’s Stanford lab: research on vision, robotics, natural language, and knowledge systems all grew from the environment he helped to start.
McCarthy won many of computer science’s highest honors, reflecting his impact. He received the A.M. Turing Award in 1971 (often called the “Nobel Prize of Computing”) for his work on non-numerical programming and AI. He later won the Kyoto Prize (1988) and the U.S. National Medal of Science (1990). He was elected to the National Academy of Sciences and the National Academy of Engineering. These awards cemented his reputation as a visionary thinker.
After retirement and his death in 2011, colleagues and students remembered McCarthy as a brilliant but accessible scholar. Many of his papers and notes remain available on his Stanford web archive for students and researchers. Tributes emphasize his humor and focus – Dale Feigenbaum remembered him as generous with ideas, and others recall his encouraging of creativity balanced by insistence on rigor.
In education, many AI textbooks and courses still cite McCarthy’s early results. Courses on programming languages teach Lisp history, and on AI teach about logical reasoning – all because of him. Even though modern AI also relies on statistical machine learning, many researchers recognize that McCarthy’s emphasis on clear definitions and rational debate remains important.
In short, John McCarthy’s legacy is that he gave AI a name, a foundation in logic and language, and a culture of bold vision. He made computers into tools for exploring what intelligence means. Today’s advances in AI (from game-playing programs to self-driving cars) stand on a base that McCarthy helped construct. Whether or not we achieve “thinking machines,” his influence is felt in every corner of AI and in the very idea that machines can emulate human thought.
Selected Works
- “Proposal for the Dartmouth Summer Research Project on Artificial Intelligence” (1955): McCarthy’s document coining “artificial intelligence” and sketching the first AI research workshop.
- “The Advice Taker” (1958, concept paper): Early idea for an AI program using formal logic statements as its input (though not formally published until later).
- “Programs with Common Sense” (1960): Paper outlining how a program could reason with everyday knowledge expressed as logical sentences.
- “Recursive Functions of Symbolic Expressions and Their Computation by Machine” (1960): Foundational paper describing Lisp and its use of symbolic expressions (lists) to perform computation.
- “Circumscription: A Form of Non-Monotonic Reasoning” (1980): Research paper developing circumscription, a formal method for enabling computers to draw commonsense conclusions by default.
Timeline
- 1927: Born in Boston, Massachusetts.
- 1948: B.S. in Mathematics from Caltech.
- 1951: Ph.D. in Mathematics from Princeton University.
- 1956: Co-organizes the Dartmouth workshop that introduces the term “artificial intelligence.”
- 1958: Invents the Lisp programming language at MIT/Stanford.
- 1959: Publishes concept of computer time-sharing.
- 1960: Publishes “Programs with Common Sense”; Lisp is established as an AI tool.
- 1965: Founds Stanford Artificial Intelligence Laboratory (SAIL).
- 1971: Receives the ACm A.M. Turing Award for contributions to AI.
- 1978–86: Develops circumscription for non-monotonic reasoning in AI.
- 1988: Awarded the Kyoto Prize in Advanced Technology.
- 1990: Awarded the U.S. National Medal of Science.
- 2001: Retires as Professor Emeritus at Stanford.
- 2011: Passes away at age 84 in Stanford, California.