This is Part 1 of a four-part series on the history of artificial intelligence.
Part 1 — Thinking automata and the dawn of computation (1920s–1960s)
Part 2 — AI winter, expert systems, and the rise of machine learning (1970s–2000s)
Part 3 — Deep learning, AlphaGo, and the language renaissance (2010s)
Part 4 — ChatGPT, generative AI, and the Web4 horizon (2020s and beyond)
Before the Machines Spoke
If we could flip through newspapers from 1948, we might find a headline reading "The Clicking Brain Is Smarter Than a Man." It sounds like science fiction. That's precisely what fueled the idea.
Long before transistors or mainframes existed, the idea of artificial intelligence was already alive in literature. In 1920, Czech writer Karel Čapek introduced the word robot in his play R.U.R., referring to synthetic laborers. In 1942, Isaac Asimov set down his Three Laws of Robotics — an ethical blueprint for machines that didn't yet exist but already felt inevitable.
While writers theorized, engineers built. Not artificial minds, exactly. Strange, electromechanical organisms that acted, responded, and sometimes won.
El Ajedrecista — A Chess Machine Before Computers
As early as 1912, Spanish inventor Leonardo Torres Quevedo unveiled El Ajedrecista — a chess-playing machine that could checkmate a human opponent without guidance. The board was wired: each square contained metal contacts, completed a circuit when a human moved a piece, and the machine processed the signal through relays. It calculated its next move and physically shifted its own rook or king with magnetic arms beneath the surface.
It wasn't just functional. It was dramatic. If an opponent made an illegal move, a lamp lit up in protest. Three illegal moves in a row, and the machine stopped the game. When it delivered checkmate, a phonograph announced "Check to the king!"
This wasn't a gimmick. It was the first time the logic of a complex game — rules, strategy, enforcement — was encoded into a thinking object.
Walter's Electric Turtles
Three decades later, across the Channel, British neurophysiologist W. Grey Walter introduced two curious machines: Elmer and Elsie. Small, round-bodied, tricycle-like creatures covered in translucent domes. They looked like toys. They moved with eerie purpose.
Each had two vacuum tubes acting as a simple analog "brain." A photocell sensed light. A bump switch sensed collisions. That was it. But it was enough. The robots moved toward light sources, slowed near harsh beams, bumped into walls, reversed, adjusted. When their batteries ran low, they returned to a "nest" marked by a soft lamp, recharged themselves, then rolled back out into the world.
People watched, mesmerized. The machines seemed curious, even cautious — almost as if they wanted something. Walter wasn't trying to simulate the brain. He was exploring how a tiny system with minimal internal complexity could display what looked like lifelike behavior. He called the creatures machina speculatrix — machines that observed and reacted. They weren't programmed. They were wired to wander.
Ashby's Homeostat — A Machine That Persisted
Then came Ross Ashby, whose creation didn't walk, speak, or play games. It balanced.
In 1948, Ashby built the Homeostat — a black metal box with four interlinked units of coils, fluid, and pivoting metal arms. Each unit constantly sent and received voltages from the others. If one was pushed out of equilibrium, the others compensated, recalibrated, shifted. Eventually the system found a new point of balance. If disturbed again, it adapted again. No software. Just feedback.
It looked modest. To observers, it felt alive. Time magazine called it "the closest thing to a synthetic brain yet built." Ashby called it ultrastable. The Homeostat didn't think. It persisted.
Together, these three machines didn't simulate intelligence. They made behavior visible. They showed that intention, adaptation, even a kind of personality could emerge not from cognition, but from feedback, sensors, and structure. Long before Turing's test, they were already asking the real question:
How little does it take for a machine to appear alive?
The Word That Started It All
In 1950, Alan Turing threw a question into the wind that's still echoing: "Can machines think?"
He proposed a test that was disarmingly simple. If a machine can respond in a way indistinguishable from a human, then we must, reluctantly or not, admit its intelligence. The Turing Test was born.
While some asked the question, others rolled up their sleeves and built — sometimes literal mousetraps.
Shannon's Theseus — A Mouse That Learned
One of the strangest experiments came from Claude Shannon, already famous as the father of information theory. In the early 1950s, Shannon introduced Theseus — a squat, wheeled "mouse" crawling through a plexiglass maze. To the casual observer, it looked like a science-fair curiosity. But inside the wooden cabinet beneath the maze, nearly 90 electromechanical relays clicked and buzzed, remembering the route the mouse had taken.
On its first run, Theseus bumped into walls and wandered blindly. With each successful turn, the machine recorded its steps — not in code, but in metal and wire. When placed back at the start, it glided toward the goal as if it had always known the way.
It hadn't been programmed. It had learned.
Dartmouth and the Birth of a Discipline
In 1956, at Dartmouth College, John McCarthy convened a summer research workshop with a bold title: The Dartmouth Conference on Artificial Intelligence. He gathered a cadre of visionaries — Shannon, Marvin Minsky, Nathaniel Rochester — and gave the field its name.
From that moment on, AI was no longer a philosophical side quest. It was an emerging science.
Newell and Simon's Logic Theorist
In that same pivotal year, Allen Newell and Herbert Simon created the Logic Theorist, widely regarded as the first true AI program. It didn't live in a robot's body. It had no wheels, no whiskers. It existed inside the hulking frame of an IBM mainframe, where it quietly proved theorems from Whitehead and Russell's Principia Mathematica.
Not replicate them — prove them, from scratch. It succeeded on 38 out of 52. In one case, it discovered a proof shorter and more elegant than the published version.
Logic Theorist didn't know anything. But it manipulated abstract symbols according to rules, combining logic, heuristics, and backtracking. It wasn't mimicry. It was reasoning, in digital form. And it suggested a radical idea: cognition might be mechanizable.
Rosenblatt's Perceptron
While Newell and Simon simulated thought, Frank Rosenblatt tried to recreate something more primal — sight.
In 1957, he introduced the Perceptron. Not a metaphor. Not a simulation. A real, physical device with a grid of 400 photosensors — a 20×20 array mimicking a retina — wired to a layer of "neurons" and then to a set of output units. The hidden connections between layers could be adjusted physically, using motors and potentiometers. When the system failed to recognize a pattern, it would literally twist itself into a new configuration. Trial by adjustment.
It learned to recognize simple shapes — lines, letters, geometric patterns. After each failure, it modified its weights. Over time, accuracy improved.
The Perceptron was presented to the press in 1960 as the future of machine vision. Newspapers buzzed with talk of a machine that would soon "walk, talk, and reproduce itself."
That didn't happen. But the architecture Rosenblatt built laid the foundations for every neural network that followed — including the ones being used today to recognize objects in photos, translate languages, and drive cars.
Minsky's SNARC
Meanwhile, McCarthy was writing Lisp, a programming language as strange and beautiful as the ideas it was built to express. Recursive, expressive, symbolic — it would become the lingua franca of AI research for decades.
And then came a machine that didn't just adjust weights or run logic trees. It learned by doing.
In 1951, a young Marvin Minsky, along with Dean Edmonds, created SNARC — the Stochastic Neural Analog Reinforcement Calculator. They salvaged military-grade components from old B-24 autopilot systems, wired together over 3,000 vacuum tubes, and built a neural network that modeled a rat learning to navigate a maze.
Each "neuron" was an analog component. The learning was Hebbian: if a connection led to success, it was reinforced. Not in code. In current, voltage, and real-time electrical feedback. The virtual rat didn't know what it was doing. But it learned what worked, and repeated it.
The machine overheated. It failed often. But when it worked, it adapted.
This was a different paradigm. Not command and control. Not rational planning. Trial, error, reinforcement — a conceptual ancestor of modern reinforcement learning. It would take decades for the field to circle back to this insight. But SNARC was already there, clicking in the dark.
Optimism, Chatbots, and the First Robots
The 1960s were charged with the spirit of progress. Humanity was aiming for the Moon, and on Earth, scientists were building electronic minds.
ELIZA — The First Chatbot
In 1961, MIT's Joseph Weizenbaum unveiled ELIZA — one of the first chatbots, designed to simulate a Rogerian psychotherapist. ELIZA implemented the DOCTOR script, mimicking therapy with keyword matching and templated prompts.
Remarkably, users — even when told it was code — often sensed empathy and comprehension. The phenomenon became known as the ELIZA effect: the tendency to attribute human qualities to software that merely reflects them back.
Weizenbaum's intent wasn't to birth chatbots. It was to study human–machine dialogue. Ironically, ELIZA became the unintended ancestor of every conversational agent that followed — including the ones living in your phone today.
DENDRAL — The First Expert System
In 1965, Stanford researchers built DENDRAL — a system capable of analyzing chemical spectra and predicting molecular structures. Edward Feigenbaum led the work, with biochemist Joshua Lederberg and chemist Carl Djerassi.
DENDRAL's architectural insight was to separate the knowledge base from the inference engine. This separation defined the expert system paradigm that would dominate AI for decades, eventually leading to descendants like MYCIN that could diagnose infectious diseases.
Shakey — A Robot That Reasoned
In 1966, Shakey arrived. A wheeled robot at SRI International, equipped with cameras and sensors. Shakey wasn't just mobile — it reasoned about movement. It combined computer vision, language parsing, and action planning into a layered decision architecture: reflex responses, intermediate actions (push an object), and higher-level goal planning (navigate a path).
Shakey was a foundational bridge to autonomous robotics. The lessons learned there would influence route planning, autopilot systems, and robotic explorers for the next fifty years.
Samuel's Checkers and the Birth of "Machine Learning"
Arthur Samuel's self-learning checkers program quietly carried out one of the most influential experiments of the decade. By the early 1960s, Samuel had implemented alpha-beta pruning to optimize game-tree searches. The program played thousands of games against itself, refining its evaluation heuristics over time.
In 1962, the program defeated Robert Nealy, a self-described checkers master. Beyond the headline, Samuel had pioneered something larger: he popularized the phrase machine learning — and demonstrated that a program could improve by playing itself.
The Predictions
Famous thinkers made big claims. Herbert Simon promised that within ten years, a computer would become world chess champion. Marvin Minsky claimed that creating human-level intelligence would be solved "within a generation."
It felt like a breakthrough was just around the corner.
It wasn't.
Machines performed well in narrow domains and stumbled on tasks requiring common sense. AI could play — but not think. By the end of the decade, the cracks in techno-utopia began to show. Promises were made too often. Results came too slowly.
The first tremors of disillusionment had begun. The next chapter of the story would be much darker than the first.
Continue to Part 2 — AI winter, expert systems, and the rise of machine learning (1970s–2000s).