Vancouver, BC alek@aleksystem.com

Artificial Intelligence

Artificial Intelligence and what is the history of it’s inception from hardware to software

Artificial Intelligence (AI) is a field of computer science dedicated to creating systems capable of performing tasks that typically require human intelligence. These tasks include learning, reasoning, problem-solving, perception, and language understanding.+1

The inception of AI is a story of moving from physical “thinking” machines (hardware) to abstract, data-driven algorithms (software).

1. The Pre-Hardware Era: Philosophical Roots

Before computers existed, AI was a philosophical concept.

  • Ancient History: Myths of “automatons” (mechanical beings) existed in Greek and Chinese legends.
  • 17th–19th Century: Philosophers like Gottfried Leibniz and George Boole (inventor of Boolean algebra) attempted to formalize human thought into mathematical logic, suggesting that if thinking followed rules, it could be replicated by a machine.

2. The Early Hardware Era (1940s–1950s): Mechanical Brains

The birth of modern AI coincided with the invention of the first electronic, programmable computers.

  • 1943 (The Neural Model): Warren McCulloch and Walter Pitts created a mathematical model of biological neurons, proving that simple “nerve nets” could compute any logical function.
  • 1950 (The Turing Test): Alan Turing published “Computing Machinery and Intelligence,” proposing the “Imitation Game” (Turing Test) as a benchmark for machine intelligence.+1
  • 1951 (The First Neural Network): Marvin Minsky and Dean Edmonds built SNARC, the first neural network machine, using 3,000 vacuum tubes and a flight simulator autopilot to mimic the brain of a rat.
  • 1956 (The Official Birth): The term “Artificial Intelligence” was coined at the Dartmouth Workshop, marking the start of AI as an academic discipline.

3. The Symbolic & Software Era (1960s–1980s): Rule-Based Logic

As hardware became more reliable, the focus shifted to “Symbolic AI”—writing software that followed complex “If-Then” rules.

  • 1966 (The First Chatbot): Joseph Weizenbaum created ELIZA, a program that used pattern matching to simulate a psychotherapist.+1
  • 1970s–1980s (Expert Systems): Businesses began using “Expert Systems”—software that encoded the knowledge of human experts (like doctors or engineers) into thousands of hard-coded rules to solve specific problems.
  • The AI Winters: Progress stalled during two periods (the mid-70s and late 80s) because computers lacked the memory and processing power to handle the complexity of the real world.

4. The Data & Machine Learning Era (1990s–2010s): From Rules to Learning

Instead of humans writing every rule, researchers began building software that could learn its own rules from data.

  • 1997 (Deep Blue): IBM’s Deep Blue defeated chess champion Garry Kasparov. While it used massive hardware power to “brute force” calculations, it signaled a shift toward specialized AI.
  • 2012 (The Deep Learning Breakthrough): Using GPUs (Graphics Processing Units) originally designed for video games, researchers were able to train “Deep Neural Networks” on massive datasets. This led to a “Big Bang” in image and speech recognition.+1

5. The Modern Era (2020s–Present): Generative AI & Transformers

We are currently in the era of Large Language Models (LLMs) and Generative AI.

  • 2017 (The Transformer): Google researchers introduced the “Transformer” architecture, a software breakthrough that allowed AI to understand the context and relationship between words in a way never before possible.+1
  • Present Day: AI has moved from a specialized lab tool to a ubiquitous service (like ChatGPT or Gemini) that can generate text, code, images, and video in real-time.

Summary: Hardware vs. Software Evolution

  • Hardware Evolution: Moved from massive vacuum-tube machines (ENIAC) to integrated circuits, then to specialized GPUs and TPUs (Tensor Processing Units) designed specifically to handle the trillions of calculations required by modern AI.
  • Software Evolution: Moved from Hard-Coded Logic (telling the computer exactly what to do) to Machine Learning (giving the computer data and letting it figure out the patterns itself).