Home
Skip to main content
xStryk™

Decision Intelligence for AI in production — guardrails, traceability & evaluation.

xSingular · AI Authority · 1936 - Present

The History
of AI

From the Turing Machine to large language models. A complete journey through the ideas, papers, pioneers, and breakthroughs that shaped the age of artificial intelligence.

0
Years of history
0
Key milestones
0
Pioneers
0
Landmark papers
A
J
M
C
G
Y
Y
F
D
I
1936 – 1955

Foundations of Computation

"Can machines think?" - The question that started everything. Turing, 1950.

1936
Theory

The Turing Machine

Alan Turing publishes "On Computable Numbers," introducing the theoretical Turing Machine—an abstract mathematical model that defines the limits of what can be computed. Every processor on Earth is a physical realization of this idea.

1943
Neuroscience

McCulloch–Pitts Neuron

Warren McCulloch and Walter Pitts publish the first mathematical model of a neuron in "A Logical Calculus of Ideas Immanent in Nervous Activity." The idea that cognition could be formalized mathematically ignites a revolution that leads directly to modern neural networks.

1945
Architecture

Von Neumann Architecture

John von Neumann formalizes the stored-program computer in his First Draft on EDVAC. The separation of processor and memory—with programs stored as data—remains the foundation of virtually every computer built since.

1948
Cybernetics

Cybernetics

Norbert Wiener publishes "Cybernetics: Control and Communication in the Animal and the Machine," founding the science of regulatory systems and introducing feedback loops as a core mechanism of intelligent behavior—decades before reinforcement learning.

1950
Philosophy

The Turing Test

Alan Turing publishes "Computing Machinery and Intelligence" in Mind. Asking "Can machines think?", he proposes the Imitation Game as a practical test for machine intelligence. A philosophical challenge that echoes in every conversation with a chatbot today.

1956 – 1969

AI is Named & Born

At Dartmouth 1956, McCarthy coined "Artificial Intelligence" - optimism was at its peak.

1956
Historic

Dartmouth Conference

John McCarthy, Marvin Minsky, Claude Shannon, and Nathaniel Rochester organize the Dartmouth Summer Research Project. McCarthy coins "Artificial Intelligence" and the discipline is formally born. The group predicted human-level AI within a generation—an optimism that proved premature by decades.

1957
Reasoning

Logic Theorist & GPS

Allen Newell and Herbert Simon build the Logic Theorist—the first program to prove mathematical theorems—and the General Problem Solver. Simon controversially predicts a computer will beat the world chess champion within 10 years. He was off by 40.

1958
Neural Nets

The Perceptron

Frank Rosenblatt introduces the Perceptron at Cornell—the first trainable neural network. It learns from examples by adjusting weights. The New York Times reports it will soon walk, talk, see, and reproduce itself. The hype begins here.

1959
ML Term

"Machine Learning" Coined

Arthur Samuel of IBM coins the term "machine learning" in a seminal paper on game-playing programs. He demonstrates the first self-improving program—a checkers engine that beats its creator by playing against itself through self-play reinforcement, decades before AlphaGo.

1965
NLP

ELIZA: The First Chatbot

Joseph Weizenbaum at MIT creates ELIZA, the first natural language processing program. Using simple pattern matching to mimic a psychotherapist, ELIZA fools users into believing they are talking to a real person. Weizenbaum is disturbed—his warnings about anthropomorphizing AI are ignored.

1969
Crisis

The XOR Problem Kills Neural Networks

Minsky and Papert publish "Perceptrons," mathematically proving that single-layer perceptrons cannot solve XOR. Their authoritative critique eliminates neural network funding for over a decade. The first generation of neural network researchers pivots to other careers. The First AI Winter approaches.

1969 – 1980

The First AI Winter

The Lighthill Report (1973) concluded AI research had failed to meet its grand promises. Funding collapsed.

1973
UK Report

The Lighthill Report

Sir James Lighthill delivers a devastating critique of AI research to the UK Science Research Council. Concluding that AI has failed to live up to its promises, the report triggers massive funding cuts in the UK and US. AI researchers struggle to find support for basic work.

1979
Robotics

Stanford Cart

Despite the funding winter, the Stanford Cart navigates a 20-meter obstacle course over 5 hours—the first demonstration of autonomous mobile robotics. Hardware capabilities are quietly advancing even as AI theory stalls. The seeds of self-driving technology are planted.

1980 – 1987

The Expert Systems Renaissance

XCON saved DEC $40M/year. Rule-based AI finally made money - and everyone wanted a piece.

1980
Industry

Expert Systems Boom

Expert systems like XCON (DEC) and MYCIN (Stanford Medicine) generate massive commercial value. XCON alone saves DEC $40M per year configuring VAX computers. The first real AI companies are founded. Japan launches its ambitious $400M Fifth Generation Computer project.

1986
Neural Nets

Backpropagation Revived

Rumelhart, Hinton, and Williams publish "Learning representations by back-propagating errors" in Nature—making neural network training practical at scale. Backpropagation, an efficient algorithm for computing gradients, becomes the foundation of all modern deep learning. Neural networks are reborn.

1986
Robotics

First Autonomous Driving

Ernst Dickmanns at Bundeswehr University Munich drives VaMoRs—his autonomous van—on the German autobahn using real-time computer vision. This is 30 years before Tesla Autopilot. A remarkable demonstration that applied neural networks can work in the physical world.

1987 – 1997

Second Winter & the Probabilistic Turn

Japan's ¥850B "Fifth Generation" project ended without delivering. Investors lost faith again.

1987
Collapse

Second AI Winter

The expert systems market collapses. Lisp machine vendors go bankrupt. DARPA cuts strategic AI computing funding. The AI industry loses billions. The Second AI Winter forces a generation of researchers to abandon grand claims and focus on rigorous, measurable, statistical approaches to intelligence.

1989
CNN

Convolutional Neural Networks

Yann LeCun, Léon Bottou, Yoshua Bengio, and Patrick Haffner apply CNNs to handwriting recognition for the US Postal Service. LeNet achieves 99%+ accuracy on digit recognition. By the late 1990s, it processes approximately 10% of all checks in the United States.

1995
SVM

Support Vector Machines

Corinna Cortes and Vladimir Vapnik publish Support Vector Machines—a theoretically elegant algorithm with strong generalization guarantees from statistical learning theory. SVMs dominate applied ML for 15 years in spam filters, medical diagnosis, and financial modeling.

1997
RNN

Long Short-Term Memory (LSTM)

Sepp Hochreiter and Jürgen Schmidhuber publish LSTM—a recurrent architecture that solves the vanishing gradient problem through gated memory cells. LSTMs power speech recognition, machine translation, and NLP for the next two decades until transformers arrive.

1997
Historic

Deep Blue Defeats Kasparov

IBM's Deep Blue defeats reigning world chess champion Garry Kasparov 3.5–2.5 in a landmark match. Kasparov alleged IBM cheated with human intervention—beginning AI's complex, adversarial relationship with the world champions it defeats.

1997 – 2006

Statistical Machine Learning Matures

ImageNet (2009): 14 million labeled images that ignited the deep learning revolution 3 years later.

1998
Data Economy

PageRank & the Web Data Revolution

Sergey Brin and Larry Page publish the PageRank algorithm and found Google. Web-scale data begins accumulating—tens of billions of text documents, images, and interactions. This vast dataset is quietly building the training corpus that will fuel the deep learning revolution a decade later.

2001
Ensemble

Random Forests

Leo Breiman publishes Random Forests in Machine Learning—one of the most robust and widely-used ML algorithms ever created. Ensemble methods like boosting, bagging, and stacking become the dominant tools of production ML throughout the 2000s.

2004
Vision

ImageNet Project Begins

Fei-Fei Li at Princeton (later Stanford) begins building ImageNet—a massive labeled image database. The idea: to give neural networks what the brain has, an enormous amount of visual experience. Five years of labeling work by hundreds of thousands of Amazon Mechanical Turk workers follows.

2006 – 2014

The Deep Learning Revolution

AlexNet 2012: 10.9% error rate vs 26% for the runner-up. The gap was so large the field had no choice but to pivot.

2006
Breakthrough

Deep Belief Networks

Geoffrey Hinton and Ruslan Salakhutdinov publish "Reducing the Dimensionality of Data with Neural Networks" in Science—demonstrating that deep networks can be trained effectively with greedy layer-wise pretraining. This single paper single-handedly reignites neural network research after a 15-year drought.

2009
Dataset

ImageNet Released

Fei-Fei Li releases ImageNet—14 million labeled images across 20,000 categories. The annual ImageNet Large Scale Visual Recognition Challenge (ILSVRC) begins. For three years, handcrafted feature engineering approaches dominate. Then 2012 changes everything.

2012
Historic

AlexNet: The Moment AI Changed Forever

Alex Krizhevsky, Ilya Sutskever, and Geoffrey Hinton win ImageNet with AlexNet—a deep CNN trained on two GPUs with dropout and ReLU. Error rate: 15.3% vs. 26.1% for the runner-up. The largest gap in the competition's history. This is the official start of the modern AI era.

2013
Embeddings

Word2Vec

Tomáš Mikolov et al. at Google publish Word2Vec—dense vector embeddings that capture semantic relationships. "King − Man + Woman ≈ Queen" demonstrates that language meaning can be encoded geometrically in continuous space. The foundation of modern NLP is laid.

2014
Generative

Generative Adversarial Networks

Ian Goodfellow et al. publish GANs—perhaps the most creative idea in recent deep learning history. A generator creates fake data; a discriminator judges authenticity. They play a minimax game until the generator fools the discriminator. Result: machines that can generate photorealistic images.

2015 – 2020

Reinforcement Learning & Transformers

"Attention Is All You Need" (2017) scrapped recurrence entirely. 8 researchers changed how every AI model works.

2015
Deep RL

Deep Q-Networks (DQN)

DeepMind publishes DQN in Nature—a deep RL agent that learns to play 49 Atari games from raw pixels at superhuman level, using only the game score as reward. No hand-crafted features. No domain knowledge. Pure end-to-end reinforcement learning. The path to AlphaGo begins here.

2016
Historic

AlphaGo Defeats Lee Sedol

DeepMind's AlphaGo defeats 9-dan world Go champion Lee Sedol 4–1. Move 37 in Game 2—a move no human player would ever make, yet which proves decisive—stuns the Go community as evidence that AI can discover strategies that transcend human intuition.

2017
Architecture

"Attention Is All You Need"

Vaswani et al. at Google Brain publish the Transformer architecture—replacing recurrence with self-attention for massive parallelization. The paper that makes GPT, BERT, and every large language model possible. The most cited ML paper of the decade. 2017 is Year Zero for modern AI.

2018
NLP

BERT: Bidirectional Pre-training

Devlin et al. at Google publish BERT—bidirectional transformer pre-trained on masked language modeling and next-sentence prediction, then fine-tuned for specific tasks. BERT sets new state-of-the-art on 11 NLP benchmarks simultaneously, making pre-training + fine-tuning the dominant paradigm.

2020
Biology

AlphaFold 2 Solves Protein Folding

DeepMind's AlphaFold 2 achieves near-experimental accuracy on CASP14—solving a 50-year grand challenge in biology. For the first time, AI demonstrably accelerates foundational scientific discovery at a civilizational scale. The database of 200M+ predicted protein structures is released free to the world.

2020 – 2023

Foundation Models & Generative AI

ChatGPT reached 100M users in 60 days - faster than any product in history. The AI era went mainstream overnight.

2020
LLM

GPT-3: 175 Billion Parameters

OpenAI's GPT-3 with 175 billion parameters demonstrates surprising few-shot learning: given 2–3 examples in a prompt, the model performs new tasks without updating a single weight. Its text is indistinguishable from human writing. The era of prompting begins.

2021
Multimodal

DALL-E, CLIP & the Multimodal Turn

OpenAI releases DALL-E (text-to-image generation) and CLIP (contrastive vision-language learning). For the first time, a single model bridges text and vision without task-specific training. The multimodal era begins—AI that sees, reads, and draws simultaneously.

2022
Inflection

ChatGPT: AI Goes Mainstream

OpenAI releases ChatGPT. It reaches 100 million users in 2 months—the fastest-growing consumer product in history. Every major company in every industry begins restructuring their AI strategy overnight. AI is no longer for specialists. The decade of AI democratization begins.

2023
Competition

GPT-4, Claude, Llama: The Racing Era

GPT-4 passes the Bar exam at the 90th percentile. Anthropic releases Claude with Constitutional AI. Meta open-sources Llama. Google launches Gemini. A global race for frontier AI ignites—with open source, safety, and capabilities on a collision course.

2024 – Present

The Frontier: Towards AGI

For the first time, AI systems score above 90th percentile on bar exams, code competitions, and graduate-level science.

2024
Safety

AI Safety Enters the Boardroom

The EU AI Act comes into force—the first comprehensive AI regulation globally. AI Safety, AI Ethics, AI Governance, and AI Liability become board-level priorities. Leading labs establish safety teams. The question is no longer whether AI is powerful, but whether it is trustworthy and auditable in production.

2024
Production AI

AI in Critical Infrastructure

Production AI systems deploy at scale in mining, banking, healthcare, energy, and critical infrastructure. Decision Intelligence—the discipline of connecting ML predictions to auditable, traceable, per-decision accountability—becomes the key competency separating AI leaders from laggards in regulated industries.

2024+
Quantum ML

Quantum Machine Learning

Quantum neural networks, variational quantum circuits (VQC), and quantum-enhanced optimization begin delivering measurable advantages on specific high-complexity optimization and simulation problems. The boundary between classical and quantum ML is actively moving. The next chapter of AI is being written now.

Exponential Scale

The Parameter Explosion

From ~300 weights in the Perceptron to ~1.8 trillion in GPT-4. Logarithmic scale; each grid line is 1,000x larger than the one below.

The Minds Behind AI

Pioneers

The people whose ideas, obsessions, and daring built artificial intelligence.

Alan Turing

1912–1954 · UK
Father of Computer Science & AI

Formalized computation with the Turing Machine, proposed the Turing Test, cracked the Enigma cipher in WWII, and laid the mathematical foundations of every computer built since.

John McCarthy

1927–2011 · USA
Father of Artificial Intelligence

Coined the term "Artificial Intelligence," organized the 1956 Dartmouth Conference, and invented LISP—the programming language that defined the first generation of AI research.

Marvin Minsky

1927–2016 · USA
Co-founder, MIT AI Lab

Pioneered neural networks, robotics, cognitive architectures, and the theory of frames in knowledge representation. Co-founded the MIT AI Lab—the most influential AI research institution of the 20th century.

Claude Shannon

1916–2001 · USA
Father of Information Theory

Founded information theory, formalized binary code and digital logic, wrote the first paper on computer chess, and established the mathematical foundations of all digital communication.

Geoffrey Hinton

1947– · UK / Canada
"Godfather of Deep Learning"

Co-invented backpropagation, Boltzmann machines, dropout, and deep belief networks. Turing Award 2019. Resigned from Google in 2023 to warn about existential AI risks. His students built GPT and AlexNet.

Yann LeCun

1960– · France / USA
Pioneer of Convolutional Neural Networks

Invented CNNs and applied them to the first practical deep learning system. His LeNet powered check reading in US banks. Chief AI Scientist at Meta. Turing Award 2019. Advocates for autonomous AI over LLM-centric approaches.

Yoshua Bengio

1964– · Canada
Deep Learning Pioneer, AI Safety Advocate

Foundational work on recurrent nets, attention mechanisms, generative models, and representation learning. Turing Award 2019. Among the most vocal senior researchers warning about AI Safety and existential risk.

Fei-Fei Li

1976– · China / USA
Creator of ImageNet

Created ImageNet (14M labeled images)—the dataset that catalyzed the deep learning revolution. Co-directed Stanford AI Lab. Former Chief AI Scientist at Google Cloud. A leading voice for human-centered AI and diversity in technology.

Vladimir Vapnik

1936– · Russia / USA
Inventor of Support Vector Machines

Developed statistical learning theory and invented SVMs with Corinna Cortes. His Vapnik-Chervonenkis (VC) dimension theory provides fundamental theoretical bounds on generalization—the core of all rigorous ML.

Demis Hassabis

1976– · UK
CEO & Co-founder, Google DeepMind

Led AlphaGo, AlphaZero, AlphaFold, and Gemini. Turing Award 2024. Combines neuroscience-inspired RL with massive scale to advance general intelligence—while championing responsible AI development and safety research.

Ilya Sutskever

1986– · Russia / Canada
Co-founder, OpenAI & SSI

Co-designed AlexNet with Hinton. Co-founded OpenAI and oversaw GPT development through GPT-4. Founded Safe Superintelligence Inc. (SSI) in 2024—a company focused exclusively on building safe superintelligent AI.

Arthur Samuel

1901–1990 · USA
Pioneer of Machine Learning

Coined "Machine Learning" at IBM in 1959. Built the first self-improving program—a checkers engine that improved through self-play reinforcement, demonstrating that machines can learn without being explicitly programmed for a task.

Papers That Changed the World

Landmark Papers

Eleven publications that redefined what is possible.

1950
Computing Machinery and Intelligence
Alan Turing · Mind

Proposes the Turing Test. Asks "Can machines think?" Defines the problem AI still works on today.

Read paper →
1958
The Perceptron: A Probabilistic Model for Information Storage
Frank Rosenblatt · Psychological Review

First trainable neural network. Machines that learn from examples. The first spark of modern AI.

Read paper →
1986
Learning Representations by Back-propagating Errors
Rumelhart, Hinton & Williams · Nature

Backpropagation makes deep networks trainable at scale. The algorithm underlying all modern deep learning.

Read paper →
1989
Backpropagation Applied to Handwritten Zip Code Recognition
LeCun, Boser, Denker et al. · Neural Computation

First practical CNN. 99%+ accuracy on digit recognition. Processes 10% of all US checks by the 1990s.

Read paper →
1997
Long Short-Term Memory
Hochreiter & Schmidhuber · Neural Computation

Solves vanishing gradients with gated memory. Powers speech recognition and NLP for two decades.

Read paper →
2006
A Fast Learning Algorithm for Deep Belief Nets
Hinton, Osindero & Teh · Neural Computation

Greedy layer-wise pretraining reopens the deep learning frontier after 15 years. The paper that started the revolution.

Read paper →
2012
ImageNet Classification with Deep Convolutional Neural Networks
Krizhevsky, Sutskever & Hinton · NeurIPS

AlexNet. 15.3% error vs 26.1% runner-up. The moment the modern AI era began. The most consequential ML paper ever written.

Read paper →
2014
Generative Adversarial Networks
Goodfellow, Pouget-Abadie, Mirza et al. · NeurIPS

Generator vs discriminator: machines that fake reality. Foundation of DALL-E, Stable Diffusion, and all generative AI.

Read paper →
2017
Attention Is All You Need
Vaswani, Shazeer, Parmar et al. · NeurIPS

The Transformer architecture. Replaces recurrence with self-attention. Makes GPT, BERT, LLaMA, Claude, and Gemini possible.

Read paper →
2020
Language Models are Few-Shot Learners
Brown, Mann, Ryder et al. · NeurIPS

GPT-3 with 175B parameters demonstrates emergent few-shot learning. The birth of in-context learning and the prompting paradigm.

Read paper →
2021
An Image is Worth 16×16 Words: Transformers for Image Recognition at Scale
Dosovitskiy, Beyer, Kolesnikov et al. · ICLR

Vision Transformer (ViT): the same architecture conquers both language and vision. The Transformer unified AI.

Read paper →
From Symbolic to Quantum

The AI Continuum

Five paradigms. Eighty-eight years. One trajectory toward general intelligence.

1950–1985

Symbolic AI

Logic, rules, expert systems. Knowledge encoded by hand. LISP. PROLOG.

1990–2005

Statistical ML

Probabilistic models, SVMs, Random Forests, boosting. Data-driven.

2006–2016

Deep Learning

CNNs, RNNs, representation learning. GPUs unlock massive scale.

2017–2023

Foundation Models

Transformers, LLMs, RLHF, multimodal. Few-shot. Scale is all you need.

2024–

Frontier AI

Agents, reasoning models, AI Safety, quantum ML, production AI.

From history to critical production AI

xSingular applies 88 years of collective learning to production AI systems for mining, banking, and critical infrastructure. Decision Intelligence, AI Safety, continuous evaluation, and per-decision traceability.