All Guides
Débutant • 80 h80 h estimées (cours, ateliers, projets)Free GuideSection 1/50

Free Generative AI Course — 80h Complete Training

Maîtrisez l'IA générative en 80 h : prompt engineering, ChatGPT, Claude, Copilot, automatisation avec Make, créativité multimédia et agents IA. Du débutant au praticien confiant.

💡

This training is 100%% free. To keep it that way, this page contains ads. Thank you for your understanding and support!

Section 10.1.1 : History of AI — From Origins to Today

🎯 Learning Objective

Understand the evolution of artificial intelligence from its theoretical beginnings to today's generative models. You will be able to place the major milestones of AI in their historical context and explain why generative AI represents a major technological breakthrough.


Why Learn the History of AI?

Understanding the history of AI is not an academic exercise: it is a practical necessity. The "AI winters" — those periods of massive disillusionment — teach us to distinguish hype from reality. Technological breakthroughs show us what conditions are needed for an innovation to succeed. And the overall trajectory helps us anticipate the likely developments of the coming years.

In 80 years, AI has gone from a philosophical idea to a tool that transforms every industry. This section gives you the keys to understanding this trajectory — and to avoid repeating the mistakes of the past.

The Roots of Artificial Intelligence

The history of artificial intelligence begins long before the invention of computers. Since antiquity, humans have dreamed of creating artificial beings capable of thought. The Greek myth of Talos — a bronze giant protecting Crete — is one of the first "robots" in literature. In the Middle Ages, alchemists sought to create an artificial "homunculus." The mechanical automata of the 18th century, such as the famous Vaucanson Duck (1739) that simulated digestion, or the Mechanical Turk (1770) that played chess (actually a hoax with a human hidden inside), illustrate this ancestral fascination with simulating intelligence.

These historical attempts reveal a fundamental question that still drives research in 2026: what is intelligence? Is it the ability to reason? To create? To feel emotions? The answer we give to this question determines what we expect from AI — and what we consider success or failure.

1943-1956: The Theoretical Foundations

Modern AI truly begins in the 1940s with the work of Warren McCulloch and Walter Pitts (1943), who proposed the first mathematical model of an artificial neuron. Their paper "A Logical Calculus of the Ideas Immanent in Nervous Activity" laid the foundations of neural networks. This model, extremely simplified compared to a real biological neuron, would become the building block of all modern deep learning — from GPT-3's 175 billion parameters to the estimated trillions of GPT-5.

In 1950, Alan Turing published his foundational paper "Computing Machinery and Intelligence" in which he proposed the famous Turing test: can a machine pass as a human during a written conversation? This philosophical question still guides part of AI research today. By March 2026, some models like GPT-5 and Claude Opus 4.6 convincingly pass simplified versions of the Turing test — a threshold that even Turing had not foreseen before the year 2000.

The term "Artificial Intelligence" was officially coined at the Dartmouth Conference in 1956, organized by John McCarthy, Marvin Minsky, Nathaniel Rochester, and Claude Shannon. This two-month conference brought together the brightest minds of the era and established the discipline as a field of research in its own right. The four organizers would become legendary figures in AI: McCarthy would invent the Lisp language, Minsky would co-found the MIT AI Lab, and Shannon would lay the foundations of information theory.

The Dartmouth Prophecy

The organizers wrote in their proposal: "We propose a two-month study (...) proceeding on the basis of the conjecture that every aspect of learning or any other feature of intelligence can, in principle, be so precisely described that a machine can be made to simulate it." — seventy years later, this vision is becoming reality.

1956-1974: The Golden Age and the First Programs

The next two decades saw an explosion of optimism. Researchers created programs that were impressive for their time:

  • ELIZA (1966, Joseph Weizenbaum, MIT): a chatbot simulating a psychotherapist. ELIZA rephrased the user's sentences as questions, creating the illusion of understanding. Some users developed genuine emotional attachment to the program — a phenomenon we see again today with modern chatbots. The ELIZA effect (the human tendency to attribute emotions and intelligence to a program) is more relevant than ever with ChatGPT and Claude.

  • SHRDLU (1970, Terry Winograd, MIT): a program capable of understanding English instructions to manipulate colored blocks in a virtual world. SHRDLU demonstrated the possibility of contextual human-machine dialogue — but only within an extremely limited "micro-world."

  • The General Problem Solver (1957, Newell and Simon): an ambitious program meant to solve any formalizable problem. While it worked on simple puzzles, it failed when faced with real-world complexity.

  • The Perceptron (1958, Frank Rosenblatt): the first neural network capable of learning. The Perceptron could classify simple images, generating immense enthusiasm. But in 1969, Minsky and Papert published Perceptrons, mathematically demonstrating its limitations — which would slow neural network research for 15 years.

Funding poured in, promises were bold. Herbert Simon predicted in 1965 that "within twenty years, machines will be capable of doing any work a man can do." Marvin Minsky declared in 1967 that "the problem of artificial intelligence will be solved within a generation." These unfulfilled promises triggered the first crisis of confidence.

1974-1980: The First AI Winter

Reality caught up with the promises. In 1973, the Lighthill Report commissioned by the British government concluded that AI had not delivered on its promises and recommended cutting funding. The problems were multiple:

  • Combinatorial explosion: programs did not scale. What worked for 10 elements became impossible for 1,000.
  • Lack of common sense: machines could not understand context, ambiguity, or the implicit nature of human language. The "common sense problem" remains partially unsolved in 2026.
  • Insufficient computing power: 1970s computers simply did not have the capacity to process complex problems. A 2026 smartphone is more than 100 million times more powerful than a 1975 computer.

This was the first "AI winter" — a period of disillusionment, budget cuts, and academic discredit that lasted nearly a decade. AI researchers were stigmatized, funding collapsed, and the term "artificial intelligence" became almost taboo in grant applications.

Lesson for Today: The Hype-Disillusionment Cycle

The historical pattern of AI winters — excessive promises → unrealistic expectations → disappointment → budget cuts — should be kept in mind in 2026. The spectacular announcements about imminent AGI (artificial general intelligence) echo Simon's predictions from 1965. History teaches us caution without pessimism: AI progresses, but not always at the pace announced.

1980-1987: Expert Systems and the Renaissance

AI was reborn in the 1980s in a different form: expert systems. Instead of simulating general intelligence, these programs codified human expert knowledge in the form of if-then rules.

MYCIN (Stanford, 1976) diagnosed bacterial infections with 69% accuracy — better than some general practitioners. XCON (DEC, 1980) automatically configured computer orders and saved DEC more than $40 million per year. DENDRAL (Stanford) analyzed mass spectrometry data to identify chemical molecules — one of the first examples of AI serving science.

Japan launched its ambitious Fifth Generation Project in 1982, aiming to create "intelligent" computers based on logic programming (Prolog). This $850 million project stimulated an international AI race, with responses from the UK (Alvey Programme), the European Union (ESPRIT), and the United States (MCC, SCI).

Global AI funding exceeded $1 billion by the mid-1980s. Thousands of companies created AI departments or purchased expert systems. The market for Lisp machines — specialized computers for AI — reached $400 million annually.

1987-1993: The Second AI Winter

Expert systems showed their limits: they were fragile (a missing rule broke everything), expensive to maintain (a human expert had to constantly update the rules), and incapable of learning (no automatic adaptation to new data). The Fifth Generation Project failed to meet its objectives. The Lisp machine market collapsed in the face of rising PCs. This was the second AI winter.

But this time, a different approach survived quietly: neural networks. The backpropagation algorithm, popularized by Geoffrey Hinton, David Rumelhart, and Ronald Williams in 1986, made it possible to train multi-layer networks. Yann LeCun developed LeNet (1989) for handwritten digit recognition — the first successfully applied convolutional neural network. The foundations of deep learning were laid, even though their full potential would not be exploited for another 25 years.

1993-2011: The Rise of Machine Learning

AI reinvented itself by abandoning promises of general intelligence in favor of statistical and pragmatic approaches. Machine Learning took over. Researchers stopped saying "artificial intelligence" and spoke of "machine learning," "data mining," "pattern recognition" — less emotionally charged terms.

  • 1997: Deep Blue (IBM) defeated world chess champion Garry Kasparov in 6 games (2 wins, 3 draws, 1 loss). It was a symbolic moment that made world headlines, even though the program relied more on brute force (200 million positions/second) than on intelligence. Kasparov would accuse IBM of cheating — a controversy that foreshadowed today's ethical debates about AI.

  • 2006: Geoffrey Hinton published a foundational paper on deep belief networks, inaugurating the era of Deep Learning. He showed that deep (multi-layer) networks could be efficiently trained through layer-by-layer pre-training. This marked the beginning of the third wave of interest in neural networks.

  • 2011: Watson (IBM) won the TV game show Jeopardy! against the best human players, demonstrating natural language understanding and information retrieval capabilities. Siri (Apple) was launched the same year, marking AI's entry into every consumer's pocket.

Loading diagram…

2012-2017: The Deep Learning Revolution

Everything accelerated in 2012 with AlexNet, a deep neural network created by Alex Krizhevsky, Ilya Sutskever, and Geoffrey Hinton. AlexNet won the ImageNet competition with a 15.3% error rate — compared to 26.2% for the best classical method. This was a massive improvement that proved deep learning's superiority for computer vision. This result triggered a massive influx of researchers and investment into deep learning.

Three factors converged to make deep learning possible:

  1. Big Data: The internet generated massive amounts of training data. ImageNet alone contained 14 million manually labeled images.
  2. GPUs: NVIDIA graphics cards enabled parallel computation. A GPU could be 50× faster than a CPU for training neural networks.
  3. Algorithms: Advances in optimization (Adam, dropout, batch normalization, residual connections) made training stable on very deep networks.

In 2014, Generative Adversarial Networks (GANs) by Ian Goodfellow opened the door to image generation. The principle was elegant: two networks competed — a "generator" that created fake images and a "discriminator" that tried to distinguish them from real ones. Through this competition, both networks improved mutually.

In 2016, AlphaGo (DeepMind) defeated world Go champion Lee Sedol 4-1 — an achievement considered impossible a decade earlier, as Go has more possible configurations (10^170) than atoms in the observable universe (10^80). Move 37 of game 2, never seen in 3,000 years of professional Go, was hailed as a moment of artificial creativity.

2017-2022: The Transformer Era

In June 2017, Google researchers published the paper that would change everything: "Attention Is All You Need." They proposed the Transformer architecture, based on an attention mechanism that allowed the model to simultaneously consider the entire input sequence, rather than processing it word by word.

Transformers made possible:

  • BERT (Google, 2018): bidirectional language understanding — revolutionized Google Search
  • GPT (OpenAI, 2018): autoregressive text generation — the precursor to ChatGPT
  • GPT-2 (OpenAI, February 2019): texts so convincing that OpenAI hesitated to release it (1.5 billion parameters)
  • GPT-3 (OpenAI, June 2020): 175 billion parameters, surprising emergent capabilities
  • DALL-E (OpenAI, January 2021): first text-to-image generation model using Transformers
  • ChatGPT (OpenAI, November 2022): the tipping point for the general public

2022-Today: The Generative AI Explosion

On November 30, 2022, OpenAI launched ChatGPT. In just 5 days, the service reached 1 million users — a historic record. For comparison, Netflix took 3.5 years to reach this milestone, Instagram 2.5 months, and TikTok 9 months. By January 2023, ChatGPT surpassed 100 million users — the fastest-growing technology product in history.

The years 2023-2024 saw an unprecedented acceleration:

  • GPT-4 (March 2023): first mainstream multimodal model (text + images), scored in the top 10% on the US bar exam
  • Claude (Anthropic, 2023): focused on safety, helpfulness, and honesty
  • Gemini (Google, December 2023): directly integrated into the Google ecosystem
  • Llama 2 and 3 (Meta, 2023-2024): powerful open-source models that democratized access
  • Mistral (France, 2023-2024): the French AI gem, founded by former Google DeepMind and Meta researchers
  • DeepSeek-R1 (China, January 2025): AI's "Sputnik moment" — performance rivaling GPT-4 at a training cost of $5.6 million (vs. ~$100 million for GPT-4)
  • o1 and o3 (OpenAI, 2024-2025): first reasoning models ("thinking models")

2025-2026: The Era of Agentic AI and Reasoning

The year 2025 marks a major turning point with the emergence of models capable of deep reasoning and autonomous actions (AI agents). Models are no longer simple "text completers" — they can think, plan, use tools, and execute complex multi-step tasks:

  • GPT-5 (OpenAI, August 2025): hybrid architecture combining a fast model and a reasoning module. Followed by GPT-5.2 (December 2025) and GPT-5.4
  • Claude 4 (Anthropic, May 2025): advanced reasoning and coding capabilities. Followed by Claude Opus 4.6 and Sonnet 4.6 (February 2026)
  • Gemini 3 (Google, November 2025): integrates Deep Think for complex reasoning. Gemini 3.1 Pro (February 2026) rivals GPT-5
  • Llama 4 (Meta, April 2025): Mixture-of-Experts architecture with 400 billion parameters (Maverick) and a 10-million-token context window (Scout)
  • Grok 4 (xAI, July 2025): native integration with the X network. Followed by Grok 4.1 and 4.20 Beta (February 2026)
  • DeepSeek V3.1 (August 2025): hybrid open-source model (MIT License) outperforming its predecessors by 40% on certain benchmarks
  • Sora 2 (OpenAI, September 2025): second-generation video generation with an integrated social app
  • Midjourney v7 (April 2025): new standard for photorealistic image generation
  • Claude Code and GitHub Copilot: coding agents become daily tools for developers

Section Summary

PeriodNameKey InnovationLimitation
1943-1956FoundationsArtificial neuron, Turing testTheoretical only
1956-1974Golden AgeELIZA, Perceptron, General Problem SolverExcessive promises
1974-19801st WinterCombinatorial explosion, common sense
1980-1987Expert SystemsMYCIN, XCON, DENDRALFragile, non-adaptive
1987-19932nd WinterBackpropagation, LeNet (seeds planted)Insufficient power
1993-2011Machine LearningDeep Blue, WatsonLimited data and compute
2012-2017Deep LearningAlexNet, AlphaGo, GANsTask-specific
2017-2022TransformersBERT, GPT-3, DALL-ETraining cost
2022-2024Generative AIChatGPT, GPT-4, SoraHallucinations, ethics
2025-2026Agentic AIGPT-5, Claude 4.6, Gemini 3.1, DeepSeekCost, safety, control

Practical Exercise: Personal Timeline

Duration: 15 minutes

  1. Ask ChatGPT or Claude: "Summarize the history of AI in 10 key milestones with dates and impacts"
  2. Compare the response with this section — did the model miss any milestones? Add relevant details?
  3. Identify a historical event you did not know about and look it up on Wikipedia
  4. Personal reflection: in your opinion, are we in a new "golden age" or on the verge of a third "AI winter"? Argue in 3 sentences.

Learn AI — From Prompts to Agents

10 Free Interactive Guides120+ Hands-On Exercises100% Free

Frequently Asked Questions

Is this generative AI course really free?

Yes, 100% free. All 50 sections and 8 chapters are accessible without any account, payment, or registration. We sustain the platform through non-intrusive ads.

How long does the complete training take?

The full course is designed for approximately 80 hours of study, spread across 8 chapters and 50 sections. You can learn at your own pace.

Do I need programming experience?

No. The course starts from absolute zero and covers everything from basic prompting to advanced AI agent design. No coding required for most modules.

What tools will I learn to use?

ChatGPT, Claude, GitHub Copilot, Make (Integromat), Zapier, n8n, Midjourney, DALL-E, Stable Diffusion, and many more AI tools.

What is prompt engineering?

Prompt engineering is the practice of crafting effective instructions for AI models to produce desired outputs. It's the most in-demand AI skill in 2025.

Newsletter

Weekly AI Insights

Tools, techniques & news — curated for AI practitioners. Free, no spam.

Free, no spam. Unsubscribe anytime.