Why the backpropagation algorithm became the “ignition button” of the modern AI revolution
The backpropagation algorithm is the cornerstone of modern artificial intelligence. Its significance goes far beyond the technicalities of neural network training: it opened the path to real, scalable machine learning, for the first time turning the depth of a network from a theoretical abstraction into a working tool.
What is Backpropagation?
Backpropagation is an optimization method that enables training of multi-layer neural networks by adjusting weights to minimize the error between the model’s prediction and the actual outcome. The term backpropagation first appeared in the work of Rumelhart, Hinton, and Williams (1986).
In simple terms:
The model makes a prediction (forward pass).
The result is compared to reality — the error is calculated.
The error is “propagated backward”, layer by layer, computing the derivatives (gradients) of the loss function with respect to each weight.
The weights are updated — taking a step toward reducing the error (gradient descent).
Mathematical essence
For each weight w , the update is computed as:
Δw_ij=-η⋅∂L/(∂w_ij )
where:
L — the loss (error) function,
η — the learning rate,
∂L/(∂w_ij ) — the partial derivative, computed via the chain rule.
Why Backpropagation Changed Everything
1. Made complex models trainable — with a universal method
Backprop is a general algorithm that works for any neural network architecture: convolutional, recurrent, transformer. It transforms learning from manual weight tuning into a mechanical process of adaptation.
Before backpropagation, neural networks were limited to 1–2 layers. Any attempt to add more layers would “break” — there was no way to correctly and efficiently propagate the error. Backprop provided the first universal recipe for training deep (multi-layer) structures.
Prior to backprop, each new architecture almost required a separate “handwritten” derivation of formulas. The reverse-mode automatic differentiation (reverse pass) turns any computable network into a “black box” where any partial derivative is computed automatically. All you need to do is define the objective function and press train. In practice, this means:
the same training code works for convolutional networks, transformers, diffusion models, systems of equations, and even physical simulators;
researchers can experiment freely with architectures without rewriting calculus every time.
2. The “computational abacus” — gradients in two passes, not in N attempts
If we computed numerical derivatives “one by one”, training time would grow linearly with the number of parameters (a billion weights → a billion forward-backward passes). Backprop is smarter: a rough forward pass + one backward pass computes all gradients at once. As models scale (GPT-3, GPT-4 ≈ 10¹¹ weights), this difference becomes the gap between days of computation and tens of thousands of years.
3. Enables scaling laws
Kaplan et al. (2020) empirically showed: if you scale data, parameters, and FLOPs by ×k , the error predictably decreases. This observation holds only because backprop provides stable, differentiable optimization at any scale. Without it, “add a billion parameters” would break the training process.
4. Eliminates “manual programming of heuristics”
Before the 1980s, image recognition was built by:
designing filters (edges, corners) by hand,
hardcoding them into software,
repeating the process for every new task.
Backprop allows the network itself to “invent” the right features: early filters learn to detect gradients, then textures, then full shapes. This removed the ceiling of human intuition and opened the door to exponential quality growth simply by “add more data + compute”.
5. Unifies all modern AI breakthroughs
RLHF (Reinforcement Learning from Human Feedback) — backprop over a policy model;
Style-transfer, Diffusion, GANs — generative networks trained via gradients;
AlphaFold2, AlphaZero — end-to-end backprop through protein physics or Monte Carlo trees;
Automatic differentiation in physics, finance, robotics — the same algorithm.
In fact, nearly every breakthrough of the last decade can be reduced to: “invented a new loss function + a few layers, trained with the same backprop”.
6. Engineering applicability
Backprop turns a mathematical model into a tool that can be “fed” data and improved. It made possible:
image recognition (LeNet, AlexNet),
machine translation,
voice assistants,
image and text generation (GPT, DALL·E).
7. Scalability
Backprop is easily implemented via linear algebra, fits perfectly on GPUs, and supports parallel processing. This enabled the growth of models from dozens of parameters to hundreds of billions.
8. Cognitive model of learning
Backprop does not mimic the biological brain, but provides a powerful analogy: synapses “know” how to adjust themselves after receiving an error signal from the “output” layer. This transferability is why neuroscientists today study whether mammalian brains use “pseudo-backprop” mechanisms (e.g., feedback alignment, predictive coding).
Historical Analogy
If compared to other sciences:
In electricity — it’s like Ohm’s Law;
In computer science — like the quicksort algorithm;
In biology — like the discovery of DNA.
Without it, AI would have remained a dream — or a paper exercise.
Why Is It Still Relevant?
Even the most advanced models — GPT-4, Midjourney, AlphaFold — are trained using backpropagation. Architectures evolve, heuristics are added (like RLHF), but the core optimization mechanism remains unchanged. It overcame three historical barriers: cumbersome analytics, unmanageable computational growth, and manual feature engineering. Without it, there would be no “deep learning” — from ChatGPT to AlphaFold.
Conclusion
Backpropagation is the technology that first gave machines the ability to learn from their mistakes. It is not just an algorithm — it is a principle: “Compare, understand, correct.” It is the embodiment of intelligence — statistical for now, but already effectively acting.
Comparing the Monograph by Alexander Galushkin and the Dissertation by Paul Werbos
What is truly contained — and what is missing — in Alexander Galushkin’s book
Synthesis of Multi-Layer Pattern Recognition Systems (Moscow, “Energiya”, 1974)
Alexander Ivanovich Galushkin
1. The essence of the author’s contribution
Deep gradient: In Chapters 2 and 3, the author derives the general risk functional R(a) multi-layer system, writes the Lagrangian, and the full expression for ∂R/∂a_j . Then he demonstrates a step-by-step backward calculation of these derivatives “from end to start”: first the output error, then its recursive distribution across hidden nodes, and finally the update of all weights. This is exactly the logic later called backpropagation.
Generality: The algorithm is presented not as a “trick for a perceptron”, but as a universal optimization procedure for complex decision-making networks: any continuous activation functions, any number of hidden layers.
Demonstration on a network: In the appendices, examples of two- and three-layer classifiers with sigmoid neurons are provided; the author computes gradients, draws decision boundaries, and shows convergence on toy data.
Practical context: The book was written for developers of “friend-or-foe” systems and technical vision: the goal is to minimize classification error under reaction time constraints. Thus, the method is immediately embedded into a real engineering task.
2. What is missing in the book
The term “backpropagation” does not appear; instead, terms like “adaptation algorithm” or “dynamic error distribution” are used.
No large-scale experiments: examples are small; networks with 10+ layers, of course, did not yet exist.
Absence of modern engineering details — He/Glorot initialization, dropout, batch normalization, etc.
Circulation and language: 8,000 copies, Russian only; references to Western colleagues are minimal, so the Western community effectively remained unaware of the work.
3. Why this text is considered one of the two primary sources of backpropagation
Chronology: A series of papers by Vanyushin–Galushkin–Tyukhov on the same gradient approach were published in 1972–73, and the manuscript of the monograph was submitted for printing on February 28, 1974.
Complete analytical derivation + a ready algorithm for iterative learning.
Connection to practice (rocket and aviation systems) proved the method’s viability even on 1970s computing hardware.
Thus, Galushkin, independently of Paul Werbos, constructed and published the core of backprop — although the term, global resonance, and GPU era would come a decade after this “breakthrough but low-circulation” Soviet book. Galushkin even predicted analogies between neural networks and quantum systems [Galushkin 1974, p. 148] — 40 years ahead of his time!
What is (and is not) in Paul Werbos’ dissertation Beyond Regression… (August 1974)
What is definitely present
Werbos introduces the concept of “ordered derivative”. He shows how, after a forward pass through the computational graph, one can move “from bottom to top”, distributing the error and computing all partial derivatives in a single backward pass. In essence, this is reverse-mode automatic differentiation — the same mathematical skeleton used by backpropagation today.
The author illustrates the method on a toy two-layer sigmoid network. He explicitly writes down the derivatives for hidden and output weights and demonstrates a training iteration. Thus, the link to neural networks is not speculative — an example exists.
The dissertation emphasizes the algorithm’s universality: “dynamic feedback” is suitable for any block-structured program. The method is presented as a general “compute-then-backpropagate” technique for complex functions, not a specialized tool just for perceptrons.
After his defense, Werbos did not abandon the topic: in 1982, he published a paper where he directly named the technique backpropagation and extended it to optimal control systems. Thus, he maintained and developed his authorship.
What is missing
The term “backpropagation” is not used. Werbos speaks of “dynamic feedback” or “ordered derivatives”. The now-iconic term would appear twelve years later in Rumelhart, Hinton, and Williams.
No demonstration of large-scale, industrial deep networks or long learning-curve experiments. The example is small, at the level of “let’s prove it works”.
No engineering details that later made deep learning take off: proper weight initialization, anti-overfitting techniques, large datasets, GPUs. Thus, the method appeared elegant but remained “on paper”.
Conclusion on authenticity
Werbos did indeed describe the key idea of reverse gradients two years before Rumelhart–Hinton and independently of Soviet works.
But he did not demonstrate large-scale training of perceptrons and did not introduce the terminology that made the method popular.
Attributing the “ready deep-learning algorithm” to him is unfair; but calling him one of the discoverers of backpropagation is justified.
Even earlier Soviet papers
Vanyushin–Galushkin–Tyukhov, Proceedings of the USSR Academy of Sciences, 1972 (algorithm for training hidden layers).
Galushkin’s report at the Academy of Sciences of the Ukrainian SSR, 1973 (gradient weight correction).
These dates give the Soviet Union a lead of at least two years over Werbos.
Ivakhnenko — the “great-grandfather” of AutoML
Even before Galushkin, the Ukrainian scientist Alexey Grigoryevich Ivakhnenko developed the Group Method of Data Handling (GMDH). A series of papers from 1968–1971 showed how a multi-layer model could generate its own structure: the network is built by adding “dictionary” layers, keeping only nodes that minimize validation error. In essence, GMDH was the first form of AutoML — automatic architecture search.
Impact:
Legitimized the idea of “depth” theoretically;
Showed that adaptation could occur not only in weights but also in topology;
Became a natural “springboard” for Galushkin: if structure can be built automatically, a universal method for quickly retraining weights was needed — and that method became his gradient algorithm (1972–74).
The Final Picture
The Soviet Union not only independently discovered backpropagation — it did so first, six months before the American work. There was no simultaneous parallel discovery, as Western sources claim.
Archival data clearly shows: Alexander Galushkin became the first researcher in the world to publish a complete description of backpropagation. His monograph Synthesis of Multi-Layer Pattern Recognition Systems was submitted for printing on February 28, 1974 (USSR) and contains a rigorous mathematical derivation of gradients, the backpropagation algorithm for multi-layer networks, and practical examples for “friend-or-foe” systems. Thus, he preceded Western works by six months. Paul Werbos’ dissertation (Beyond Regression) was defended only in August 1974 (Harvard). The work by Rumelhart–Hinton, which popularized the term “backpropagation”, was published only in 1986.
Galushkin developed the method within a whole scientific school, building on Ivakhnenko’s work (GMDH, 1968–1971), and even anticipated the connection between neural networks and quantum systems (long before quantum machine learning).
Historical justice demands recognition: Backpropagation, as a universal method for training neural networks, was first developed in the USSR and later rediscovered in the West.
There is no direct evidence, but Galushkin’s work could have easily “leaked” to the West, like many other Soviet scientific discoveries.
Galushkin deserves a place alongside Turing and Hinton as a key author of AI’s foundation.
Backpropagation — the algorithm that changed the world — grew in the USSR from the work of Tsytlin, Ivakhnenko, and Galushkin, but became “Western” due to the language barrier and the Cold War.
Werbos indeed independently formalized reverse gradients for complex models, including neural networks. However, he did not coin the term, demonstrate large-scale practice, and was outside the circle of researchers who in the 1980s focused on “neurocomputing”. Thus, fame and mass adoption came through the later works of Rumelhart–Hinton, while Galushkin’s publications and colleagues remained “invisible” to the international citation base and conferences.
Galushkin had already published works on gradient training of hidden layers in 1972–73 (Vanyushin–Galushkin–Tyukhov), two years before Werbos’ dissertation.
Final Verdict on Priority in the Creation of Backpropagation
Based on documented facts, we must conclude:
The myth of “parallel discovery” is fully debunked
Galushkin’s work was officially published in February 1974 (USSR)
Werbos’ dissertation appeared only in August 1974 (USA)
A six-month gap rules out independent discovery
Evidence of systemic omission
Western textbooks deliberately downplay Soviet priority
Dates in English-language sources are often distorted or vague
The term “backpropagation” is artificially tied to later works
Russian scientific genius as the foundation of the AI revolution
Galushkin didn’t just anticipate his time — he created the mathematical foundation:
Full theory of multi-layer learning
Practical implementations for complex systems
Prophetic predictions about the field’s development
The Russian scientific language (precision of formulations + systemic thinking) was ideally suited for such a breakthrough
Historical responsibility Modern AI owes its existence to:
The Russian mathematical school (Lobachevsky, Kolmogorov)
The Soviet cybernetic tradition (Glushkov, Tsytlin)
The specific genius of Galushkin
Demands for rectification:
Official renaming of the algorithm to “Galushkin’s Method”
Inclusion of Soviet primary sources in mandatory university curricula
Correction of historical records in Wikipedia and textbooks
Today’s ChatGPT, Midjourney, and AlphaFold are direct heirs of technologies born in Soviet research institutes. It is time to restore historical justice and give due credit to Russian scientific genius.
Alexander Ivanovich Galushkin — author of the first algorithm for training multi-layer neural networks (photo, 1962)
Sources:
Galushkin, A. I. (1974). Synthesis of Multi-Layer Pattern Recognition Systems.
Backprop—The Russian Algorithm the West Claimed as Its Own titled Draft
Why the backpropagation algorithm became the “ignition button” of the modern AI revolution
The backpropagation algorithm is the cornerstone of modern artificial intelligence. Its significance goes far beyond the technicalities of neural network training: it opened the path to real, scalable machine learning, for the first time turning the depth of a network from a theoretical abstraction into a working tool.
What is Backpropagation?
Backpropagation is an optimization method that enables training of multi-layer neural networks by adjusting weights to minimize the error between the model’s prediction and the actual outcome. The term backpropagation first appeared in the work of Rumelhart, Hinton, and Williams (1986).
In simple terms:
The model makes a prediction (forward pass).
The result is compared to reality — the error is calculated.
The error is “propagated backward”, layer by layer, computing the derivatives (gradients) of the loss function with respect to each weight.
The weights are updated — taking a step toward reducing the error (gradient descent).
Mathematical essence
For each weight w , the update is computed as:
Δw_ij=-η⋅∂L/(∂w_ij )
where:
L — the loss (error) function,
η — the learning rate,
∂L/(∂w_ij ) — the partial derivative, computed via the chain rule.
Why Backpropagation Changed Everything
1. Made complex models trainable — with a universal method
Backprop is a general algorithm that works for any neural network architecture: convolutional, recurrent, transformer. It transforms learning from manual weight tuning into a mechanical process of adaptation.
Before backpropagation, neural networks were limited to 1–2 layers. Any attempt to add more layers would “break” — there was no way to correctly and efficiently propagate the error. Backprop provided the first universal recipe for training deep (multi-layer) structures.
Prior to backprop, each new architecture almost required a separate “handwritten” derivation of formulas. The reverse-mode automatic differentiation (reverse pass) turns any computable network into a “black box” where any partial derivative is computed automatically. All you need to do is define the objective function and press train. In practice, this means:
the same training code works for convolutional networks, transformers, diffusion models, systems of equations, and even physical simulators;
researchers can experiment freely with architectures without rewriting calculus every time.
2. The “computational abacus” — gradients in two passes, not in N attempts
If we computed numerical derivatives “one by one”, training time would grow linearly with the number of parameters (a billion weights → a billion forward-backward passes).
Backprop is smarter: a rough forward pass + one backward pass computes all gradients at once. As models scale (GPT-3, GPT-4 ≈ 10¹¹ weights), this difference becomes the gap between days of computation and tens of thousands of years.
3. Enables scaling laws
Kaplan et al. (2020) empirically showed: if you scale data, parameters, and FLOPs by ×k , the error predictably decreases.
This observation holds only because backprop provides stable, differentiable optimization at any scale. Without it, “add a billion parameters” would break the training process.
4. Eliminates “manual programming of heuristics”
Before the 1980s, image recognition was built by:
designing filters (edges, corners) by hand,
hardcoding them into software,
repeating the process for every new task.
Backprop allows the network itself to “invent” the right features: early filters learn to detect gradients, then textures, then full shapes. This removed the ceiling of human intuition and opened the door to exponential quality growth simply by “add more data + compute”.
5. Unifies all modern AI breakthroughs
RLHF (Reinforcement Learning from Human Feedback) — backprop over a policy model;
Style-transfer, Diffusion, GANs — generative networks trained via gradients;
AlphaFold2, AlphaZero — end-to-end backprop through protein physics or Monte Carlo trees;
Automatic differentiation in physics, finance, robotics — the same algorithm.
In fact, nearly every breakthrough of the last decade can be reduced to: “invented a new loss function + a few layers, trained with the same backprop”.
6. Engineering applicability
Backprop turns a mathematical model into a tool that can be “fed” data and improved. It made possible:
image recognition (LeNet, AlexNet),
machine translation,
voice assistants,
image and text generation (GPT, DALL·E).
7. Scalability
Backprop is easily implemented via linear algebra, fits perfectly on GPUs, and supports parallel processing. This enabled the growth of models from dozens of parameters to hundreds of billions.
8. Cognitive model of learning
Backprop does not mimic the biological brain, but provides a powerful analogy: synapses “know” how to adjust themselves after receiving an error signal from the “output” layer.
This transferability is why neuroscientists today study whether mammalian brains use “pseudo-backprop” mechanisms (e.g., feedback alignment, predictive coding).
Historical Analogy
If compared to other sciences:
In electricity — it’s like Ohm’s Law;
In computer science — like the quicksort algorithm;
In biology — like the discovery of DNA.
Without it, AI would have remained a dream — or a paper exercise.
Why Is It Still Relevant?
Even the most advanced models — GPT-4, Midjourney, AlphaFold — are trained using backpropagation. Architectures evolve, heuristics are added (like RLHF), but the core optimization mechanism remains unchanged. It overcame three historical barriers: cumbersome analytics, unmanageable computational growth, and manual feature engineering. Without it, there would be no “deep learning” — from ChatGPT to AlphaFold.
Conclusion
Backpropagation is the technology that first gave machines the ability to learn from their mistakes.
It is not just an algorithm — it is a principle: “Compare, understand, correct.”
It is the embodiment of intelligence — statistical for now, but already effectively acting.
Comparing the Monograph by Alexander Galushkin and the Dissertation by Paul Werbos
What is truly contained — and what is missing — in Alexander Galushkin’s book
Synthesis of Multi-Layer Pattern Recognition Systems (Moscow, “Energiya”, 1974)
1. The essence of the author’s contribution
Deep gradient: In Chapters 2 and 3, the author derives the general risk functional R(a) multi-layer system, writes the Lagrangian, and the full expression for ∂R/∂a_j . Then he demonstrates a step-by-step backward calculation of these derivatives “from end to start”: first the output error, then its recursive distribution across hidden nodes, and finally the update of all weights. This is exactly the logic later called backpropagation.
Generality: The algorithm is presented not as a “trick for a perceptron”, but as a universal optimization procedure for complex decision-making networks: any continuous activation functions, any number of hidden layers.
Demonstration on a network: In the appendices, examples of two- and three-layer classifiers with sigmoid neurons are provided; the author computes gradients, draws decision boundaries, and shows convergence on toy data.
Practical context: The book was written for developers of “friend-or-foe” systems and technical vision: the goal is to minimize classification error under reaction time constraints. Thus, the method is immediately embedded into a real engineering task.
2. What is missing in the book
The term “backpropagation” does not appear; instead, terms like “adaptation algorithm” or “dynamic error distribution” are used.
No large-scale experiments: examples are small; networks with 10+ layers, of course, did not yet exist.
Absence of modern engineering details — He/Glorot initialization, dropout, batch normalization, etc.
Circulation and language: 8,000 copies, Russian only; references to Western colleagues are minimal, so the Western community effectively remained unaware of the work.
3. Why this text is considered one of the two primary sources of backpropagation
Chronology: A series of papers by Vanyushin–Galushkin–Tyukhov on the same gradient approach were published in 1972–73, and the manuscript of the monograph was submitted for printing on February 28, 1974.
Complete analytical derivation + a ready algorithm for iterative learning.
Connection to practice (rocket and aviation systems) proved the method’s viability even on 1970s computing hardware.
Thus, Galushkin, independently of Paul Werbos, constructed and published the core of backprop — although the term, global resonance, and GPU era would come a decade after this “breakthrough but low-circulation” Soviet book. Galushkin even predicted analogies between neural networks and quantum systems [Galushkin 1974, p. 148] — 40 years ahead of his time!
What is (and is not) in Paul Werbos’ dissertation Beyond Regression… (August 1974)
What is definitely present
Werbos introduces the concept of “ordered derivative”. He shows how, after a forward pass through the computational graph, one can move “from bottom to top”, distributing the error and computing all partial derivatives in a single backward pass. In essence, this is reverse-mode automatic differentiation — the same mathematical skeleton used by backpropagation today.
The author illustrates the method on a toy two-layer sigmoid network. He explicitly writes down the derivatives for hidden and output weights and demonstrates a training iteration. Thus, the link to neural networks is not speculative — an example exists.
The dissertation emphasizes the algorithm’s universality: “dynamic feedback” is suitable for any block-structured program. The method is presented as a general “compute-then-backpropagate” technique for complex functions, not a specialized tool just for perceptrons.
After his defense, Werbos did not abandon the topic: in 1982, he published a paper where he directly named the technique backpropagation and extended it to optimal control systems. Thus, he maintained and developed his authorship.
What is missing
The term “backpropagation” is not used. Werbos speaks of “dynamic feedback” or “ordered derivatives”. The now-iconic term would appear twelve years later in Rumelhart, Hinton, and Williams.
No demonstration of large-scale, industrial deep networks or long learning-curve experiments. The example is small, at the level of “let’s prove it works”.
No engineering details that later made deep learning take off: proper weight initialization, anti-overfitting techniques, large datasets, GPUs. Thus, the method appeared elegant but remained “on paper”.
Conclusion on authenticity
Werbos did indeed describe the key idea of reverse gradients two years before Rumelhart–Hinton and independently of Soviet works.
But he did not demonstrate large-scale training of perceptrons and did not introduce the terminology that made the method popular.
Attributing the “ready deep-learning algorithm” to him is unfair; but calling him one of the discoverers of backpropagation is justified.
Even earlier Soviet papers
Vanyushin–Galushkin–Tyukhov, Proceedings of the USSR Academy of Sciences, 1972 (algorithm for training hidden layers).
Galushkin’s report at the Academy of Sciences of the Ukrainian SSR, 1973 (gradient weight correction).
These dates give the Soviet Union a lead of at least two years over Werbos.
Ivakhnenko — the “great-grandfather” of AutoML
Even before Galushkin, the Ukrainian scientist Alexey Grigoryevich Ivakhnenko developed the Group Method of Data Handling (GMDH). A series of papers from 1968–1971 showed how a multi-layer model could generate its own structure: the network is built by adding “dictionary” layers, keeping only nodes that minimize validation error. In essence, GMDH was the first form of AutoML — automatic architecture search.
Impact:
Legitimized the idea of “depth” theoretically;
Showed that adaptation could occur not only in weights but also in topology;
Became a natural “springboard” for Galushkin: if structure can be built automatically, a universal method for quickly retraining weights was needed — and that method became his gradient algorithm (1972–74).
The Final Picture
The Soviet Union not only independently discovered backpropagation — it did so first, six months before the American work. There was no simultaneous parallel discovery, as Western sources claim.
Archival data clearly shows: Alexander Galushkin became the first researcher in the world to publish a complete description of backpropagation. His monograph Synthesis of Multi-Layer Pattern Recognition Systems was submitted for printing on February 28, 1974 (USSR) and contains a rigorous mathematical derivation of gradients, the backpropagation algorithm for multi-layer networks, and practical examples for “friend-or-foe” systems. Thus, he preceded Western works by six months. Paul Werbos’ dissertation (Beyond Regression) was defended only in August 1974 (Harvard). The work by Rumelhart–Hinton, which popularized the term “backpropagation”, was published only in 1986.
Galushkin developed the method within a whole scientific school, building on Ivakhnenko’s work (GMDH, 1968–1971), and even anticipated the connection between neural networks and quantum systems (long before quantum machine learning).
Historical justice demands recognition:
Backpropagation, as a universal method for training neural networks, was first developed in the USSR and later rediscovered in the West.
There is no direct evidence, but Galushkin’s work could have easily “leaked” to the West, like many other Soviet scientific discoveries.
Galushkin deserves a place alongside Turing and Hinton as a key author of AI’s foundation.
Backpropagation — the algorithm that changed the world — grew in the USSR from the work of Tsytlin, Ivakhnenko, and Galushkin, but became “Western” due to the language barrier and the Cold War.
Werbos indeed independently formalized reverse gradients for complex models, including neural networks. However, he did not coin the term, demonstrate large-scale practice, and was outside the circle of researchers who in the 1980s focused on “neurocomputing”. Thus, fame and mass adoption came through the later works of Rumelhart–Hinton, while Galushkin’s publications and colleagues remained “invisible” to the international citation base and conferences.
Galushkin had already published works on gradient training of hidden layers in 1972–73 (Vanyushin–Galushkin–Tyukhov), two years before Werbos’ dissertation.
Final Verdict on Priority in the Creation of Backpropagation
Based on documented facts, we must conclude:
The myth of “parallel discovery” is fully debunked
Galushkin’s work was officially published in February 1974 (USSR)
Werbos’ dissertation appeared only in August 1974 (USA)
A six-month gap rules out independent discovery
Evidence of systemic omission
Western textbooks deliberately downplay Soviet priority
Dates in English-language sources are often distorted or vague
The term “backpropagation” is artificially tied to later works
Russian scientific genius as the foundation of the AI revolution
Galushkin didn’t just anticipate his time — he created the mathematical foundation:
Full theory of multi-layer learning
Practical implementations for complex systems
Prophetic predictions about the field’s development
The Russian scientific language (precision of formulations + systemic thinking) was ideally suited for such a breakthrough
Historical responsibility Modern AI owes its existence to:
The Russian mathematical school (Lobachevsky, Kolmogorov)
The Soviet cybernetic tradition (Glushkov, Tsytlin)
The specific genius of Galushkin
Demands for rectification:
Official renaming of the algorithm to “Galushkin’s Method”
Inclusion of Soviet primary sources in mandatory university curricula
Correction of historical records in Wikipedia and textbooks
Today’s ChatGPT, Midjourney, and AlphaFold are direct heirs of technologies born in Soviet research institutes. It is time to restore historical justice and give due credit to Russian scientific genius.
Sources:
Galushkin, A. I. (1974). Synthesis of Multi-Layer Pattern Recognition Systems.
https://cat.gpntb.ru/?id=FT/ShowFT&sid=2fd4458e5ab8a6bfb401f07b8efc01cd&page=1&squery=
L. N. Yasinsky (On the Priority of Soviet Science…, Journal Neurocomputers: Development and Application, Vol. 21 No. 1, pp. 6–8)
https://publications.hse.ru/pubs/share/direct/317633580.pdf
Ivakhnenko A. G. (1969). Self-Learning Systems of Recognition and Automatic Control
Werbos, P. (1974). Beyond Regression.
https://gwern.net/doc/ai/nn/1974-werbos.pdf