Neural Network History - shiffman/ML-for-Creative-Coding GitHub Wiki
Major Milestones in Neural Network History
1943 β First Artificial Neuron Model
Warren McCulloch and Walter Pitts develop the first mathematical model of a neuron, demonstrating how networks of simple binary units could compute logical functions.
1949 β Hebbβs Learning Rule
Donald Hebb proposes that neurons strengthen connections when they activate together frequently, introducing a key idea behind synaptic plasticity and learning. Hebb's Rule was an early idea about how neurons might learn, but modern multi-layer perceptrons rely on error-driven learning (backpropagation), which is different.
1957 β The Perceptron
Frank Rosenblatt introduces the perceptron, a single-layer neural network that can be trained to recognize patterns using a supervised learning rule.
1969 β Perceptron Limitations Identified
Marvin Minsky and Seymour Papert show that perceptrons cannot solve problems requiring non-linear decision boundaries (e.g., XOR). This contributed to skepticism about neural networks and played a role in the onset of the first AI Winter, a period of reduced funding and interest in neural network research.
1986 β Backpropagation Revives Neural Networks
Geoffrey Hinton, David Rumelhart, and Ronald Williams demonstrate the effectiveness of backpropagation, enabling multi-layer networks to learn complex patterns. Despite backpropagationβs success, progress slowed in the late 1980s and 1990s due to computational limitations and lack of large datasets. Skepticism about neural networks resurfaced as they struggled to outperform traditional statistical methods, contributing to a second AI Winter.
1997 β Long Short-Term Memory (LSTM)
Sepp Hochreiter and JΓΌrgen Schmidhuber introduce LSTM, a recurrent neural network variant designed to remember long-term dependencies by using gating mechanisms.
1998 β Convolutional Neural Networks (LeNet-5)
Yann LeCun and colleagues develop LeNet-5, a convolutional neural network (CNN) for handwritten digit recognition, which influences modern deep learning in vision.
2006 β Deep Learning Resurgence
Geoffrey Hinton and Ruslan Salakhutdinov introduce deep belief networks (DBNs), demonstrating that unsupervised pre-training can improve the training of deep networks.
2012 β Deep Learning Breakthrough (AlexNet)
Alex Krizhevsky, Ilya Sutskever, and Geoffrey Hinton introduce AlexNet, a deep CNN that wins the ImageNet competition and kickstarts the modern deep learning boom.
2014 β Attention Mechanism Introduced
Dzmitry Bahdanau, Yoshua Bengio, and colleagues propose attention mechanisms for sequence-to-sequence models, improving neural machine translation.
- π Bahdanau et al. (2014) β Neural Machine Translation by Jointly Learning to Align and Translate
2017 β Transformer Architecture
Google Brain researchers introduce the Transformer model, eliminating recurrence and relying entirely on self-attention, which becomes the foundation for modern NLP models.