11 Nov 2024, Taro Langner
In this post: An attempt to reconstruct Ilya Sutskever's 2020 AI reading list
(8 min read)
I recently shared a summary of a viral AI reading list attributed to Ilya Sutskever, which laid claim to covering ‘90% of what matters’ back in 2020. It boils down the reading items to barely one percent of the original word count to form the TL;DR I would have wished for before reading.
The viral version of the list as shared online is known to be incomplete, however, and includes only 27 of about 40 original reading items.
The rest allegedly fell victim to the E-Mail deletion policy at Meta¹. These missing reading items have inspired some good discussions in the past, with many different ideas as to which papers would have been important enough to include.
This post is an attempt to identify these lost reading items. It builds on clues gathered from the viral list, contemporary presentations given by Ilya Sutskever, resources shared by OpenAI and more.
¹Correction: An earlier version mistakenly referred to OpenAI here instead of Meta
Filling the Gaps
The main piece of evidence is a claim shared along with the list according to which an entire selection of meta-learning papers was lost.
Meta-learning is often said to pursue ‘learning to learn’, with neural networks being trained for a general ability to adapt more easily to new tasks for which only few training samples are available. A network should thus be able benefit from its existing weights without requiring an entirely new training from scratch on the new data. One-shot learning provides just a single training sample to a model from which it is expected to learn a new downstream task, whereas zero-shot settings provide no annotated training samples at all.
For some of the candidate papers listed below, the case can be strengthened further by evidence in the form of an endorsement straight from OpenAI itself. Ilya Sutskever was chief scientist at a time when OpenAI published the educational resource ‘Spinning Up in Deep RL’ which includes several of these candidates in an entirely separate reading list of 105 ‘Key Papers in Deep RL’. Below, the papers which also appear in that list are marked with a symbol (⚛).
Clues from the Preserved Reading Items
Some meta-learning concepts can be found even in the known parts of the list.
The preserved reading items can be arranged into a narrative arc around a related branch of research on Memory-Augmented Neural Networks (MANNs). Following the ‘Neural Turing Machine’ (NTM) paper, ‘Set2Set’ and ‘Relational RNNs’ experimented with external memory banks that an RNN could read and write information on. They directly cite or closely relate to several papers which may well have been part of the original list:
Potential Reading Items (Part 1):
Clues from Contemporary Presentations
Certain papers about meta-learning and competitive self-play also feature repeatedly in a series of presentations held by Ilya Sutskever around this time and may well have eventually been included in the reading list too.
Recorded Presentations:
- Meta Learning and Self Play - Ilya Sutskever, OpenAI (YouTube), 2017
- OpenAI - Meta Learning & Self Play - Ilya Sutskever (YouTube), 2018
- Ilya Sutskever: OpenAI Meta-Learning and Self-Play (YouTube), 2018
These presentations largely overlap and repeatedly reference known contents of the reading list. They open with a fundamental motivation of why deep learning works, framing backpropagation with neural networks as a search for small circuits that relate to the Minimum Description Length principle, according to which the shortest program that can explain given data will reach the best generalization possible.
Next, all three presentations reference the following meta-learning papers:
Potential Reading Items (Part 2):
Reinforcement Learning (RL) also features heavily in all three presentations, with close links to meta-learning. One key concept is competitive self-play in which agents interact in a simulated environment to reach specific, typically adversarial objectives. As a way to ‘turn compute into data’, this approach enabled simulated agents to outperform human champions and invent new moves in rule-based games. Ilya Sutskever presents an evolutionary biology perspective that relates competitive self-play to the impact of social interaction on brain size (pay-walled link).
He goes on to suggest that rapid competence gain in a simulated ‘agent society’ may ultimately, according to his judgement, provide a plausible path towards a form of AGI.
Given the significance he ascribes to these concepts, it seems plausible that some of the cited papers on self-play may have later also been included in the reading list. They may form a sizeable chunk of the missing items, especially as RL is otherwise mentioned by only one of the preserved reading items.
Potential Reading Items (Part 3):
- ‘Hindsight Experience Replay’⚛
as Andrychowicz et al., 2017
- ‘Continuous control with deep reinforcement learning’⚛
as DDPG: Deep Deterministic Policy Gradients, 2015
- ‘Sim-to-Real Transfer of Robotic Control with Dynamics Randomization’
as Peng et al., 2017
- ‘Meta Learning Shared Hierarchies’
as Frans et al., 2017
- ‘Temporal Difference Learning and TD-Gammon [1995]’
as Tesauro et al., 1992
- ‘Karl Sims - Evolved Virtual Creatures, Evolution Simulation, 1994’
as Carl Sims, 1994 (YouTube video [4:09])
- ‘Emergent Complexity via Multi-Agent Competition’
as Bansal et al., 2017
- ‘Deep reinforcement learning from human preferences’⚛
as Christiano et al., 2017 (Note: Introduces RLHF)
Even today, these presentations from around 2018 are still worth watching. Next to fascinating bits of knowledge, they also include gems such as the statement:
‘Just like in the human world: The reason humans find life difficult is because of other humans’
-Ilya Sutskever
While some concepts in computer science accordingly appear timeless, other points may seem surprising today, like the casual remark of an audience member in the Q&A session:
‘It seems like an important sub-problem on the path to AGI will be understanding language, and the state of generative language modelling right now is pretty abysmal.’
-Audience member
To which Ilya Sutskever responds:
‘Even without any particular innovations beyond models that exist today, simply scaling up models that exist today on larger datasets is going to go surprisingly far.’
-Ilya Sutskever (in 2018)
This response was later confirmed by experimental results in the reading item ‘Scaling Laws for Neural Language Models’ (which echoes the ‘Bitter Lesson’ by Rich Sutton). It was ultimately proven true, as he would oversee Transformer architectures scaled up to an estimated 1.8 trillion parameters and costing over $60 million to train on 128 GPUs forming Large Language Models (LLMs) which are today capable of generating text that is increasingly difficult to distinguish from human writing.
Honorable Mentions
Many other works and authors may have featured on the original list, but the evidence wears increasingly thin from here on.
Overall, the preserved reading items manage to strike an impressive balance between covering different model classes, applications and theory while also including many famous authors of the field. Perhaps the exceptions to this rule are worth noting, even if they may have slipped among the ‘10% of what matters’ that didn’t make the original list.
As such, it would have seemed plausible to include:
Conclusion
This post will remain largely speculative until more becomes known. After all, even the viral list itself was never officially confirmed to be authentic. Nonetheless, the potential candidates for the lost reading items listed above seemed worth sharing. Taken together, they may well fill a gap in the viral version of the list that would, in the words of the author, corresponded roughly to a missing ‘30% of what matters’ at its time.
24 Sep 2024, Taro Langner
In this post: Ilya Sutskever's AI Reading list in ~120 words per item
(15 min read)
Earlier this year, a reading list with about 30 papers was shared on Twitter.
It reportedly forms part of a longer version originally compiled by Ilya Sutskever, co-founder and chief scientist of OpenAI at the time, for John Carmack in 2020 with the remark:
‘If you really learn all of these, you’ll know 90% of what matters’.
While the list is fragmentary and much has happened in the field since, this endorsement and the claim that it was part of onboarding at OpenAI quickly made it go somewhat viral.
At about 300,000 words total, the combined content nonetheless corresponds to around one thousand book pages of dense, technical text and requires a decent investment in time and energy for self-study.
After doing just that, I therefore dedicate this blog post to all those of us who provisionally bookmarked it (“for later”) and are still curious. What follows is my own condensed and structured summary with about 120 words per item, free of mathematical notation, to capture the essential key points, context and some perspective gained from reading it with the surrounding literature.
In a Nutshell
The list contains 27 reading items, with papers, blog posts, courses, one dissertation and two book chapters, all originally dating from 1993 to 2020.
The contents can be roughly broken down as follows:
Methodology |
Items |
Share* |
Topics |
Convolutional Neural Networks (CNNs) |
5 |
25% |
image recognition, semantic segmentation |
Recurrent Neural Networks (RNNs) |
10 |
19% |
language modeling, speech-to-text, machine translation, combinatorial optimization, visual question answering, content-based attention |
Transformers |
3 |
6% |
multi-head and dot-product attention, language model scaling |
Information Theory |
5 |
42% |
Kolmogorov complexity, compression, Minimum Description Length |
Miscellaneous |
4 |
8% |
variational inference, representation learning, graph neural networks, distributed training |
Using these categories, the next sections summarize the gist of each item, roughly sorted by how they build on each other.
Convolutional Neural Networks
CS231, 2017
Stanford University Course
Length: ~50,000 words, forming 11 blocks of 2 modules
Instructors: Fei-Fei Li, Andrej Karpathy and Justin Johnson
🔗
[CS231, 2017] is a classic course on deep learning fundamentals from Stanford University. It builds up from linear classifiers and their ability to learn a given task based on mathematical optimization, or training, which adjusts their internal parameter weights such that applying them to input data will produce more desirable outputs. This basic concept is developed into backpropagation for training of neural networks, in which trainable parameters are typically arranged into multiple layers together with other modules such as activation functions and pooling layers. Convolutional Neural Networks (CNNs) are introduced as a specialized architecture for image recognition, as used in modern computer vision systems to this day. Extended video lectures are available on youtube.
Note: If you are starting from zero, this course and newer resources by e.g. DeepLearning.AI on Coursera or FastAI will help you to get more out of the remaining list.
AlexNet, 2012
Paper
Length: ~6,000 words
Authors: Alex Krizhevsky, Ilya Sutskever and Geoffrey E. Hinton
🔗
[AlexNet, 2012] established CNNs as state of the art for image recognition and arguably initiated the widespread hype around deep learning. It outperformed its competitors in the 2012 ImageNet benchmark challenge, predicting whether a given input image contained e.g. a cat, dog, ship or any other of 1,000 possible classes, so conclusively that the real-world dominance of deep learning became commonly accepted. An important factor was its early* CUDA implementation that enabled unusually fast training on GPUs.
*Note: Earlier GPU implementations are documented in section 12.1.2 of the book Deep Learning.
ResNet, 2015
Paper
Length: ~6,000 words
Authors: Kaiming He, Xiangyu Zhang, Shaoqing Ren and Jian Sun
🔗
</details>
[ResNet, 2015] succeeded AlexNet as a more modern CNN architecture, reaching first place on the ImageNet challenge in 2015. It remains a popular CNN architecture to this day and is subject of ongoing research. It introduced residual connections into CNN architectures that had become ever deeper, stacking more convolutional layers to achieve higher representational power. By allowing residual connections to skip or bypass entire blocks of layers, ResNet architectures suffered less from gradient degradation effects in training and could thus be robustly trained at previously unseen depth.
ResNet identity mappings, 2016
Paper
Length: ~6,000 words
Authors: Kaiming He, Xiangyu Zhang, Shaoqing Ren and Jian Sun
🔗
[ResNet identity mappings, 2016] were later proposed by the ResNet authors as a ‘clean’ information path and best design for the skip connections, so that their contents are merely added to the results of a bypassed block without any further modification. Whereas earlier designs placed an activation layer on the skip path after the addition, the proposed pre-activation design moves this layer to the start of the bypassed block instead. The skip connections can thus form a shortcut through the entire neural network that is only interrupted by additions, allowing improved propagation of gradient signals that make it possible for even deeper neural networks to be trained.
Dilated convolutions, 2015
Paper
Length: ~6,000 words
Authors: Fisher Yu and Vladlen Koltun
🔗
[Dilated convolutions, 2015] (or à trous convolutions) were proposed as a new type of module for dense prediction with CNNs in tasks like semantic image segmentation, where class labels are assigned to any given pixel of an input image. Architectures such as AlexNet and ResNet condense input images to lower-dimensional representations via strided convolutions or pooling layers to predict one class label for an entire image. Related architectures for dense prediction therefore typically restore the original input image resolution from these downsampled, intermediate representations via upsampling operations. Whereas e.g. transpose convolutions achieve this with competitive results, dilated convolutions avoid downsampling entirely. Instead, they space out the filter kernel of a convolutional layer to skip one or more neighboring input pixels, thereby providing a larger receptive field without any reduction in resolution.
Recurrent Neural Networks
Today, Recurrent Neural Networks (RNNs) have been largely superseded by Transformers and date from what Ilya Sutskever himself would later call the “[pre-2017] stone age” of machine learning. They nonetheless remain subject of active research and see continued use in certain applications. Forming a substantial part of the reading list, they showcase the evolution of early insights and architectural developments that lead up to the systems of today. Most of the RNNs listed below are Long Short-Term Memory (LSTM) architectures. Some designs furthermore include Feedforward Networks with no recurrent connections, usually trained end-to-end as part of the model.
Understanding LSTM Networks, 2015
Blog Post
Length: ~2,000 words
Author: Christopher Olah
🔗
[Understanding LSTM Networks, 2015] provides a brief introduction to RNNs and LSTMs in particular. RNNs can process a sequence of inputs, one step at a time, while evolving a hidden state vector that is (re-)ingested, updated and returned again at each step along the input sequence. The hidden state vector thereby allows for information to persist and be passed to subsequent processing steps. Nonetheless, simpler RNNs typically struggle with long-term dependencies. LSTMs alleviate this by introducing a cell state as additional recurrent in- and output, acting as a memory pathway for addition, update or removal of information along each processing step via trainable gating mechanisms.
The Unreasonable Effectiveness of RNNs, 2015
Blog Post
Length: ~6,000 words
Author: Andrej Karpathy
🔗
[The Unreasonable Effectiveness of RNNs, 2015] shows use cases and results of RNNs in action. They are distinguished by their ability to both process and predict variable-sized sequences while also maintaining an internal state, prompting the author Andrej Karpathy to state “If training vanilla neural nets is optimization over functions, training recurrent nets is optimization over programs.”. He showcases results for image captioning and character-level language modeling that enables RNNs to automatically generate prose and articles. Early code generation capabilities are noted, with convincing syntax but failure to compile and a tendency to suffer from hallucinations where the model provides outputs as most probable that are evidently incorrect. The blog post also includes a minimal RNN code example.
RNN Regularization, 2014
Paper
Length: ~3,500 words
Authors: Wojciech Zaremba, Ilya Sutskever and Oriol Vinyals
🔗
[RNN regularization, 2014] addresses the challenge of training large RNNs without overfitting, where a model would excessively adapt to, or even memorize, its training samples and fail to generalize to new data. A technique for regularization, which aims to reduce this effect, was proposed that applies dropout, which omits randomly selected outputs of a given neural network layer. Dropout had been known for several years and was used e.g. in AlexNet. Here, the key insight was to utilize dropout only within a given RNN cell, but to avoid it on the recurrent connections that carried the hidden state vector. In this way, larger RNNs could avoid overfitting while preserving long-term dependencies.
Neural Turing Machines, 2014
Paper
Length: ~7,500 words
Authors: Alex Graves, Greg Wayne and Ivo Danihelka
🔗
[Neural Turing Machines, 2014] were proposed as a form of memory-augmented neural network, with an external memory bank on which an RNN controller could write or erase information with a ‘blurry’, differentiable, attention-based focus. Equipped with this working memory, the Neural Turing Machine outperformed a baseline RNN in experiments involving associative recall, copying and sorting sequences and generalized more robustly to sequence lengths that exceeded those encountered in training.
Deep Speech 2, 2016
Paper
Length: ~7,000 words
Authors: Dario Amodei, Sundaram Ananthanarayanan, Rishita Anubhai, Jingliang Bai, Eric Battenberg, Carl Case, Jared Casper, Bryan Catanzaro, Qiang Cheng, Guoliang Chen, Jie Chen, Jingdong Chen, Zhijie Chen, Mike Chrzanowski, Adam Coates, Greg Diamos, Ke Ding, Niandong Du, Erich Elsen, Jesse Engel, Weiwei Fang, Linxi Fan, Christopher Fougner, Liang Gao, Caixia Gong, Awni Hannun, Tony Han, Lappi Johannes, Bing Jiang, Cai Ju, Billy Jun, Patrick LeGresley, Libby Lin, Junjie Liu, Yang Liu, Weigao Li, Xiangang Li, Dongpeng Ma, Sharan Narang, Andrew Ng, Sherjil Ozair, Yiping Peng, Ryan Prenger, Sheng Qian, Zongfeng Quan, Jonathan Raiman, Vinay Rao, Sanjeev Satheesh, David Seetapun, Shubho Sengupta, Kavya Srinet, Anuroop Sriram, Haiyuan Tang, Liliang Tang, Chong Wang, Jidong Wang, Kaifu Wang, Yi Wang, Zhijian Wang, Zhiqian Wang, Shuang Wu, Likai Wei, Bo Xiao, Wen Xie, Yan Xie, Dani Yogatama, Bin Yuan, Jun Zhan, Zhenyao Zhu
🔗
[Deep Speech 2, 2016] proposed an automatic speech recognition system to convert audio recordings into text by processing log-spectrograms representing the audio with RNNs to predict sequences of characters of either English or Mandarin. The authors utilized batch normalization instead of dropout for regularization on the non-recurrent layers and Gated Recurrent Units (GRUs) as a somewhat simplified alternative to LSTMs used in most of the other papers examined so far, together with a plethora of other engineering tweaks, including batched processing for low-latency streaming output.
RNNsearch, 2015
Paper
Length: ~8,000 words
Authors: Dzmitry Bahdanau, KyungHyun Cho and Yoshua Bengio
🔗
[RNNsearch] is credited with introducing the first attention mechanism into Natural Language Processing (NLP), proposing additive, content-based attention for neural machine translation. Its encoder-decoder architecture encodes an input sequence of English words with an RNN encoder into a context vector used by an RNN decoder to predict an output sequence of French words. In prior work this context vector was simply the final hidden state of the encoder, which therefore had to contain all relevant information about the input sequence. RNNsearch addresses this bottleneck by making the context vector a weighted sum over all encoder hidden states, or annotations. When predicting a target word, the decoder can thereby rely on context from arbitrary parts of the encoded input sequence by (re-)calculating the context vector as a weighted sum of annotations. The weighting is determined by an alignment model, a feedforward network that receives the current decoder hidden state together with an annotation and assigns a score to the latter.
Pointer Networks, 2015
Paper
Length: ~4,500 words
Authors: Oriol Vinyals, Meire Fortunato and Navdeep Jaitly
🔗
[Pointer Networks] repurpose the concept of content-based attention to solve combinatorial optimization problems. Here, content-based attention is used to ‘point’ at elements of the input sequence in a specific order. The output sequence is therefore an indexing of the input elements. Given a set of two-dimensional points as input, Pointer Net was trained to solve for their convex hull, Delaunay triangulation or Traveling Salesman Problem by predicting in which order these points should be visited. With no limitation to the length of the output sequence or dictionary, this approach was found to generalize beyond the longest sequence length encountered in training.
Set2Set, 2016
Paper
Length: ~6,500 words
Authors: Oriol Vinyals, Samy Bengio and Manjunath Kudlur
🔗
[Set2Set] extends sequence-to-sequences methods as examined above to enable order-invariant processing of sets. These methods are shown to strongly depend on the specific order of both an input and output sequence (e.g. the exact order of random points provided to Pointer Networks for convex hull prediction). The authors propose Set2Set as a solution, with the encoder forming a memory bank (that resembles annotations of RNNSearch) to create a context vector. This memory bank, however, is sampled more than just once. Instead, a process block introduces a new LSTM, which evolves a query vector for repeated, content-based attention readouts of the memory. Finally, the write block (a Pointer Network) can add even more attention steps in the form of glimpses.
Relation Networks, 2017
Paper
Length: 5,000 words
Authors: Adam Santoro, David Raposo, David G. Barrett, Mateusz Malinowski, Razvan Pascanu, Peter Battaglia, Timothy Lillicrap
🔗
[Relation Network, 2017] modules were proposed as a method for relational inference tasks such as visual and text-based question answering. The Relation Network module ingests a pair of feature vectors, for example an LSTM hidden state for a word or sentence in text or the values at a specific pixel position in feature maps produced by a CNN for image data. A given pairing is processed with one or more neural network layers before forming an element-wise sum and creating an output with a second stack of layers. By doing this for all pairs of inputs, this approach outperformed the human baseline in answering textual questions regarding the size, position and color of 3D generated shapes relative to each other.
Relational Recurrent Neural Networks, 2018
Paper
Length: 6,000 words
Authors: Adam Santoro, Ryan Faulkner, David Raposo, Jack Rae, Mike Chrzanowski, Theophane Weber, Daan Wierstra, Oriol Vinyals, Razvan Pascanu, Timothy Lillicrap
🔗
[Relational Recurrent Neural Networks, 2018] proposed a Relational Memory Core module in which an attention mechanism allows memories to interact with each other and be recurrently refined as a fixed-size matrix. This approach was adapted for several tasks requiring relational reasoning and outperformed multiple baseline methods. Given a random set of vectors of which an arbitrary one was marked, it predicted which other vector had the highest Euclidean distance to it. It also learned to execute short code snippets involving variable manipulation, performed language modeling and scored well in a toy reinforcement learning task. The self-attention mechanism that enabled its memory interactions is described in the following section.
The previous section tracks the rise of attention mechanisms as an increasingly potent tool for providing context in sequence-to-sequence prediction tasks. Eventually, these developments yielded the Transformer as a neural network architecture that predominantly relies on attention and discards both recurrent and convolutional layers entirely. The excellent scalability of this approach, together with growing compute resources and extensive training data, established Transformers as dominant method for language modeling, forming the backbone of systems like ChatGPT and performing well even on image and multimodal data.
New attention mechanisms enabled substantial speed and efficiency advantages:
The additive attention mechanism of the previous section compared an encoder and decoder hidden state to each other by applying an alignment model to each such pair for scoring. Internally, the alignment model formed linear projections of both vectors and added them together to calculate a score, which was normalized over all encoder hidden states to form a context vector as their weighted average.
With multiplicative attention, Transformers compare multiple pairings of hidden states at once by forming a dot product of their linear projections, which can be implemented with faster, highly optimized matrix multiplications.
Attention Is All You Need, 2017
Paper
Length: ~4,500 words
Authors: Ashish Vaswani, Noam Shazeer, Niki Parmar, Jakob Uszkoreit, Llion Jones, Aidan N. Gomez, Łukasz Kaiser and Illia Polosukhin
🔗
[Attention Is All You Need] proposed the Transformer architecture. In an encoder-decoder structure for machine translation, embedding layers convert each input token into a feature vector, to which positional encodings are added. The proposed Scaled Dot-Product Attention computes a weighted average over multiple value vectors, each weighted by comparing its associated key vector to a given query vector using the dot product. The result is scaled (for numerical stability) and then normalized over all keys with a softmax function. Multi-head attention conducts this process in parallel with different, learned projections of each input.
Three variants of this mechanism are used. In self-attention used by the encoder, the query, key and value are distinct linear projections of the same output vector from the previous layer. In masked self-attention the decoder furthermore masks out the weights for future tokens. Finally, in encoder-decoder attention, each decoder block obtains only the query from the preceding decoder layer, whereas key and value originate from the final encoder layer. The experiments exceeded state-of-the-art results, with two orders of magnitude lower compute resources than previous approaches.
The Annotated Transformer, 2020
Blog Post (2022 version)
Length: ~6,000 words
Authors: Austin Huang, Suraj Subramanian, Jonathan Sum, Khalid Almubarak, and Stella Biderman (2020 original by Sasha Rush)
🔗
[The Annotated Transformer] implements the Transformer as described in ‘Attention is All You Need’ line by line as a fully functional Jupyter Notebook using PyTorch, with all code available on GitHub. Text segments of the original paper feature alongside the code, together with comments and visualizations that clarify various aspects of the architecture beyond the contents of the paper. The notebook also implements examples for data formatting, training and inference that show the Transformer applied in practice.
Note: The Illustrated Transformer by Jay Alammar is yet another in-depth guide.
Scaling Laws for Neural Language Models, 2020
Paper
Length: ~9,000 words
Authors: Jared Kaplan, Sam McCandlish, Tom Henighan, Tom B. Brown, Benjamin Chess, Rewon Child, Scott Gray, Alec Radford, Jeffrey Wu and Dario Amodei
🔗
[Scaling Laws for Neural Language Models] explores the predictive performance of Transformers for language modeling as a function of model size, data quantity and available compute resources. Extensive empirical results enable the authors to establish formulas that relate these factors to each other over seven orders of magnitude and enable several recommendations as to their optimal configuration. While each of these can form a bottleneck, model size (i.e. the number of trainable parameters) forms the single most impactful factor. Larger models reach higher sample-efficiency and better generalization earlier on in training. The specific model architecture has little effect. An eight-fold increase in model size requires a five-fold increase in training data. Given a fixed compute budget, the authors accordingly recommend to prioritize first model size, then batch size and only then the number of training steps, with early stopping before convergence typically providing the best trade-off in their experiments.
A substantial portion of the reading list is dedicated to more abstract material on theoretical informatics. Rather than proposing specific architectures or engineering solutions for concrete applications, these works are concerned with more fundamental study of the limits of computability, probability and intelligence. Recurring themes are principles for inductive inference such as Occam’s razor, which states a preference for simplicity when choosing between competing explanations (be they theories, hypotheses or models) for some given evidence or data. Another core concept is Kolmogorov complexity* for quantifying the amount of information, or potential for compression, of a given input.
*Note: Kolmogorov complexity of a sequence can be defined as the length of the shortest program that prints it and then halts. While uncomputable in practice, it can be approximated with compression software such as gzip.
A Tutorial Introduction to the Minimum Description Length Principle, 2004
Book Chapter
Length: ~30,000 words
Author: Peter Grünwald
🔗
[A Tutorial Introduction to the Minimum Description Length Principle, 2004]
describes an approach for model selection that mathematically formalizes Occam’s razor, defining a preference for the most simple model among all those that explain the available data. The principle relates learning to data compression, as the ability to exploit regularity for achieving a shortest possible description. This description is defined by codes, and codelength functions are noted as corresponding to probability mass functions. The two-part code version of the Minimum Description Length (MDL) principle measures the simplicity of a model instance as the length of its description (in bits) added to the length of the data description as encoded with it. The refined, one-part code version examines entire families of models based on their goodness-of-fit and complexity.
Kolmogorov Complexity and Algorithmic Randomness
(Chapter 14), 2017
Book Chapter
Length: ~35,000 words
Authors: Alexander Shen, Vladimir A. Uspensky and Nikolay Vereshchagin
🔗
[Kolmogorov Complexity and Algorithmic Randomness] features a final chapter on algorithmic statistics. In this framework, a given sequence of observations is encoded as one binary string. Kolmogorov complexity provides formal means of quantifying its randomness and regularity, as well as the expected and desired properties of a theory or model that can explain it. Such a model should preferably be simple, as indicated by low Kolmogorov complexity. It should also explain as much regularity in the data as possible, making the data “typical” for the model. This property is formally quantified by low randomness deficiency of the data relative to the model. Together, these two properties are also related to the two-part code of the Minimum Description Length principle. The chapter closes by drawing parallels between good models and good compressors, together with the potential of lossy compression to perform effective denoising.
The First Law of Complexodynamics, 2011
Blog Post
Length: ~2,000 words
Author: Scott Aaronson
🔗
[The First Law of Complexodynamics] explores the relationship between entropy and complexity. Whereas the second law of thermodynamics dictates that entropy of closed systems increases over time, their complexity of ‘interestingness’ is noted to first rise and then fall again. Giving the example of coffee and milk mixing in a glass, the highest such ‘complextropy’ is noted to occur midway, when tendrils of milk result from both liquids no longer being cleanly separated but also not yet forming a homogenous blend. Kolmogorov complexity is explored as a way to express both entropy and this ‘complextropy’, with the conjecture that a resource-bounded definition could provide a suitable theoretical framework.
Quantifying the Rise and Fall of Complexity in Closed Systems: The Coffee Automaton, 2014
Paper
Length: ~8,500 words
Authors: Scott Aaronson, Sean M. Carroll and Lauren Ouellette
🔗
[Quantifying the Rise and Fall of Complexity] explores these ideas in further depth. Covering various theoretical notions of complexity, it eventually settles on ‘apparent complexity’ as a way of modeling the separate phenomena of entropy and the ‘interestingness’ of a closed system. Practical experiments inspired by the blending of coffee and milk fill a 2D array with a clean split of binary values and perturb these over multiple time steps to represent random mixing. This array forms an image which is compressed by gzip to approximate Kolmogorov complexity via file size. At each time step, this is done with the image itself to approximate entropy, but also with a coarse-grained, blurred version to estimate its apparent complexity. As envisioned, the increasingly noisy image values yield rising entropy whereas their blurred representation first raises and then decreases the apparent complexity measure as the mix gets more homogeneous.
Machine Super Intelligence, 2008
Dissertation
Length: ~50,000 words
Author: Shane Legg, supervised by Marcus Hutter
🔗
[Machine Super Intelligence, 2008] explores universal artificial intelligence under aspects of algorithmic complexity, probability and information theory. It covers inductive inference from Epicurus principle of multiple explanations, Occam’s razor, Bayes rule and priors to complexity measures and agent-environment models as examined in reinforcement learning. Discussing various definitions and established tests for intelligence, it proposes a formal definition and measure for universal intelligence* as the ability of an agent to achieve specific goals in a wide range of environments. While the proposed measure itself is uncomputable in practice, it enables theoretical conclusions, such as the requirement that powerful agents be proportionally complex, and motivates several practical experiments in which a downscaled version of a hypothetically optimal agent is deployed for reinforcement learning.
*Note: This measure would accordingly score the universal intelligence of specialized, ‘narrow’ machine learning systems that form the bulk of the papers examined in this blog post as comparatively low.
Miscellaneous
Keeping Neural Networks Simple by Minimizing the Description Length of the Weights, 1993
Paper
Length: ~6,000 words
Authors: Geoffrey E. Hinton and Drew van Camp
🔗
[Keeping Neural Networks Simple by Minimizing the Description Length of the Weights] introduced the concept of Variational Inference with neural networks. This approach enables neural network training to approximate the otherwise computationally prohibitive concept of Bayesian inference. The authors propose a regularization technique that represents each weight of a neural network as a Gaussian probability distribution described by a mean and a variance value. Inspired by the Minimum Description Length principle, the cost function used during training penalizes the description length of the weights and the data misfits. The authors argue that this representation allows for a substantial reduction in the description length of the weights. Their Bits-Back Coding argument states that the distribution of each weight can be sampled with random bits at no additional cost, as the random bits can be reconstructed given a fixed learning algorithm, architecture and initial probability distribution for each weight.
Variational Lossy Autencoder, 2017
Paper
Length: ~6,000 words
Authors: Xi Chen, Diederik P. Kingma, Tim Salimans, Yan Duan, Prafulla Dhariwal, John Schulman, Ilya Sutskever and Pieter Abbeel
🔗
[Variational Lossy Autoencoders] provide a way for data compression with control over which aspects of the data should be retained or discarded. In experimental results, this enables 2D image compression that discards local texture while retaining global structure. Autencoders use an inference model to compresses input data to a compact latent code, from which a generative model decodes the original input. This latent code should accordingly represent all information relevant for describing the input. When using sufficiently powerful autoregressive models like RNNs however, decoders had been previously found capable of predicting the output while ignoring the latent code entirely. Here, a theoretical explanation for this phenomenon is provided based on Bits Back Coding. The proposed approach weakens the decoder (e.g. limiting it to reconstruct small receptive fields) such that it depends on the missing information (e.g. global structure) being fully provided by the latent code to which the input is compressed.
GPipe, 2018
Paper
Length: ~5,000 words
Authors: Yanping Huang, Youlong Cheng, Ankur Bapna, Orhan Firat, Mia Xu Chen, Dehao Chen, HyoukJoong Lee, Jiquan Ngiam, Quoc V. Le, Yonghui Wu and Zhifeng Chen
🔗
[GPipe] is* a library for distributed training of neural networks on more than one accelerator (e.g. GPUs). It subdivides the neural network architecture into cells formed by one or more consecutive layers and assigns each cell to a separate accelerator. It furthermore employs pipeline parallelism by also splitting each mini-batch of training samples into several micro-batches that are pipelined through, so that multiple accelerators can work on different micro batches concurrently. The gradients for all micro-batches are aggregated for one synchonous update per mini-batch. Training thus remains consistent regardless of cell count or micro-batch size.
*Note: As for the library itself, GitHub shows that its most recent commit for the final version v0.0.7 occurred in September 2020.
Neural Message Passing for Quantum Chemistry, 2017
Paper
Length: ~6,000 words
Authors: Justin Gilmer, Samuel S. Schoenholz, Patrick F. Riley, Oriol Vinyals and George E. Dahl
🔗
[Neural Message Passing for Quantum Chemistry] explores the application of graph neural networks to predict quantum mechanical properties of organic molecules. The commonalities between several related works that utilize neural networks for graph data are first discussed and abstracted into a new concept of Message Passing Neural Networks. This framework considers undirected graphs composed of edges and nodes, both of which can have features. Each forward pass performs one or more steps of a message passing phase in which the hidden state of each given node is updated based on messages that depend on the hidden states and connecting edges with all adjacent nodes. Next, a readout phase calculates one hidden state for the entire graph using a readout function that is invariant to the order of graph nodes. Using the Gated Graph Neural Network architecture, the authors present experimental results on graph data of molecular structures that achieved state-of-the-art results at the time.
Concluding Thoughts
Whereas this summary was written by hand, the described technology is approaching a point where its language modeling capabilities are hard to distinguish from human writing already now in 2024. Eventually, Large Language Models may indeed become self-explanatory in a literal sense. Prompting e.g. ChatGPT to summarize this material nonetheless still yields explanations that seem very convincing but can also be largely hallucinated and often misleading. Perhaps that will already improve once this article is ingested into the training data?
Although low-quality, generated content was a recurring theme I encountered while researching this list, there are also several independent summaries of the reading list worth sharing here:
With this blog post, the known contents of the reading list are compressed to barely more than one percent of the original word count. This leaves a lot more to be discussed, but hopefully it still has something to offer for the interested reader. My own, subjective review will be saved for another post in the future.