Date:

Circuit Analysis Fundamentals

Circuits: A Living Document

Introduction

In the original narrative of deep learning, each neuron builds progressively more abstract, meaningful features by composing features in the preceding layer. In recent years, there has been some skepticism about this view, but what happens if you take it seriously?

The Circuits Thread

The Circuits thread is a collection of short articles, experiments, and critical commentary around a narrow or unusual research topic, along with a slack channel for real-time discussion and collaboration. It is intended to be an earlier stage than a full Distill paper, allowing for more fluid publishing, feedback, and discussion.

Articles and Comments

The natural unit of publication for investigating circuits seems to be short papers on individual circuits or small families of features. Compared to normal machine learning papers, this is a small and unusual topic for a paper.

Zoom In: An Introduction to Circuits

Does it make sense to treat individual neurons and the connections between them as a serious object of study? This essay proposes three claims which, if true, might justify serious inquiry into them: the existence of meaningful features, the existence of meaningful circuits between features, and the universality of those features and circuits.

An Overview of Early Vision in InceptionV1

An overview of all the neurons in the first five layers of InceptionV1, organized into a taxonomy of "neuron groups." This article sets the stage for future deep dives into particular aspects of early vision.

Curve Detectors

Every vision model we’ve explored in detail contains neurons which detect curves. Curve detectors is the first in a series of three articles exploring this neuron family in detail.

Naturally Occurring Equivariance in Neural Networks

Neural networks naturally learn many transformed copies of the same feature, connected by symmetric weights.

High-Low Frequency Detectors

A family of early-vision neurons reacting to directional transitions from high to low spatial frequency.

Curve Circuits

We reverse-engineer a non-trivial learned algorithm from the weights of a neural network and use its core ideas to craft an artificial neural network from scratch that reimplements it.

Visualizing Weights

We present techniques for visualizing, contextualizing, and understanding neural network weights.

Branch Specialization

When a neural network layer is divided into multiple branches, neurons self-organize into coherent groupings.

Weight Banding

Weights in the final layer of common visual models appear as horizontal bands. We investigate how and why.

Get Involved

The Circuits thread is open to articles exploring individual features, circuits, and their organization within neural networks. Critical commentary and discussion of existing articles is also welcome. The thread is organized through the open #circuits channel on the Distill slack.

About the Thread Format

Part of Distill’s mandate is to experiment with new forms of scientific publishing. We believe that reconciling faster and more continuous approaches to publication with review and discussion is an important open problem in scientific publishing.

Citation Information

If you wish to cite this thread as a whole, citation information can be found below. The author order is all participants in the thread in alphabetical order. Since this is a living document, the citation may add additional authors as it evolves. You can also cite individual articles using the citation information provided at the bottom of the corresponding article.

Updates and Corrections

If you see mistakes or want to suggest changes, please create an issue on GitHub.

Reuse

Diagrams and text are licensed under Creative Commons Attribution CC-BY 4.0 with the source available on GitHub, unless noted otherwise. The figures that have been reused from other sources don’t fall under this license and can be recognized by a note in their caption: "Figure from …".

Citation

For attribution in academic contexts, please cite this work as:

Cammarata, et al., "Thread: Circuits", Distill, 2020.

BibTeX citation:

@article{cammarata2020thread:,
author = {Cammarata, Nick and Carter, Shan and Goh, Gabriel and Olah, Chris and Petrov, Michael and Schubert, Ludwig and Voss, Chelsea and Egan, Ben and Lim, Swee Kiat},
title = {Thread: Circuits},
journal = {Distill},
year = {2020},
note = {https://distill.pub/2020/circuits},
doi = {10.23915/distill.00024}
}

Latest stories

Read More

LEAVE A REPLY

Please enter your comment!
Please enter your name here