Implementing neural nets


The internet is full of guides to training neural nets. Here are some selected highlights.

Michael Nielson has a free online textbook with code examples in python. Christopher Olah’s visual explanations make many things clear.

Andrej’s popular unromantic messy guide to training neural nets in practice has a lot of tips that people tend to rediscover the hard way if they do not get them from him. (I did)

It is allegedly easy to get started with training neural nets. Numerous libraries and frameworks take pride in displaying 30-line miracle snippets that solve your data problems, giving the (false) impression that this stuff is plug and play. … Unfortunately, neural nets are nothing like that. They are not “off-the-shelf” technology the second you deviate slightly from training an ImageNet classifier.

NN Software

I have used

I could use any of the other autodiff systems, such as…


See edge ml for a discussion of compiled NNs.

Tracking experiments

See experiment tracking in ML.

Configuring experiments

See configuring experiments; in practice I use hydra for everything.

pre-computed/trained models

Managing axes

A lot of the time managing deep learning is remembering which axis is which. Practically, I have found Einstein convention to solve all my needs.

However, there are alternatives. Alexander Rush argues for NamedTensor. Implementations:

No comments yet. Why not leave one?

GitHub-flavored Markdown & a sane subset of HTML is supported.