This project features an autograd engine that follow patterns from Andrej Karpathy's micrograd. It has two abstractions:
- A minimal scalar-valued class that provides a clear understanding of basic neural network operations.(Start here!)
- An optimized tensor class that demonstrates improved performance techniques.
Both classes implement backpropagation through dynamically constructed computational graphs (DAGs). The project also includes neural network implementations built on top of both classes.
- Install with pip install -e . The only runtime dependency is numpy.
