Stochastic Optimization

Gradient Checkpointing

Introduced by Chen et al. in Training Deep Nets with Sublinear Memory Cost

Gradient Checkpointing is a method used for reducing the memory footprint when training deep neural networks, at the cost of having a small increase in computation time.

Source: Training Deep Nets with Sublinear Memory Cost

Papers


Paper Code Results Date Stars

Components


Component Type
🤖 No Components Found You can add them if they exist; e.g. Mask R-CNN uses RoIAlign

Categories