This repository is mainly dedicated to t-Adam, i.e. the t-momentum based Adam optimizer, and will no longer be updated.
Please visit the following official t-momentum (https://ieeexplore.ieee.org/document/9296551) repository: here. The t-momentum repository will be updated with the latest advances concerning the t-momentum algorithm in general.
TAdam (http://arxiv.org/abs/2003.00179) and RoAdam optimizers for neural networks.
Implemented using the pytorch library.
- install by pip
git clone https://github.com/Mahoumaru/torchoptim.git
cd torchoptim
pip install -e .
- replace optimizer!
from torchoptim.optimizers.TAdam import TAdam
optimizer = TAdam(net.parameters())
- adjust hyperparameters (same as Adam: lr, betas, eps, weight_decay, amsgrad. - N.B. RoAdam has a third beta. -)
Contains the codes used for the results in the paper (http://arxiv.org/abs/2003.00179)
In order to use the reinforcement learning code, you need to install the rlpyt library, developped by the Berkley A.I. research team: https://github.com/astooke/rlpyt.git Follow the instructions on the link to install it.
Note that a modified version of PPO and A2C is included under the reinforcement learning demo folder. The only difference with the BAIR implementation is the fact that this modified version does not use the gradient norm clipping scheme (More details in the paper).
Arxiv Preprint
@article{ilboudo2020tadam,
title={TAdam: A Robust Stochastic Gradient Optimizer},
author={Ilboudo, Wendyam Eric Lionel and Kobayashi, Taisuke and Sugimoto, Kenji},
journal={arXiv preprint arXiv:2003.00179},
year={2020}
}
t-momentum Journal Paper
@ARTICLE{9296551,
author={W. E. L. {Ilboudo} and T. {Kobayashi} and K. {Sugimoto}},
journal={IEEE Transactions on Neural Networks and Learning Systems},
title={Robust Stochastic Gradient Descent With Student-t Distribution Based First-Order Momentum},
year={2020},
volume={},
number={},
pages={1-14},
doi={10.1109/TNNLS.2020.3041755}}