Awesome
Hyperbolic Learning Rate Scheduler
This repository contains the implementation and experimental code for the paper "HyperbolicLR: Epoch Insensitive Learning Rate Scheduler". It includes the novel learning rate schedulers, HyperbolicLR and ExpHyperbolicLR, designed to address the learning curve decoupling problem in deep learning.
Overview
HyperbolicLR and ExpHyperbolicLR are learning rate schedulers that maintain consistent initial learning rate changes, regardless of the total number of epochs. This property helps mitigate the learning curve decoupling problem observed in conventional schedulers, potentially leading to more robust and efficient training of deep neural networks.
Installation
To use the HyperbolicLR and ExpHyperbolicLR schedulers in your project:
-
Clone this repository:
git clone https://github.com/Axect/HyperbolicLR
-
Copy
hyperbolic_lr.py
to your project directory or add this repository to your Python path.
Usage
Here's a basic example of how to use HyperbolicLR in your PyTorch project:
from hyperbolic_lr import HyperbolicLR # or ExpHyperbolicLR
import torch
# Define your model and optimizer
model = YourModel()
optimizer = torch.optim.AdamW(model.parameters(), lr=1e-3)
num_epochs = 100
# Create the HyperbolicLR scheduler
scheduler = HyperbolicLR(optimizer, upper_bound=250, max_iter=num_epochs, infimum_lr=1e-5)
# Training loop
for epoch in range(num_epochs):
train(model, optimizer)
scheduler.step()
For more detailed usage and examples, please refer to the paper and experimental code in the paper/
directory.
Citing
If you use HyperbolicLR or ExpHyperbolicLR in your research, please cite our paper:
@misc{kim2024hyperboliclr,
title={HyperbolicLR: Epoch insensitive learning rate scheduler},
author={Tae-Geun Kim},
year={2024},
eprint={2407.15200},
archivePrefix={arXiv},
primaryClass={cs.LG},
url={https://arxiv.org/abs/2407.15200},
}
License
This project is licensed under the MIT License.
Contributing
We welcome contributions to improve HyperbolicLR and ExpHyperbolicLR. Please feel free to submit issues or pull requests.
Contact
For any questions or discussions regarding this project, please open an issue in this repository.