Ignite
.. image:: https://travis-ci.org/pytorch/ignite.svg?branch=master
:target: https://travis-ci.org/pytorch/ignite
.. image:: https://codecov.io/gh/pytorch/ignite/branch/master/graph/badge.svg
:target: https://codecov.io/gh/pytorch/ignite
.. image:: https://pepy.tech/badge/pytorch-ignite
:target: https://pepy.tech/project/pytorch-ignite
.. image:: https://img.shields.io/badge/dynamic/json.svg?label=docs&url=https%3A%2F%2Fpypi.org%2Fpypi%2Fpytorch-ignite%2Fjson&query=%24.info.version&colorB=brightgreen&prefix=v
:target: https://pytorch.org/ignite/index.html
.. image:: https://anaconda.org/pytorch/ignite/badges/version.svg
:target: https://anaconda.org/pytorch/ignite
.. image:: https://img.shields.io/badge/dynamic/json.svg?label=PyPI&url=https%3A%2F%2Fpypi.org%2Fpypi%2Fpytorch-ignite%2Fjson&query=%24.info.version&colorB=brightgreen&prefix=v
:target: https://pypi.org/project/pytorch-ignite/
.. image:: https://img.shields.io/badge/Optuna-integrated-blue
:target: https://optuna.org
Ignite is a high-level library to help with training neural networks in PyTorch.
- ignite helps you write compact but full-featured training loops in a few lines of code
- you get a training loop with metrics, early-stopping, model checkpointing and other features without the boilerplate
Below we show a side-by-side comparison of using pure pytorch and using ignite to create a training loop
to train and validate your model with occasional checkpointing:
.. image:: assets/ignite_vs_bare_pytorch.png
:target: https://raw.githubusercontent.com/pytorch/ignite/master/assets/ignite_vs_bare_pytorch.png
As you can see, the code is more concise and readable with ignite. Furthermore, adding additional metrics, or
things like early stopping is a breeze in ignite, but can start to rapidly increase the complexity of
your code when "rolling your own" training loop.
Installation
From pip <https://pypi.org/project/pytorch-ignite/>
_:
.. code:: bash
pip install pytorch-ignite
From conda <https://anaconda.org/pytorch/ignite>
_:
.. code:: bash
conda install ignite -c pytorch
From source:
.. code:: bash
pip install git+https://github.com/pytorch/ignite
Nightly releases
From pip:
.. code:: bash
pip install --pre pytorch-ignite
From conda (this suggests to install pytorch nightly release <https://anaconda.org/pytorch-nightly/pytorch>
_ instead
of stable version as dependency):
.. code:: bash
conda install ignite -c pytorch-nightly
Why Ignite?
Ignite's high level of abstraction assumes less about the type of network (or networks) that you are training, and we require the user to define the closure to be run in the training and validation loop. This level of abstraction allows for a great deal more of flexibility, such as co-training multiple models (i.e. GANs) and computing/tracking multiple losses and metrics in your training loop.
Ignite also allows for multiple handlers to be attached to events, and a finer granularity of events in the engine loop.
Documentation
API documentation and an overview of the library can be found here <https://pytorch.org/ignite/index.html>
_.
Structure
-
ignite: Core of the library, contains an engine for training and evaluating, all of the classic machine learning metrics and a variety of handlers to ease the pain of training and validation of neural networks!
-
ignite.contrib: The Contrib directory contains additional modules contributed by Ignite users. Modules vary from TBPTT engine, various optimisation parameter schedulers, logging handlers and a metrics module containing many regression metrics (
ignite.contrib.metrics.regression <https://github.com/pytorch/ignite/tree/master/ignite/contrib/metrics/regression>
_)!
The code in ignite.contrib is not as fully maintained as the core part of the library. It may change or be removed at any time without notice.
Examples
We provide several examples ported from pytorch/examples <https://github.com/pytorch/examples>
_ using ignite
to display how it helps to write compact and full-featured training loops in a few lines of code:
MNIST example
Basic neural network training on MNIST dataset with/without ignite.contrib
module:
MNIST with ignite.contrib TQDM/Tensorboard/Visdom loggers <https://github.com/pytorch/ignite/tree/master/examples/contrib/mnist>
_MNIST with native TQDM/Tensorboard/Visdom logging <https://github.com/pytorch/ignite/tree/master/examples/mnist>
_
Distributed CIFAR10 example
Training a small variant of ResNet on CIFAR10 in various configurations: 1) single gpu, 2) single node multiple gpus, 3) multiple nodes and multilple gpus.
CIFAR10 <https://github.com/pytorch/ignite/tree/master/examples/contrib/cifar10>
_
Other examples
DCGAN <https://github.com/pytorch/ignite/tree/master/examples/gan>
_Reinforcement Learning <https://github.com/pytorch/ignite/tree/master/examples/reinforcement_learning>
_Fast Neural Style <https://github.com/pytorch/ignite/tree/master/examples/fast_neural_style>
_
Notebooks
Text Classification using Convolutional Neural Networks <https://github.com/pytorch/ignite/blob/master/examples/notebooks/TextCNN.ipynb>
_Variational Auto Encoders <https://github.com/pytorch/ignite/blob/master/examples/notebooks/VAE.ipynb>
_Training Cycle-GAN on Horses to Zebras <https://github.com/pytorch/ignite/blob/master/examples/notebooks/CycleGAN.ipynb>
_Finetuning EfficientNet-B0 on CIFAR100 <https://github.com/pytorch/ignite/blob/master/examples/notebooks/EfficientNet_Cifar100_finetuning.ipynb>
_Convolutional Neural Networks for Classifying Fashion-MNIST Dataset <https://github.com/pytorch/ignite/blob/master/examples/notebooks/FashionMNIST.ipynb>
_Hyperparameters tuning with Ax <https://github.com/pytorch/ignite/blob/master/examples/notebooks/Cifar10_Ax_hyperparam_tuning.ipynb>
_
Reproducible trainings <examples/references>
_
Inspired by torchvision/references <https://github.com/pytorch/vision/tree/master/references>
_, we provide several
reproducible baselines for vision tasks:
ImageNet <examples/references/classification/imagenet>
_Pascal VOC2012 <examples/references/segmentation/pascal_voc2012>
_
Features:
- Distributed training with mixed precision by
nvidia/apex <https://github.com/NVIDIA/apex/>
_ - Experiments tracking with
MLflow <https://mlflow.org/>
_ orPolyaxon <https://polyaxon.com/>
_
Contributing
We appreciate all contributions. If you are planning to contribute back bug-fixes, please do so without any further discussion. If you plan to contribute new features, utility functions or extensions, please first open an issue and discuss the feature with us.
Please see the contribution guidelines <https://github.com/pytorch/ignite/blob/master/CONTRIBUTING.md>
_ for more information.
As always, PRs are welcome :)
They use Ignite
State-of-the-Art Conversational AI with Transfer Learning <https://github.com/huggingface/transfer-learning-conv-ai>
_Tutorial on Transfer Learning in NLP held at NAACL 2019 <https://github.com/huggingface/naacl_transfer_learning_tutorial>
_Implementation of "Attention is All You Need" paper <https://github.com/akurniawan/pytorch-transformer>
_Implementation of DropBlock: A regularization method for convolutional networks in PyTorch <https://github.com/miguelvr/dropblock>
_Deep-Reinforcement-Learning-Hands-On-Second-Edition, published by Packt <https://github.com/PacktPublishing/Deep-Reinforcement-Learning-Hands-On-Second-Edition>
_Kaggle Kuzushiji Recognition: 2nd place solution <https://github.com/lopuhin/kaggle-kuzushiji-2019>
_Unsupervised Data Augmentation experiments in PyTorch <https://github.com/vfdev-5/UDA-pytorch>
_Hyperparameters tuning with Optuna <https://github.com/pfnet/optuna/blob/master/examples/pytorch_ignite_simple.py>
_
See other projects at "Used by" <https://github.com/pytorch/ignite/network/dependents?package_id=UGFja2FnZS02NzI5ODEwNA%3D%3D>
_
If your project implements a paper, represents other use-cases not covered in our official tutorials,
Kaggle competition's code or just your code presents interesting results and uses Ignite. We would like to add your project
in this list, so please send a PR with brief description of the project.
User feedback
We have created a form for "user feedback" <https://github.com/pytorch/ignite/issues/new/choose>
_.
We appreciate any type of feedback and this is how we would like to see our community:
-
If you like the project and want to say thanks, this the right place.
-
If you do not like something, please, share it with us and we can see how to improve it.
Thank you !