Awesome
CoreNet: A library for training deep neural networks
CoreNet is a deep neural network toolkit that allows researchers and engineers to train standard and novel small and large-scale models for variety of tasks, including foundation models (e.g., CLIP and LLM), object classification, object detection, and semantic segmentation.
Table of contents
- What's new?
- Research efforts at Apple using CoreNet
- Installation
- Directory Structure
- Maintainers
- Contributing to CoreNet
- License
- Relationship with CVNets
- Citation
What's new?
- October 2024: Version 0.1.1 of the CoreNet library includes
Research efforts at Apple using CoreNet
Below is the list of publications from Apple that uses CoreNet. Also, training and evaluation recipes, as well as links to pre-trained models, can be found inside the projects folder. Please refer to it for further details.
- KV Prediction for Improved Time to First Token
- OpenELM: An Efficient Language Model Family with Open Training and Inference Framework
- CatLIP: CLIP-level Visual Recognition Accuracy with 2.7x Faster Pre-training on Web-scale Image-Text Data
- Reinforce Data, Multiply Impact: Improved Model Accuracy and Robustness with Dataset Reinforcement
- CLIP meets Model Zoo Experts: Pseudo-Supervision for Visual Enhancement
- FastVit: A Fast Hybrid Vision Transformer using Structural Reparameterization
- Bytes Are All You Need: Transformers Operating Directly on File Bytes
- MobileOne: An Improved One millisecond Mobile Backbone
- RangeAugment: Efficient Online Augmentation with Range Learning
- Separable Self-attention for Mobile Vision Transformers (MobileViTv2)
- CVNets: High performance library for Computer Vision, ACM MM'22
- MobileViT: Light-weight, General-purpose, and Mobile-friendly Vision Transformer, ICLR'22
Installation
You will need Git LFS (instructions below) to run tests and Jupyter notebooks (instructions) in this repository, and to contribute to it so we recommend that you install and activate it first.
On Linux we recommend to use Python 3.10+ and PyTorch (version >= v2.1.0), on macOS system Python 3.9+ should be sufficient.
Note that the optional dependencies listed below are required if you'd like to make contributions and/or run tests.
For Linux (substitute apt
for your package manager):
sudo apt install git-lfs
git clone git@github.com:apple/corenet.git
cd corenet
git lfs install
git lfs pull
# The following venv command is optional, but recommended. Alternatively, you can create and activate a conda environment.
python3 -m venv venv && source venv/bin/activate
python3 -m pip install --editable .
To install optional dependencies for audio and video processing:
sudo apt install libsox-dev ffmpeg
For macOS, assuming you use Homebrew:
brew install git-lfs
git clone git@github.com:apple/corenet.git
cd corenet
cd \$(pwd -P) # See the note below.
git lfs install
git lfs pull
# The following venv command is optional, but recommended. Alternatively, you can create and activate a conda environment.
python3 -m venv venv && source venv/bin/activate
python3 -m pip install --editable .
To install optional dependencies for audio and video processing:
brew install sox ffmpeg
Note that on macOS the file system is case insensitive, and case sensitivity
can cause issues with Git. You should access the repository on disk as if the
path were case sensitive, i.e. with the same capitalization as you see when you
list the directories ls
. You can switch to such a path with the cd $(pwd -P)
command.
Directory Structure
This section provides quick access and a brief description for important CoreNet directories.
<table> <thead> <tr> <th> Description </th> <th> Quick Access </th> </tr> </thead> <tbody> <!-- Row boilerplate (copy-paste the following commented snippet for adding a new row to the table.) <tr> <td> <h3> title </h3> description </td> <td> <pre> folders </pre> </td> </tr> --> <tr> <td> <h3> Getting Started </h3> Working with the examples is an easy way to get started with CoreNet. </td> <td> <pre> └── tutorials ├── <a href="tutorials/train_a_new_model_on_a_new_dataset_from_scratch.ipynb">train_a_new_model_on_a_new_dataset_from_scratch.ipynb</a> ├── <a href="tutorials/guide_slurm_and_multi_node_training.md">guide_slurm_and_multi_node_training.md</a> ├── <a href="tutorials/clip.ipynb">clip.ipynb</a> ├── <a href="tutorials/semantic_segmentation.ipynb">semantic_segmentation.ipynb</a> └── <a href="tutorials/object_detection.ipynb">object_detection.ipynb</a> </pre> </td> </tr> <tr> <td> <h3> Training Recipes </h3> CoreNet provides reproducible training recipes, in addition to the pretrained model weights and checkpoints for the publications that are listed in <code>projects/</code> directory.Publication project directories generally contain the following contents:
README.md
provides documentation, links to the pretrained weights, and citations.<task_name>/<model_name>.yaml
provides configuration for reproducing the trainings and evaluations.
Each model class is decorated by a
@MODEL_REGISTRY.register(name="<model_name>", type="<task_name>")
decorator.
To use a model class in CoreNet training or evaluation,
assign models.<task_name>.name = <model_name>
in the YAML configuration.
Maintainers
This code is developed and maintained by <a href="https://mchorton.com" target="_blank">Maxwell Horton</a>, <a href="https://www.mohammad.pro" target="_blank">Mohammad Sekhavat</a> Yanzi Jin, and <a href="https://huggingface.co/depthwise" target="_blank">Dmitry Belenko</a>.
Previous Maintainers
- <a href="https://sacmehta.github.io" target="_blank">Sachin Mehta</a>
- <a href="https://farzadab.github.io" target="_blank">Farzad Abdolhosseini</a>
Contributing to CoreNet
We welcome PRs from the community! You can find information about contributing to CoreNet in our contributing document.
Please remember to follow our Code of Conduct.
License
For license details, see LICENSE.
Relationship with CVNets
CoreNet evolved from CVNets, to encompass a broader range of applications beyond computer vision. Its expansion facilitated the training of foundational models, including LLMs.
Citation
If you find our work useful, please cite the following paper:
@inproceedings{mehta2022cvnets,
author = {Mehta, Sachin and Abdolhosseini, Farzad and Rastegari, Mohammad},
title = {CVNets: High Performance Library for Computer Vision},
year = {2022},
booktitle = {Proceedings of the 30th ACM International Conference on Multimedia},
series = {MM '22}
}