Project Name | Stars | Downloads | Repos Using This | Packages Using This | Most Recent Commit | Total Releases | Latest Release | Open Issues | License | Language |
---|---|---|---|---|---|---|---|---|---|---|
Transformers | 112,383 | 64 | 1,869 | 19 hours ago | 114 | July 18, 2023 | 825 | apache-2.0 | Python | |
🤗 Transformers: State-of-the-art Machine Learning for Pytorch, TensorFlow, and JAX. | ||||||||||
Stable Diffusion Webui | 103,367 | a day ago | 2 | January 17, 2022 | 1,516 | agpl-3.0 | Python | |||
Stable Diffusion web UI | ||||||||||
Pytorch | 71,061 | 3,341 | 6,728 | 19 hours ago | 37 | May 08, 2023 | 12,782 | other | Python | |
Tensors and Dynamic neural networks in Python with strong GPU acceleration | ||||||||||
Keras | 59,408 | 578 | 19 hours ago | 80 | June 27, 2023 | 90 | apache-2.0 | Python | ||
Deep Learning for humans | ||||||||||
Real Time Voice Cloning | 46,865 | a day ago | 167 | other | Python | |||||
Clone a voice in 5 seconds to generate arbitrary speech in real-time | ||||||||||
Yolov5 | 41,813 | 2 days ago | 8 | September 21, 2021 | 227 | agpl-3.0 | Python | |||
YOLOv5 🚀 in PyTorch > ONNX > CoreML > TFLite | ||||||||||
Annotated_deep_learning_paper_implementations | 36,223 | 1 | 5 days ago | 78 | September 24, 2022 | 27 | mit | Jupyter Notebook | ||
🧑🏫 60 Implementations/tutorials of deep learning papers with side-by-side notes 📝; including transformers (original, xl, switch, feedback, vit, ...), optimizers (adam, adabelief, sophia, ...), gans(cyclegan, stylegan2, ...), 🎮 reinforcement learning (ppo, dqn), capsnet, distillation, ... 🧠 | ||||||||||
Made With Ml | 34,182 | 4 days ago | 5 | May 15, 2019 | 2 | mit | Jupyter Notebook | |||
Learn how to design, develop, deploy and iterate on production-grade ML applications. | ||||||||||
Gfpgan | 32,185 | 9 | 13 days ago | 11 | September 20, 2022 | 271 | other | Python | ||
GFPGAN aims at developing Practical Algorithms for Real-world Face Restoration. | ||||||||||
Mockingbird | 30,784 | 21 days ago | 2 | February 28, 2022 | 446 | other | Python | |||
🚀AI拟声: 5秒内克隆您的声音并生成任意语音内容 Clone a voice in 5 seconds to generate arbitrary speech in real-time |
Split-Attention Network, A New ResNet Variant. It significantly boosts the performance of downstream models such as Mask R-CNN, Cascade R-CNN and DeepLabV3.
# using github url
pip install git+https://github.com/zhanghang1989/ResNeSt
# using pypi
pip install resnest --pre
crop size | PyTorch | Gluon | |
---|---|---|---|
ResNeSt-50 | 224 | 81.03 | 81.04 |
ResNeSt-101 | 256 | 82.83 | 82.81 |
ResNeSt-200 | 320 | 83.84 | 83.88 |
ResNeSt-269 | 416 | 84.54 | 84.53 |
3rd party implementations are available: Tensorflow, Caffe, JAX.
Extra ablation study models are available in link
import torch
# get list of models
torch.hub.list('zhanghang1989/ResNeSt', force_reload=True)
# load pretrained models, using ResNeSt-50 as an example
net = torch.hub.load('zhanghang1989/ResNeSt', 'resnest50', pretrained=True)
# using ResNeSt-50 as an example
from resnest.torch import resnest50
net = resnest50(pretrained=True)
# using ResNeSt-50 as an example
from resnest.gluon import resnest50
net = resnest50(pretrained=True)
We provide a wrapper for training Detectron2 models with ResNeSt backbone at d2. Training configs and pretrained models are released. See details in d2.
The ResNeSt backbone has been adopted by MMDetection.
Note: the inference speed reported in the paper are tested using Gluon implementation with RecordIO data.
Here we use raw image data format for simplicity, please follow GluonCV tutorial if you would like to use RecordIO format.
cd scripts/dataset/
# assuming you have downloaded the dataset in the current folder
python prepare_imagenet.py --download-dir ./
# use resnest50 as an example
cd scripts/torch/
python verify.py --model resnest50 --crop-size 224
# use resnest50 as an example
cd scripts/gluon/
python verify.py --model resnest50 --crop-size 224
For object detection and instance segmentation models, please visit our detectron2-ResNeSt fork.
ResNeSt: Split-Attention Networks [arXiv]
Hang Zhang, Chongruo Wu, Zhongyue Zhang, Yi Zhu, Zhi Zhang, Haibin Lin, Yue Sun, Tong He, Jonas Muller, R. Manmatha, Mu Li and Alex Smola
@article{zhang2020resnest,
title={ResNeSt: Split-Attention Networks},
author={Zhang, Hang and Wu, Chongruo and Zhang, Zhongyue and Zhu, Yi and Zhang, Zhi and Lin, Haibin and Sun, Yue and He, Tong and Muller, Jonas and Manmatha, R. and Li, Mu and Smola, Alexander},
journal={arXiv preprint arXiv:2004.08955},
year={2020}
}