Gpt Neox

An implementation of model parallel autoregressive transformers on GPUs, based on the DeepSpeed library.
Alternatives To Gpt Neox
Project NameStarsDownloadsRepos Using ThisPackages Using ThisMost Recent CommitTotal ReleasesLatest ReleaseOpen IssuesLicenseLanguage
Gpt Neox6,366
3 months ago81apache-2.0Python
An implementation of model parallel autoregressive transformers on GPUs, based on the DeepSpeed library.
Pranet344
9 months ago1Python
PraNet: Parallel Reverse Attention Network for Polyp Segmentation, MICCAI 2020 (Oral). Code using Jittor Framework is available.
Prime80
9 months ago1otherPython
A simple module consistently outperforms self-attention and Transformer model on main NMT datasets with SoTA performance.
Alternatives To Gpt Neox
Select To Compare


Alternative Project Comparisons
Popular Parallel Projects
Popular Attention Projects
Popular Control Flow Categories
Related Searches

Get A Weekly Email With Trending Projects For These Categories
No Spam. Unsubscribe easily at any time.
Python
Parallel
Attention
Language Model