Task Aware Distillation

Less is More: Task-aware Layer-wise Distillation for Language Model Compression (ICML2023)
Alternatives To Task Aware Distillation
Project NameStarsDownloadsRepos Using ThisPackages Using ThisMost Recent CommitTotal ReleasesLatest ReleaseOpen IssuesLicenseLanguage
Pretrained Language Model2,912
4 months ago108Python
Pretrained language model and its related optimization techniques developed by Huawei Noah's Ark Lab.
Ld Net145
4 years ago3apache-2.0Python
Efficient Contextualized Representation: Language Model Pruning for Sequence Labeling
Causal Distill12
2 years ago1mitPython
The Codebase for Causal Distillation for Language Models
Task Aware Distillation10
9 months ago1Python
Less is More: Task-aware Layer-wise Distillation for Language Model Compression (ICML2023)
Lm Vocab Trimmer9
7 months agomitPython
Vocabulary Trimming (VT) is a model compression technique, which reduces a multilingual LM vocabulary to a target language by deleting irrelevant tokens from its vocabulary. This repository contains a python-library vocabtrimmer, that remove irrelevant tokens from a multilingual LM vocabulary for the target language.
Alternatives To Task Aware Distillation
Select To Compare


Alternative Project Comparisons
Popular Model Compression Projects
Popular Language Model Projects
Popular Machine Learning Categories
Related Searches

Get A Weekly Email With Trending Projects For These Categories
No Spam. Unsubscribe easily at any time.
Python
Language Model
Model Compression