#
model-compression
Here are 170 public repositories matching this topic...
Awesome Knowledge Distillation
deep-learning
knowledge-distillation
teacher-student
knowledge-transfer
co-training
model-compression
distillation
kd
knowldge-distillation
distillation-model
model-distillation
-
Updated
Mar 12, 2022
An Automatic Model Compression (AutoMC) framework for developing smaller and faster AI applications.
-
Updated
Jan 15, 2021 - Python
Efficient AI Backbones including GhostNet, TNT and MLP, developed by Huawei Noah's Ark Lab.
tensorflow
pytorch
transformer
imagenet
convolutional-neural-networks
pretrained-models
model-compression
efficient-inference
ghostnet
vision-transformer
-
Updated
Jul 8, 2022 - Python
Pretrained language model and its related optimization techniques developed by Huawei Noah's Ark Lab.
-
Updated
Jul 5, 2022 - Python
Awesome Knowledge-Distillation. 分类整理的知识蒸馏paper(2014-2021)。
-
Updated
May 23, 2022
micronet, a model compression and deploy lib. compression: 1、quantization: quantization-aware-training(QAT), High-Bit(>2b)(DoReFa/Quantization and Training of Neural Networks for Efficient Integer-Arithmetic-Only Inference)、Low-Bit(≤2b)/Ternary and Binary(TWN/BNN/XNOR-Net); post-training-quantization(PTQ), 8-bit(tensorrt); 2、 pruning: normal、regular and group convolutional channel pruning; 3、 group convolution structure; 4、batch-normalization fuse for quantization. deploy: tensorrt, fp32/fp16/int8(ptq-calibration)、op-adapt(upsample)、dynamic_shape
pytorch
pruning
convolutional-networks
quantization
xnor-net
tensorrt
model-compression
bnn
neuromorphic-computing
group-convolution
onnx
network-in-network
tensorrt-int8-python
dorefa
twn
network-slimming
integer-arithmetic-only
quantization-aware-training
post-training-quantization
batch-normalization-fuse
-
Updated
Oct 6, 2021 - Python
A curated list of neural network pruning resources.
-
Updated
Jun 18, 2022
NLP DNN Toolkit - Building Your NLP DNN Models Like Playing Lego
natural-language-processing
deep-learning
text-classification
dnn
pytorch
artificial-intelligence
question-answering
knowledge-distillation
sequence-labeling
text-matching
qna
model-compression
-
Updated
Jun 21, 2022 - Python
A PyTorch implementation for exploring deep and shallow knowledge distillation (KD) experiments with flexibility
deep-neural-networks
computer-vision
pytorch
knowledge-distillation
cifar10
dark-knowledge
model-compression
-
Updated
Jun 21, 2022 - Python
A toolkit to optimize ML models for deployment for Keras and TensorFlow, including quantization and pruning.
machine-learning
sparsity
compression
deep-learning
tensorflow
optimization
keras
ml
pruning
quantization
model-compression
quantized-training
quantized-neural-networks
quantized-networks
-
Updated
Jul 7, 2022 - Python
Channel Pruning for Accelerating Very Deep Neural Networks (ICCV'17)
deep-neural-networks
acceleration
image-classification
image-recognition
object-detection
model-compression
channel-pruning
-
Updated
Feb 28, 2022 - Python
PaddleSlim is an open-source library for deep model compression and architecture search.
pruning
quantization
nas
knowledge-distillation
evolution-strategy
model-compression
neural-architecture-search
hyperparameter-search
autodl
-
Updated
Jul 8, 2022 - Python
Pytorch implementation of various Knowledge Distillation (KD) methods.
knowledge-distillation
teacher-student
knowledge-transfer
model-compression
distillation
kd
kd-methods
-
Updated
Nov 25, 2021 - Python
A list of high-quality (newest) AutoML works and lightweight models including 1.) Neural Architecture Search, 2.) Lightweight Structures, 3.) Model Compression, Quantization and Acceleration, 4.) Hyperparameter Optimization, 5.) Automated Feature Engineering.
tensorflow
pytorch
hyperparameter-optimization
awesome-list
quantization
nas
automl
model-compression
neural-architecture-search
meta-learning
architecture-search
quantized-training
model-acceleration
automated-feature-engineering
quantized-neural-network
-
Updated
Jun 19, 2021
Collection of recent methods on (deep) neural network compression and acceleration.
-
Updated
Mar 31, 2022
Lightweight and Scalable framework that combines mainstream algorithms of Click-Through-Rate prediction based computational DAG, philosophy of Parameter Server and Ring-AllReduce collective communication.
distributed-systems
machine-learning
deep-learning
factorization-machines
computational-graphs
parameter-server
model-compression
-
Updated
Jun 17, 2019 - C++
Structural Pruning for Model Acceleration
-
Updated
Jul 7, 2022 - Python
knowledge distillation papers
-
Updated
Jul 1, 2022
[CVPR2020] GhostNet: More Features from Cheap Operations
-
Updated
Aug 8, 2020 - Python
Filter Pruning via Geometric Median for Deep Convolutional Neural Networks Acceleration (CVPR 2019 Oral)
-
Updated
Jun 17, 2021 - Python
avishreekh
commented
May 7, 2021
We also need to benchmark the Lottery-tickets Pruning algorithm and the Quantization algorithms. The models used for this would be the student networks discussed in #105 (ResNet18, MobileNet v2, Quantization v2).
Pruning (benchmark upto 40, 50 and 60 % pruned weights)
- Lottery Tickets
Quantization
- Static
- QAT
Open
Benchmarking KD
Open
Update README.rst
10
Archai accelerates Neural Architecture Search (NAS) through fast, reproducible and modular research.
python
machine-learning
deep-learning
pytorch
darts
nas
automated-machine-learning
model-compression
neural-architecture-search
petridish
-
Updated
Jul 5, 2022 - Jupyter Notebook
[ECCV 2018] AMC: AutoML for Model Compression and Acceleration on Mobile Devices
-
Updated
Feb 20, 2021 - Python
Papers for deep neural network compression and acceleration
-
Updated
Jun 21, 2021
Soft Filter Pruning for Accelerating Deep Convolutional Neural Networks
-
Updated
Oct 2, 2019 - Python
TinyNeuralNetwork is an efficient and easy-to-use deep learning model compression framework.
-
Updated
Jul 8, 2022 - Python
Infrastructures™ for Machine Learning Training/Inference in Production.
kubernetes
machine-learning
apache-spark
deep-learning
artificial-intelligence
awesome-list
pruning
quantization
knowledge-distillation
deep-learning-framework
model-compression
apache-arrow
federated-learning
machine-learning-systems
apache-mesos
-
Updated
May 24, 2019
-
Updated
Jan 10, 2021 - Python
Improve this page
Add a description, image, and links to the model-compression topic page so that developers can more easily learn about it.
Add this topic to your repo
To associate your repository with the model-compression topic, visit your repo's landing page and select "manage topics."
Bug with GPU Model
Currently, while using pruning methods like
TaylorFOWeightPruner, If I use a model on GPU for getting the metrics (as calculated for getting masks), it fails on line while creating masks. The reason why it fails i