#
model-compression
Here are 103 public repositories matching this topic...
An Automatic Model Compression (AutoMC) framework for developing smaller and faster AI applications.
-
Updated
Nov 16, 2020 - Python
Pretrained language model and its related optimization techniques developed by Huawei Noah's Ark Lab.
-
Updated
Oct 25, 2020 - Python
NLP DNN Toolkit - Building Your NLP DNN Models Like Playing Lego
natural-language-processing
deep-learning
text-classification
dnn
pytorch
artificial-intelligence
question-answering
knowledge-distillation
sequence-labeling
text-matching
qna
model-compression
-
Updated
Oct 26, 2020 - Python
[CVPR2020] GhostNet: More Features from Cheap Operations
tensorflow
pytorch
imagenet
convolutional-neural-networks
cvpr
model-compression
mobilenet
efficient-inference
fbnet
mobilenetv3
cvpr2020
-
Updated
Dec 17, 2020 - Python
model compression and deploy. compression: 1、quantization: quantization-aware-training, 16/8/4/2-bit(dorefa/Quantization and Training of Neural Networks for Efficient Integer-Arithmetic-Only Inference)、ternary/binary value(twn/bnn/xnor-net); post-training-quantization, 8-bit(tensorrt); 2、 pruning: normal、regular and group convolutional channel pruning; 3、 group convolution structure; 4、batch-normalization folding for quantization. deploy: tensorrt, fp32/fp16/int8(ptq-calibration)、op-adapt(upsample)、dynamic_shape
pytorch
pruning
convolutional-networks
quantization
xnor-net
tensorrt
model-compression
bnn
neuromorphic-computing
group-convolution
onnx
network-in-network
tensorrt-int8-python
dorefa
twn
network-slimming
integer-arithmetic-only
quan-batch-normalization-folding
quantization-aware-training
post-training-quantization
-
Updated
Dec 17, 2020 - Python
Channel Pruning for Accelerating Very Deep Neural Networks (ICCV'17)
deep-neural-networks
acceleration
image-classification
image-recognition
object-detection
model-compression
channel-pruning
-
Updated
Nov 26, 2019 - Python
A toolkit to optimize ML models for deployment for Keras and TensorFlow, including quantization and pruning.
machine-learning
sparsity
compression
deep-learning
tensorflow
optimization
keras
ml
pruning
quantization
model-compression
quantized-training
quantized-neural-networks
quantized-networks
-
Updated
Dec 17, 2020 - Python
A curated list of neural network pruning resources.
-
Updated
Dec 4, 2020
A PyTorch implementation for exploring deep and shallow knowledge distillation (KD) experiments with flexibility
deep-neural-networks
computer-vision
pytorch
knowledge-distillation
cifar10
dark-knowledge
model-compression
-
Updated
Dec 10, 2020 - Python
Awesome Knowledge-Distillation. 分类整理的知识蒸馏paper(2014-2020)。
-
Updated
Dec 16, 2020
A list of high-quality (newest) AutoML works and lightweight models including 1.) Neural Architecture Search, 2.) Lightweight Structures, 3.) Model Compression, Quantization and Acceleration, 4.) Hyperparameter Optimization, 5.) Automated Feature Engineering.
tensorflow
pytorch
hyperparameter-optimization
awesome-list
quantization
nas
automl
model-compression
neural-architecture-search
meta-learning
architecture-search
quantized-training
model-acceleration
automated-feature-engineering
quantized-neural-network
-
Updated
Oct 8, 2020
Lightweight and Scalable framework that combines mainstream algorithms of Click-Through-Rate prediction based computational DAG, philosophy of Parameter Server and Ring-AllReduce collective communication.
distributed-systems
machine-learning
deep-learning
factorization-machines
computational-graphs
parameter-server
model-compression
-
Updated
Jun 17, 2019 - C++
PaddleSlim is an open-source library for deep model compression and architecture search.
pruning
quantization
nas
knowledge-distillation
evolution-strategy
model-compression
neural-architecture-search
hyperparameter-search
autodl
-
Updated
Dec 17, 2020 - Python
[CVPR2020] GhostNet: More Features from Cheap Operations
-
Updated
Aug 8, 2020 - Python
knowledge distillation papers
-
Updated
Jun 9, 2020
Pytorch implementation of various Knowledge Distillation (KD) methods.
knowledge-distillation
teacher-student
knowledge-transfer
model-compression
distillation
kd
kd-methods
-
Updated
Nov 13, 2020 - Python
Data Efficient Model Compression
-
Updated
Sep 28, 2020 - Python
Filter Pruning via Geometric Median for Deep Convolutional Neural Networks Acceleration (CVPR 2019 Oral)
-
Updated
Sep 10, 2019 - Python
Papers for deep neural network compression and acceleration
-
Updated
May 24, 2018
[ECCV 2018] AMC: AutoML for Model Compression and Acceleration on Mobile Devices
-
Updated
Nov 25, 2020 - Python
Soft Filter Pruning for Accelerating Deep Convolutional Neural Networks
-
Updated
Oct 2, 2019 - Python
Infrastructures™ for Machine Learning Training/Inference in Production.
kubernetes
machine-learning
apache-spark
deep-learning
artificial-intelligence
awesome-list
pruning
quantization
knowledge-distillation
deep-learning-framework
model-compression
apache-arrow
federated-learning
machine-learning-systems
apache-mesos
-
Updated
May 24, 2019
Deep Face Model Compression
-
Updated
Aug 21, 2018 - MATLAB
Java interface for fastText
-
Updated
Nov 22, 2020 - Java
-
Updated
Nov 27, 2020 - Python
PyTorch Implementation of Weights Pruning
-
Updated
Jan 24, 2018 - Python
Model Compression CLI Tool for Keras.
-
Updated
May 9, 2019 - Python
[ECCV 2018] AMC: AutoML for Model Compression and Acceleration on Mobile Devices
-
Updated
Oct 25, 2019 - Python
Awesome machine learning model compression research papers, tools, and learning material.
-
Updated
May 10, 2020
Improve this page
Add a description, image, and links to the model-compression topic page so that developers can more easily learn about it.
Add this topic to your repo
To associate your repository with the model-compression topic, visit your repo's landing page and select "manage topics."
What would you like to be added: As title
Why is this needed: All pruning schedule except AGPPruner only support level, L1, L2. While there are FPGM, APoZ, MeanActivation and Taylor, it would be much better if we can choose any pruner with any pruning schedule.
**Without this feature, how does current nni