#
model-compression
Here are 94 public repositories matching this topic...
An Automatic Model Compression (AutoMC) framework for developing smaller and faster AI applications.
-
Updated
May 28, 2019 - Python
NLP DNN Toolkit - Building Your NLP DNN Models Like Playing Lego
natural-language-processing
deep-learning
text-classification
dnn
pytorch
artificial-intelligence
question-answering
knowledge-distillation
sequence-labeling
text-matching
qna
model-compression
-
Updated
Mar 27, 2020 - Python
Pretrained language model and its related optimization techniques developed by Huawei Noah's Ark Lab.
-
Updated
Sep 6, 2020 - Python
[CVPR2020] Surpassing MobileNetV3: "GhostNet: More Features from Cheap Operations"
tensorflow
imagenet
convolutional-neural-networks
cvpr
model-compression
mobilenet
efficient-inference
fbnet
mobilenetv3
cvpr2020
-
Updated
Jun 22, 2020 - Python
Channel Pruning for Accelerating Very Deep Neural Networks (ICCV'17)
deep-neural-networks
acceleration
image-classification
image-recognition
object-detection
model-compression
channel-pruning
-
Updated
Nov 26, 2019 - Python
A toolkit to optimize ML models for deployment for Keras and TensorFlow, including quantization and pruning.
machine-learning
sparsity
compression
deep-learning
tensorflow
optimization
keras
ml
pruning
quantization
model-compression
quantized-training
quantized-neural-networks
quantized-networks
-
Updated
Sep 11, 2020 - Python
model compression based on pytorch (1、quantization: 16/8/4/2 bits(dorefa/Quantization and Training of Neural Networks for Efficient Integer-Arithmetic-Only Inference)、ternary/binary value(twn/bnn/xnor-net);2、 pruning: normal、regular and group convolutional channel pruning;3、 group convolution structure;4、batch-normalization folding for quantization)
pytorch
pruning
convolutional-networks
quantization
xnor-net
model-compression
bnn
neuromorphic-computing
group-convolution
network-in-network
dorefa
twn
network-slimming
integer-arithmetic-only
quan-batch-normalization-folding
quantization-aware-training
-
Updated
Jul 27, 2020 - Python
A PyTorch implementation for exploring deep and shallow knowledge distillation (KD) experiments with flexibility
deep-neural-networks
computer-vision
pytorch
knowledge-distillation
cifar10
dark-knowledge
model-compression
-
Updated
Jan 28, 2020 - Python
A curated list of neural network pruning resources.
-
Updated
Sep 2, 2020
Awesome Knowledge-Distillation. 分类整理的知识蒸馏paper(2014-2020)。
-
Updated
Aug 20, 2020
A list of high-quality (newest) AutoML works and lightweight models including 1.) Neural Architecture Search, 2.) Lightweight Structures, 3.) Model Compression, Quantization and Acceleration, 4.) Hyperparameter Optimization, 5.) Automated Feature Engineering.
tensorflow
pytorch
hyperparameter-optimization
awesome-list
quantization
nas
automl
model-compression
neural-architecture-search
meta-learning
architecture-search
quantized-training
model-acceleration
automated-feature-engineering
quantized-neural-network
-
Updated
Sep 3, 2020
Lightweight and Scalable framework that combines mainstream algorithms of Click-Through-Rate prediction based computational DAG, philosophy of Parameter Server and Ring-AllReduce collective communication.
distributed-systems
machine-learning
deep-learning
factorization-machines
computational-graphs
parameter-server
model-compression
-
Updated
Jun 17, 2019 - C++
PaddleSlim is an open-source library for deep model compression and architecture search.
pruning
quantization
nas
knowledge-distillation
evolution-strategy
model-compression
neural-architecture-search
hyperparameter-search
autodl
-
Updated
Sep 11, 2020 - Python
[CVPR2020] GhostNet: More Features from Cheap Operations
-
Updated
Aug 8, 2020 - Python
knowledge distillation papers
-
Updated
Jun 9, 2020
Data Efficient Model Compression
-
Updated
Aug 17, 2020 - Python
Filter Pruning via Geometric Median for Deep Convolutional Neural Networks Acceleration (CVPR 2019 Oral)
-
Updated
Sep 10, 2019 - Python
Papers for deep neural network compression and acceleration
-
Updated
May 24, 2018
Soft Filter Pruning for Accelerating Deep Convolutional Neural Networks
-
Updated
Oct 2, 2019 - Python
[ECCV 2018] AMC: AutoML for Model Compression and Acceleration on Mobile Devices
-
Updated
Oct 25, 2019 - Python
Infrastructures™ for Machine Learning Training/Inference in Production.
kubernetes
machine-learning
apache-spark
deep-learning
artificial-intelligence
awesome-list
pruning
quantization
knowledge-distillation
deep-learning-framework
model-compression
apache-arrow
federated-learning
machine-learning-systems
apache-mesos
-
Updated
May 24, 2019
Deep Face Model Compression
-
Updated
Aug 21, 2018 - MATLAB
Java interface for fastText
-
Updated
Feb 24, 2020 - Java
PyTorch Implementation of Weights Pruning
-
Updated
Jan 24, 2018 - Python
Model Compression CLI Tool for Keras.
-
Updated
May 9, 2019 - Python
[ECCV 2018] AMC: AutoML for Model Compression and Acceleration on Mobile Devices
-
Updated
Oct 25, 2019 - Python
Efficient Contextualized Representation: Language Model Pruning for Sequence Labeling
pytorch
named-entity-recognition
ner
language-model
sequence-labeling
model-compression
contextualized-representation
-
Updated
Feb 29, 2020 - Python
Awesome machine learning model compression research papers, tools, and learning material.
-
Updated
May 10, 2020
-
Updated
Jul 20, 2020 - Python
Improve this page
Add a description, image, and links to the model-compression topic page so that developers can more easily learn about it.
Add this topic to your repo
To associate your repository with the model-compression topic, visit your repo's landing page and select "manage topics."
What would you like to be added: As title
Why is this needed: All pruning schedule except AGPPruner only support level, L1, L2. While there are FPGM, APoZ, MeanActivation and Taylor, it would be much better if we can choose any pruner with any pruning schedule.
**Without this feature, how does current nni