| IntelLabs/distiller |
4,252 |
|
0 |
0 |
almost 3 years ago |
0 |
|
65 |
apache-2.0 |
Jupyter Notebook |
| Neural Network Distiller by Intel AI Lab: a Python package for neural network compression research. https://intellabs.github.io/distiller |
| AutoGPTQ/AutoGPTQ |
3,206 |
|
0 |
0 |
about 2 years ago |
0 |
|
174 |
mit |
Python |
| An easy-to-use LLMs quantization package with user-friendly apis, based on GPTQ algorithm. |
| PINTO0309/PINTO_model_zoo |
3,121 |
|
0 |
0 |
about 2 years ago |
0 |
|
11 |
mit |
Python |
| A repository for storing models that have been inter-converted between various frameworks. Supported frameworks are TensorFlow, PyTorch, ONNX, OpenVINO, TFJS, TFTRT, TensorFlowLite (Float32/16/INT8), EdgeTPU, CoreML. |
| IntelLabs/nlp-architect |
2,924 |
|
0 |
0 |
over 3 years ago |
10 |
April 12, 2020 |
14 |
apache-2.0 |
Python |
| A model library for exploring state-of-the-art deep learning topologies and techniques for optimizing Natural Language Processing neural networks |
| aaron-xichen/pytorch-playground |
2,366 |
|
0 |
0 |
over 3 years ago |
0 |
|
9 |
mit |
Python |
| Base pretrained models and datasets in pytorch (MNIST, SVHN, CIFAR10, CIFAR100, STL10, AlexNet, VGG16, VGG19, ResNet, Inception, SqueezeNet) |
| 666DZY666/micronet |
2,129 |
|
0 |
0 |
over 4 years ago |
46 |
October 06, 2021 |
70 |
mit |
Python |
| micronet, a model compression and deploy lib. compression: 1、quantization: quantization-aware-training(QAT), High-Bit(>2b)(DoReFa/Quantization and Training of Neural Networks for Efficient Integer-Arithmetic-Only Inference)、Low-Bit(≤2b)/Ternary and Binary(TWN/BNN/XNOR-Net); post-training-quantization(PTQ), 8-bit(tensorrt); 2、 pruning: normal、regular and group convolutional channel pruning; 3、 group convolution structure; 4、batch-normalization fuse for quantization. deploy: tensorrt, fp32/fp16/int8(ptq-calibration)、op-adapt(upsample)、dynamic_shape |
| dvmazur/mixtral-offloading |
1,943 |
|
0 |
0 |
about 2 years ago |
0 |
|
12 |
mit |
Python |
| Run Mixtral-8x7B models in Colab or consumer desktops |
| huggingface/optimum |
1,908 |
|
0 |
53 |
about 2 years ago |
53 |
December 06, 2023 |
295 |
apache-2.0 |
Python |
| 🚀 Accelerate training and inference of 🤗 Transformers and 🤗 Diffusers with easy to use hardware optimization tools |
| lucidrains/vector-quantize-pytorch |
1,627 |
|
0 |
25 |
about 2 years ago |
160 |
December 06, 2023 |
27 |
mit |
Python |
| Vector Quantization, in Pytorch |
| open-mmlab/mmrazor |
1,231 |
|
0 |
2 |
over 2 years ago |
8 |
May 04, 2022 |
133 |
apache-2.0 |
Python |
| OpenMMLab Model Compression Toolbox and Benchmark. |