66 skills found · Page 1 of 3
666DZY666 / Micronetmicronet, a model compression and deploy lib. compression: 1、quantization: quantization-aware-training(QAT), High-Bit(>2b)(DoReFa/Quantization and Training of Neural Networks for Efficient Integer-Arithmetic-Only Inference)、Low-Bit(≤2b)/Ternary and Binary(TWN/BNN/XNOR-Net); post-training-quantization(PTQ), 8-bit(tensorrt); 2、 pruning: normal、regular and group convolutional channel pruning; 3、 group convolution structure; 4、batch-normalization fuse for quantization. deploy: tensorrt, fp32/fp16/int8(ptq-calibration)、op-adapt(upsample)、dynamic_shape
princeton-nlp / LLM Shearing[ICLR 2024] Sheared LLaMA: Accelerating Language Model Pre-training via Structured Pruning
tianyic / Only Train Once Personal FootprintOTOv1-v3, NeurIPS, ICLR, TMLR, DNN Training, Compression, Structured Pruning, Erasing Operators, CNN, Diffusion, LLM
VainF / Diff Pruning[NeurIPS 2023] Structural Pruning for Diffusion Models
princeton-nlp / CoFiPruning[ACL 2022] Structured Pruning Learns Compact and Accurate Models https://arxiv.org/abs/2204.00408
ghimiredhikura / Awasome PruningAwesome Pruning. ✅ Curated Resources for Neural Network Pruning.
marcoancona / TorchPrunerOn-the-fly Structured Pruning for PyTorch models. This library implements several attributions metrics and structured pruning utils for neural networks in PyTorch.
lmbxmu / ABCPrunerPytorch implementation of our paper accepted by IJCAI 2020 -- Channel Pruning via Automatic Structure Search
alecwangcq / EigenDamage PytorchCode for "EigenDamage: Structured Pruning in the Kronecker-Factored Eigenbasis" https://arxiv.org/abs/1905.05934
jordddan / Pruning LLMsThe framework to prune LLMs to any size and any config.
polarizationpruning / PolarizationPruningImplementation of Neuron-level Structured Pruning using Polarization Regularizer
necludov / Group Sparsity SbpStructured Bayesian Pruning, NIPS 2017
CASIA-LMC-Lab / FLAP[AAAI 2024] Fluctuation-based Adaptive Structured Pruning for Large Language Models
microsoft / Only Train OnceOTOv1-v3, NeurIPS, ICLR, TMLR, DNN Training, Compression, Structured Pruning, Erasing Operators, CNN, Diffusion, LLM
MMorafah / Sub FedAvgPersonalized Federated Learning by Structured and Unstructured Pruning under Data Heterogeneity
Eclipsess / CHIP NeurIPS2021Code for CHIP: CHannel Independence-based Pruning for Compact Neural Networks (NeruIPS 2021).
vantienpham / CORING:ring: Efficient tensor decomposition-based filter pruning
StijnVerdenius / SNIP ItThis repository is the official implementation of the paper Pruning via Iterative Ranking of Sensitivity Statistics and implements novel pruning / compression algorithms for deep learning / neural networks. Amongst others it implements structured pruning before training, its actual parameter shrinking and unstructured before/during training.
peremartra / OptipfairStructured pruning and bias visualization for Large Language Models. Tools for LLM optimization and fairness analysis.
ldery / BonsaiCode for "Everybody Prune Now: Structured Pruning of LLMs with only Forward Passes"