DeepSpeed is a deep learning optimization library that makes distributed training and inference easy, efficient, and effective.
-
Updated
Nov 8, 2024 - Python
DeepSpeed is a deep learning optimization library that makes distributed training and inference easy, efficient, and effective.
👑 Easy-to-use and powerful NLP and LLM library with 🤗 Awesome model zoo, supporting wide-range of NLP tasks from research to industrial applications, including 🗂Text Classification, 🔍 Neural Search, ❓ Question Answering, ℹ️ Information Extraction, 📄 Document Intelligence, 💌 Sentiment Analysis etc.
Deduplicating archiver with compression and authenticated encryption.
Insane(ly slow but wicked good) PNG image optimization
Extract files from any kind of container formats
AIMET is a library that provides advanced quantization and compression techniques for trained neural network models.
Simple, configuration-driven backup software for servers and workstations
PaddleSlim is an open-source library for deep model compression and architecture search.
A toolkit to optimize ML models for deployment for Keras and TensorFlow, including quantization and pruning.
A curated list for Efficient Large Language Models
A PyTorch library and evaluation platform for end-to-end compression research
Code for CRATE (Coding RAte reduction TransformEr).
[CVPR 2020] GAN Compression: Efficient Architectures for Interactive Conditional GANs
PyTorch Implementation of "Lossless Image Compression through Super-Resolution"
Neural Network Compression Framework for enhanced OpenVINO™ inference
Access large archives as a filesystem efficiently, e.g., TAR, RAR, ZIP, GZ, BZ2, XZ, ZSTD archives
[NeurIPS 2023] LLM-Pruner: On the Structural Pruning of Large Language Models. Support Llama-3/3.1, Llama-2, LLaMA, BLOOM, Vicuna, Baichuan, TinyLlama, etc.
Data compression in TensorFlow
Transformers-compatible library for applying various compression algorithms to LLMs for optimized deployment with vLLM
High Octane Triage Analysis
Add a description, image, and links to the compression topic page so that developers can more easily learn about it.
To associate your repository with the compression topic, visit your repo's landing page and select "manage topics."