![DeepSpeed Compression: A composable library for extreme compression and zero-cost quantization - Microsoft Research DeepSpeed Compression: A composable library for extreme compression and zero-cost quantization - Microsoft Research](https://www.microsoft.com/en-us/research/uploads/prod/2022/07/DeepSpeed_Compression_Fig3.png)
DeepSpeed Compression: A composable library for extreme compression and zero-cost quantization - Microsoft Research
![DeepSpeed: Accelerating large-scale model inference and training via system optimizations and compression - Microsoft Research DeepSpeed: Accelerating large-scale model inference and training via system optimizations and compression - Microsoft Research](https://www.microsoft.com/en-us/research/uploads/prod/2021/05/DeepSpeed5_fig9_final.jpg)
DeepSpeed: Accelerating large-scale model inference and training via system optimizations and compression - Microsoft Research
![DeepSpeed: Accelerating large-scale model inference and training via system optimizations and compression - Microsoft Research DeepSpeed: Accelerating large-scale model inference and training via system optimizations and compression - Microsoft Research](https://www.microsoft.com/en-us/research/uploads/prod/2021/05/Fig1_DeepSpeed5_Blog.jpg)
DeepSpeed: Accelerating large-scale model inference and training via system optimizations and compression - Microsoft Research
![DeepSpeed: Accelerating large-scale model inference and training via system optimizations and compression - Microsoft Research DeepSpeed: Accelerating large-scale model inference and training via system optimizations and compression - Microsoft Research](https://www.microsoft.com/en-us/research/uploads/prod/2021/05/DeepSpeed5_fig4_blog_final.jpg)
DeepSpeed: Accelerating large-scale model inference and training via system optimizations and compression - Microsoft Research
![Model compression and optimization: Why think bigger when you can think smaller? | by David Williams | Data Science at Microsoft | Medium Model compression and optimization: Why think bigger when you can think smaller? | by David Williams | Data Science at Microsoft | Medium](https://miro.medium.com/v2/resize:fit:1200/1*4_mqsLRPr6BZdRlv4OK1jg.png)
Model compression and optimization: Why think bigger when you can think smaller? | by David Williams | Data Science at Microsoft | Medium
![Introduction to scaling Large Model training and inference using DeepSpeed | by mithil shah | Medium Introduction to scaling Large Model training and inference using DeepSpeed | by mithil shah | Medium](https://miro.medium.com/v2/resize:fit:1400/1*pXBax7uZYWAieu9flhIuKg.jpeg)
Introduction to scaling Large Model training and inference using DeepSpeed | by mithil shah | Medium
![Microsoft's Open Sourced a New Library for Extreme Compression of Deep Learning Models | by Jesus Rodriguez | Medium Microsoft's Open Sourced a New Library for Extreme Compression of Deep Learning Models | by Jesus Rodriguez | Medium](https://miro.medium.com/v2/resize:fit:1400/1*v7ClDr-_AnCA4SlU8824Bw.png)
Microsoft's Open Sourced a New Library for Extreme Compression of Deep Learning Models | by Jesus Rodriguez | Medium
![REQUEST] Add more device-agnostic compression algorithms · Issue #2894 · microsoft/DeepSpeed · GitHub REQUEST] Add more device-agnostic compression algorithms · Issue #2894 · microsoft/DeepSpeed · GitHub](https://user-images.githubusercontent.com/16394660/221105239-87edd628-30d6-4ab3-a3e9-6a013993a25d.png)
REQUEST] Add more device-agnostic compression algorithms · Issue #2894 · microsoft/DeepSpeed · GitHub
![DeepSpeed: Accelerating large-scale model inference and training via system optimizations and compression - Microsoft Research DeepSpeed: Accelerating large-scale model inference and training via system optimizations and compression - Microsoft Research](https://www.microsoft.com/en-us/research/uploads/prod/2021/05/1400x788_deepspeed_no_logo_still-1-scaled.jpg)
DeepSpeed: Accelerating large-scale model inference and training via system optimizations and compression - Microsoft Research
GitHub - microsoft/DeepSpeed: DeepSpeed is a deep learning optimization library that makes distributed training and inference easy, efficient, and effective.
![DeepSpeed Compression: A composable library for extreme compression and zero-cost quantization - Microsoft Research DeepSpeed Compression: A composable library for extreme compression and zero-cost quantization - Microsoft Research](https://www.microsoft.com/en-us/research/uploads/prod/2023/09/DeepSpeed4Science-BlogHeroFeature-no-text-1400x788-1.jpg)
DeepSpeed Compression: A composable library for extreme compression and zero-cost quantization - Microsoft Research
![DeepSpeed Compression: A composable library for extreme compression and zero-cost quantization - Microsoft Research DeepSpeed Compression: A composable library for extreme compression and zero-cost quantization - Microsoft Research](https://www.microsoft.com/en-us/research/uploads/prod/2023/10/ICCV-SpaceEVO-2023-BlogHeroFeature-1400x788-1.png)
DeepSpeed Compression: A composable library for extreme compression and zero-cost quantization - Microsoft Research
![DeepSpeed: Accelerating large-scale model inference and training via system optimizations and compression - Microsoft Research DeepSpeed: Accelerating large-scale model inference and training via system optimizations and compression - Microsoft Research](https://www.microsoft.com/en-us/research/uploads/prod/2021/05/DeepSpeed_fig2_5blog.jpg)
DeepSpeed: Accelerating large-scale model inference and training via system optimizations and compression - Microsoft Research
GitHub - microsoft/DeepSpeed: DeepSpeed is a deep learning optimization library that makes distributed training and inference easy, efficient, and effective.
![DeepSpeed: Accelerating large-scale model inference and training via system optimizations and compression - Microsoft Research DeepSpeed: Accelerating large-scale model inference and training via system optimizations and compression - Microsoft Research](https://www.microsoft.com/en-us/research/uploads/prod/2021/05/DeepSpeed5_fig6_final.jpg)
DeepSpeed: Accelerating large-scale model inference and training via system optimizations and compression - Microsoft Research
![DeepSpeed Compression: A composable library for extreme compression and zero-cost quantization - Microsoft Research DeepSpeed Compression: A composable library for extreme compression and zero-cost quantization - Microsoft Research](https://www.microsoft.com/en-us/research/uploads/prod/2022/01/1400x788_Deepspeed_MoE_no_logo_still-1-scaled.jpg)
DeepSpeed Compression: A composable library for extreme compression and zero-cost quantization - Microsoft Research
![DeepSpeed: Accelerating large-scale model inference and training via system optimizations and compression - Microsoft Research DeepSpeed: Accelerating large-scale model inference and training via system optimizations and compression - Microsoft Research](https://www.microsoft.com/en-us/research/uploads/prod/2021/05/DeepSpeed5_fig7_final.jpg)