Contents

deepspeed 0.15.2

0

DeepSpeed library

DeepSpeed is a Python package developed by Microsoft that provides a deep learning optimization library designed to scale across multiple GPUs and servers. It is capable of training models with billions or even trillions of parameters, achieving excellent system throughput and efficiently scaling to thousands of GPUs.

DeepSpeed is particularly useful for training and inference of large language models, and it falls under the category of Machine Learning Frameworks and Libraries. It is designed to work with PyTorch and offers system innovations such as Zero Redundancy Optimizer (ZeRO), 3D parallelism, and model-parallelism to enable efficient training of large models.

Stars: 35068, Watchers: 35068, Forks: 4062, Open Issues: 1109

The microsoft/DeepSpeed repo was created 4 years ago and the last code push was Yesterday.
The project is extremely popular with a mindblowing 35068 github stars!

How to Install deepspeed

You can install deepspeed using pip

pip install deepspeed

or add it to a project with poetry

poetry add deepspeed

Package Details

Author
DeepSpeed Team
License
Apache Software License 2.0
Homepage
http://deepspeed.ai
PyPi:
https://pypi.org/project/deepspeed/
Documentation:
https://deepspeed.readthedocs.io
GitHub Repo:
https://github.com/microsoft/DeepSpeed

Classifiers

No  deepspeed  pypi packages just yet.

Errors

A list of common deepspeed errors.

Code Examples

Here are some deepspeed code examples and snippets.

GitHub Issues

The deepspeed package has 1109 open issues on GitHub

  • [BUG] matmul_ext_update_autotune_table atexit error
  • [BUG] Unexpected caculations at backward pass with ZeRO-Infinity SSD offloading
  • update ut/doc for glm/codegen
  • Multi-node and multi-GPU fine-tuning error: ncclInternalError
  • Zero Stage-2 Frozen Layers[BUG]
  • [PROBLEM] P2p recv waiting for data will cause other threads under the same process to be unable to perform any operations
  • Spread layers more uniformly when using partition_uniform
  • Issue with DeepSpeed Inference - Multiple Processes for Model Loading and Memory Allocation
  • [BUG] CPU Adam failing
  • [BUG] Cannot increase batch size more than 1 with ZeRO-Infinity SSD offloading
  • [REQUEST] please provide clear working installation guide
  • load linear layer weight with dtype from ckpt
  • [QNA] How can i choose adam between fused and cpu?
  • Refactor autoTP inference for HE
  • [BUG] No runnable example for MoE / PR-MoE GPT inference

See more issues on GitHub

Related Packages & Articles

datasets 3.0.1

HuggingFace community-driven open-source library of datasets

thinc 9.1.1

A refreshing functional take on deep learning, compatible with your favorite libraries

nlp 0.4.0

HuggingFace/NLP is an open library of NLP datasets.

keras 3.6.0

Keras is a deep learning API written in Python, running on top of the machine learning platform TensorFlow. The core data structures of Keras are layers and models. The philosophy is to keep simple things simple, while allowing the user to be fully in control when they need to (the ultimate control being the easy extensibility of the source code via subclassing).