Log in
Enquire now
DeepSpeed

DeepSpeed

DeepSpeed, part of Microsoft AI at Scale, is a deep learning optimization library.

OverviewStructured DataIssuesContributors

Contents

deepspeed.ai
Is a
Software
Software

Software attributes

License
MIT License
First Release
February 11, 2020
Latest Release
September 2020
Created/Discovered by
Microsoft
Microsoft

Other attributes

Source Code
github.com/microsoft/DeepSpeed
Written in
Python
CUDA
C++

DeepSpeed is an open source deep learning optimization library for PyTorch. The library is designed to reduce computing power and memory use and to train large distributed models with better parallelism on existing computer hardware. DeepSpeed is part of Microsoft's AI at Scale initiative to enable artificial intelligence capabilities at scale.

DeepSpeed can train DL models with trillions of parameters on current generation GPU clusters. Adopters of DeepSpeed have produced a language model with over 17 billion paramters called Turing-NLG.

DeepSpeed provides memory-efficient data parallelism and enables training models without model parallelism. For example, DeepSpeed can train models with up to 13 billion parameters on NVIDIA V100 GPUs with 32GB of device memory.

Part of DeepSpeed's effectiveness is its reduction of the training memory footprint through a solution called Zero Redundancy Optimizer (ZeRO). ZeRO partitions model states and gradients to save significant memory. Furthermore, it reduces activation memory and fragmented memory.

Timeline

No Timeline data yet.

Further Resources

Title
Author
Link
Type
Date

DeepSpeed - Microsoft Research

https://www.microsoft.com/en-us/research/project/deepspeed/

Web

Microsoft Releases Latest Version Of DeepSpeed, Its Python Library For Deep Learning Optimisation

Ambika Choudhury, Ambika Choudhury, Ambika Choudhury

https://analyticsindiamag.com/microsoft-releases-latest-version-of-deepspeed-its-python-library-for-deep-learning-optimisation/

Web

September 15, 2020

ZeRO: Memory Optimizations Toward Training Trillion Parameter Models - Microsoft Research

https://www.microsoft.com/en-us/research/publication/zero-memory-optimizations-toward-training-trillion-parameter-models/

Web

October 7, 2019

References

Find more entities like DeepSpeed

Use the Golden Query Tool to find similar entities by any field in the Knowledge Graph, including industry, location, and more.
Open Query Tool
Access by API
Golden Query Tool
Golden logo

Company

  • Home
  • Press & Media
  • Blog
  • Careers
  • WE'RE HIRING

Products

  • Knowledge Graph
  • Query Tool
  • Data Requests
  • Knowledge Storage
  • API
  • Pricing
  • Enterprise
  • ChatGPT Plugin

Legal

  • Terms of Service
  • Enterprise Terms of Service
  • Privacy Policy

Help

  • Help center
  • API Documentation
  • Contact Us
By using this site, you agree to our Terms of Service.