Английская Википедия:DeepSpeed

Материал из Онлайн справочника
Версия от 23:50, 25 февраля 2024; EducationBot (обсуждение | вклад) (Новая страница: «{{Английская Википедия/Панель перехода}} {{short description|Microsoft open source library}} {{Infobox software | name = DeepSpeed | logo = DeepSpeed logo.svg | screenshot = | screenshot size = | caption = | author = Microsoft Research | developer = Microsoft | released = {{Start date and age|2020|05|18}} | latest release version = v0.12.3 | latest release date = {{Start date and age|2023|11|10}} | repo = {{URL|htt...»)
(разн.) ← Предыдущая версия | Текущая версия (разн.) | Следующая версия → (разн.)
Перейти к навигацииПерейти к поиску

Шаблон:Short description Шаблон:Infobox software DeepSpeed is an open source deep learning optimization library for PyTorch.[1] The library is designed to reduce computing power and memory use and to train large distributed models with better parallelism on existing computer hardware.[2][3] DeepSpeed is optimized for low latency, high throughput training. It includes the Zero Redundancy Optimizer (ZeRO) for training models with 1 trillion or more parameters.[4] Features include mixed precision training, single-GPU, multi-GPU, and multi-node training as well as custom model parallelism. The DeepSpeed source code is licensed under MIT License and available on GitHub.[5]

The team claimed to achieve up to a 6.2x throughput improvement, 2.8x faster convergence, and 4.6x less communication.[6]

See also

Шаблон:Portal

References

Шаблон:Reflist

Further reading

External links

Шаблон:Deep Learning Software Шаблон:Microsoft FOSS Шаблон:Microsoft Research


Шаблон:Microsoft-software-stub