Here are
30 public repositories
matching this topic...
DeepSpeed is a deep learning optimization library that makes distributed training and inference easy, efficient, and effective.
Updated
Aug 24, 2022
Python
Colossal-AI: A Unified Deep Learning System for Big Model Era
Updated
Aug 24, 2022
Python
Distributed Deep Learning, with a focus on distributed training, using Keras and Apache Spark.
Updated
Jul 25, 2018
Python
A state-of-the-art multithreading runtime: message-passing based, fast, scalable, ultra-low overhead
Paddle Distributed Training Examples. 飞桨分布式训练示例 Resnet Bert GPT MOE DataParallel ModelParallel PipelineParallel HybridParallel AutoParallel Zero Sharding Recompute GradientMerge Offload AMP DGC LocalSGD Wide&Deep
Updated
Aug 24, 2022
Python
Distributed Keras Engine, Make Keras faster with only one line of code.
Updated
Oct 3, 2019
Python
LiBai(李白): A Toolbox for Large-Scale Distributed Parallel Training
Updated
Aug 23, 2022
Python
Ternary Gradients to Reduce Communication in Distributed Deep Learning (TensorFlow)
Updated
Nov 19, 2018
Python
Easy Parallel Library (EPL) is a general and efficient deep learning framework for distributed model training.
Updated
Jul 21, 2022
Python
Orkhon: ML Inference Framework and Server Runtime
Multi-GPU training for Keras
Updated
Jun 23, 2017
Python
WIP. Veloce is a low-code Ray-based parallelization library that makes machine learning computation novel, efficient, and heterogeneous.
Updated
Aug 4, 2022
Python
☕ Implement of Parallel Matrix Multiplication Methods Using FOX Algorithm on Peking University's High-performance Computing System
Understanding the effects of data parallelism and sparsity on neural network training
Updated
Jul 27, 2021
Python
OpenCL Data Parallel Primitives
Dependence-Based Code Transformation for Coarse-Grained Parallelism
Accelerating Ascon: Light Weight Cryptography
Development of Project HPGO | Hybrid Parallelism Global Orchestration
Torch Automatic Distributed Neural Network (TorchAD-NN) training library. Built on top of TorchMPI, this module automatically parallelizes neural network training.
A mostly POSIX-compliant utility that scans a given interval for vampire numbers.
Improve this page
Add a description, image, and links to the
data-parallelism
topic page so that developers can more easily learn about it.
Curate this topic
Add this topic to your repo
To associate your repository with the
data-parallelism
topic, visit your repo's landing page and select "manage topics."
Learn more
You can’t perform that action at this time.
You signed in with another tab or window. Reload to refresh your session.
You signed out in another tab or window. Reload to refresh your session.