DeepSpeed (research) developer at @microsoft
-
Microsoft
- Seattle, WA
- https://rasley.io
- @jeffra45
Highlights
- Pro
Block or Report
Block or report jeffra
Report abuse
Contact GitHub support about this user’s behavior. Learn more about reporting abuse.
Report abusePinned
-
microsoft/DeepSpeed Public
DeepSpeed is a deep learning optimization library that makes distributed training and inference easy, efficient, and effective.
-
-
microsoft/Megatron-DeepSpeed Public
Forked from NVIDIA/Megatron-LM
Ongoing research training transformer language models at scale, including: BERT & GPT-2
1,225 contributions in the last year
Less
More
Activity overview
Contributed to
microsoft/DeepSpeed,
microsoft/DeepSpeed-MII,
jeffra/sandbox
and 13 other
repositories
Contribution activity
April 2023
Created 10 commits in 2 repositories
Reviewed 14 pull requests in 2 repositories
microsoft/DeepSpeed
13 pull requests
- Update DS-Chat docs for v0.9.0
- fix license badge
- Chatgpt chinese blog
- DeepSpeed Chat
- [DRAFT] Tentative implementation of MiCS
- op_builder: conditionally compute relative path for hip compiled files
- Update torch version check in building sparse_attn
- Remove benchmark code
- Update curriculum-learning.md
- [CPU support] Optionally bind each rank to different cores on host
-
[ci]
nv-transformers-v100- use the same torch version as transformers CI - Fix launch issue
- Update SD triton version in requirements-sd.txt
microsoft/DeepSpeedExamples
1 pull request
28
contributions
in private repositories
Apr 2 – Apr 13





