-
Notifications
You must be signed in to change notification settings - Fork 1.1k
Comparing changes
Open a pull request
base repository: deepspeedai/DeepSpeedExamples
base: master
head repository: zhangsmallshark/DeepSpeedExamples
compare: master
- 14 commits
- 8 files changed
- 13 contributors
Commits on Feb 12, 2025
-
Εnable reward model offloading option (#930)
* enable reward model offloading option * fixed code formatting * more formatting fixes * Pre-commit formatting fix --------- Co-authored-by: Olatunji Ruwase <[email protected]> Co-authored-by: Logan Adams <[email protected]> Co-authored-by: Logan Adams <[email protected]> Signed-off-by: zhangsmallshark <[email protected]>
Configuration menu - View commit details
-
Copy full SHA for 18450b0 - Browse repository at this point
Copy the full SHA 18450b0View commit details -
Remove the fixed
eot_token
mechanism for SFT (#927)Not all pretrained LLMs use `<|endoftext|>` as the `eot_token`, therefore it's inappropriate to fix it. Co-authored-by: Olatunji Ruwase <[email protected]> Co-authored-by: Logan Adams <[email protected]> Signed-off-by: zhangsmallshark <[email protected]>
Configuration menu - View commit details
-
Copy full SHA for 216e147 - Browse repository at this point
Copy the full SHA 216e147View commit details -
* add domino * use transformer from deepspeed * clean args * mega opt * add opt & timer * add opt * fix loss * folder name * Change arguent in pretrain script * Add readme for domino * Update readme for domino * Fixing usage issues * update dataset * megatron dependencies * path * Update README.md * remove imports * update import * Update README.md * Minor example script changes * train bash * require * Update README.md --------- Co-authored-by: chengming-zhang <[email protected]> Co-authored-by: Zheyu SHEN <[email protected]> Co-authored-by: root <[email protected]> Co-authored-by: Olatunji Ruwase <[email protected]> Co-authored-by: Logan Adams <[email protected]> Signed-off-by: zhangsmallshark <[email protected]>
Configuration menu - View commit details
-
Copy full SHA for 8d143c9 - Browse repository at this point
Copy the full SHA 8d143c9View commit details -
Update DeepSpeed version requirement to >=0.16.0 in requirements.txt …
…for Domino (#939) Signed-off-by: zhangsmallshark <[email protected]>
Configuration menu - View commit details
-
Copy full SHA for 600db32 - Browse repository at this point
Copy the full SHA 600db32View commit details -
Example and benchmark of APIs to offload states (#942)
* add benchmarking for offloading states * fix api names Signed-off-by: zhangsmallshark <[email protected]>
Configuration menu - View commit details
-
Copy full SHA for 2b906c9 - Browse repository at this point
Copy the full SHA 2b906c9View commit details -
Signed-off-by: zhangsmallshark <[email protected]>
Configuration menu - View commit details
-
Copy full SHA for 937f452 - Browse repository at this point
Copy the full SHA 937f452View commit details -
Configuration menu - View commit details
-
Copy full SHA for 6b8109e - Browse repository at this point
Copy the full SHA 6b8109eView commit details -
Signed-off-by: zhangsmallshark <[email protected]>
Configuration menu - View commit details
-
Copy full SHA for 0be49e3 - Browse repository at this point
Copy the full SHA 0be49e3View commit details -
Add DPO support for DeepSpeed-Chat (#828)
* Add label_smoothing while calculating step2 DPO loss in DeepSpeed-Chat. * Add training scripts for step2 DPO in DeepSpeed-Chat. * Remove unused packages and format the code of step2 DPO in DeepSpeed-Chat. * Update training scripts of step2 DPO in DeepSpeed-Chat. * Follow upstream fixes. * Update README.md for Step2 DPO finetuning. * Add opt 350M training log demo for step 2 dpo finetuning in DeepSpeed-Chat. * Address the formatting issue in step2 dpo finetuning in DeepSpeed-Chat. --------- Co-authored-by: Logan Adams <[email protected]> Co-authored-by: Olatunji Ruwase <[email protected]> Signed-off-by: zhangsmallshark <[email protected]>
Configuration menu - View commit details
-
Copy full SHA for f7628f4 - Browse repository at this point
Copy the full SHA f7628f4View commit details -
Update references to torchvision (#949)
Signed-off-by: zhangsmallshark <[email protected]>
Configuration menu - View commit details
-
Copy full SHA for b436bd4 - Browse repository at this point
Copy the full SHA b436bd4View commit details -
Configuration menu - View commit details
-
Copy full SHA for abfedb9 - Browse repository at this point
Copy the full SHA abfedb9View commit details -
Signed-off-by: zhangsmallshark <[email protected]>
Configuration menu - View commit details
-
Copy full SHA for d6047a5 - Browse repository at this point
Copy the full SHA d6047a5View commit details -
fix: the json format of the training imagenet configuration file (#954)
Signed-off-by: zhangsmallshark <[email protected]>
Configuration menu - View commit details
-
Copy full SHA for df387f5 - Browse repository at this point
Copy the full SHA df387f5View commit details
Commits on Feb 19, 2025
-
Configuration menu - View commit details
-
Copy full SHA for 9e128ed - Browse repository at this point
Copy the full SHA 9e128edView commit details
This comparison is taking too long to generate.
Unfortunately it looks like we can’t render this comparison for you right now. It might be too big, or there might be something weird with your repository.
You can try running this command locally to see the comparison on your machine:
git diff master...master