Skip to content
Permalink

Comparing changes

Choose two branches to see what’s changed or to start a new pull request. If you need to, you can also or learn more about diff comparisons.

Open a pull request

Create a new pull request by comparing changes across two branches. If you need to, you can also . Learn more about diff comparisons here.
base repository: deepspeedai/DeepSpeedExamples
Failed to load repositories. Confirm that selected base ref is valid, then try again.
Loading
base: master
Choose a base ref
...
head repository: zhangsmallshark/DeepSpeedExamples
Failed to load repositories. Confirm that selected head ref is valid, then try again.
Loading
compare: master
Choose a head ref
Checking mergeability… Don’t worry, you can still create the pull request.
  • 14 commits
  • 8 files changed
  • 13 contributors

Commits on Feb 12, 2025

  1. Εnable reward model offloading option (#930)

    * enable reward model offloading option
    
    * fixed code formatting
    
    * more formatting fixes
    
    * Pre-commit formatting fix
    
    ---------
    
    Co-authored-by: Olatunji Ruwase <[email protected]>
    Co-authored-by: Logan Adams <[email protected]>
    Co-authored-by: Logan Adams <[email protected]>
    Signed-off-by: zhangsmallshark <[email protected]>
    4 people authored and zhangsmallshark committed Feb 12, 2025
    Configuration menu
    Copy the full SHA
    18450b0 View commit details
    Browse the repository at this point in the history
  2. Remove the fixed eot_token mechanism for SFT (#927)

    Not all pretrained LLMs use `<|endoftext|>` as the `eot_token`, therefore it's inappropriate to fix it.
    
    Co-authored-by: Olatunji Ruwase <[email protected]>
    Co-authored-by: Logan Adams <[email protected]>
    Signed-off-by: zhangsmallshark <[email protected]>
    3 people authored and zhangsmallshark committed Feb 12, 2025
    Configuration menu
    Copy the full SHA
    216e147 View commit details
    Browse the repository at this point in the history
  3. DeepSpeed-Domino (#929)

    * add domino
    
    * use transformer from deepspeed
    
    * clean args
    
    * mega opt
    
    * add opt & timer
    
    * add opt
    
    * fix loss
    
    * folder name
    
    * Change arguent in pretrain script
    
    * Add readme for domino
    
    * Update readme for domino
    
    * Fixing usage issues
    
    * update dataset
    
    * megatron dependencies
    
    * path
    
    * Update README.md
    
    * remove imports
    
    * update import
    
    * Update README.md
    
    * Minor example script changes
    
    * train bash
    
    * require
    
    * Update README.md
    
    ---------
    
    Co-authored-by: chengming-zhang <[email protected]>
    Co-authored-by: Zheyu SHEN <[email protected]>
    Co-authored-by: root <[email protected]>
    Co-authored-by: Olatunji Ruwase <[email protected]>
    Co-authored-by: Logan Adams <[email protected]>
    Signed-off-by: zhangsmallshark <[email protected]>
    6 people committed Feb 12, 2025
    Configuration menu
    Copy the full SHA
    8d143c9 View commit details
    Browse the repository at this point in the history
  4. Update DeepSpeed version requirement to >=0.16.0 in requirements.txt …

    …for Domino (#939)
    
    Signed-off-by: zhangsmallshark <[email protected]>
    shenzheyu authored and zhangsmallshark committed Feb 12, 2025
    Configuration menu
    Copy the full SHA
    600db32 View commit details
    Browse the repository at this point in the history
  5. Example and benchmark of APIs to offload states (#942)

    * add benchmarking for offloading states
    
    * fix api names
    
    Signed-off-by: zhangsmallshark <[email protected]>
    tohtana authored and zhangsmallshark committed Feb 12, 2025
    Configuration menu
    Copy the full SHA
    2b906c9 View commit details
    Browse the repository at this point in the history
  6. add checkpoint

    Signed-off-by: zhangsmallshark <[email protected]>
    zhangsmallshark committed Feb 12, 2025
    Configuration menu
    Copy the full SHA
    937f452 View commit details
    Browse the repository at this point in the history
  7. fix args

    Signed-off-by: zhangsmallshark <[email protected]>
    zhangsmallshark committed Feb 12, 2025
    Configuration menu
    Copy the full SHA
    6b8109e View commit details
    Browse the repository at this point in the history
  8. remove-redundant-code (#947)

    Signed-off-by: zhangsmallshark <[email protected]>
    simonJJJ authored and zhangsmallshark committed Feb 12, 2025
    Configuration menu
    Copy the full SHA
    0be49e3 View commit details
    Browse the repository at this point in the history
  9. Add DPO support for DeepSpeed-Chat (#828)

    * Add label_smoothing while calculating step2 DPO loss in DeepSpeed-Chat.
    
    * Add training scripts for step2 DPO in DeepSpeed-Chat.
    
    * Remove unused packages and format the code of step2 DPO in DeepSpeed-Chat.
    
    * Update training scripts of step2 DPO in DeepSpeed-Chat.
    
    * Follow upstream fixes.
    
    * Update README.md for Step2 DPO finetuning.
    
    * Add opt 350M training log demo for step 2 dpo finetuning in DeepSpeed-Chat.
    
    * Address the formatting issue in step2 dpo finetuning in DeepSpeed-Chat.
    
    ---------
    
    Co-authored-by: Logan Adams <[email protected]>
    Co-authored-by: Olatunji Ruwase <[email protected]>
    Signed-off-by: zhangsmallshark <[email protected]>
    3 people authored and zhangsmallshark committed Feb 12, 2025
    Configuration menu
    Copy the full SHA
    f7628f4 View commit details
    Browse the repository at this point in the history
  10. Update references to torchvision (#949)

    Signed-off-by: zhangsmallshark <[email protected]>
    loadams authored and zhangsmallshark committed Feb 12, 2025
    Configuration menu
    Copy the full SHA
    b436bd4 View commit details
    Browse the repository at this point in the history
  11. save args

    Signed-off-by: zhangsmallshark <[email protected]>
    zhangsmallshark committed Feb 12, 2025
    Configuration menu
    Copy the full SHA
    abfedb9 View commit details
    Browse the repository at this point in the history
  12. Cleanup CODEOWNERS (#953)

    Signed-off-by: zhangsmallshark <[email protected]>
    loadams authored and zhangsmallshark committed Feb 12, 2025
    Configuration menu
    Copy the full SHA
    d6047a5 View commit details
    Browse the repository at this point in the history
  13. fix: the json format of the training imagenet configuration file (#954)

    Signed-off-by: zhangsmallshark <[email protected]>
    navanis authored and zhangsmallshark committed Feb 12, 2025
    Configuration menu
    Copy the full SHA
    df387f5 View commit details
    Browse the repository at this point in the history

Commits on Feb 19, 2025

  1. Configuration menu
    Copy the full SHA
    9e128ed View commit details
    Browse the repository at this point in the history
Loading