-
Notifications
You must be signed in to change notification settings - Fork 11.8k
[sync #10544] llama/ggml: add LLM training support #13105
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
base: master
Are you sure you want to change the base?
Conversation
@JohannesGaessler This is a tentative sync - still need to wait for #12799 to get merged. The optimization code in In #12799, the batch management is delegated to the KV cache object, so I've updated |
780d6fb
to
58115a2
Compare
this new feature is very helpful for AI beginners(such as me) to understand more details in hard-core AI tech. thanks too much! |
58115a2
to
7e79a42
Compare
more compact progress bar refactor: llama_prepare_sbatch/ubatch llama_save_model_to_file gqa_mode arg for repeat_back llama_opt_param_filter ggml_graph_dup force_grads refactor ggml_opt, fix test-opt
ggml-ci
@JohannesGaessler I've rebased this and should be good to update #10544 respectively and merge. Let me know if something does not work as expected. |
Thank you. I'll take a look when I get a chance. |
original #10544
This is a rebase of the #10544 PR by @JohannesGaessler on top of the upcoming #12799 (edit: merged now into
master
). The purpose is only to highlight the necessary changes that need to be applied to #10544.Testing with:
make -j && ./bin/llama-finetune --file ./wikitext-2-raw/wiki.test.raw --model ../models/llama-3.2-3b/ggml-model-f32.gguf -c 512 -b 512 -ub 512
TODOs:
test-backend-ops
asserts becauseggml_set_param
assertstensor->op == GGML_OP_NONE
, but does not take into account that the tensor could be a view.