[sync #10544] llama/ggml: add LLM training support #13105
Draft
+1,372
−322
Add this suggestion to a batch that can be applied as a single commit.
This suggestion is invalid because no changes were made to the code.
Suggestions cannot be applied while the pull request is closed.
Suggestions cannot be applied while viewing a subset of changes.
Only one suggestion per line can be applied in a batch.
Add this suggestion to a batch that can be applied as a single commit.
Applying suggestions on deleted lines is not supported.
You must change the existing code in this line in order to create a valid suggestion.
Outdated suggestions cannot be applied.
This suggestion has been applied or marked resolved.
Suggestions cannot be applied from pending reviews.
Suggestions cannot be applied on multi-line comments.
Suggestions cannot be applied while the pull request is queued to merge.
Suggestion cannot be applied right now. Please check back later.
original #10544
This is a rebase of the #10544 PR by @JohannesGaessler on top of the upcoming #12799 (edit: merged now into
master
). The purpose is only to highlight the necessary changes that need to be applied to #10544.Testing with:
make -j && ./bin/llama-finetune --file ./wikitext-2-raw/wiki.test.raw --model ../models/llama-3.2-3b/ggml-model-f32.gguf -c 512 -b 512 -ub 512
TODOs:
test-backend-ops
asserts becauseggml_set_param
assertstensor->op == GGML_OP_NONE
, but does not take into account that the tensor could be a view.