Open hamelsmu opened 7 months ago
To add to this:
sharegpt
typeHappy to make a PR
If you are very familiar with axoltol, this --data flag was really confusing to me, because a key parameter in my config that I am used to using is being completely ignored with an extra layer of indirection. I actually got stuck on this personally as an experienced axolotl user, so I found the need to provide these two caveats.
@hamelsmu Even as a newcomer to axolotl, the discrepancy between the data flags in the two frameworks is really confusing to me. It would be helpful to have a guide describing the difference between how the flags are being used by the different frameworks (as a start)
Agree with @JUNIORCO. It would be great to have a conversational dataset example that works with a model like Llama3-8B-Instruct. I made a few attempts based on axolotl's example config and the example configs provided in this repo, but none seem to work with Llama3-8B-Instruct's format.
Additionally, it would also be great to have more details about the docker container and axolotl version used by Modal.
Carrying over discussion with @mwaskom from this thread
axoltol
, this--data
flag was really confusing to me, because a key parameter in my config that I am used to using is being completely ignored with an extra layer of indirection. I actually got stuck on this personally as an experienced axolotl user, so I found the need to provide these two caveats.cc: @charlesfrye @winglian curious what you think
_Originally posted by @hamelsmu in https://github.com/modal-labs/llm-finetuning/pull/48#discussion_r1575225245_