Closed mmcclean-aws closed 1 month ago
Specifically for Llama-2-7B can we have one with TP and Zero-1. similar to the one in Neuron tutorials but we need a corresponding optimum-neuron version.
Tutorial for Llama-2-70B Training using (TP+PP)
This issue has been automatically marked as stale because it has not had recent activity. If you think this still needs to be addressed please comment on this thread. Thank you!
This issue has been automatically marked as stale because it has not had recent activity. If you think this still needs to be addressed please comment on this thread. Thank you!
This issue has been automatically marked as stale because it has not had recent activity. If you think this still needs to be addressed please comment on this thread. Thank you!
This issue has been automatically marked as stale because it has not had recent activity. If you think this still needs to be addressed please comment on this thread. Thank you!
This issue has been automatically marked as stale because it has not had recent activity. If you think this still needs to be addressed please comment on this thread. Thank you!
This issue has been automatically marked as stale because it has not had recent activity. If you think this still needs to be addressed please comment on this thread. Thank you!
LLaMa 2 is highly requested by customers. Can we ensure we have LLaMa 2 fine-tuning working with neuronx-distributed including sample code and tutorials for the 7B, 13B and 70B models ?