Open kemaloksuz opened 1 week ago
We train Senna on 8xA800 GPUs (80G), with a training duration of approximately 40 hours using 800k QA samples. By reducing the batch size, training on GPUs with 40G or 24G of memory should be feasible.
Thanks for the response. Does this include Senna E2E part or only Senna VLM? If not, how long does it take to train SennaE2E in addition to Senna VLM?
Senna-E2E and Senna-VLM can be trained independently in parallel. The training time for Senna-E2E depends on the training data and model used. With VAD as the base model on the nuScenes dataset, the training duration is approximately 12 hours (on 8 * 4090 GPUs).
Thanks for the answer. I think the details will be more clear once you release the code. Before then, can I ask 2 more questions please:
Hi, thanks for the great work. Could you please specify the time and compute requirements to train Senna?