time-series-foundation-models / lag-llama

Lag-Llama: Towards Foundation Models for Probabilistic Time Series Forecasting
Apache License 2.0
1.08k stars 121 forks source link

different prerpint has different computing power (Summit supercomputer) and dataloader PatchTST #61

Closed m6129 closed 1 week ago

m6129 commented 1 month ago

Hello, Dear developer.

Question about the computing power expended: in the 20 November, 2023 year preprint you write

We acknowledge the support from the Canada CIFAR AI Chair Program and from the Canada Excellence Research Chairs (CERC) Program. This research was made possible thanks to the computing resources on the Summit supercomputer, provided as a part of the INCITE program award “Scalable Foundation Models for Transferable Generalist AI”. These resources were provided by the Oak Ridge Leadership Computing Facility at the Oak Ridge National Laboratory, which is supported by the Office of Science of the U.S. Department of Energy under Contract No. DE-AC05-00OR22725.

in the latest version of the preprint you state

We elaborate on our training procedure in Appendix B. For all the models trained in this paper, we use a single Nvidia Tesla-P100 GPU with 12 GB of memory, 4 CPU cores, and 24 GB of RAM

Could you explain this difference?

I also wanted to ask you if your model's prediction would be affected by the data_loader from the PatchTST repository (https://github.com/yuqinie98/PatchTST/tree/main/PatchTST_supervised/data_provider)?

ashok-arjun commented 3 weeks ago

Hi!!

The acknowledgement is also present in the latest version of the preprint, in the acknowledgements section. We uses the same compute in all versions of our paper; which is a single Nvidia Tesla-P100 GPU with 12 GB of memory, 4 CPU cores, and 24 GB of RAM. This was probably not clearly mentioned in the older Nov 2023 preprint.

In any case, I'd suggest referring to the newer version of the preprint for the details.

As for the PatchTST dataloader, I expect our model will not work with it, since we do not rely on the concept of patches. I'd suggest using our provided dataloaders (you can adapt any dataset as the tutorials suggest) to benchmark our model.

ashok-arjun commented 1 week ago

Hi! Just checking if this is resolved? @m6129

m6129 commented 1 week ago

Thanks!!!