KimMeen / Time-LLM

[ICLR 2024] Official implementation of " 🦙 Time-LLM: Time Series Forecasting by Reprogramming Large Language Models"
https://arxiv.org/abs/2310.01728
Apache License 2.0
1.02k stars 179 forks source link

Performance check between LLM vs No LLM backbone #118

Open SajayR opened 1 week ago

SajayR commented 1 week ago

Was the performance gap between using an LLM in the backbone vs no LLM in the backbone ever tested to verify the excellent eval results to actually be originating from the LLMs? I tried to run tests on my own, and the metrics on the runs without the LLM backbone seem to be performing better or sometimes equivalent to the ones with LLM backbones, and I just wanted to get the author's perspective on the case

kwuking commented 6 hours ago

Thank you very much for your attention to our work and valuable feedback. I'm quite puzzled by this result, as it seems to be inconsistent with our experimental findings. However, given the complexity of training a large model and the relatively small size of the current dataset, insufficient training may lead to ineffective outcomes. We are currently reflecting on this issue. It is important to emphasize that TimeLLM is clearly more important and valuable as a model that integrates text and time series data across modes. We are making every effort in this direction, hoping to achieve semantic integration of time series and text data. Our new work is in progress, and we'll keep everyone updated if we make progress.