Closed zhaihaotian closed 1 week ago
Hello, thank you for your interest in our work and for taking the time to read the paper! Let me provide some clarification.
Thank you for your detailed response and clarification!
It seems that your experimental setup is based on transductive learning, which makes perfect sense in this context. My earlier question likely stemmed from my limited familiarity with this specific domain. The few-shot fine-tuning or test-time adaptation methods I’ve been reviewing recently primarily rely on historical information from test samples, such as [1] and [2].
Including a comparison with other transductive learning methods in the updated version of your paper would indeed make your approach even more comprehensible and compelling. Such a comparison would not only highlight the unique strengths of KCL but also position it more clearly within the broader landscape of transductive methods.
Thank you again for addressing my questions so thoroughly. I look forward to seeing more exciting developments from your future work!
[1]: Dual Memory Networks: A Versatile Adaptation Approach for Vision-Language Models [2]: Dual Prototype Evolving for Test-Time Generalization of Vision-Language Models
Hello, thank you for sharing this interesting work and for open-sourcing the code! While reading the paper, I have some questions regarding the assumption of test sample visibility in the KCL method:
In the paper, the KCL method assumes that all test samples $f$ can be accessed at once, allowing the model to iteratively select high-confidence samples to supplement the few-shot knowledge. However, in practical scenarios, most mainstream few-shot learning methods or test-time fine-tuning methods typically restrict the model to process test samples sequentially, without granting access to the entire test dataset at once.
This assumption seems to diverge from the common streaming scenarios in few-shot learning or test-time fine-tuning. Having full access to all test samples undoubtedly provides the advantage of leveraging global information, which may significantly boost classification accuracy. Based on this, I have the following specific questions:
Looking forward to your response! Thank you again for your contributions to this field and for open-sourcing the code!