Closed ArcherShirou closed 2 months ago
the same question!
We recently added supported to QWen1.5! Have a try!
This issue has been automatically marked as inactive due to lack of recent activity. Should you believe it remains unresolved and warrants attention, kindly leave a comment on this thread. 此问题由于长期未有新进展而被系统自动标记为不活跃。如果您认为它仍有待解决,请在此帖下方留言以补充信息。
In the paper LLM Maybe LongLM: Self-Extend LLM Context Window Without Tuning, the authors describe a method to extend the context-window of any rope-based model without fine-tuning at inference time. The results reported in the paper seem game-changing.I apply it in Qwen by myself, but I failed to modify modeling_qwen.py. How could we add support for this in Qwen?