I am facing issue while giving prompt to LLM models like oasst-pythia-12b, fastchat-t5-3b.
I provided a prompt question along with context but I see that prompt is getting clipped off while providing input to model in Chat Arena.
For ex. - I provided a prompt which of around 1010 token size to fastchat-t5-3b model in Arena (I understand that context length for fastchat-t5-3b is 2048) but the eventually the input taken by Arena tool was striped to around 672 tokens.
Can you please help me understand why this is happening or if I am missing something here.
Hi,
I am facing issue while giving prompt to LLM models like oasst-pythia-12b, fastchat-t5-3b. I provided a prompt question along with context but I see that prompt is getting clipped off while providing input to model in Chat Arena. For ex. - I provided a prompt which of around 1010 token size to fastchat-t5-3b model in Arena (I understand that context length for fastchat-t5-3b is 2048) but the eventually the input taken by Arena tool was striped to around 672 tokens.
Can you please help me understand why this is happening or if I am missing something here.