Closed giyaseddin closed 8 months ago
t5 is currently supported officially by TGI, this model falls back on the transformers implementation which we cannot guarantee will work 100% of the time.
We do depend on transformers internals (which may vary on model per model, like last_hidden_state here)).
This issue is stale because it has been open 30 days with no activity. Remove stale label or comment or this will be closed in 5 days.
System Info
I am trying to run an encoder-decoder kind of seq2seq model using TGI I read the docs in here https://huggingface.co/docs/transformers/model_doc/bert-generation
however it seems like there's something with the support of the seq2seq models is happening, I really appreciate any note if it has anything that I misunderstand or miss in the usage.
Information
Tasks
Reproduction
BERT2BERT model is configured in the following docker-compose file
Example model: bert2bert_shared-spanish-finetuned-summarization
From the openAPI swagger docs, a regular
/generate
call is yielding to thisExpected behavior
It is supposed to run and generate the answer just like Llama Models, Mistral, T5, etc.