-
Hi!
Let's bring the documentation to all the Spanish-speaking community 🌐
Who would want to translate? Please follow the 🤗 [TRANSLATING guide](https://github.com/huggingface/transformers/blob/m…
-
### Model description
MiniCPM-V is a series of Openbmb's vision language models.
We want to add support for MiniCPM-V-2 and later models
### Open source status
- [x] The model implementation is av…
-
I must start by saying that is an awesome library that even simplify training cross-encoder.
Quick question, are there any plan to add support for tabular data in cross-encoder ? This recently bec…
-
I noticed that the pretrain section seems to only include a basic Transformer network without any attention weights exploration, as mentioned in the paper. I was wondering if the absence of attention …
-
Thank you for sharing this collection of papers
I also made a paper collection list about vision attention and transformer:
https://github.com/cmhungsteve/Awesome-Transformer-Attention
Feel fre…
-
Hi, thanks for this awesome project! I've been using it successfully with an ollama instance hosted on a remote machine, and an anthropic api key.
I have many equations in my code, and the LLMs fre…
-
Have you ever considered hosting the pre-trained SSAST model checkpoints on the Huggingface Hub?
-
Hi~ Thanks for your awesome contribution, it looks so cool for CNN related models!!!
but, can you give some examples like RNN/Stacked RNN/Transformer etc. really appreciate it
-
Hi there, 1st of all thanks for this awesome work !
Since we've 'doxed' it in our [HyMPS project](https://forart.it/HyMPS) (under [AUDIO section](https://github.com/FORARTfe/HyMPS#- ) \ [Effects pa…
-
### Feature request
Paged attention has been enabled by a lot of server engine, e.g., [vllm](https://github.com/vllm-project/vllm), [tensorrt-llm](https://github.com/NVIDIA/TensorRT-LLM/blob/relea…