-
### Feature request
We want to standardize the logic flow through Processor classes. Since processors can have different kwargs depending on the model and modality, we are adding a `TypedDict` fo…
-
I am trying to export LayoutLMv2 model to onnx but there is no support for that available in transformers library.
I have tried to follow the method available for layoutLM but that is not working.
…
-
**Is your feature request related to a problem? Please describe.**
*TLDR*: Lock bounding boxes to better support token classification with a new `editable` option. Only a very small feature in code t…
-
Following what was done by @ChainYo in Transformers, in the [ONNXConfig: Add a configuration for all available models](https://github.com/huggingface/transformers/issues/16308) issue, the idea is to a…
-
Hi there,
the model I am using is LayoutLMV3 (**LayoutLMv3ForTokenClassification**).
I want that the model can take more than 512 tokens because when the text is very large it does not classify…
-
Looking through the [recently posted codebase](https://github.com/microsoft/unilm/tree/master/layoutlmft) I noted that [modeling_layoutlmv2.py](https://github.com/microsoft/unilm/blob/master/layoutlmf…
-
Let's say I wish to merge LLAMA-3-8B with Mistral 7B creating a MOE.
How should I proceed?
Or for example 2 small models (3B/4B) of different architectures.
-
Hi,
I've added LayoutLMv2 and LayoutXLM to HuggingFace Transformers. I've also created several notebooks to fine-tune the model on custom data, as well as to use it for inference. Demo notebooks ca…
-
### Overview
The current table extraction model doesn't perform well on "paragraph" layout types. This may be an error in the way that the filing is converted to HTML. Try excluding these filings e…
-
I am getting error
For all sequences i have len(label_ids)==512 but for 1 example i am getting a length of 513.
Assertion Error, assert(len(label_ids)==max_seq_length)
What is the reason behind thi…