A fast, easy-to-use, production-ready inference server for computer vision supporting deployment of many popular model architectures and fine-tuned models.
[X] I have searched the Inference issues and found no similar feature requests.
Description
DocTR produces not only text, but also location of the text. Our implementation linearises all of text data, skipping its location and not providing blocks-of-texts info.
We should think about:
making results of OCR models more generic (and apply that for inference server)
making changes into workflows blocks with OCR models to reflect those
Search before asking
Description
DocTR
produces not only text, but also location of the text. Our implementation linearises all of text data, skipping its location and not providing blocks-of-texts info.We should think about:
inference
server)workflows
blocks with OCR models to reflect thoseUse case
No response
Additional
No response
Are you willing to submit a PR?