Replies: 1 comment 3 replies
-
Hi @dasantosa 👋, It's the highest prio this year to make docTR multilingual but this will still take some time. We have already a ticket for this: #1699 and #988 You are right I trained only parseq on this dataset but the good thing the used dataset for fine tuning is 100% synth generated so I can share it and you can fine tune vitstr on it on your own :) dataset: https://drive.google.com/file/d/1TNQN8uBMiGjzf2GM41BWDICefLubau5q/view?usp=sharing |
Beta Was this translation helpful? Give feedback.
3 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Uh oh!
There was an error while loading. Please reload this page.
-
🚀 The feature
I'm using DocTR for text detection and recognition and I'm having trouble with the recognition of the "ñ"/"Ñ" characters. I've been using the vitstr_small model and its default weights which have been trained with the French alphabet.
I've read the issues page for a similar issue and found that there is a multilingual model loaded in HuggingFace. The problem is that these weights are for the ParseQ model and it's a bit slower than VitStr.
My question is, is it possible to load weights for each language instead of a multilingual model to get more efficiency in inferences? Or is there a plan to load multilingual weights for other models?
Thanks!.
Motivation, pitch
Use VitStr models for each language efficiently.
Alternatives
No response
Additional context
No response
Beta Was this translation helpful? Give feedback.
All reactions