Extension of the question discussed in #1943 #1992
Replies: 1 comment 1 reply
-
|
Hi @rithika2035 👋 The issue is still that the pretrained models tend to "forget" very quickly. Training for fewer epochs (1–3) and possibly freezing some layers can help. There are two common strategies: either freeze only the backbone, or freeze everything up to the head - freezing up to the head depends on the architecture you want to fine tune. Ideally, you'd combine a subset of the pretraining dataset with your new data. Unfortunately, that's not possible in this case, as the original dataset is a private, internal Mindee resource. As for point 3 - that sounds more like a detection issue. QR codes shouldn’t be detected as text, so fine-tuning the detection model on a few targeted samples might help. Best regards, |
Beta Was this translation helpful? Give feedback.
Uh oh!
There was an error while loading. Please reload this page.
-
Hi @felixdittrich92,
So I was going through all the chats in discussion #1943, and I did follow all the steps and I am able to get the result as intended. But adding the "₹" symbol in the vocab and then training it, just messed up the confidence scores. So when I ran inference for an invoice image the following were the issues I faced:
So is there any way to fix this?
Beta Was this translation helpful? Give feedback.
All reactions