-
I've been struggling to improve my model's performance using a real-world dataset, but no matter what I try, the results are disappointing:
It seems like using words from documents that sometimes include non-alphanumeric characters messes everything up. The only slight improvement was with backbone freezing, but even then, the accuracy remained low. Any suggestions or similar experiences? |
Beta Was this translation helpful? Give feedback.
Replies: 1 comment 2 replies
-
Well, I finally figured out why this might be happening. My dataset contains very clean and neat images of words, but I forgot to consider the impact of noise. When running inference on noisy documents, the model struggles to recognize the words properly, which makes sense because it was only trained on pristine examples. After testing inference on a clean document, the model performed well. Lesson learned on my part! Hope this helps someone facing similar issues in the future. |
Beta Was this translation helpful? Give feedback.
Well, I finally figured out why this might be happening. My dataset contains very clean and neat images of words, but I forgot to consider the impact of noise.
When running inference on noisy documents, the model struggles to recognize the words properly, which makes sense because it was only trained on pristine examples. After testing inference on a clean document, the model performed well.
Lesson learned on my part! Hope this helps someone facing similar issues in the future.