Poor text detection performance after training fast_base on custom dataset #1704
Replies: 3 comments 4 replies
-
Another additional detail is that during the training process, the loss will drop from 0.8 to 0.4 after 2-3 epochs, and then stop decreasing. |
Beta Was this translation helpful? Give feedback.
-
Hi @GrayChan813 👋🏼, Best regards |
Beta Was this translation helpful? Give feedback.
-
Beta Was this translation helpful? Give feedback.
-
I prepared my custom data according to the training tutorial and trained the fast_base model using the following script:
python references/detection/train_pytorch.py data/train/ data/test/ fast_base --name fast_base_20240819 --epochs 100 --batch_size 16 --device 0 --lr 0.0001 --workers 16 --rotation
After iterating for multiple epochs, I evaluated the performance of the model and the indicators obtained are as follows:
(Validation loss: 0.3952 (Recall: 41.16% | Precision: 1.34% | Mean IoU: 2.00%))
I don’t know where the problem lies. I followed the training instructions completely when preparing the dataset. I also visualized the dataset in text lines's bounding box to confirm that there was no problem with the data format. In addition, I also used the evaluation script "references/detection/evaluate_pytorch.py" to evaluate the performance of my trained model on the FUNSD dataset, and also got a very poor detection performance.
Therefore, I would like to ask if I have overlooked any key steps in the training process, or if some training parameters will affect the performance, such as --ratation, etc. I really hope someone can answer my doubts, thank you!
Beta Was this translation helpful? Give feedback.
All reactions