Can one train a top down centered instance model without retraining an outstanding centroid model every time? #1291
-
Hello, I am doing a lot of training and retraining of a top down model with only a single instance on Google Colab. The centroid model is accurately trained enough every time I go through training, but of course the centered instance model is what needs more training. I understand that if I generate new training data by correcting predicted instances, that must be added to the centroid model or else the centered instance model will have no centroid on those new training frames to crop and use as input for training. However, sometimes if I find that the previous training session did not improve the predictions, then I might simple change some hyperparameters (like different types of augmentation, adjustment of hard keypoints for OHKM, or the number of filters in the model, etc.) on the centered instance model and run the training again to see if that helps improve accuracy. I ask two adjacent questions: When retraining the top down model with no new training frames: If I have an outstanding centroid model, and I have generated additional training data in order to retrain the centered instance model, is there a way to use the previous centroid model to simply predict (rather than train) the centroids in the new training subset and add those predicted centroid frames to the input group of training frames for the centered instance model training set? This idea may be flawed, but ultimately, I want to not have to 'train' an already trained centroid model and still be able to retrain the centered instance model with either different hyperparameters or new training data. Thanks for your help! |
Beta Was this translation helpful? Give feedback.
Replies: 2 comments 5 replies
-
Hi @amblypatty, Yes! You can use opt to just use the trained model instead of retraining. There should be a "Use Trained Model" checkbox that pops up if you select an already trained model. Let us know if this helps! Thanks, |
Beta Was this translation helpful? Give feedback.
-
Hi @roomrys, I am still having some trouble with retraining my model when I have new annotations, where the training and validation loss doesn't descend. However, I am using the same hyperparameters as the previous training session. I feel confident that the new annotations I've made should help the model learn and make more accurate predictions (I specifically worked on fixing the erroneous predictions, nearly doubling the number of training annotations (from 164 labeled frames to 300 labeled frames) as the previous training session). The following are my general steps:
Am I doing something wrong, like not actually enacting the centroid model? Additionally, should I be adjusting the learning rate when retraining the centered_instance model to enable better loss and val_loss evaluation? If so, in which direction and by how much would one expect to adjust it? For example, if the first training session had an initial learning rate of 0.0001 and stopped at 1.5e^-06, would I adjust the initial learning rate of the retraining configuration up to 0.0005, 0.001, or down to something like 0.00005? |
Beta Was this translation helpful? Give feedback.
Hi @amblypatty,
Apologies for the delay! Let me tackle some of these:
👍
We're chasing down a bug right now where it looks like this functionality might not be working from the GUI due to how we handle the CLI vs config specification of resuming training. We're working on a hotfix that we hope will be ready for release next week.
In the meantime, you can try using the
--base_ch…