-
Hi, I am a little bit confused about the relationship between max_iters and batch_size when using
And I set the train_cfg like this:
In addition, I auto scale the learning rate, as the overall batch = 2 * 2 = 4, not 16:
When I run this experiment, I used dist_train.sh
In the training process, the log prints like this (exp1):
So I think for exp1, the real total iterations (the number of visiting data sample) is 40000 * 4 = 160000, and for exp2, it is 40000 * 2 = 80000. In addition, I find in the log file mask2former_cityscapes, it has the key |
Beta Was this translation helpful? Give feedback.
Replies: 2 comments 1 reply
-
Hi, the best practice for training the model with a smaller batch size or fewer GPUs is:
As for the step arguments you mentioned, you need to get it from the |
Beta Was this translation helpful? Give feedback.
-
Hi @HAOCHENYE , thank you for your answers! I will play with that I understand the relationship about the learning rate and batch size. Suppose I don't care much about the learning rate (e.g. I am training my own model, and I don't know the optimal learning rate yet), and I want to fine tune the value of
I think 1 should be correct. Therefore, in my previous example, I should manually increase the |
Beta Was this translation helpful? Give feedback.
Correct!