Hi, I use setting
single_pass: true to train in a epoch.
My vocab size is about:
In the begining, I set batch_size about 2048 to matche the ram of the GPU (RTX2070 * 2, 8G RAM) and avoid the OOM problem which will happen if batch_size is more than 2048
And then I can set the batch_size up to 2500 and training without OOM in the next epoch.
So can I change the batch_size when trainning in the next epoch? why the batch_size can’t set much more in the beginning?