r=1; batch_size=32;
| > Number of instances : 9489
with a max len of 200,
| > Num. instances discarded by max-min seq limits: 684
-OOM
with a max len of 150,
| > Num. instances discarded by max-min seq limits: 1610
-OOM
with a max len of 100,
| > Num. instances discarded by max-min seq limits: 3591
-OOM
with a max len of 50,
| > Num. instances discarded by max-min seq limits: 7338
works but i’ve lost 80% of my data.
I think I am going to have to look into the dynamic batch size hack.
For reference:
Using 0,1,2 at max len 50(will go as high as i can)
+-----------------------------------------------------------------------------+
| NVIDIA-SMI 430.26 Driver Version: 430.26 CUDA Version: 10.2 |
|-------------------------------+----------------------+----------------------+
| GPU Name Persistence-M| Bus-Id Disp.A | Volatile Uncorr. ECC |
| Fan Temp Perf Pwr:Usage/Cap| Memory-Usage | GPU-Util Compute M. |
|===============================+======================+======================|
| 0 TITAN V Off | 00000000:18:00.0 Off | N/A |
| 56% 77C P2 105W / 250W | 8813MiB / 12066MiB | 67% Default |
+-------------------------------+----------------------+----------------------+
| 1 TITAN V Off | 00000000:3B:00.0 Off | N/A |
| 59% 82C P2 131W / 250W | 5823MiB / 12066MiB | 58% Default |
+-------------------------------+----------------------+----------------------+
| 2 TITAN Xp Off | 00000000:86:00.0 Off | N/A |
| 51% 83C P2 91W / 250W | 5981MiB / 12196MiB | 83% Default |
+-------------------------------+----------------------+----------------------+
| 3 TITAN Xp Off | 00000000:AF:00.0 On | N/A |
| 32% 51C P5 26W / 250W | 1095MiB / 12193MiB | 0% Default |
+-------------------------------+----------------------+----------------------+