You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
I am using TAPIR as backbone architecture, and starting from the checkpoint version. I have to further train it for my case. Do you have any suggestions for what training times to expect from the checkpoint version? or have any fine-tuning stats?
For example, it seems that BootTAP uses pre-trained TAPIR as backbone, I wonder what were the training time stats for it, in case I missed on the paper.
Sorry, we haven't been releasing much on the training loss curves because it's difficult to maintain them. BootsTAPIR took about 2 weeks to train on YouTube data on 256 A100 GPUs, but that's due to the diversity of YouTube. Further finetuning it on Libero took about 3 days (50K steps) on 128 GPUs, but it only took that long because we were jointly training on youtube and kubric as well. I suspect that a much shorter training run will produce similar results, but we haven't done rigorous experiments.
Unfortunately we haven't yet released BootsTAP training code, and it's unclear when we'll be able get to it considering how busy the team is. I'm hoping it'll be before ECCV, but I can't make any guarantees.
thank you for sharing these! I will leave this issue open, as I may follow back here with further questions, plus share my training time stats results, and in case that helps anyone.
I am using TAPIR as backbone architecture, and starting from the checkpoint version. I have to further train it for my case. Do you have any suggestions for what training times to expect from the checkpoint version? or have any fine-tuning stats?
For example, it seems that BootTAP uses pre-trained TAPIR as backbone, I wonder what were the training time stats for it, in case I missed on the paper.
Thank you!
(also noted this #59)
The text was updated successfully, but these errors were encountered: