clarification on augreg2 models #2420
-
Hi, |
Beta Was this translation helpful? Give feedback.
Replies: 1 comment 3 replies
-
@mueller-mp yes, I re-did the fine-tune from the original in21k checkpoint, mostly to show lucas that they could be better :) The biggest difference was that these fine-tunes used the timm scripts & augmentations (original pretrained & fine-tunes were using the google jax train code). Using layer-wise LR decay was the biggest single hparam change, will see if I have that config files somewhere... |
Beta Was this translation helpful? Give feedback.
@mueller-mp yes, I re-did the fine-tune from the original in21k checkpoint, mostly to show lucas that they could be better :)
The biggest difference was that these fine-tunes used the timm scripts & augmentations (original pretrained & fine-tunes were using the google jax train code). Using layer-wise LR decay was the biggest single hparam change, will see if I have that config files somewhere...