Description
Is your feature request related to a problem? Please describe.
Currently when changing ViT img size from a rectangular size, resample_abs_pos_embed()
does not work correctly since it does not know the original rectangular size and assume a square.
pytorch-image-models/timm/models/vision_transformer.py
Lines 1096 to 1103 in 5dce710
pytorch-image-models/timm/layers/pos_embed.py
Lines 32 to 34 in 5dce710
Describe the solution you'd like
It should work out of the box.
Describe alternatives you've considered
Manually resize it.
Additional context
Apparently dynamic img size also will not work when original img size is rectangle.
pytorch-image-models/timm/models/vision_transformer.py
Lines 603 to 609 in 5dce710
This is a rare problem since most image ViT use square inputs. The particular model I'm using is my previously ported AudioMAE (https://huggingface.co/gaunernst/vit_base_patch16_1024_128.audiomae_as2m), which uses rectangular input (mel-spectrogram).
I understand it is not so straight-forward to support this, since once the model is created (with updated image size), the original image size is lost. Some hacks can probably bypass this, but not so nice
- Propagate the original image size to the
_load_weights()
function - Create a model with the original image size, load weights as usual. Add a new method like
.set_img_size()
which will update the internalimg_size
attribute and resamle pos embed.
Perhaps an easier solution is to fix dynamic img size to pass the original img size (which I tested locally and works)
if self.dynamic_img_size:
B, H, W, C = x.shape
pos_embed = resample_abs_pos_embed(
self.pos_embed,
(H, W),
self.patch_embed.grid_size,
num_prefix_tokens=0 if self.no_embed_class else self.num_prefix_tokens,
)