MINR: Implicit Neural Representations with Masked Image Modelling
By: Sua Lee, Joonhun Lee, Myungjoo Kang
Potential Business Impact:
Teaches computers to see better, even new things.
Self-supervised learning methods like masked autoencoders (MAE) have shown significant promise in learning robust feature representations, particularly in image reconstruction-based pretraining task. However, their performance is often strongly dependent on the masking strategies used during training and can degrade when applied to out-of-distribution data. To address these limitations, we introduce the masked implicit neural representations (MINR) framework that synergizes implicit neural representations with masked image modeling. MINR learns a continuous function to represent images, enabling more robust and generalizable reconstructions irrespective of masking strategies. Our experiments demonstrate that MINR not only outperforms MAE in in-domain scenarios but also in out-of-distribution settings, while reducing model complexity. The versatility of MINR extends to various self-supervised learning applications, confirming its utility as a robust and efficient alternative to existing frameworks.
Similar Papers
MINR: Efficient Implicit Neural Representations for Multi-Image Encoding
CV and Pattern Recognition
Saves space by sharing computer brain parts.
Learning Mask Invariant Mutual Information for Masked Image Modeling
CV and Pattern Recognition
Helps computers learn to see better by focusing on important details.
MIMRS: A Survey on Masked Image Modeling in Remote Sensing
CV and Pattern Recognition
Fixes cloudy satellite pictures for better maps.