Abstract

Representing images or videos as object-level feature vectors, rather than pixel-level feature maps, facilitates advanced visual tasks. Object-Centric Learning (OCL) primarily achieves this by reconstructing the input under the guidance of Variational Autoencoder (VAE) intermediate representation to drive so-called slots to aggregate as much object information as possible. However, existing VAE guidance does not explicitly address that objects can vary in pixel sizes while models typically excel at specific pattern scales. We propose Multi-Scale Fusion (MSF) to enhance VAE guidance for OCL training. To ensure objects of all sizes fall within VAE’s comfort zone, we adopt the image pyramid, which produces intermediate representations at multiple scales; To foster scale-invariance/variance in object super-pixels, we devise inter/intra-scale fusion, which augments low-quality object super-pixels of one scale with corresponding high-quality superpixels from another scale. On standard OCL benchmarks, our technique improves mainstream methods, including state-of-the-art diffusion-based ones. The source code is available on https://github.com/Genera1Z/MultiScaleFusion.
Original languageEnglish
Title of host publicationThe Thirteenth International Conference on Learning Representations
Number of pages17
Publication statusAccepted/In press - 2025
MoE publication typeA4 Conference publication
EventInternational Conference on Learning Representations - Singapore, Singapore, Singapore
Duration: 24 Apr 202528 Apr 2025
Conference number: 13
https://iclr.cc/

Conference

ConferenceInternational Conference on Learning Representations
Abbreviated titleICLR
Country/TerritorySingapore
CitySingapore
Period24/04/202528/04/2025
Internet address

Fingerprint

Dive into the research topics of 'Multi-Scale Fusion for Object Representation'. Together they form a unique fingerprint.
  • Science-IT

    Hakala, M. (Manager)

    School of Science

    Facility/equipment: Facility

Cite this