Method Overview
PE generation
To generate high-quality implicit neural representations, we propose the generation of adaptive positional embeddings (PEs) using a diffusion model instead of generating the weights of INRs. This approach shifts the primary expressive power from MLPs to PEs, which we have observed to result in finer detailed generation results. To further enhance expressive capacity, we hierarchically decompose the PEs into multiple scales (HDBFs) and modulate the MLPs in a coarse-to-fine manner (CFC).
DDMI
In the first stage, we learn the latent space of continuous signals using our D2C-VAE framework. We then approximate the distribution of this latent space with the latent diffusion model. After training, DDMI generates hierarchically decomposed positional embeddings (HDBFs), from which the MLPs read out the signal values for given coordinates.