A disentangled representation is a latent vector where distinct, semantically meaningful factors of variation in the data are encoded in separate and independent dimensions. This is a primary objective in models like variational autoencoders (VAEs), where the goal is to learn a structured latent space where single latent units correspond to single generative factors, such as object color, size, or orientation in an image dataset. Achieving this separation allows for more interpretable, controllable, and generalizable models, as manipulating one latent dimension yields a predictable and isolated change in the generated output.
