The overall idea behind generative models is to learn a distribution from some training samples, in order to generate new samples by sampling from that distribution.
For instance if I sample an image from the space of all possible images in , most of the time I will get an image that’s only noise. The images that are natural and make sense are way less than the noisy images, and we assume they live in a portion of the entire space, and we are only interested in this subspace (this is the Manifold hypothesis).
Learning a distribution means that natural images will have more probability to be sampled with respect to noisy images that do not make sense in the real world.
Examples of generative models are: