5/19/2025 at 6:59:58 PM
One of the key intuitions: If you take a natural image and add random noise, you will get a different random noise image every time you do this. However, all of these (different!) random noise images will be lined up in the direction perpendicular to the natural images manifold.So you will always know where to go to restore the original image: shortest distance to the natural image manifold.
How all these random images end up perpendicular to the manifold? High dimensional statistics and the fact that the natural image manifold has much lower dimension than the overall space.
by g42gregory
5/19/2025 at 10:49:07 PM
TIL.Generative Visual Manipulation on the Natural Image Manifold
https://arxiv.org/abs/1609.03552
For me, the most intriguing aspect of LLMs (and friends) are the embedding space and the geometry of the embedded manifolds. Curious if anyone has looked into comparative analysis of the geometry of the manifolds corresponding to distinct languages. Intuitively I see translations as a mapping from one language manifold to another, with expressions being paths on that manifold, which makes me wonder if there is a universal narrative language manifold that captures 'human expression semantics' in the same way as a "natural image manifold".
by yubblegum
5/23/2025 at 5:33:15 AM
I think this is related: https://news.ycombinator.com/item?id=44054425by Ey7NFZ3P0nzAe