5/21/2025
at
8:02:58 AM
My understanding is that your levers are roughly better / more diverse embeddings or computing more embeddings (embed chunks / groups / etc) + aggregating more cosine similarities / scores. More flops = better search w/ steep diminishing returnsColbert being a good google-able application of utilizing more embeddings.
Search ends up often being a funnel of techniques. Cheap and high recall for phase 1 and ratchet up the flops and precision in
subsequent passes on the previous result set.
by forrestp
5/21/2025
at
2:25:02 PM
Exactly! A near property of the matryoshka embeddings is that you can compute a low dimension embedding similarity really fast and then refine afterwards.
by 0101111101