3 Comments
User's avatar
Ahmed Besbes's avatar

you could retrain the two embedding models jointly by minimzing the L2 distance between the embeddings of the same input and maximizing it for different inputs.

this training can be achieved using a contrastive loss.

Expand full comment
Eddy Giusepe's avatar

That's an interesting point... Thank you, AVI CHAWLA !

Expand full comment
Benison Sam's avatar

Would you do a dimensional reduction, say PCA before or after concatenation?

Expand full comment