(no title)
mlpro
|
2 months ago
Not really. If the models are trained on different dataset - like one ViT trained on satellite images and another on medical X-rays - one would expect their parameters, which were randomly initialized to be completely different or even orthogonal.
crooked-v|2 months ago
energy123|2 months ago
The surprising thing is inter-modality shared variation. I wouldn't have bet against it but I also wouldn't have guessed it.
I would like to see model interpretability work into whether these subspace vectors can be interpreted as low level or high level abstractions. Are they picking up low level "edge detectors" that are somehow invariant to modality (if so, why?) or are they picking up higher level concepts like distance vs. closeness?
TheOtherHobbes|2 months ago
The "human" part of that matters. This is all human-made data, collected from human technology, which was created to assist human thinking and experience.
So I wonder if this isn't so much about universals or Platonic ideals. More that we're starting to see the outlines of the shapes that define - perhaps constrict - our own minds.