top | item 22945690

Understanding BERT and Search Relevance (2019)

78 points| martinlaz | 5 years ago |opensourceconnections.com | reply

18 comments

order
[+] binarymax|5 years ago|reply
Hey I wrote this about 6 months ago, nice to see it here! AMA, but please note that this is SoTA territory and things have changed significantly since then. Notably, folks are now seeing good preliminary results with SBERT (sentence level encodings instead of token-level): https://www.aclweb.org/anthology/D19-1410.pdf
[+] petulla|5 years ago|reply
How did you end up handling the query/document asymmetry issue? Seems like query sentenceBERT/averaged document vectors?
[+] lootsauce|5 years ago|reply
Any experience clustering or classifying documents based on these high dimensional vectors? Also what have you found of dimension reduction techniques such as UMAP or good old PCA?