This approach seems to require knowledge of which LLM was used to generate the given text. I wonder if e.g. model fine-tuning - as already provided by Open AI [0] - could evade this detection approach.
Awesome thread by one of the authors of paper, concise and insightful, thanks for sharing !
Research was aiming to bring a tool/approach on distinguishing text from LLM and other sources but in the end of the day it will only benefit those with non-open-source LLMs adjust to such technique and "fool" better everyone else (cause we need log-probs out of a model on each of the sample text).
It seems kinda ironic for me, maybe i missed some crucial point here.
[+] [-] eclipsetheworld|3 years ago|reply
[0] https://beta.openai.com/docs/guides/fine-tuning
[+] [-] mdorazio|3 years ago|reply
[+] [-] O__________O|3 years ago|reply
https://towardsdatascience.com/understanding-auc-roc-curve-6...
[+] [-] gault8121|3 years ago|reply
I created AIwritingcheck.org to provide teachers with a user friendly interface for this model.
[+] [-] SachinDSI|3 years ago|reply
[+] [-] andai|3 years ago|reply
It's been a while since I used it but I very rarely got plausible output from it.
[+] [-] eh9|3 years ago|reply
[+] [-] O__________O|3 years ago|reply
https://arxiv.org/abs/2301.11305
Additional explanation:
https://twitter.com/_eric_mitchell_/status/16188203614199152...
[+] [-] m00viin_pics|3 years ago|reply
Research was aiming to bring a tool/approach on distinguishing text from LLM and other sources but in the end of the day it will only benefit those with non-open-source LLMs adjust to such technique and "fool" better everyone else (cause we need log-probs out of a model on each of the sample text).
It seems kinda ironic for me, maybe i missed some crucial point here.