(no title)
Timsky | 4 months ago
No, it does not. It only produces a highly convincing counterfeit. I am honestly happy for people who are satisfied with its output: life is way easier for them than for me. Obviously, the machine discriminates me personally. When I spend hours in the library looking for some engineering-related math made in the 70s-80s, as a last resort measure, I can try to play this gambling with chat, hoping for any tiny clue to answer my question. And then for the following hours, I am trying to understand what is wrong with the chat output. Most often, I experience the "it simply can't be" feeling, and I know I am not the only one having it.
crazygringo|4 months ago
Of the other 50% that are real, it's often ~evenly split into sources I'm familiar with and sources I'm not.
So it's hugely useful in surfacing papers that I may very well never have found otherwise using e.g. Google Scholar. It's particularly useful in finding relevant work in parallel subfields -- e.g. if you work in physics but it turns out their are math results, or you work in political science and it turns out there are relevant findings from anthropology. And also just obscure stuff -- a random thesis that never got published or cited but the PDF is online and turns out to be relevant.
It doesn't matter if 75% of the results are not useful to me or hallucinated. Those only waste me minutes. The other 25% more than make up for it -- they're things I simply might never find otherwise.
andrewflnr|4 months ago
macrolime|4 months ago
scosman|4 months ago
Should you blindly trust the summary? No. Should you verify key claims by clicking through to the source? Yes. Is it still incredibly useful as a search tool and productivity booster? Absolutely.
scruple|4 months ago
Timsky|4 months ago
bootsmann|4 months ago
andai|4 months ago
But then they flip to the next page and they read a story on a subject they're not an expert on and they just accept all of it without question.
I think people might have a similar relationship with ChatGPT.
btrettel|4 months ago
lukev|4 months ago
happy_dog1|4 months ago
I don't know if something like this already exists and I'm just not aware of it to be fair.
Timsky|4 months ago
Currently, I am applying RDF/OWL to describe some factual information and contradictions in the scientific literature. On an amateur level. Thus I do it mostly manually. The GPT-discourse somehow brings up not only the human-related perception problems, such as cognitive biases, but also truly philosophical questions of epistemology that should be resolved beforehand. LLM developers cannot solve this because it is not under their control. They can only choose what to learn from. For instance, when we consider a scientific text, it is not an absolute truth but rather a carefully verified and reviewed opinion that is based on the previous authorized opinions and subject to change in the future. So the same author may have various opinions over time. More recent opinions are not necessarily more "truthful" ones. Now imagine a corresponding RDF triple (subject-predicate-object tuple) that describes that. Pretty heavy thing, and no NLTK can decide for us what the truth is and what is not.
opdahl|4 months ago
Basically we are trying to combine the benefits of chat with normal academic search results using semantic search and keyword search. That way you get the benefit of LLMs but you’re actually engaging with sources like a normal search.
Hope it was what you were looking for!
unknown|4 months ago
[deleted]
astrange|4 months ago
malshe|4 months ago
I have a recent example where it helped me locate a highly relevant paper for my research. It was from an obscure journal and wouldn't show up in the first few pages of Google Scholar search. The paper was real and recently published.
However, using LLMs for doing lit review has been fraught with peril. LLMs often misinterpret the research findings or extrapolate them to make incorrect inferences.
kianN|4 months ago
Example: https://platform.sturdystatistics.com/deepdive?search_type=e...
Timsky|4 months ago
glenstein|4 months ago
khuston|4 months ago
As to not trusting the generated text, you’re totally right. That’s why I use it as a search tool but mostly ignore the content of what the LLM has to say and go to the source.
cj|4 months ago
It correctly described the locations in text, then it offered to provide a diagram.
I said “sure”, and it generated an image saying the chest location is on the neck, and a bunch of other clearly incorrect locations for the other measurement sites.
It’s gotten better. But it’s still bad.