this post was submitted on 26 Oct 2024
1068 points (97.1% liked)
Science Memes
10826 readers
2127 users here now
Welcome to c/science_memes @ Mander.xyz!
A place for majestic STEMLORD peacocking, as well as memes about the realities of working in a lab.
Rules
- Don't throw mud. Behave like an intellectual and remember the human.
- Keep it rooted (on topic).
- No spam.
- Infographics welcome, get schooled.
Research Committee
Other Mander Communities
Science and Research
Biology and Life Sciences
- [email protected]
- [email protected]
- [email protected]
- [email protected]
- [email protected]
- [email protected]
- [email protected]
- [email protected]
- [email protected]
- [email protected]
- [email protected]
- [email protected]
- [email protected]
- [email protected]
- [email protected]
- [email protected]
- [email protected]
- [email protected]
- [email protected]
- [email protected]
- [email protected]
- [email protected]
- [email protected]
- [email protected]
- !reptiles and [email protected]
Physical Sciences
- [email protected]
- [email protected]
- [email protected]
- [email protected]
- [email protected]
- [email protected]
- [email protected]
- [email protected]
- [email protected]
Humanities and Social Sciences
Practical and Applied Sciences
- !exercise-and [email protected]
- [email protected]
- !self [email protected]
- [email protected]
- [email protected]
- [email protected]
Memes
Miscellaneous
founded 2 years ago
MODERATORS
you are viewing a single comment's thread
view the rest of the comments
view the rest of the comments
LLMs can cite. It's called Retrival-Augmented Generation. Basically LLM that can do Information Retrival, which is just academic term for search engines.
You can just print retrival logs into references. Well, kinda stretching definition of "just".
My question is that the thing they are citing actually exists and if it does exist, contains the information it claims.
In case of RAGs it exists in searched dataset.
Not guaranteed.
Depends. In my experience, it usually does exist. Now there are hallucinations where GPT makes up stuff or just misinterprets what it read. But it's super easy to read the GPT output, look at the cited work, skim works for relevance, then tweak the wording and citing to match.
If you just copy/paste and take GPT's word for it without the minimal amount of checking, you're digging your own grave.