XerXesXu
Virgin' on literate.
- Joined
- Oct 18, 2011
- Posts
- 1,865
Further to which, apparently requiring COT models to show their working and rewarding them for the correct procedure, even if they give an incorrect result reduces hallucinations as well as providing other benefits.Here's a useful recent paper, including a literature review, on 'hallucinations', how they're caused, and how they can be prevented.
Why Language Models Hallucinate
For the full 90 pages:
Stress Testing Deliberative Alignment
Last edited: