How can it never hallucinates when retrieval accuracy is at 95% instead of 100% ? #244
-
|
I'm curious on how you did it. |
Beta Was this translation helpful? Give feedback.
Answered by
ArnavAgrawal03
Sep 22, 2025
Replies: 1 comment 1 reply
-
|
We grade hallucinations and abstentions separately. So for example: saying "I don't know" is not a hallucination since it doesn't give the user the wrong answer. However, ideally the system would just respond with the correct answer. As a result, for some questions, "I don't know" gets a 0/1, even though it is not a hallucination. Let me know if this makes sense @vikyw89! |
Beta Was this translation helpful? Give feedback.
1 reply
Answer selected by
ArnavAgrawal03
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
We grade hallucinations and abstentions separately. So for example: saying "I don't know" is not a hallucination since it doesn't give the user the wrong answer. However, ideally the system would just respond with the correct answer.
As a result, for some questions, "I don't know" gets a 0/1, even though it is not a hallucination.
Let me know if this makes sense @vikyw89!