Self-RAG System
Collection
Instead of having a LLM generate reflection tokens, what if we have a system of models generate reflection tokens?
•
4 items
•
Updated
•
1
This generates the IsSupported
token as descirbed in Self-RAG.
We are testing to see if a generated LLM answer is supported by the document. This is similar to testing for a hallucination in the model result.
The expected input to the model is shown here:
Context: {'doc'}\nAnswer: {answer}"
{'eval_loss': 0.11030498147010803,
'eval_mse': 0.11030498147010803,
'eval_mae': 0.14249496161937714,
'eval_r2': 0.6906673524053266,
'eval_accuracy': 0.9117161716171617}