Jaahas Jaahas

jaahas
Β·

AI & ML interests

None yet

Recent Activity

Organizations

None yet

jaahas's activity

New activity in barisaydin/ollama 4 months ago

Create scripts/

#1 opened 4 months ago by
jaahas

Create scripts/

#1 opened 4 months ago by
jaahas
reacted to clefourrier's post with 🀯 10 months ago
view post
Post
2223
Fun fact about evaluation, part 2!

How much do scores change depending on prompt format choice?

Using different prompts (all present in the literature, from Prompt question? to Question: prompt question?\nChoices: enumeration of all choices\nAnswer: ), we get a score range of...

10 points for a single model!
Keep in mind that we only changed the prompt, not the evaluation subsets, etc.
Again, this confirms that evaluation results reported without their details are basically bullshit.

Prompt format on the x axis, all these evals look at the logprob of either "choice A/choice B..." or "A/B...".

Incidentally, it also changes model rankings - so a "best" model might only be best on one type of prompt...