Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

You are misreading the benchmark.

https://artificialanalysis.ai/#aa-omniscience-hallucination-...

If you look at the results 3.0 hallucinates an awful lot, when it's wrong.

It's just not wrong that often.

(And it looks like 3.1 does better on both fronts)

 help



Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: