Reasoning models are now supposed to be the cream of the crop when it comes to accuracy, but that isn’t necessarily the case for the latest buzzy AI.
DeepSeek’s R1 model makes up answers at a significantly higher rate than comparable reasoning and open-source models, according to enterprise AI startup Vectara, which tested R1 and others on its own hallucination-measuring model.
Källa: DeepSeek hallucinates alarmingly more than other AI models
