spot_img

Date:

Share:

Ranked: AI Models With the Lowest Hallucination Rates

As AI-powered tools and applications become more integrated into our daily lives, it’s important to keep in mind that models may sometimes generate incorrect information.

This phenomenon, known as “hallucinations,” is described by IBM as occurring when a large language model (LLM)—such as a generative AI chatbot or computer vision tool—detects patterns or objects that do not exist or are imperceptible to humans, leading to outputs that are inaccurate or nonsensical.

This chart visualizes the top 15 AI large language models with the lowest hallucination rates.

The hallucination rate is the frequency that an LLM generates false or unsupported information in its outputs.

The data comes from Vectara and is updated as of Dec. 11, 2024. Hallucination rates were calculated by summarizing 1,000 short documents with each LLM and using a model to detect hallucinations, yielding a percentage of factually inconsistent summaries.

Which AI Models Have the Lowest Hallucination Rates?

Below, we show the top 15 AI models with the lowest hallucination rates, their company, and their country of origin.

ModelCompanyCountryHallucination Rate
Zhipu AI GLM-4-9B-ChatZhipu AI🇨🇳 China1.3%
Google Gemini-2.0-Flash-ExpGoogle🇺🇸 United States1.3%
OpenAI-o1-miniOpenAI🇺🇸 United States1.4%
GPT-4oOpenAI🇺🇸 United States1.5%
GPT-4o-miniOpenAI🇺🇸 United States1.7%
GPT-4-TurboOpenAI🇺🇸 United States1.7%
GPT-4OpenAI🇺🇸 United States1.8%
GPT-3.5-TurboOpenAI🇺🇸 United States1.9%
DeepSeek-V2.5DeepSeek🇨🇳 China2.4%
Microsoft Orca-2-13bMicrosoft🇺🇸 United States2.5%
Microsoft Phi-3.5-MoE-instructMicrosoft🇺🇸 United States2.5%
Intel Neural-Chat-7B-v3-3Intel🇺🇸 United States2.6%
Qwen2.5-7B-InstructAlibaba Cloud🇨🇳 China2.8%
AI21 Jamba-1.5-MiniAI21 Labs🇮🇱 Israel2.9%
Snowflake-Arctic-InstructSnowflake🇺🇸 United States3.0%

Smaller or more specialized models, such as Zhipu AI GLM-4-9B-Chat, OpenAI-o1-mini, and OpenAI-4o-mini have some of the lowest hallucination rates among all models. Intel’s Neural-Chat 7B is also a smaller model.

According to Vectara, small-size models can “achieve hallucination rates comparable or even better (lower) than LLMs that are much larger in size.”

Measuring hallucination rates is becoming increasingly critical as AI systems are deployed in high-stakes applications across fields such as medicine, law, and finance.

While larger models generally outperform smaller ones and are continually scaled up for better results, they come with drawbacks like high costs, slow inference, and complexity.

Smaller models, however, are closing the gap, with many performing well on specific tasks. For example, a study showed that the smaller Mistral 8x7B model successfully reduced hallucinations in AI-generated text.

In terms of foundational models, Google’s Gemini 2.0 slightly outperforms OpenAI GPT-4 with a hallucination rate difference of just 0.2%.

However overall, several variants of GPT-4 (e.g., Turbo, Mini, Standard) fall within the 1.5%–1.8% range, highlighting a strong focus on accuracy across different tiers of the same architecture.

This article was original published by Visual Capitalist By Kayla Zhu

Article/Editing:

Graphics/Design:

Sabrina Lam

spot_img
spot_img

━ More like this

Samsung, DBE Announce Top 20 Schools Selected to Advance to the Next Level of 2026 Solve For Tomorrow Competition

At an event that took place on 14 April 2026, Samsung in partnership with the Department of Basic Education (DBE) officially announced the Top 20 schools...

Education for a connected world: Preparing students for global careers

The world of work is changing fast. Careers no longer sit neatly within a single industry, city, or even country; they span disciplines, time...

New SA platform takes on AI’s biggest classroom problem – students getting answers without understanding them

South Africa’s newest AI tutoring platform is designed to help students think, understand, and grow – rather than simply find answers. Ask any South African teacher...

Kaspersky warns of digital medicine risks on the occasion of World Health Day

On World Health Day, Kaspersky warns of risks tied to the digitisation of healthcare and use of telemedicine. Recent incidents show that medical services...

500 young women equipped with future ready AI and cloud computing skills

Cell C has again partnered with GirlCode to train and certify 500 unemployed South African youth in Artificial Intelligence (AI) and cloud computing, strengthening access...
spot_img