AI ‘Hallucinations’ Explained: Frequency Bias in 200M GPT Model

AI 'Hallucinations' Explained: Frequency Bias in 200M GPT Model

Photo by Google DeepMind on Pexels

Researchers are exploring the intricacies of AI model behavior by focusing on its limitations and responses to specific scenarios. A recent example involves a 200M GPT model, trained on synthetic data, which exhibits a tendency to project knowledge from well-known areas when confronted with unfamiliar concepts. This phenomenon, commonly called ‘hallucinations,’ can be more accurately characterized as ‘frequency bias,’ where the model favors more common associations. A Reddit user highlighted this insightful observation from a live Discord demo, encouraging others to experiment with the model and share their findings. To participate in the discussion and test the model firsthand, join the Discord server: [https://discord.gg/aTbRrQ67ju]. The original Reddit post can be found here: [https://old.reddit.com/r/artificial/comments/1ma2t8a/understanding_model_behavior_through_limitations/]