Dr. Alok Aggarwal

AI Systems and Their Hallucinations

Introduction

OpenAI’s Transformer, GPT-3 with 175 billion parameters, was trained on approximately 500 billion tokens (i.e., approximately 365 billion words in English). Keeping this in view, we posed the following question: “Is there anything distinctive about the roof of the Massachusetts Institute of Technology?” To this, GPT-3’s answer was surprising — “there is nothing distinctive about the roof of the Massachusetts Institute of Technology building. It is made of grass.” When we asked the same question again, GPT-3 replied, “the roof of the Massachusetts Institute of Technology is made of red clay tiles.”
The figure given below shows the dome that is distinctive of MIT’s roof, but neither grass nor red clay tiles are anywhere to be seen.

What are Machine Hallucinations?

Indeed, we would have been satisfied with a typical human response like, “I don’t know” or “I have never seen this roof even in pictures,” but the wrong and yet confident response from GPT-3 was shocking.

If humans provided answers like those provided above by ChatGPT, then such humans would be referred to as hallucinating. Using this analogy, AI professionals have called such responses from AI systems as “Machine Hallucinations”, of which another example is given below.

Examples of Machine Hallucinations

Falcon-40B is an open-source Generative Pre-trained Transformer (GPT) with 40 billion parameters. Unfortunately, like most other GPTs of similar ilk, it occasionally provides the right answer the first time and a wrong one the second time (or vice versa). For example, we asked the following question:

  • Question: “Which of the following is a mammal: a) snake, b) eagle, c) dolphin, or d) frog?”
  • Falcon-40B’s first answer is correct: “A mammal is a member of the class of vertebrate animals that
    have hair or fur and nourish their young with milk produced by mammary glands. Therefore, the
    correct answer is ‘c) dolphin’.”
  • Falcon-40B’s second answer to the same question is incorrect: “All of the animals listed are
    mammals except for the eagle. Eagles are birds.”

The Risks of Machine Hallucinations

Unfortunately, all DLNs including GPTs are unexplainable and uninterpretable. Hence, they become highly untrustworthy when they provide inconsistent or wrong answers. This characteristic hinders their extensive use especially because it forces humans to spend substantial time to check the veracity of their answers or be duped into accepting wrong ones. Unsurprisingly, researchers are working hard to develop ways to reduce brittleness, hallucinations, and the inconsistent behavior of DLNs; and these are discussed further in the upcoming book titled, “The Fourth Industrial Revolution and 100 Years of AI (1950-2050)”. One approach is to create an ensemble of two or more different DLNs (for example two 2 GPTs) whereas another is to combine a Transformer and an Internet search engine that can provide a reference for GPT’s answer; it is unclear whether such rudimentary approaches would work.

The book titled “The Fourth Industrial Revolution and 100 Years of AI (1950-2050) will be published in December 2023. For details, see www.scryai.com/book

Blog Written by

Dr. Alok Aggarwal

CEO, Chief Data Scientist at Scry AI
Author of the book The Fourth Industrial Revolution
and 100 Years of AI (1950-2050)