Skip to main content

Model Limitations

Updated Jan 01, 2025 ·

Overview

Every model, no matter how advanced, has limitations due to the data it's trained on. Recognizing these helps in crafting prompts that avoid these pitfalls and evaluating outputs critically.

The Reversal Curse

The reversal curse shows that ChatGPT’s knowledge is often one-dimensional, requiring questions from a specific direction for accurate answers

  • Example: Ask ChatGPT, "Who is Tom Cruise’s mother?" It correctly answers "Mary Lee Pfeiffer"
  • Issue: Ask "Who is Mary Lee Pfeiffer’s son?" and ChatGPT may not know

Biases

Language models learn from extensive internet data and may reflect societal biases

  • Example: Asking "Who typically cooks in a household?" might yield a gendered response
  • Solution: Encourage neutral responses, recognizing that anyone can perform tasks regardless of gender

Hallucinations

Hallucinations occur when the model provides inaccurate information confidently.

  • Example: "Who was the only survivor of the sinking of the Titanic?" ChatGPT might incorrectly respond with "Violet Jessop"
  • Solution: Ask for sources to prompt the model to correct itself. Always cross-reference answers

Overfitting - Echoing the Data

Overfitting means the model mirrors its training data too closely, reducing its ability to generalize.

  • Example: Asking for a joke often results in the same 25 jokes out of many attempts
  • Insight: The model struggles with tasks requiring creative leaps, like writing new jokes or scientific hypotheses

Focus on incremental tasks like writing summaries, answering questions, or imitating writing styles using one-shot or few-shot learning techniques