00:00:00

Share Your Feedback 🏝️

A is B, B is not A**

A is B, B is not A**

MinWoo(Daniel) Park | Tech Blog

Read more
Previous: Star Coder Next: Logic CoT

A is B, B is not A**

  • Related Project: Private
  • Category: Paper Review
  • Date: 2023-09-26

The Reversal Curse: LLMs trained on “A is B” fail to learn “B is A”

  • url: https://arxiv.org/abs/2309.12288v2
  • pdf: https://arxiv.org/pdf/2309.12288v2
  • abstract: We expose a surprising failure of generalization in auto-regressive large language models (LLMs). If a model is trained on a sentence of the form “A is B”, it will not automatically generalize to the reverse direction “B is A”. This is the Reversal Curse. For instance, if a model is trained on “Olaf Scholz was the ninth Chancellor of Germany”, it will not automatically be able to answer the question, “Who was the ninth Chancellor of Germany?”. Moreover, the likelihood of the correct answer (“Olaf Scholz”) will not be higher than for a random name. Thus, models exhibit a basic failure of logical deduction and do not generalize a prevalent pattern in their training set (i.e. if “A is B’’ occurs, “B is A” is more likely to occur). We provide evidence for the Reversal Curse by fine-tuning GPT-3 and Llama-1 on fictitious statements such as “Uriah Hawthorne is the composer of ‘Abyssal Melodies’” and showing that they fail to correctly answer “Who composed ‘Abyssal Melodies?’”. The Reversal Curse is robust across model sizes and model families and is not alleviated by data augmentation. We also evaluate ChatGPT (GPT-3.5 and GPT-4) on questions about real-world celebrities, such as “Who is Tom Cruise’s mother? [A: Mary Lee Pfeiffer]” and the reverse “Who is Mary Lee Pfeiffer’s son?”. GPT-4 correctly answers questions like the former 79% of the time, compared to 33% for the latter. This shows a failure of logical deduction that we hypothesize is caused by the Reversal Curse. Code is available at this https URL.

TL;DR

  • Generalization Failure in LLMs: Identified as “curse of inversion.”
  • Issue: Models don’t infer “B is A” from “A is B.”
  • Example: Can’t deduce “9th Chancellor of Germany” from “Olaf Scholz is the 9th Chancellor.”
  • Implication: Lacks basic logical reasoning and pattern generalization.
  • Scope: Widespread across various model sizes and families.
  • Data Augmentation: Ineffective in mitigating the issue.
  • Test Case: Lower accuracy in inverse questions about public figures in GPT-3.5 and GPT-4.
  • Code: Available at provided URL.


Previous: Star Coder Next: Logic CoT

post contain ""

    No matching posts found containing ""