00:00:00

Share Your Feedback 🏝️

Scaling Up using RL on Synthesized Data

Scaling Up using RL on Synthesized Data

MinWoo(Daniel) Park | Tech Blog

Read more
Previous: Self-Tuning Next: Magpie Scratch Data Synthesis

Scaling Up using RL on Synthesized Data

  • Related Project: Private
  • Category: Paper Review
  • Date: 2024-06-17

Beyond Model Collapse: Scaling Up with Synthesized Data Requires Reinforcement

  • url: https://arxiv.org/abs/2406.07515v1
  • pdf: https://arxiv.org/pdf/2406.07515v1
  • html https://arxiv.org/html/2406.07515v1
  • abstract: Synthesized data from generative models is increasingly considered as an alternative to human-annotated data for fine-tuning Large Language Models. This raises concerns about model collapse: a drop in performance of models fine-tuned on generated data. Considering that it is easier for both humans and machines to tell between good and bad examples than to generate high-quality samples, we investigate the use of feedback on synthesized data to prevent model collapse. We derive theoretical conditions under which a Gaussian mixture classification model can achieve asymptotically optimal performance when trained on feedback-augmented synthesized data, and provide supporting simulations for finite regimes. We illustrate our theoretical predictions on two practical problems: computing matrix eigenvalues with transformers and news summarization with large language models, which both undergo model collapse when trained on model-generated data. We show that training from feedback-augmented synthesized data, either by pruning incorrect predictions or by selecting the best of several guesses, can prevent model collapse, validating popular approaches like RLHF.
Previous: Self-Tuning Next: Magpie Scratch Data Synthesis

post contain ""

    No matching posts found containing ""