×
AI models improve with less human oversight, new study finds
Written by
Published on
Join our daily newsletter for breaking news, product launches and deals, research breakdowns, and other industry-leading AI coverage
Join Now

Artificial intelligence researchers at Hong Kong University and UC Berkeley have discovered that language models perform better when allowed to develop their own solutions through reinforcement learning rather than being trained on human-labeled examples. This finding challenges conventional wisdom about how to best train large language models (LLMs) and vision language models (VLMs).

Key research findings: The study compared supervised fine-tuning (SFT) with reinforcement learning (RL) approaches across both textual and visual reasoning tasks.

  • Models trained primarily through reinforcement learning showed superior ability to generalize to new, unseen scenarios
  • Excessive use of hand-crafted training examples can actually impair a model’s ability to handle novel situations
  • The research used two benchmark tasks: GeneralPoints for testing arithmetic reasoning and V-IRL for evaluating spatial reasoning capabilities

Technical methodology: The researchers conducted their experiments using Llama-3.2-Vision-11B as their base model, implementing a hybrid training approach.

  • Models received initial “warm-up” training using a small supervised fine-tuning dataset
  • Separate versions were created for each task and training method
  • Training was scaled independently for both RL and SFT approaches to compare their effectiveness

Critical results: The study revealed clear advantages of reinforcement learning over traditional supervised training methods.

  • RL-trained models consistently outperformed SFT models when faced with out-of-distribution examples
  • SFT-trained models showed signs of memorizing training rules rather than truly learning to generalize
  • These findings held true across both text-only and multimodal (text and vision) scenarios

Practical implications: The research suggests important considerations for future AI model development and deployment.

  • A small amount of supervised fine-tuning remains valuable for stabilizing model output format
  • Pure reinforcement learning approaches may be particularly valuable for tasks with clearly verifiable results
  • This approach could reduce the cost and effort required to create extensive hand-labeled training datasets

Looking ahead: While these findings challenge established training paradigms, the researchers note important nuances in their results that differ from other recent developments like DeepSeek-R1-Zero, suggesting that the effectiveness of pure RL training may depend on the specific architecture of the base model being used.

Less supervision, better results: Study shows AI models generalize more effectively on their own

Recent News

AI-powered agents poised to upend US auto industry in customers’ favor

Car buyers show strong interest in AI assistance for maintenance alerts and repair verification as dealerships aim to restore consumer confidence.

Eaton’s AI data center stock dips on the arrival of DeepSeek

Market jitters over AI efficiency gains overlook tech giants' continued commitment to data center expansion.

Long story short: Top AI summarizers for articles and documents in 2025

Enterprise-grade AI document summarizers are gaining traction as companies seek to cut down the 20% of work time spent organizing information.