×
Databricks founder offers $1M to solve AI coding challenges
Written by
Published on
Join our daily newsletter for breaking news, product launches and deals, research breakdowns, and other industry-leading AI coverage
Join Now

The artificial intelligence community has a new challenge with significant implications for the future of coding, as Databricks and Perplexity co-founder Andy Konwinski announces a major prize for advancing AI coding capabilities.

Prize details and objectives: Konwinski is offering a $1 million reward to researchers who can achieve a 90% score on a new AI coding benchmark, with the competition specifically targeting open-source language models.

  • The contest, named K Prize, aims to encourage smaller, independent researchers to develop innovative approaches to AI model training
  • A minimum prize of $50,000 is guaranteed for the top submission, even if it falls short of the 90% threshold
  • The competition focuses exclusively on open models, excluding closed systems from companies like OpenAI and Anthropic

Technical challenge and methodology: The competition addresses a fundamental issue in AI benchmarking while establishing a more rigorous testing framework.

  • Current AI coding benchmarks often suffer from inflated scores due to training data contamination (similar to having test answers in advance)
  • The existing SWE-bench coding test, which uses real-world GitHub problems, has proven challenging for AI models, with the best performers only achieving 55% accuracy
  • To prevent gaming the system, Konwinski is collaborating with SWE-bench and Kaggle to create new test problems that won’t exist until after model submissions

Infrastructure support: The competition organizers are implementing measures to ensure fair participation regardless of resources.

  • Kaggle will provide computing resources to developers who lack access to sufficient GPU power
  • This democratization of resources aims to level the playing field for smaller teams and individual researchers
  • The initiative is personally funded by Konwinski, leveraging his success with Databricks (valued at $62 billion) and Perplexity

Broader vision for AI development: The competition represents a strategic push toward more efficient AI systems that don’t rely solely on massive computing power.

  • Konwinski advocates for “small AI,” emphasizing elegant innovation over raw computational scale
  • Some researchers point to the human brain’s efficiency as evidence that more compact AI models are possible
  • The initiative aims to reinvigorate research into more efficient AI approaches rather than simply pursuing larger models

Future implications: While the prize amount might seem modest compared to the potential value of achieving such capabilities, the competition’s true significance lies in its potential to catalyze innovation in AI efficiency and accessibility.

  • As noted on Hacker News, an AI system capable of achieving the 90% benchmark would likely be worth significantly more than the prize money
  • The competition could help bridge the gap between current AI coding capabilities and human-level programming proficiency
  • The focus on open-source models could accelerate the democratization of advanced AI technologies
Databricks co-founder offers $1 million prize to solve AI coding problems

Recent News

AI courses from Google, Microsoft and more boost skills and résumés for free

As AI becomes critical to business decision-making, professionals can enhance their marketability with free courses teaching essential concepts and applications without requiring technical backgrounds.

Veo 3 brings audio to AI video and tackles the Will Smith Test

Google's latest AI video generation model introduces synchronized audio capabilities, though still struggles with realistic eating sounds when depicting the celebrity in its now-standard benchmark test.

How subtle biases derail LLM evaluations

Study finds language models exhibit pervasive positional preferences and prompt sensitivity when making judgments, raising concerns for their reliability in high-stakes decision-making contexts.