AutoCode: A New AI Framework that Lets LLMs Create and Verify Competitive Programming Problems, Mirroring the Workflow of Human Problem Setters
Are your LLM code benchmarks really rejecting wrong-complexity options and interactive-protocol violations, or are they passing under-specified unit assessments? A crew of researchers from UCSD, NYU, University of Washington, Princeton University, Canyon Crest Academy, OpenAI, UC Berkeley, MIT, University of Waterloo, and Sentient Labs introduce AutoCode, a brand new AI framework that lets LLMs create…
