Innovation & Dev Culture

Building the 3D Challenge System: Inside Code Olympics

November 27, 2025
.
6 min

How We Created a Constraint-Based Programming Championship from Scratch

Introduction

Code Olympics 2025 was not just another hackathon. It was an experiment in stripping away the comfortable scaffolding of modern development—the frameworks, the libraries, the endless npm packages—and asking a simple question: what can developers actually build when they are forced back to fundamentals? From October 31 to November 10, 2025, we ran what might be the most unusual programming competition in recent memory, and this article pulls back the curtain on how we made it happen.

The premise was deliberately uncomfortable. Participants received randomly generated constraints across three dimensions: what they could not use, how many lines they could write, and what domain they had to build in. The result was 640 possible challenge combinations, ensuring no two competitors faced identical obstacles. With 66 submissions and over 230 individual evaluations from 12 expert judges, the competition delivered exactly what we hoped: a proving ground for genuine programming skill.

The Technical Architecture

Challenge Generation System

The heart of Code Olympics was the constraint generator—an AI-powered system that created personalized challenges by combining elements from three distinct dimensions. The first dimension defined core constraints: limitations like "No-Import Rookie" (only built-in functions allowed), "Few-Variable Hero" (maximum eight variables in the entire program), or "One-Loop Warrior" (exactly one loop permitted). The second dimension set line budgets ranging from a brutal 50-line "Tiny Scripter" tier to a more generous 650-line "Enterprise Creator" level. The third dimension assigned project domains: simple games, text processing, file management, system utilities, and more.

The generator operated as an independent service, architecturally decoupled from the main competition platform. This separation ensured optimal performance during peak usage and allowed us to scale challenge generation independently. Participants could re-roll their initial constraints up to three times, introducing a strategic element—do you stick with a challenging but familiar combination, or gamble on something potentially easier but less known? The machine learning algorithms behind the system prevented clustering around popular constraint combinations while maintaining fair difficulty distribution across all participants.

Evaluation Infrastructure

Scoring followed a weighted formula: Functionality and Reliability accounted for 40% of the final score, Constraint Mastery contributed 30%, Code Quality added 20%, and Innovation completed the remaining 10%. All scores were normalized on a 0-5 scale and averaged across multiple judges. The infrastructure had to handle concurrent evaluations from judges distributed across time zones spanning from Serbia to the United States, requiring careful attention to synchronization and data consistency.

Automated constraint verification formed the first line of defense. Before human evaluation began, systems checked line counts, analyzed import statements, counted variable declarations, and measured response times. Only submissions passing automated verification proceeded to human review. This two-tier approach ensured that judges spent their time on substantive code evaluation rather than mechanical compliance checking.

The Judging Team

Finding the right judges was critical. We needed evaluators who understood the difference between functional code and exceptional engineering, people who had built systems at scale and solved complex technical challenges under real-world constraints. The 12 professionals who formed our panel brought expertise from organizations including Cisco, Microsoft, First Line Software, and enterprise consulting firms across multiple continents.

  • Sergii Demianchuk, a Software Engineering Technical Leader at Cisco working in the Security and Trust Organization, brought over 16 years of experience in application development and cyber security. His background spans multiple companies including Narvar and SoftServe, where he served as Application Architect. With a Ph.D. in Computer Science from Lviv Polytechnic National University and deep expertise in Ruby, C/C++, JavaScript, and database design, Sergii provided rigorous evaluation of submissions tackling security-related challenges and system-level programming.
  • Aleksandr Meshkov, Head of QA and AI Evaluation at First Line Software, contributed a unique perspective shaped by over 12 years in quality assurance spanning manual, automation, and performance testing. His current work focuses on creating scalable testing methodologies and developing evaluation tools for AI solutions, including frameworks for assessing RAG systems and AI agents using DeepEval, Ragas, and Azure AI Studio. Aleksandr's experience building guardrails for AI solution safety proved particularly relevant when evaluating projects that pushed constraint boundaries in creative ways.
  • Artem Mukhin, a Software Engineer at Microsoft and former Yandex developer, offered a developer experience perspective honed through years of frontend development and internal tooling work. At Microsoft, he initiated and secured a dedicated DevEx role by demonstrating improvements to mobile build optimization—reducing fast refresh times from over 60 seconds to just 2-5 seconds. His four years at Yandex included designing and implementing web applications as both the sole frontend developer and product manager, giving him deep appreciation for what can be achieved under resource constraints.
  • Sagar Gupta, a Research-Driven Technology Leader and member of the Forbes Technology Council, brought expertise in transforming advanced AI research into scalable enterprise solutions. His work centers on multi-agent orchestration, retrieval-augmented generation, and explainable AI frameworks. With a background spanning ERP implementation leadership and AI integration architecture, Sagar evaluated submissions through the lens of practical applicability and real-world scalability.

The judging process generated some memorable feedback. Roman Kamyshnikov praised the Best Constraint Mastery winner with particular enthusiasm, noting that the submission embodied the spirit of the hackathon—well-written, readable code that achieved high utility despite severe limitations. Naman Jain commended the champion project for transforming a simple terminal into a full visual and interactive environment, recognizing both the technical depth and creative approach. Kishore Subramanya Hebbar highlighted how the third-place project demonstrated deliberate craftsmanship through recursive design patterns, showing that elegant architecture remained possible even under tight constraints. Ishu Anand Jaiswal noted that TextMind excelled with a powerful suite of offline analysis tools packed into an ultra-light footprint, calling it a best-in-class lightweight AI text analyzer.

Projects and Engineering Challenges

The winning projects demonstrated that artificial limitations genuinely breed creativity. CuteBunny's VIST (Visual Terminal Tool) claimed the championship with a final score of 4.480 out of 5.00 and took home a combined $1,100 for also winning Best Functionality and Reliability. The project transformed a terminal into a multimedia canvas supporting image-to-ASCII conversion with GPU acceleration, data visualization, a 35 FPS game called Raptors Go, and real-time video playback with synchronized audio—all in under 500 lines of Python with zero crashes.

The second-place project, Code Auditor by team 3s, packed a professional security scanner into 300 lines of dependency-free Python. It detected hardcoded secrets, API keys, SQL injection patterns, and code complexity hotspots while generating interactive HTML dashboards. The tool could scan over 10,000 files in under three seconds. Third place went to PILYT, a Go-based terminal assistant that transformed YouTube videos into structured, searchable knowledge libraries complete with AI summaries, subtitle downloads, and cross-platform playback support.

Category winners revealed different facets of excellence. CalC's Omni Convertor achieved the highest individual criterion score in the entire competition—4.88 out of 5.00 for Constraint Mastery—by delivering a feature-rich, zero-dependency Python toolkit that included a Morse code converter. BeTheNOOB's TextMind won Best Code Quality by creating an offline text analyzer under extreme constraints of only eight variables and 400 lines that could summarize documents, detect sentiment, and generate ML-ready datasets with complete privacy. Ctrl+Alt+Victory's Sigma Clean took Best Innovation by combining AI-powered file management via Groq API with error-proof architecture in just 149 lines—a project that never crashed regardless of input.

Patterns in the Submissions

Analyzing all 66 submissions revealed interesting trends. Python dominated, accounting for over 60% of projects, though we saw creative implementations in Go, Rust, C++, and JavaScript. Twelve teams achieved true zero-dependency status, avoiding all external libraries. Terminal and CLI applications represented 76% of submissions, suggesting that when forced to work without visual frameworks, developers naturally gravitated toward text-based interfaces where they could exercise maximum control.

The scoring distribution told its own story. Eight teams achieved scores above 4.3 out of 5.00, representing the top 17% of submissions. Half of all participants scored between 4.0 and 4.3, indicating that the constraint system successfully created challenges that were difficult but achievable. The highest score of 4.480 and an overall average of 4.025 suggested our calibration was close to optimal—hard enough to differentiate skill levels but not so brutal as to discourage quality submissions.

Running the Competition: What Went Right and What We Learned

The evaluation period ran from November 3 to November 15, 2025, during which 12 judges completed over 230 individual project reviews. Coordinating evaluators across time zones from the United States to Serbia required careful scheduling and clear protocols. We established standardized evaluation templates that guided judges through each criterion while allowing space for qualitative feedback. The most heavily evaluated projects—Team Jhotta's Code Scraper, BeTheNOOB's TextMind, and Sanjay Sah's entry—each received assessments from eight different judges, providing robust scoring data.

The three-day build window from October 31 to November 3 created natural pressure points. Participants had to balance ambition against execution risk, knowing that a partially working sophisticated project would score worse than a fully functional simple one. The re-roll mechanic for constraints proved more strategic than expected, with some participants using all three re-rolls while others stuck with their first combination. Post-competition surveys indicated that constraint selection strategy was itself a significant source of mental challenge for competitors.

Several technical lessons emerged. The automated constraint verification system needed refinement—detecting variable counts in languages with implicit declarations proved trickier than anticipated. Some constraints like "Error-Proof Coder" required human judgment to verify properly, as automated crash testing could not cover all edge cases. Future iterations will likely implement more sophisticated static analysis for preliminary verification while reserving behavioral assessment for human evaluators.

Looking Forward

Code Olympics 2025 validated a hypothesis: that constraint-based programming competitions can reveal dimensions of skill that traditional algorithmic contests miss. The projects submitted demonstrated that elegant solutions emerge from limitation, that minimalism can coexist with power, and that real-world utility does not require thousands of dependencies.

The total prize pool of $2,400 distributed across main placements and category awards, with a Community Choice Award of $500 still pending the LinkedIn poll results. Beyond prizes, participants gained something harder to quantify: proof of their fundamental capabilities when the safety net of external tools disappears.

Planning for the next edition is already underway. We are considering expanded constraint categories, possibly including memory limitations and I/O restrictions. The judging infrastructure will evolve to handle larger participant pools while maintaining evaluation quality. Most importantly, we will continue refining the core premise: in an industry where dependencies have become a crutch, Code Olympics offers a space to discover what developers can really build.

True engineering mastery lies not in writing more code, but in writing the right code. Every line counts.

Related Blogs

No items found.