Back to BlogTechnology

Introducing Switchboard AI: Democratising Access to Advanced AI

Anthony Duncalf
25 June 2025
8 min read

We're transforming how users access AI by solving the fundamental challenge of model selection. Instead of forcing users to navigate an overwhelming maze of AI options, Switchboard AI intelligently routes requests to the most capable model for each specific task.

The AI landscape has exploded with options—from OpenAI's GPT models to Anthropic's Claude, Google's Gemini, and countless specialised alternatives. This abundance creates what researchers call the "paradox of choice": when too many options lead to decision paralysis and suboptimal outcomes. Users waste time experimenting with different models, often settling for mediocre results simply because they lack the expertise to navigate the complex ecosystem effectively.

The Problem: AI's Unprecedented Challenge

Artificial intelligence represents a unique technology in human history. Unlike previous innovations that augmented human capabilities through tools and infrastructure, AI has the potential to replace human decision-making entirely. This unprecedented capability drives rapid adoption across vast populations for countless purposes, but it also creates a fundamental mismatch between the technology's complexity and users' ability to harness it effectively.

The challenge is compounded by the fact that AI systems can excel in unexpected combinations. A model might demonstrate PhD-level reasoning in physics whilst struggling with basic common sense, or produce brilliant creative writing whilst failing simple arithmetic. Traditional evaluation frameworks, designed for more predictable systems, break down when confronted with such multifaceted and sometimes contradictory capabilities.

Current solutions force users into suboptimal scenarios. They either stick with a single familiar model—sacrificing performance on tasks where other models excel—or manually experiment across providers, requiring technical knowledge and significant time investment. Both approaches fail to harness the collective potential of our diverse AI ecosystem.

Why Current Evaluation Systems Fail

For years, researchers attempted to solve this challenge through static benchmarks. Tests like MMLU-Pro, Humans Last Exam, and GPQA were designed to measure knowledge breadth and multi-step reasoning capabilities across diverse domains. These benchmarks represented sophisticated attempts to capture AI capability through comprehensive, standardised evaluation.

Yet static benchmarks fundamentally fail to capture the full extent of AI systems, inevitably introducing measurement bias whilst suffering from critical structural flaws. The presence of a static metric can lead AI companies to iterate towards optimising for adversarial competitiveness instead of high utility in real-world applications. These benchmarks prove difficult to update, don't scale well with rapid model advancement, and remain vulnerable to data contamination when test materials leak into training sets.

Most critically, the static approach breaks down precisely when we need it most—as AI systems become increasingly sophisticated and diverse in their applications. A benchmark measuring mathematical reasoning reveals nothing about creative writing capability, whilst tests optimised for factual recall may entirely miss the nuanced judgement required for real-world decision-making.

LMArena's Revolutionary Insight and Its Problems

Recognising these fundamental limitations, LMArena introduced a paradigm shift: pairwise human comparisons evaluating model outputs side-by-side. This approach eliminated rigid rubrics whilst harnessing collective human judgement, allowing statistical patterns to emerge from numerous trials where individual biases theoretically average out.

The innovation was profound—rather than forcing AI evaluation into predetermined categories designed by researchers, LMArena let human users directly compare outputs and decide which better served their actual needs. This approach acknowledged a fundamental truth: the ultimate arbiters of AI utility are the humans who use these systems, not the academics who design abstract tests.

However, LMArena's influence has not gone unchallenged. Substantial scrutiny emerged from researchers across multiple prestigious institutions who published "The Leaderboard Illusion," presenting the strongest challenge to Arena's methodology and fairness. Their critique reveals that LMArena's implementation, despite its innovative approach, has created systematic advantages for well-resourced organisations whilst disadvantaging smaller players and open-source alternatives.

  • Private Testing Advantage: Major AI companies like Meta, OpenAI, Google, and Amazon received special treatment through undisclosed private testing policies, allowing them to test multiple variants before selecting their best performer for public submission
  • Data Distribution Imbalance: Proprietary models received a disproportionate share of Arena's crowdsourced data, with Google and OpenAI capturing significant portions whilst numerous open-weight models combined received far less
  • Unequal Model Treatment: Of 243 public models, 205 had been "silently deprecated" without public notice, with 64% of these being open-weight or open-source models

When evaluation systems become high-stakes competitions, they distort reality itself, creating artificial advantages, misdirecting resources, and ultimately serving corporate interests over genuine progress in AI capability.

The Switchboard Solution

Switchboard is an AI prompt routing service that directs user prompts to the most capable model for each specific task, ensuring maximum utility responses. Rather than creating another flawed ranking system, we eliminate the incentive misalignment through systematic design choices:

  • No Public Leaderboards: By removing the prize of topping rankings, we eliminate gamification incentives whilst ensuring AI companies focus on building genuinely useful models rather than optimising for benchmark performance
  • Rewarding User Contributions: Our routing service relies on human-evaluated pairwise comparisons, but we supercharge data collection by rewarding users with credits to access frontier models in exchange for providing comparison data
  • Democratised Insights: We release community insights developed from comparison data, allowing all model developers to learn how to improve their models' real-world performance based on actual user preferences
  • Accessible Models Only: We focus exclusively on models that users can actually access, testing only models people can use, not hypothetical variants that never reach production

How Our Technology Works

Our routing service leverages human preference data collected through pairwise comparisons to learn which models perform best for specific types of requests. The system works through three key components:

  • Prompt Analysis: Our routing model analyses incoming inputs (text, text combined with images, or requests involving specific file types) to identify task characteristics such as mathematical reasoning requirements, creative elements, coding complexity, factual accuracy needs, and other dimensions that influence model performance
  • Model Capability Learning: Using continuously updated preference comparison data, we employ the Bradley-Terry statistical model to understand relative model strengths across different task types, allowing us to build dynamic profiles of each model's capabilities that evolve as we gather more feedback
  • Intelligent Routing: Rather than creating static rankings, we learn dynamic selection functions that adapt to the characteristics of each individual input, automatically connecting users with the most capable model for their specific request

What This Means for Users

With Switchboard AI, users can focus on their work rather than worrying about which AI model to use. Whether you're a developer seeking coding assistance, a writer looking for creative inspiration, or a researcher analysing complex data, our platform ensures you're always using the most appropriate model for your specific needs.

We maintain complete transparency by showing users which model was selected and why, building trust through clear communication about our routing decisions. This approach transforms the AI experience from a technical challenge into a seamless tool that adapts to user requirements.

Looking Ahead

Our immediate next steps follow a clear three-phase roadmap:

  • Launch comprehensive data collection across 35 models through human-evaluated pairwise comparisons, enabling users to earn credits for their contributions
  • Build our team and release transparent, actionable insights to help all model developers improve their systems based on real user preferences
  • Deploy our intelligent routing service with scalable API infrastructure, connecting users seamlessly with optimal AI capabilities for their needs

Switchboard measures success by how well we serve users, not by how effectively we rank providers. Through collaborative progress over zero-sum competition, we aim to create an ecosystem that prioritises human utility above all else.

We're excited to democratise access to advanced AI technology and work with our community to shape the future of intelligent model routing. The age of choosing between AI models is ending—the age of AI that chooses itself is just beginning.