Main Menu
  • Tools
  • Developers
  • Topics
  • Discussions
  • News
  • Blogs
  • Builds
  • Contests
Create
Sign In
    EveryDev.ai
    Sign inSubscribe
    Home
    Tools

    1,715+ AI tools

    • New
    • Trending
    • Featured
    • Compare
    Categories
    • Agents891
    • Coding869
    • Infrastructure377
    • Marketing357
    • Design302
    • Research276
    • Projects271
    • Analytics266
    • Testing160
    • Integration157
    • Data150
    • Security131
    • MCP125
    • Learning124
    • Extensions108
    • Communication107
    • Prompts100
    • Voice90
    • Commerce89
    • DevOps70
    • Web66
    • Finance17
    Sign In
    1. Home
    2. Tools
    3. LM Arena
    LM Arena icon

    LM Arena

    Performance Metrics

    Web platform for comparing, running, and deploying large language models with hosted inference and API access.

    Visit Website

    At a Glance

    Pricing

    Free

    Free platform for comparing AI models and contributing to crowdsourced leaderboards.

    Engagement

    Available On

    Web
    API

    Resources

    WebsiteDocsGitHubllms.txt

    Topics

    Performance MetricsUser ResearchLLM Evaluations

    Alternatives

    LLM StatsDesign ArenaIsItNerfed?

    Developer

    LM ArenaSan Francisco, CAEst. 2023$250M raised

    Updated Feb 2026

    About LM Arena

    LM Arena provides a web-based environment for running, comparing, and deploying large language models. It focuses on making model evaluation, hosted inference, and simple deployment workflows accessible from a browser and via an API. The platform supports uploading or connecting models, running evaluation workloads, and exposing inference endpoints for applications.

    • Model comparison run side-by-side evaluations and benchmarks across models.
    • Hosted inference deploy models to managed endpoints for production usage.
    • API access programmatically invoke models and integrate into applications.
    • Custom model uploads bring your own model artifacts for testing and deployment.
    • Usage monitoring track usage metrics and performance of deployed endpoints.

    To get started, sign up on the web app, upload or connect a model, run a comparison job, and create an inference endpoint; use the provided API keys to integrate inference into your applications.

    LM Arena - 1

    Community Discussions

    Create a new discussion about LM Arena
    Joe Seifi's avatar
    Joe Seifi
    January 23, 2026·Apple, Disney, Adobe, Eventbrite,…

    Video Arena is out on LM Arena and it's pretty wild

    So LM Arena shipped their Video Arena feature and I've been messing around with it. You can now compare AI video models side by side, things like Sora, Hailuo, Veo 3.1, and a bunch of others. The cool part is it runs through their Discord server and you get to vote on which model output you like bet…

    0
    news

    Pricing

    FREE

    Free Plan Available

    Free platform for comparing AI models and contributing to crowdsourced leaderboards.

    • Compare answers across all AI models
    • Side-by-side model battles
    • Vote and contribute to public leaderboard
    • Save chat history with account
    • Access to all leaderboard categories (Text, Vision, WebDev, Image, Video, etc.)
    View official pricing

    Capabilities

    Key Features

    • Model comparison
    • Hosted inference endpoints
    • API access
    • Custom model uploads
    • Usage monitoring and metrics

    Integrations

    Hugging Face
    OpenAI
    Docker
    API Available
    View Docs

    Demo Video

    LM Arena Demo Video
    Watch on YouTube

    Reviews & Ratings

    No ratings yet

    Be the first to rate LM Arena and help others make informed decisions.

    Developer

    LM Arena Team

    LM Arena builds a web platform for evaluating and deploying large language models. The team builds tools for model comparison, hosted inference, and API integration. They focus on simplifying model workflows for engineers and researchers. The company emphasizes accessible deployment and monitoring features.

    Founded 2023
    San Francisco, CA
    $250M raised
    40 employees

    Used by

    OpenAI
    Google DeepMind
    Anthropic
    Meta
    +9 more
    Read more about LM Arena Team
    WebsiteGitHubX / Twitter
    1 tool in directory

    Similar Tools

    LLM Stats icon

    LLM Stats

    Public leaderboards and benchmark site that publishes verifiable evaluations, scores, and performance metrics for large language models and AI providers.

    Design Arena icon

    Design Arena

    Crowdsourced benchmark for AI‑generated design. Users vote on head‑to‑head outputs (web UI, images, video, audio) to rank models by human preference.

    IsItNerfed? icon

    IsItNerfed?

    Continuous LLM evaluation platform that tracks AI model performance over time through community voting and automated coding task metrics.

    Browse all tools

    Related Topics

    Performance Metrics

    Specialized tools for measuring, evaluating, and optimizing AI model performance across accuracy, speed, resource utilization, and other critical parameters.

    33 tools

    User Research

    AI-enhanced platforms for conducting usability testing, gathering feedback, and analyzing user behavior with automated insights and pattern recognition.

    10 tools

    LLM Evaluations

    Platforms and frameworks for evaluating, testing, and benchmarking LLM systems and AI applications. These tools provide evaluators and evaluation models to score AI outputs, measure hallucinations, assess RAG quality, detect failures, and optimize model performance. Features include automated testing with LLM-as-a-judge metrics, component-level evaluation with tracing, regression testing in CI/CD pipelines, custom evaluator creation, dataset curation, and real-time monitoring of production systems. Teams use these solutions to validate prompt effectiveness, compare models side-by-side, ensure answer correctness and relevance, identify bias and toxicity, prevent PII leakage, and continuously improve AI product quality through experiments, benchmarks, and performance analytics.

    48 tools
    Browse all topics
    Back to all tools
    Explore AI Tools
    • AI Coding Assistants
    • Agent Frameworks
    • MCP Servers
    • AI Prompt Tools
    • Vibe Coding Tools
    • AI Design Tools
    • AI Database Tools
    • AI Website Builders
    • AI Testing Tools
    • LLM Evaluations
    Follow Us
    • X / Twitter
    • LinkedIn
    • Reddit
    • Discord
    • Threads
    • Bluesky
    • Mastodon
    • YouTube
    • GitHub
    • Instagram
    Get Started
    • About
    • Editorial Standards
    • Corrections & Disclosures
    • Community Guidelines
    • Advertise
    • Contact Us
    • Newsletter
    • Submit a Tool
    • Start a Discussion
    • Write A Blog
    • Share A Build
    • Terms of Service
    • Privacy Policy
    Explore with AI
    • ChatGPT
    • Gemini
    • Claude
    • Grok
    • Perplexity
    Agent Experience
    • llms.txt
    Theme
    With AI, Everyone is a Dev. EveryDev.ai © 2026
    Sign in
    147views
    1discussion