EveryDev.ai
Sign inSubscribe
  1. Home
  2. Tools
  3. LM Arena
LM Arena icon

LM Arena

Performance Metrics

Web platform for comparing, running, and deploying large language models with hosted inference and API access.

Visit Website

At a Glance

Pricing

Open Source

Free platform for comparing AI models and contributing to crowdsourced leaderboards.

Engagement

Available On

Web
API

Resources

WebsiteDocsGitHubllms.txt

Topics

Performance MetricsUser ResearchLLM Evaluations

About LM Arena

LM Arena provides a web-based environment for running, comparing, and deploying large language models. It focuses on making model evaluation, hosted inference, and simple deployment workflows accessible from a browser and via an API. The platform supports uploading or connecting models, running evaluation workloads, and exposing inference endpoints for applications.

  • Model comparison run side-by-side evaluations and benchmarks across models.
  • Hosted inference deploy models to managed endpoints for production usage.
  • API access programmatically invoke models and integrate into applications.
  • Custom model uploads bring your own model artifacts for testing and deployment.
  • Usage monitoring track usage metrics and performance of deployed endpoints.

To get started, sign up on the web app, upload or connect a model, run a comparison job, and create an inference endpoint; use the provided API keys to integrate inference into your applications.

LM Arena

Community Discussions

Create a new discussion about LM Arena
Joe Seifi's avatar
Joe Seifi
13d·Apple, Disney, Adobe, Eventbrite,…

Video Arena is out on LM Arena and it's pretty wild

So LM Arena shipped their Video Arena feature and I've been messing around with it. You can now compare AI video models side by side, things like Sora, Hailuo, Veo 3.1, and a bunch of others. The cool part is it runs through their Discord server and you get to vote on which model output you like bet…

0
news

Pricing

OPEN SOURCE

Open Source

Free platform for comparing AI models and contributing to crowdsourced leaderboards.

  • Compare answers across all AI models
  • Side-by-side model battles
  • Vote and contribute to public leaderboard
  • Save chat history with account
  • Access to all leaderboard categories (Text, Vision, WebDev, Image, Video, etc.)
View official pricing

Capabilities

Key Features

  • Model comparison
  • Hosted inference endpoints
  • API access
  • Custom model uploads
  • Usage monitoring and metrics

Integrations

Hugging Face
OpenAI
Docker
API Available
View Docs

Demo Video

LM Arena Demo Video
Watch on YouTube

Reviews & Ratings

No ratings yet

Be the first to rate LM Arena and help others make informed decisions.

Developer

LM Arena Team

LM Arena builds a web platform for evaluating and deploying large language models. The team builds tools for model comparison, hosted inference, and API integration. They focus on simplifying model workflows for engineers and researchers. The company emphasizes accessible deployment and monitoring features.

Founded 2023
San Francisco, CA
$250M raised
40 employees

Used by

OpenAI
Google DeepMind
Anthropic
Meta
+9 more
Read more about LM Arena Team
WebsiteGitHubX / Twitter
1 tool in directory

Similar Tools

LLM Stats icon

LLM Stats

Public leaderboards and benchmark site that publishes verifiable evaluations, scores, and performance metrics for large language models and AI providers.

Design Arena icon

Design Arena

Crowdsourced benchmark for AI‑generated design. Users vote on head‑to‑head outputs (web UI, images, video, audio) to rank models by human preference.

Artificial Analysis icon

Artificial Analysis

Independent AI model benchmarking platform providing comprehensive performance analysis across intelligence, speed, cost, and quality metrics

Browse all tools

Related Topics

Performance Metrics

Specialized tools for measuring, evaluating, and optimizing AI model performance across accuracy, speed, resource utilization, and other critical parameters.

23 tools

User Research

AI-enhanced platforms for conducting usability testing, gathering feedback, and analyzing user behavior with automated insights and pattern recognition.

4 tools

LLM Evaluations

Platforms and frameworks for evaluating, testing, and benchmarking LLM systems and AI applications. These tools provide evaluators and evaluation models to score AI outputs, measure hallucinations, assess RAG quality, detect failures, and optimize model performance. Features include automated testing with LLM-as-a-judge metrics, component-level evaluation with tracing, regression testing in CI/CD pipelines, custom evaluator creation, dataset curation, and real-time monitoring of production systems. Teams use these solutions to validate prompt effectiveness, compare models side-by-side, ensure answer correctness and relevance, identify bias and toxicity, prevent PII leakage, and continuously improve AI product quality through experiments, benchmarks, and performance analytics.

25 tools
Browse all topics
Back to all tools
Explore AI Tools
  • AI Coding Assistants
  • Agent Frameworks
  • MCP Servers
  • AI Prompt Tools
  • Vibe Coding Tools
  • AI Design Tools
  • AI Database Tools
  • AI Website Builders
  • AI Testing Tools
  • LLM Evaluations
Follow Us
  • X / Twitter
  • LinkedIn
  • Reddit
  • Discord
  • Threads
  • Bluesky
  • Mastodon
  • YouTube
  • GitHub
  • Instagram
Get Started
  • About
  • Advertise
  • Contact Us
  • Newsletter
  • Submit a Tool
  • Start a Discussion
  • Write A Blog
  • Share A Build
  • Terms of Service
  • Privacy Policy
Explore with AI
  • ChatGPT
  • Gemini
  • Claude
  • Grok
  • Perplexity
Agent Experience
  • llms.txt
Theme
With AI, Everyone is a Dev. EveryDev.ai © 2026
Main Menu
  • Tools
  • Developers
  • Topics
  • Discussions
  • News
  • Blogs
  • Builds
  • Contests
Create
Sign In
    Sign in
    78views
    0saves
    1discussions