Main Menu
  • Tools
  • Developers
  • Topics
  • Discussions
  • News
  • Blogs
  • Builds
  • Contests
Create
Sign In
    EveryDev.ai
    Sign inSubscribe
    Home
    Tools

    1,711+ AI tools

    • New
    • Trending
    • Featured
    • Compare
    Categories
    • Agents891
    • Coding869
    • Infrastructure377
    • Marketing357
    • Design302
    • Research276
    • Projects271
    • Analytics266
    • Testing160
    • Integration157
    • Data150
    • Security131
    • MCP125
    • Learning124
    • Extensions108
    • Communication107
    • Prompts100
    • Voice90
    • Commerce89
    • DevOps70
    • Web66
    • Finance17
    Sign In
    1. Home
    2. Tools
    3. Galileo
    Galileo icon

    Galileo

    LLM Evaluations

    End-to-end platform for generative AI evaluation, observability, and real-time protection that helps teams test, monitor, and guard production AI applications.

    Visit Website

    At a Glance

    Pricing

    Free tier available

    Developer tier for experimenting, iterating, and building with Galileo.

    Pro: $100/mo
    Enterprise: Custom/contact

    Engagement

    Available On

    Web
    API
    SDK

    Resources

    WebsiteDocsGitHubllms.txt

    Topics

    LLM EvaluationsObservability PlatformsApplication Security

    Alternatives

    Confident AIPatronus AIOpik

    Developer

    GalileoSan Francisco, CAEst. 2021$68M raised

    Updated Feb 2026

    About Galileo

    Galileo provides an enterprise-focused GenAI Studio for evaluating, observing, and protecting generative AI systems in development and production. It centralizes evaluation metrics, real-time observability, and runtime guardrails so teams can iterate on prompts, models, and retrieval strategies with measurable feedback. Galileo runs low-latency evaluators (Luna models), hosts inference for live monitoring, and offers SDKs and APIs to integrate logging and traces into existing applications.

    • Evaluate — Rapidly run and compare combinations of prompts, models, embedding params, and chain nodes to find the right configuration; use the platform UI or SDKs to log experiments and golden test sets.
    • Observe — Monitor live traffic, traces, and sessions with pre-built and custom metrics to detect drift, latency, and accuracy issues in production systems; connect via SDKs or API to stream logs and traces.
    • Protect — Intercept requests and responses in real time with guardrail policies and threat detection to block harmful outputs or attacks before they reach users.
    • Luna evaluation models — Use Galileo’s low-latency evaluator models to run automated judgements (<200ms typical) for production monitoring and inexpensive continuous evaluation.
    • SDKs & integrations — Install Python or TypeScript SDKs, initialize with an API key, and add the log decorator or GalileoLogger to capture prompts, responses, traces, and spans.

    Getting started: sign up for the hosted console, install the Python or TypeScript SDK, set GALILEO_API_KEY, and either use the log decorator or GalileoLogger to begin sending traces and running evaluations.

    Galileo - 1

    Community Discussions

    Be the first to start a conversation about Galileo

    Share your experience with Galileo, ask questions, or help others learn from your insights.

    Pricing

    FREE

    Free Plan Available

    Developer tier for experimenting, iterating, and building with Galileo.

    • 5,000 traces per month
    • Unlimited users
    • Unlimited custom evals

    Pro

    Popular

    Plan for teams launching production AI with higher trace quotas and enterprise features.

    $100
    per month
    • Everything in Free
    • 50,000 traces per month
    • Standard RBAC
    • Advanced analytics & insights
    • Real-time guardrails
    • Dedicated support via Slack

    Enterprise

    Custom plans for large teams that need unlimited scale, security, and premium support.

    Custom
    contact sales
    • Unlimited traces
    • Custom rate limits
    • Deploy: Hosted, VPC, or on-prem
    • Enterprise-grade security (RBAC, SSO)
    • Dedicated CSM and 24/7 support
    • Low-latency dedicated inference servers
    View official pricing

    Capabilities

    Key Features

    • AI evaluation workflows for prompts, models, and RAG systems
    • Real-time observability of traces, sessions, and metrics
    • Runtime protection and guardrail policies
    • Prebuilt and custom evaluator metrics (including hallucination detection)
    • Luna low-latency evaluation models and hosted inference server
    • Python and TypeScript SDKs and a public API
    • Auto-tune evaluators with continuous learning (CLHF)

    Integrations

    OpenAI (wrapper)
    NVIDIA NeMo
    Python SDK
    TypeScript SDK
    API Available
    View Docs

    Demo Video

    Galileo Demo Video
    Watch on YouTube

    Reviews & Ratings

    No ratings yet

    Be the first to rate Galileo and help others make informed decisions.

    Developer

    Galileo Team

    Galileo builds an end-to-end GenAI Studio that centers reliability for production AI systems through evaluation, observability, and real-time protection. The team includes founders and engineers with backgrounds in large language models, speech recognition, and production AI systems from companies like Google and Uber. Galileo focuses on combining research-backed guardrail metrics with low-latency evaluators and developer-friendly SDKs to help teams ship safer AI.

    Founded 2021
    San Francisco, CA
    $68M raised
    175 employees

    Used by

    Comcast
    Twilio
    HP
    Magid
    +7 more
    Read more about Galileo Team
    WebsiteGitHubX / Twitter
    1 tool in directory

    Similar Tools

    Confident AI icon

    Confident AI

    End-to-end platform for LLM evaluation and observability that benchmarks, tests, monitors, and traces LLM applications to prevent regressions and optimize performance.

    Patronus AI icon

    Patronus AI

    Automated evaluation and monitoring platform that scores, detects failures, and optimizes LLMs and AI agents using evaluation models, experiments, traces, and an API/SDK ecosystem.

    Opik icon

    Opik

    Open-source platform for evaluating, testing, and monitoring LLM applications with tracing and observability features.

    Browse all tools

    Related Topics

    LLM Evaluations

    Platforms and frameworks for evaluating, testing, and benchmarking LLM systems and AI applications. These tools provide evaluators and evaluation models to score AI outputs, measure hallucinations, assess RAG quality, detect failures, and optimize model performance. Features include automated testing with LLM-as-a-judge metrics, component-level evaluation with tracing, regression testing in CI/CD pipelines, custom evaluator creation, dataset curation, and real-time monitoring of production systems. Teams use these solutions to validate prompt effectiveness, compare models side-by-side, ensure answer correctness and relevance, identify bias and toxicity, prevent PII leakage, and continuously improve AI product quality through experiments, benchmarks, and performance analytics.

    48 tools

    Observability Platforms

    Comprehensive platforms that combine metrics, logs, and traces with AI-powered analytics to provide deep insights into complex distributed systems and application behavior.

    48 tools

    Application Security

    AI tools for securing software applications and identifying vulnerabilities.

    39 tools
    Browse all topics
    Back to all tools
    Explore AI Tools
    • AI Coding Assistants
    • Agent Frameworks
    • MCP Servers
    • AI Prompt Tools
    • Vibe Coding Tools
    • AI Design Tools
    • AI Database Tools
    • AI Website Builders
    • AI Testing Tools
    • LLM Evaluations
    Follow Us
    • X / Twitter
    • LinkedIn
    • Reddit
    • Discord
    • Threads
    • Bluesky
    • Mastodon
    • YouTube
    • GitHub
    • Instagram
    Get Started
    • About
    • Editorial Standards
    • Corrections & Disclosures
    • Community Guidelines
    • Advertise
    • Contact Us
    • Newsletter
    • Submit a Tool
    • Start a Discussion
    • Write A Blog
    • Share A Build
    • Terms of Service
    • Privacy Policy
    Explore with AI
    • ChatGPT
    • Gemini
    • Claude
    • Grok
    • Perplexity
    Agent Experience
    • llms.txt
    Theme
    With AI, Everyone is a Dev. EveryDev.ai © 2026
    Sign in
    34views