Main Menu
  • Tools
  • Developers
  • Topics
  • Discussions
  • News
  • Blogs
  • Builds
  • Contests
Create
Sign In
    EveryDev.ai
    Sign inSubscribe
    Home
    Tools

    1,711+ AI tools

    • New
    • Trending
    • Featured
    • Compare
    Categories
    • Agents891
    • Coding875
    • Infrastructure377
    • Marketing357
    • Design303
    • Research276
    • Projects272
    • Analytics266
    • Testing161
    • Integration157
    • Data150
    • Security131
    • MCP126
    • Learning125
    • Extensions108
    • Communication107
    • Prompts100
    • Voice90
    • Commerce89
    • DevOps70
    • Web67
    • Finance17
    Sign In
    1. Home
    2. Tools
    3. Giskard
    Giskard icon

    Giskard

    LLM Evaluations

    Automated testing platform for LLM agents that detects hallucinations, security vulnerabilities, and quality issues through continuous red teaming.

    Visit Website

    At a Glance

    Pricing

    Free tier available

    Free open-source library for individual developers

    Enterprise: Custom/contact

    Engagement

    Available On

    Web
    API
    SDK

    Resources

    WebsiteDocsGitHubllms.txt

    Topics

    LLM EvaluationsSecurity TestingAutomated Testing

    Alternatives

    promptfooAshrPatronus AI

    Developer

    Giskard AIGiskard AI builds an automated testing platform for securing…

    Listed Feb 2026

    About Giskard

    Giskard provides an automated testing platform designed to secure LLM agents by detecting hallucinations, security vulnerabilities, and quality issues before they reach production. The platform combines continuous red teaming with expert guidance to help enterprise teams find and fix security gaps in their AI applications. Trusted by major companies including L'Oréal, AXA, Société Générale, Decathlon, Michelin, and BNP Paribas, Giskard serves as a safety net for organizations deploying conversational AI agents.

    • Continuous Red Teaming automatically generates sophisticated attack scenarios to uncover vulnerabilities that manual audits miss, delivering the largest test coverage of both security and quality issues with high domain specificity.

    • Hallucination Detection leverages internal knowledge sources like RAG knowledge bases to identify factual errors, contradictions, omissions, and unsupported information in AI responses.

    • Security Vulnerability Testing detects critical issues including prompt injection attacks, data disclosure risks, sycophancy attacks, and inappropriate content generation using security vulnerability taxonomies.

    • Quality Assurance catches business failures like inappropriate denials where bots incorrectly refuse valid questions, ensuring AI agents deliver helpful responses without being overly restrictive.

    • Human-in-the-Loop Dashboards enable business, engineering, and security teams to review, customize, and approve tests through a collaborative visual interface with annotation tools.

    • Regression Prevention transforms discovered vulnerabilities into permanent protection by automatically converting detected issues into reproducible test suites that enrich golden test datasets.

    • Python SDK & Web UI allows teams to execute tests programmatically or schedule them through the web interface to ensure AI agents meet requirements after each update.

    • Sovereign & Secure Infrastructure offers data residency choices (EU or US), role-based access control, audit trails, identity provider integration, end-to-end encryption, and compliance with GDPR, SOC 2 Type II, and HIPAA.

    To get started, teams can request a trial through the Giskard website. The platform operates as a black-box testing tool, meaning it only requires API endpoint access to the AI agent without needing knowledge of internal components like foundation models or vector databases.

    Giskard - 1

    Community Discussions

    Be the first to start a conversation about Giskard

    Share your experience with Giskard, ask questions, or help others learn from your insights.

    Pricing

    FREE

    Free Plan Available

    Free open-source library for individual developers

    • LLM evaluation
    • Basic vulnerability scanning
    • Python SDK
    • Community support

    Enterprise

    Full-featured enterprise platform with continuous red teaming

    Custom
    contact sales
    • Continuous Red Teaming
    • Human-in-the-Loop Dashboards
    • Regression Prevention
    • Data Residency Options
    • RBAC & Audit Trails
    • Identity Provider Integration
    • SOC 2 Type II Compliance
    • HIPAA Compliance
    • GDPR Compliance
    • Technical Consulting Support
    • On-premise deployment option
    View official pricing

    Capabilities

    Key Features

    • Continuous Red Teaming
    • Hallucination Detection
    • Prompt Injection Detection
    • Data Disclosure Prevention
    • Sycophancy Attack Detection
    • Quality Assurance Testing
    • Human-in-the-Loop Dashboards
    • Regression Prevention
    • Python SDK
    • Web UI Scheduling
    • Role-Based Access Control
    • Audit Trails
    • Identity Provider Integration
    • End-to-End Encryption
    • GDPR Compliance
    • SOC 2 Type II Compliance
    • HIPAA Compliance
    • Data Residency Options

    Integrations

    Hugging Face
    AWS
    Google Cloud
    Microsoft Azure
    Mistral AI
    API Available
    View Docs

    Reviews & Ratings

    No ratings yet

    Be the first to rate Giskard and help others make informed decisions.

    Developer

    Giskard AI

    Giskard AI builds an automated testing platform for securing LLM agents against hallucinations and security vulnerabilities. Founded by Alex Combessie and Jean-Marie John-Mathews, PhD, the company operates from Europe with a team of ML researchers and engineers. Giskard maintains both an open-source library with over 5,000 GitHub stars and an enterprise Hub product. The platform serves major enterprise clients including L'Oréal, AXA, BNP Paribas, and Google DeepMind.

    Read more about Giskard AI
    WebsiteGitHubLinkedIn
    1 tool in directory

    Similar Tools

    promptfoo icon

    promptfoo

    Promptfoo is an AI security testing platform that helps developers and enterprises find and fix vulnerabilities in LLM applications through automated red teaming, guardrails, and evaluations.

    Ashr icon

    Ashr

    Ashr is an AI agent evaluation platform that mimics production environments and user behavior to catch agent failures before they reach real users.

    Patronus AI icon

    Patronus AI

    Automated evaluation and monitoring platform that scores, detects failures, and optimizes LLMs and AI agents using evaluation models, experiments, traces, and an API/SDK ecosystem.

    Browse all tools

    Related Topics

    LLM Evaluations

    Platforms and frameworks for evaluating, testing, and benchmarking LLM systems and AI applications. These tools provide evaluators and evaluation models to score AI outputs, measure hallucinations, assess RAG quality, detect failures, and optimize model performance. Features include automated testing with LLM-as-a-judge metrics, component-level evaluation with tracing, regression testing in CI/CD pipelines, custom evaluator creation, dataset curation, and real-time monitoring of production systems. Teams use these solutions to validate prompt effectiveness, compare models side-by-side, ensure answer correctness and relevance, identify bias and toxicity, prevent PII leakage, and continuously improve AI product quality through experiments, benchmarks, and performance analytics.

    48 tools

    Security Testing

    Tools for automated security testing and penetration testing.

    9 tools

    Automated Testing

    AI-powered platforms that automate end-to-end testing processes with intelligent test case generation, execution, and reporting for faster, more reliable software delivery.

    76 tools
    Browse all topics
    Back to all tools
    Explore AI Tools
    • AI Coding Assistants
    • Agent Frameworks
    • MCP Servers
    • AI Prompt Tools
    • Vibe Coding Tools
    • AI Design Tools
    • AI Database Tools
    • AI Website Builders
    • AI Testing Tools
    • LLM Evaluations
    Follow Us
    • X / Twitter
    • LinkedIn
    • Reddit
    • Discord
    • Threads
    • Bluesky
    • Mastodon
    • YouTube
    • GitHub
    • Instagram
    Get Started
    • About
    • Editorial Standards
    • Corrections & Disclosures
    • Community Guidelines
    • Advertise
    • Contact Us
    • Newsletter
    • Submit a Tool
    • Start a Discussion
    • Write A Blog
    • Share A Build
    • Terms of Service
    • Privacy Policy
    Explore with AI
    • ChatGPT
    • Gemini
    • Claude
    • Grok
    • Perplexity
    Agent Experience
    • llms.txt
    Theme
    With AI, Everyone is a Dev. EveryDev.ai © 2026
    Sign in
    13views