EveryDev.ai
Sign inSubscribe
  1. Home
  2. Tools
  3. Giskard
Giskard icon

Giskard

LLM Evaluations

Automated testing platform for LLM agents that detects hallucinations, security vulnerabilities, and quality issues through continuous red teaming.

Visit Website

At a Glance

Pricing

Open Source
Free tier available

Free open-source library for individual developers

Enterprise: Custom/contact

Engagement

Available On

Web
API
SDK

Resources

WebsiteDocsGitHubllms.txt

Topics

LLM EvaluationsSecurity TestingAutomated Testing

About Giskard

Giskard provides an automated testing platform designed to secure LLM agents by detecting hallucinations, security vulnerabilities, and quality issues before they reach production. The platform combines continuous red teaming with expert guidance to help enterprise teams find and fix security gaps in their AI applications. Trusted by major companies including L'Oréal, AXA, Société Générale, Decathlon, Michelin, and BNP Paribas, Giskard serves as a safety net for organizations deploying conversational AI agents.

  • Continuous Red Teaming automatically generates sophisticated attack scenarios to uncover vulnerabilities that manual audits miss, delivering the largest test coverage of both security and quality issues with high domain specificity.

  • Hallucination Detection leverages internal knowledge sources like RAG knowledge bases to identify factual errors, contradictions, omissions, and unsupported information in AI responses.

  • Security Vulnerability Testing detects critical issues including prompt injection attacks, data disclosure risks, sycophancy attacks, and inappropriate content generation using security vulnerability taxonomies.

  • Quality Assurance catches business failures like inappropriate denials where bots incorrectly refuse valid questions, ensuring AI agents deliver helpful responses without being overly restrictive.

  • Human-in-the-Loop Dashboards enable business, engineering, and security teams to review, customize, and approve tests through a collaborative visual interface with annotation tools.

  • Regression Prevention transforms discovered vulnerabilities into permanent protection by automatically converting detected issues into reproducible test suites that enrich golden test datasets.

  • Python SDK & Web UI allows teams to execute tests programmatically or schedule them through the web interface to ensure AI agents meet requirements after each update.

  • Sovereign & Secure Infrastructure offers data residency choices (EU or US), role-based access control, audit trails, identity provider integration, end-to-end encryption, and compliance with GDPR, SOC 2 Type II, and HIPAA.

To get started, teams can request a trial through the Giskard website. The platform operates as a black-box testing tool, meaning it only requires API endpoint access to the AI agent without needing knowledge of internal components like foundation models or vector databases.

Giskard

Community Discussions

Be the first to start a conversation about Giskard

Share your experience with Giskard, ask questions, or help others learn from your insights.

Pricing

FREE

Free Plan Available

Free open-source library for individual developers

  • LLM evaluation
  • Basic vulnerability scanning
  • Python SDK
  • Community support

Enterprise

Full-featured enterprise platform with continuous red teaming

Custom
contact sales
  • Continuous Red Teaming
  • Human-in-the-Loop Dashboards
  • Regression Prevention
  • Data Residency Options
  • RBAC & Audit Trails
  • Identity Provider Integration
  • SOC 2 Type II Compliance
  • HIPAA Compliance
  • GDPR Compliance
  • Technical Consulting Support
  • On-premise deployment option
View official pricing

Capabilities

Key Features

  • Continuous Red Teaming
  • Hallucination Detection
  • Prompt Injection Detection
  • Data Disclosure Prevention
  • Sycophancy Attack Detection
  • Quality Assurance Testing
  • Human-in-the-Loop Dashboards
  • Regression Prevention
  • Python SDK
  • Web UI Scheduling
  • Role-Based Access Control
  • Audit Trails
  • Identity Provider Integration
  • End-to-End Encryption
  • GDPR Compliance
  • SOC 2 Type II Compliance
  • HIPAA Compliance
  • Data Residency Options

Integrations

Hugging Face
AWS
Google Cloud
Microsoft Azure
Mistral AI
API Available
View Docs

Reviews & Ratings

No ratings yet

Be the first to rate Giskard and help others make informed decisions.

Developer

Giskard AI

Giskard AI builds an automated testing platform for securing LLM agents against hallucinations and security vulnerabilities. Founded by Alex Combessie and Jean-Marie John-Mathews, PhD, the company operates from Europe with a team of ML researchers and engineers. Giskard maintains both an open-source library with over 5,000 GitHub stars and an enterprise Hub product. The platform serves major enterprise clients including L'Oréal, AXA, BNP Paribas, and Google DeepMind.

Read more about Giskard AI
WebsiteGitHubLinkedIn
1 tool in directory

Similar Tools

Patronus AI icon

Patronus AI

Automated evaluation and monitoring platform that scores, detects failures, and optimizes LLMs and AI agents using evaluation models, experiments, traces, and an API/SDK ecosystem.

Confident AI icon

Confident AI

End-to-end platform for LLM evaluation and observability that benchmarks, tests, monitors, and traces LLM applications to prevent regressions and optimize performance.

Latitude icon

Latitude

An AI engineering platform for product teams to build, test, evaluate, and deploy reliable AI agents and prompts.

Browse all tools

Related Topics

LLM Evaluations

Platforms and frameworks for evaluating, testing, and benchmarking LLM systems and AI applications. These tools provide evaluators and evaluation models to score AI outputs, measure hallucinations, assess RAG quality, detect failures, and optimize model performance. Features include automated testing with LLM-as-a-judge metrics, component-level evaluation with tracing, regression testing in CI/CD pipelines, custom evaluator creation, dataset curation, and real-time monitoring of production systems. Teams use these solutions to validate prompt effectiveness, compare models side-by-side, ensure answer correctness and relevance, identify bias and toxicity, prevent PII leakage, and continuously improve AI product quality through experiments, benchmarks, and performance analytics.

25 tools

Security Testing

Tools for automated security testing and penetration testing.

4 tools

Automated Testing

AI-powered platforms that automate end-to-end testing processes with intelligent test case generation, execution, and reporting for faster, more reliable software delivery.

46 tools
Browse all topics
Back to all tools
Explore AI Tools
  • AI Coding Assistants
  • Agent Frameworks
  • MCP Servers
  • AI Prompt Tools
  • Vibe Coding Tools
  • AI Design Tools
  • AI Database Tools
  • AI Website Builders
  • AI Testing Tools
  • LLM Evaluations
Follow Us
  • X / Twitter
  • LinkedIn
  • Reddit
  • Discord
  • Threads
  • Bluesky
  • Mastodon
  • YouTube
  • GitHub
  • Instagram
Get Started
  • About
  • Advertise
  • Contact Us
  • Newsletter
  • Submit a Tool
  • Start a Discussion
  • Write A Blog
  • Share A Build
  • Terms of Service
  • Privacy Policy
Explore with AI
  • ChatGPT
  • Gemini
  • Claude
  • Grok
  • Perplexity
Agent Experience
  • llms.txt
Theme
With AI, Everyone is a Dev. EveryDev.ai © 2026
Main Menu
  • Tools
  • Developers
  • Topics
  • Discussions
  • News
  • Blogs
  • Builds
  • Contests
Create
Sign In
    Sign in
    1views
    0saves
    0discussions