Explore AI Tools, Topics & Discussions in Testing
AI tools for ensuring code quality, identifying bugs, and automating QA processes.
Related AI categories in Testing
Automated Testing
AI-powered platforms that automate end-to-end testing processes with intelligent test case generation, execution, and reporting for faster, more reliable software delivery.
Bug Detection
Intelligent tools that leverage AI to identify, classify, and prioritize software defects and vulnerabilities before they reach production environments.
LLM Evaluations
Platforms and frameworks for evaluating, testing, and benchmarking LLM systems and AI applications. These tools provide evaluators and evaluation models to score AI outputs, measure hallucinations, assess RAG quality, detect failures, and optimize model performance. Features include automated testing with LLM-as-a-judge metrics, component-level evaluation with tracing, regression testing in CI/CD pipelines, custom evaluator creation, dataset curation, and real-time monitoring of production systems. Teams use these solutions to validate prompt effectiveness, compare models side-by-side, ensure answer correctness and relevance, identify bias and toxicity, prevent PII leakage, and continuously improve AI product quality through experiments, benchmarks, and performance analytics.
Performance Testing
AI-enhanced tools for load, stress, and endurance testing that analyze application performance under various conditions with predictive insights and optimization recommendations.
Test Generation
AI-powered tools that automatically generate comprehensive test cases and scenarios based on code analysis, user journeys, and historical test data.
Visual Testing
AI-driven tools for automated visual interface testing that detect UI/UX inconsistencies, layout issues, and visual regressions across different browsers and devices.
AI Tools in Testing
Hamming
19hAutomated testing and production monitoring platform for voice and chat AI agents with 50+ built-in metrics and compliance support.

Lunary
19hOpen-source platform to monitor, improve, and secure AI chatbots with observability, prompt management, evaluations, and analytics.

HoneyHive
19hAI observability and evaluation platform to monitor, evaluate, and govern AI agents and applications across any model, framework, or agent runtime.

Prompt Mixer
19hOpen source desktop app for prompt engineering that enables teams to create, test, and evaluate prompts and chains with different AI models.

Giskard
19hAutomated testing platform for LLM agents that detects hallucinations, security vulnerabilities, and quality issues through continuous red teaming.

FinetuneDB
19hAI fine-tuning platform to create custom LLMs by training models with your data in minutes, not weeks.

Klu
19hDesign, deploy, and optimize LLM apps with collaborative prompt design, evaluation workflows, and observability tools.

QualGent
6dAPI-driven mobile app testing platform that enables automated test execution for Android and iOS apps directly from CI/CD pipelines.

Ray
2dA desktop debugging app that keeps all your debug output neatly organized, supporting PHP, Laravel, JavaScript, and more.

Opik
5dOpen-source platform for evaluating, testing, and monitoring LLM applications with tracing and observability features.
