EveryDev.ai
Sign inSubscribe
  1. Home
  2. Builds
  3. AgentTrace - X-Ray Vision for AI Workflows

AgentTrace - X-Ray Vision for AI Workflows

Cosmas Mandikonza's avatar
Cosmas Mandikonza
2mo·Cosmas Mandikonza is a…
AgentTrace - X-Ray Vision for AI Workflows

Debugging AI agents sucks. Your pipeline fails at step 47, you have no idea why, so you add print statements, re-run everything from scratch, and wait. I built AgentTrace to fix this. AgentTrace gives you X-ray vision into AI workflows. See your agent pipeline as an interactive flowchart. Click any node to inspect inputs, outputs, tool calls, and errors. Every step shows tokens, cost, and latency. The killer feature is the Replay Engine. Failed at the Reviewer step? Don't re-run Planner and Coder. Click the failed node, tweak the context, replay from that exact point. AgentTrace copies prior events and re-executes only what's needed. You can also diff two sessions to see exactly what changed between a working and broken run. Tech: FastAPI with event sourcing, Next.js 14, React Flow, SQLite, Docker. Built in a weekend with Claude. Surprise: Event sourcing isn't just for databases - it's perfect for AI observability since agent workflows are naturally sequential and immutable.

Tools & Technologies Used

Claude Code icon

Claude Code

Click to learn more

→
Claude icon

Claude

Click to learn more

→
Open Demo
Contest Entry

Build Details

Build Time

1 week

Difficulty

intermediate

Comments

Sign in to join the discussion.

Sam Moore's avatar
Sam Moore2 months ago

Great build!

The Replay Engine is genuinely clever since most observability tools like Langfuse, LangSmith, and Phoenix let you see what failed, but you still have to re-run the whole pipeline from scratch. Being able to click a failed node, tweak context, and resume from that exact point is a real time-saver that the big players don’t offer yet. That space is crowded (Langfuse has 19k GitHub stars, LangSmith has LangChain’s backing), but none of them do mid-pipeline replay so if you leaned hard into that differentiator, there’s a genuine gap to fill.

Question: How does the Replay Engine handle non-determinism and side effects? When you replay from step 47, the LLM will likely give a different response than the original run and for tool calls that hit external APIs or write to databases, is there a dry-run mode to prevent duplicate side effects?​​​​​​​​​​​​​​​​

Explore AI Tools
  • AI Coding Assistants
  • Agent Frameworks
  • MCP Servers
  • AI Prompt Tools
  • Vibe Coding Tools
  • AI Design Tools
  • AI Database Tools
  • AI Website Builders
  • AI Testing Tools
  • LLM Evaluations
Follow Us
  • X / Twitter
  • LinkedIn
  • Reddit
  • Discord
  • Threads
  • Bluesky
  • Mastodon
  • YouTube
  • GitHub
  • Instagram
Get Started
  • About
  • Editorial Standards
  • Corrections & Disclosures
  • Community Guidelines
  • Advertise
  • Contact Us
  • Newsletter
  • Submit a Tool
  • Start a Discussion
  • Write A Blog
  • Share A Build
  • Terms of Service
  • Privacy Policy
Explore with AI
  • ChatGPT
  • Gemini
  • Claude
  • Grok
  • Perplexity
Agent Experience
  • llms.txt
Theme
With AI, Everyone is a Dev. EveryDev.ai © 2026
Main Menu
  • Tools
  • Developers
  • Topics
  • Discussions
  • News
  • Blogs
  • Builds
  • Contests
Create
Sign In
    Sign in