Main Menu
  • Tools
  • Developers
  • Topics
  • Discussions
  • News
  • Blogs
  • Builds
  • Contests
  • Compare
Create
    EveryDev.ai
    Sign inSubscribe
    1. Home
    2. Builds
    3. AgentTrace - X-Ray Vision for AI Workflows

    AgentTrace - X-Ray Vision for AI Workflows

    Cosmas Mandikonza's avatar
    Cosmas Mandikonza
    December 22, 2025·Cosmas Mandikonza is a…
    Discuss (1)
    AgentTrace - X-Ray Vision for AI Workflows

    Debugging AI agents sucks. Your pipeline fails at step 47, you have no idea why, so you add print statements, re-run everything from scratch, and wait. I built AgentTrace to fix this. AgentTrace gives you X-ray vision into AI workflows. See your agent pipeline as an interactive flowchart. Click any node to inspect inputs, outputs, tool calls, and errors. Every step shows tokens, cost, and latency. The killer feature is the Replay Engine. Failed at the Reviewer step? Don't re-run Planner and Coder. Click the failed node, tweak the context, replay from that exact point. AgentTrace copies prior events and re-executes only what's needed. You can also diff two sessions to see exactly what changed between a working and broken run. Tech: FastAPI with event sourcing, Next.js 14, React Flow, SQLite, Docker. Built in a weekend with Claude. Surprise: Event sourcing isn't just for databases - it's perfect for AI observability since agent workflows are naturally sequential and immutable.

    Tools & Technologies Used

    Claude Code icon

    Claude Code

    Click to learn more

    →
    Claude icon

    Claude

    Click to learn more

    →
    Open Demo
    Contest Entry

    Build Details

    Build Time

    1 week

    Difficulty

    intermediate

    Comments (1)

    Join the discussion

    Sign in to share your thoughts

    Sam Moore's avatar
    Sam Moore4 months ago

    Great build!

    The Replay Engine is genuinely clever since most observability tools like Langfuse, LangSmith, and Phoenix let you see what failed, but you still have to re-run the whole pipeline from scratch. Being able to click a failed node, tweak context, and resume from that exact point is a real time-saver that the big players don’t offer yet. That space is crowded (Langfuse has 19k GitHub stars, LangSmith has LangChain’s backing), but none of them do mid-pipeline replay so if you leaned hard into that differentiator, there’s a genuine gap to fill.

    Question: How does the Replay Engine handle non-determinism and side effects? When you replay from step 47, the LLM will likely give a different response than the original run and for tool calls that hit external APIs or write to databases, is there a dry-run mode to prevent duplicate side effects?​​​​​​​​​​​​​​​​

    Explore AI Tools
    • AI Coding Assistants
    • Agent Frameworks
    • MCP Servers
    • AI Prompt Tools
    • Vibe Coding Tools
    • AI Design Tools
    • AI Database Tools
    • AI Website Builders
    • AI Testing Tools
    • LLM Evaluations
    Follow Us
    • X / Twitter
    • LinkedIn
    • Reddit
    • Discord
    • Threads
    • Bluesky
    • Mastodon
    • YouTube
    • GitHub
    • Instagram
    Get Started
    • About
    • Editorial Standards
    • Corrections & Disclosures
    • Community Guidelines
    • Advertise
    • Contact Us
    • Newsletter
    • Submit a Tool
    • Start a Discussion
    • Write A Blog
    • Share A Build
    • Terms of Service
    • Privacy Policy
    Explore with AI
    • ChatGPT
    • Gemini
    • Claude
    • Grok
    • Perplexity
    Agent Experience
    • llms.txt
    Theme
    With AI, Everyone is a Dev. EveryDev.ai © 2026