RFTSystems Agent Forensics Suite — audit, prove, replay, diff agent runs

Hi all — I’ve been building a verification-first toolkit for agent runs because most “agent debugging” today is still opinion-based. When something goes wrong, we get partial logs, screenshots, and a postmortem story — but no way for a third party to independently prove what happened, replay it, or pinpoint the exact divergence.

RFTSystems: Agent Forensics Suite turns agent behaviour into verifiable artifacts: hash-chained timelines, tamper-evident receipts, deterministic replays, and first-divergence diffs. The goal is simple: replace “trust me” with evidence you can validate.

Start here (guided entrypoint):

Full collection:
https://huggingface.co/collections/RFTSystems/rftsystems-agent-forensics-suite

What you can do right now

The workflow

learn → generate proof → record reality → seal it → diff it → audit it → benchmark it

What I’m looking for (to harden this for real-world use)

If you’re working with agents (LangGraph/LangChain/custom), I’d value:

  • run bundles you’re willing to share (even tiny ones) so we can validate cross-machine reproducibility
  • failure cases where you can’t explain why two “similar” runs diverged
  • feedback on what you’d want in a “minimum viable audit trail” for deployment/compliance

If you try it and it breaks, tell me exactly where — I’m building this to survive professional scrutiny.

#Agents #LLMOps #MLOps #Reproducibility #Observability #Forensics #AISafety #Governance

— Liam (RFTSystems)

1 Like