Replies: 7 comments 1 reply
-
|
@jman-msc — thank you for reaching out directly, and for the generous words. Finding the ATF spec back in February was one of those "someone already articulated the threat model we've been building against" moments. Your five-pillar structure gave us a clean compliance target, and mapping our toolkit against all 15 requirements validated that we'd been converging on the same architecture from the implementation side. On the three spec additions1. Agent delegation chain verification (v0.2.0) This is the gap that shows up the moment you move from single-agent to multi-agent systems. Our AgentMesh package already implements this — cryptographically signed delegation chains where each hop in an agent-to-agent call carries:
We'd be happy to contribute our implementation patterns and test vectors as input for the v0.2.0 spec. The tricky design decision is how much of the chain each intermediate agent should be able to inspect (full chain vs. only its immediate delegator) — we currently support both modes and have opinions on when each is appropriate. 2. AI-BOM integration We built a full AI Bill of Materials system in Agent OS that tracks model provenance, training data lineage, weight checksums, and dependency graphs. This maps directly to the Data Governance pillar and is the artifact we believe regulators will eventually require — the EU AI Act's transparency obligations effectively demand something like this. The piece we think the ATF spec could formalize: a standard schema for AI-BOMs that governance frameworks can validate against. We have a working schema today; would be valuable to align it with what the CSA working group envisions. 3. Trust scoring quantification The maturity model's qualitative promotion criteria (Intern → Junior → Senior → Principal) are the right abstraction. What we've added on the implementation side is a numeric trust scoring engine that feeds those promotion decisions:
Making this quantified methodology part of the spec would give organizations a concrete, auditable basis for autonomy decisions instead of subjective judgment calls. Happy to share our scoring algorithms and the reasoning behind the weight defaults. On the conformance specificationThe co-author invitation is genuinely compelling — our implementation experience across all 15 requirements is exactly the kind of input that keeps conformance criteria practical. We're very interested in this and would like to understand the scope and timeline better before formalizing involvement. Would a call to walk through what you're envisioning for the conformance spec be a good starting point? On RSACI won't be at RSAC this year unfortunately — but I'd very much like to connect. Would you be open to a call the week after RSAC (March 30 week) once you're back? Happy to do a screen-share walkthrough of the toolkit's ATF coverage, the delegation chain implementation, and the trust scoring engine. That might be more productive than a conference hallway conversation anyway. On the webinar recordingThank you for sharing — I'll watch it this week and may have follow-up questions on the spec direction, particularly around how conformance testing intersects with the maturity model tiers. Genuinely excited about this convergence. We built the implementation; you wrote the specification; the alignment is too natural to not formalize. Looking forward to connecting. Imran Siddique |
Beta Was this translation helpful? Give feedback.
-
|
Hi @imran-siddique, This is exactly the kind of response I was hoping for. Technical depth, specific implementation details, and clear opinions on the design tradeoffs. That's what makes a conformance spec work. A few things worth calling out: Delegation chainsThe full-chain vs. immediate-delegator visibility question is one of the harder design decisions in v0.2.0. I have opinions on when each mode is appropriate too. I'd rather hash that out on a call than in a thread. Your test vectors and implementation patterns would be a huge input. AI-BOM schemaRegulators will ask for this first, and most orgs won't have a good answer. If you already have a working schema validated against the EU AI Act transparency requirements, that saves months. Let's align it with what the CSA working group is building. Trust scoringThe hysteresis to prevent promotion/demotion oscillation is a detail that only shows up after you've run this in production. The ATF maturity model uses MUST/SHOULD/MAY per requirement at each level (Intern through Principal), but promotion criteria between levels are still qualitative. A numeric scoring methodology would close that gap. I want implementation experience driving the spec here, not theory. On conformanceI've published the ATF Conformance Specification. Two tiers: ATF Compatible (self-attestation against all 25 requirements) and ATF Certified (third-party audit, 90 days production operation, additional operational controls). There's a maturity level matrix that maps each requirement to MUST/SHOULD/MAY at each level. Your toolkit covers all 25 requirements across the five elements. I want your read on whether the conformance criteria are practical from an implementer's seat. That's the blind spot a spec author can't test alone. We've also launched agentictrustframework.ai as the home for the spec and verifiedagents.ai for assessment and scoring. Both go well past the original blog post. Worth a look before our call. On timing: week of March 30 works. I'll be back from RSAC with fresh context. A screen-share walkthrough of your ATF coverage, delegation chains, and trust scoring engine is the right format. I'll follow up after RSAC to lock in a time. Thanks, |
Beta Was this translation helpful? Give feedback.
-
|
Josh — excellent, this is moving fast in the right direction. Conformance spec reviewI read through CONFORMANCE.md. A few notes from the implementer's seat: What works well:
Feedback for consideration:
I'll bring a detailed requirement-by-requirement assessment to the call. New sitesChecked out agentictrustframework.ai and verifiedagents.ai — significant upgrade from the GitHub-only presence. The assessment model on verifiedagents.ai is exactly what enterprises will ask for when evaluating agent platforms. Call logisticsWeek of March 30 works perfectly. I'll send a calendar invite — suggest 60 minutes for:
What timezone works best for you? Looking forward to it. |
Beta Was this translation helpful? Give feedback.
-
|
Quick update on logistics — my schedule is packed the next few weeks with upcoming OOF, so a live call may be hard to fit. Would you be open to doing this async instead? I'm thinking: I'll prepare a detailed requirement-by-requirement conformance assessment as a document (our ATF coverage, delegation chain implementation details, trust scoring algorithms, AI-BOM schema) and share it here or as a PR against the ATF repo. You can review at your pace, leave comments, and we iterate on the doc. That way we both get the depth of a call without the scheduling overhead. If there are specific design decisions that need real-time discussion (like the delegation chain visibility question), we can do a focused 30-min call just for those — but let's see if the doc gets us 90% there first. Let me know if that works. |
Beta Was this translation helpful? Give feedback.
-
|
Imran, congrats on the public launch. Seven packages, five language SDKs, 9,500+ tests, 20 tutorials, and framework integrations already shipping. That's serious work. Async works well for me. A few things, picking up where we left off. Your conformance feedback I've been sitting with your notes on Delegation chains as MUST at Senior/Principal. Yes. Multi-agent systems without verifiable authorization provenance are a non-starter at those maturity levels. The maturity level matrix in AI-BOM: model identifier + training data hash as MUST, full provenance as SHOULD. That's the right split. It matches how EU AI Act Art. 13 will land in practice. Organizations can hit the floor quickly and build toward full provenance over time. Adopting this. Trust scoring minimum signals. You're right that without defined signals, "trust scoring" becomes a checkbox. I'll specify task success rate, policy compliance rate, and anomaly frequency as minimum required inputs. Weight selection stays implementation-specific. That gives conformance teeth without locking out different scoring approaches, including your 0-100 composite score model. The async conformance assessment Since we last spoke, the spec has moved to
Your three conformance recommendations will land in the next point release. If you're still up for it, a PR with your toolkit's coverage mapped against all 25 requirements would be the first formal conformance document from any implementation. That carries weight. I can set up a The delegation chain design question The full-chain vs. immediate-delegator visibility question is one that needs real-time discussion. I have a position on when each mode applies, and I'd rather debate it live than in a thread. When your schedule opens up, I'd like to find 30 minutes for that one. ATF spec (v0.9.0): https://github.com/massivescale-ai/agentic-trust-framework |
Beta Was this translation helpful? Give feedback.
-
|
Josh -- thanks for the update and congrats feedback on our launch. Quick status on the conformance assessment: Since your last message we shipped v3.0.2 with several changes directly relevant to ATF alignment:
I will prepare the full requirement-by-requirement assessment against CONFORMANCE.md and share as a PR against the ATF repo. Targeting next week. On delegation chain visibility -- agree we should hash out full-chain vs immediate-delegator modes. Our current implementation exposes parent_did (immediate) but can walk the chain via recursive lookup. Happy to document both modes with pros/cons in the assessment. |
Beta Was this translation helpful? Give feedback.
-
|
Josh -- the conformance assessment is ready. Published at: https://github.com/microsoft/agent-governance-toolkit/blob/main/docs/compliance/atf-conformance-assessment.md Results against ATF v0.9.0:
7 gaps documented with specific code citations and recommended fixes:
Every requirement includes file paths, class names, and how the implementation works. Happy to iterate on any section. If you want, I can also submit this as a PR against the ATF repo with a conformance-statements/ directory. |
Beta Was this translation helpful? Give feedback.
Uh oh!
There was an error while loading. Please reload this page.
-
Hi Imran,
I'm Josh Woodruff, author of the Agentic Trust Framework and co-chair of the CSA Zero Trust Working Group.
I just found your CSA-ATF-PROPOSAL.md and wanted to reach out directly. You built full coverage across all five pillars, got this moved into the microsoft/ org, and formally proposed CSA working group engagement. All within 30 days of the blog post. That's impressive, and I appreciate it.
I fully support positioning the Agent Governance Toolkit as an ATF reference implementation. On your three proposed spec additions:
I'd like to invite you to contribute to the ATF specification through the CSA Zero Trust Working Group. I'll be at RSAC 2026 (March 23-26 in San Francisco) and would love to meet in person if you're attending. If not, let's set up a call.
I also just ran a CSA webinar on ATF. Here's the recording if you want to see the current state of the spec and where it's heading: CSA ATF Webinar Recording
One more thing: I'm developing a formal ATF Conformance specification, defining what "ATF Compatible" means as a verifiable checklist based on the five elements. You've already mapped against all 15 requirements. I'd welcome your input as a co-author. Your implementation experience is exactly what keeps a conformance spec practical instead of theoretical.
Looking forward to connecting.
Josh Woodruff
Founder and CEO, MassiveScale.AI
CSA Research Fellow | IANS Faculty | Co-Chair, CSA Zero Trust Working Group
agentictrustframework.ai
Beta Was this translation helpful? Give feedback.
All reactions