Replies: 1 comment
-
|
Namit, David, Parag, this resonates. I like the framing because it puts the TPM role where it belongs for intelligent systems: helping the team make safe decisions under uncertainty, not just reporting task progress. A project can be green on scope, schedule, and risks while still being unsafe if the team has not earned confidence in model behavior, data quality, evaluation coverage, operator workflow, rollback posture, or drift detection. A few additions I would consider as you move toward pilots:
The caution I would add is process weight. We should pilot the smallest useful pack first, probably the Confidence Dashboard, Decision Ledger, Uncertainty Framing Contract, and one evaluation/readiness artifact. If the tooling helps a crew make a safer decision faster, it belongs in HVE Core. If it mostly creates another status surface, it will give back the gains we are trying to protect. I would be happy to help review or pilot from the TPM Guild side. |
Beta Was this translation helpful? Give feedback.
Uh oh!
There was an error while loading. Please reload this page.
Uh oh!
There was an error while loading. Please reload this page.
-
As we all have realised, Intelligent systems don't behave like traditional software projects. Outcomes are probabilistic, model performance shifts over time, and progress can't be measured by task completion alone. Yet most TPM tooling (scoping, scheduling, risk tracking, RAG dashboards) still assumes deterministic delivery. This creates a structural gap: projects report green while decisions remain unsafe. We've been working on closing that gap with tools and processes built specifically for probabilistic delivery.
What We Are Doing
We have kicked off an initiative to build TPM tooling specifically for intelligent systems delivery. We studied how Google, Microsoft, and AWS define production readiness for ML systems and mapped where standard TPM practices hold up and where they break down. From that research, we are redefining 10 core TPM disciplines (scope, schedule, risk, quality, progress tracking, dependencies, stakeholder management, definition of done, operational readiness, reporting and change control) for environments where behaviour is learned from data and adapts to context. Each discipline gets purpose built tools and processes. Like, confidence dashboards instead of RAG status, hypothesis driven scope packs instead of fixed requirements, timeboxed learning cycles instead of milestone driven schedules, kill criteria and risk heatmaps that account for drift and silent failure modes, etc.
Where We Are
Research across three industry frameworks is complete. Ten TPM responsibility areas have been redefined with evidence based rationale. Twenty seven tools and processes have been documented with structure, templates, and alignment to the responsibility model. These include artifacts like the Confidence Dashboard, Decision Ledger, Uncertainty Framing Contract, Quality Assurance Pack (covering data, model, system, and drift gates), Model Operations Readiness Checklist, and Monitoring Signal Map. Everything is catalogued and queryable. Next step is piloting on live engagements to validate what works.
What's Next
We are hoping to integrate this as an accelerator into HVE Core so it becomes part of the standard toolkit available to all teams. Before that, we need to pilot on real engagements and iterate based on feedback.
If this problem resonates with you or you want to help shape the tooling, reach out to any of us. We welcome contributors, reviewers, and teams willing to pilot.
Namit (namit.t@microsoft.com)
David (david.ratcliffe@microsoft.com)
Parag (paragalurkar@microsoft.com)
Beta Was this translation helpful? Give feedback.
All reactions