Shane Deconinck Trusted AI Agents · Decentralized Trust

Agentic AI: Curated Questions for the Boardroom

Curated questions on AI agent potential and risk

AI agents bring potential and risk at a scale we haven’t dealt with before. You can’t reap the potential without taming the risk: both come from the fact that agents decide what to do given a goal.

The potential of agents is that they make decisions. The risk originates from the same.

Reliability keeps getting easier as models improve. But better models won’t solve governance. An agent that’s right 99% of the time is the one you stop watching. And when it slips, it slips at machine speed.

The move isn’t to wait until we have all the answers. It’s to be bold without being naive: govern tight, deploy, learn, loosen where you can.

If I were to list the questions I’d bring to the table today, these are the ones I’d start with, across three areas: potential, accountability, and control. By no means exhaustive, but possibly some of the most impactful ones right now.

Potential

Are your agents actually making decisions, or just automating steps humans already defined?

The value of agents is that they decide what to do given a goal. If your agents are running predefined workflows, you're getting automation, not agency. The upside comes from letting them reason, plan, and act.

What decisions are you not yet delegating to agents, and what's that costing you?

Every organization has processes where human bottlenecks slow things down. Some of those are genuinely high-stakes and need a human. Others are just habit. Knowing the difference is where the opportunity lives.

Will better models make your current setup more valuable, or obsolete?

Every workaround your team builds for a model's limitations becomes dead weight when that limitation disappears. The investments that compound are context (giving agents the right information) and permissions (governing what they can do with it). Ask your team how much of the current agent codebase they'd expect to throw away in a year.

Does the right context reach your agents at the right time?

Agent quality depends on having the right information for the task at hand. Not everything, not nothing: the relevant context, when it's needed. If your agents are underperforming, the model might not be the bottleneck. The context pipeline might be.

Are you building on established and emerging standards, or on an island?

Protocols like MCP for tool integration and A2A for agent communication are maturing fast. Building proprietary alternatives might feel faster now, but risks leaving you incompatible with the ecosystem forming around you.

How much value are you leaving on the table by over-constraining?

Agents that need human approval for every action aren't agents: they're suggestion engines. Containment by design lets agents run autonomously within safe boundaries. That's where the real productivity gain is.

Accountability

Do you know every agent running in your organization?

When employees build agents on low-code platforms, the company is still the deployer. An HR screening agent built without a compliance assessment makes you non-compliant without knowing the system exists. Shadow agents are the new shadow IT.

Can your infrastructure prevent an agent from running without being registered?

Knowing what's running today is one thing. Making it structurally impossible to deploy an unregistered agent is another. If anyone can spin one up without it showing up in a registry, visibility is a snapshot, not a guarantee.

When an agent makes a consequential decision, can you trace who authorized it and what happened?

Audit logs that show "alice@company.com" aren't enough when Alice delegated to an agent three months ago. You need to know who or what made the call, and under what authority.

If an agent causes harm, is the liability chain clear?

The human who delegated, the team that deployed, the vendor who built the model: all may share responsibility. If no one owns the answer, everyone will point at each other when it matters.

Could you explain to a regulator what your agent did and why?

The EU AI Act requires traceability, risk management, and human oversight for high-risk use cases. If an agent autonomously wandered into one, can you reconstruct the chain of decisions that got it there?

Control

Are agents restricted to what they can do, or only blocked from what they can't?

You can't list everything an agent shouldn't do: the list is infinite. The inverse works: start from zero authority, grant explicit permissions per task. A blocklist is always incomplete. An allowlist is always bounded.

When agents delegate to other agents, can authority only decrease?

If your procurement agent can approve purchases up to $5,000, any sub-agent it calls should inherit that ceiling or lower. Does the architecture enforce this, or just assume it?

What happens when an agent wanders into a use case you didn't anticipate?

A general-purpose office assistant told to "handle my inbox" might draft an email (minimal risk), then screen a job application (high-risk). The risk tier depends on how open-ended the prompt is. If you can't anticipate the use case, how do you bound it?

Are your agents contained by architecture, or only by policy?

Policy says what agents shouldn't do. Architecture limits what they can do, regardless of what they try. If a prompt injection or bad dependency triggers something unexpected, what actually stops it from spreading?

What happens when human oversight breaks down in practice?

After the 20th approval prompt, people start clicking "yes" without reading. Decades of automation research confirm that humans can't reliably monitor automated systems and then rapidly take control when needed. If your safety model depends on vigilance, what happens when vigilance fades?

How do you balance agent quality with data privacy?

Agents get better with more context, but more context means more data exposure. Do your agents see only what they need for the task at hand, or do they have broad access because it's easier to set up? And where does that data go: to an open-source model running on your infrastructure, or to a frontier model behind an API? The privacy calculus is different for each.

Does your agent setup work when agents need to cross trust boundaries?

Most current approaches work within a single trust domain. When agents act across organizations, call external APIs, or coordinate with third-party agents, identity and authority need to travel with the request: verifiable at every step, not assumed.

The Bottom Line

The organizations that capture the most value from agents won’t just be the ones with better models. They’ll be the ones that moved boldly and governed tightly from day one.

My questions keep evolving. Curious what yours would be.

Slidedeck: