Skip to content

Governance layer for agent tool calls #27

@nayname

Description

@nayname

Current implementation works well for helping LLMs reason about tools.

Been thinking about situations where trust is completely delegated - complex, security-sensitive runs etc. In those settings, trust and predictability become the bottleneck: the model picks a tool, it executes, you find out later if something went wrong (which surfaces several known security issues).

One framing I'm exploring is an external governance layer that captures how the operation needs to be executed. Model classifies intent and generates an auditable plan → plan resolves to concrete tool calls → validates → executes. The goal isn’t to change how tools are defined or invoked, but to make multi-step execution intent explicit and inspectable before side effects occur.

You're deep in this space, I've been prototyping the idea - would be curious whether it makes sense as complementary to your existing approach.

Metadata

Metadata

Assignees

No one assigned

    Labels

    No labels
    No labels

    Projects

    No projects

    Milestone

    No milestone

    Relationships

    None yet

    Development

    No branches or pull requests

    Issue actions