Skip to main content

Agent Hooks

TL;DR
  • Enforce quality gates on agent responses before they reach users
  • Audit and control tool usage with custom bash or Python scripts
  • Block dangerous operations automatically with policy enforcement hooks
  • Prevent early task completion by validating agent output meets your criteria

The problem

Your agent executes tasks autonomously—investigating incidents, running tools, generating responses. But autonomy without oversight creates risk:

  • Incomplete responses: The agent says "done" before addressing everything you asked for
  • Unaudited tool usage: You have no visibility into which tools the agent calls or what results it gets
  • No policy enforcement: Dangerous operations (destructive commands, unauthorized changes) proceed unchecked
  • Quality gaps: Responses miss critical information because there's no validation step

You need a way to intercept agent behavior at key moments—without slowing it down or removing its autonomy entirely.

How agent hooks work

Hooks are custom checkpoints you attach to specific agent events. When an event fires, your hook evaluates the situation and decides whether to allow or block the action.

Agent about to stop → Stop hook evaluates response → Allow or reject
Agent uses a tool → PostToolUse hook checks result → Allow, block, or inject context

Two levels of hooks

Hooks operate at two levels:

LevelWhere to configureScope
Agent levelBuilder → Hooks in the portalApplies to the entire agent — all threads and all custom agents
Custom agent levelAgent Canvas → Custom agent → Manage Hooks, or via the REST API v2Applies only when that specific custom agent runs

Both levels can coexist. If an agent-level hook and a custom-agent-level hook both match the same event, both run — agent-level hooks fire first.

Two hook events are currently supported:

EventTriggers whenYou can
StopAgent is about to return a final responseValidate completeness, reject and force the agent to continue
PostToolUseA tool finishes executing successfullyAudit usage, block results, inject additional context

Two execution types

You can implement hooks using either an LLM or a shell script:

TypeHow it worksBest for
PromptAn LLM evaluates your prompt and returns a JSON decisionNuanced validation ("Is this response complete?")
CommandA bash or Python script runs in a sandboxed environmentDeterministic checks, policy enforcement, auditing

Prompt hooks are powerful for subjective evaluation—checking if a response addresses all user concerns or verifying that an investigation was thorough enough. They use the $ARGUMENTS placeholder to receive the full hook context. If $ARGUMENTS is not present in the prompt, the context is appended automatically. Prompt hooks also receive ReadFile and GrepSearch tools when a conversation transcript is available, allowing the LLM to reason about the full conversation history.

Command hooks are better for deterministic checks—validating that a response contains required markers, blocking dangerous commands, or logging tool usage to an external system.


What makes this different

Without hooksWith hooks
Agent decides when it's "done"You define what "done" means
Tool usage is invisibleEvery tool call can be audited
Dangerous commands proceed silentlyPolicy enforcement blocks them automatically
Quality depends on prompt engineering aloneAutomated quality gates catch gaps

Hooks don't replace run mode safety controls—they complement them. Run modes control what the agent can do. Hooks control how well it does it and what happens with the results.

Before and after

BeforeAfter
Response qualityAgent stops when it thinks it's doneYour Stop hook validates completeness before the response reaches users
Tool visibilityNo audit trail of tool executionPostToolUse hooks log and verify every tool call
Policy enforcementDangerous commands execute uncheckedScripts block rm -rf, sudo, and other risky patterns automatically
Quality assurancePrompt engineering is your only leverLLM-based hooks evaluate nuance; scripts enforce deterministic rules

How to configure hooks

The easiest way to create hooks is through the portal UI:

  1. Agent-level hooks: Go to Builder → Hooks → click Create hook
  2. Custom-agent-level hooks: Go to Agent Canvas → click a custom agent → Manage Hooks

See the Create Hooks via Portal tutorial for step-by-step instructions.

For API users

Hooks can also be configured via the REST API v2 using PUT /api/v2/extendedAgent/agents/{agentName}. The YAML format below shows the full configuration schema. See the API tutorial for details.

Note: The Agent Canvas YAML tab displays v1 format and does not show hooks. Use the Hooks page under Builder to view and manage hooks.

api_version: azuresre.ai/v2
kind: ExtendedAgent
metadata:
name: my_hooked_agent
spec:
instructions: |
You are a helpful assistant.
handoffDescription: ""
enableVanillaMode: true
hooks:
Stop:
- type: prompt
prompt: |
Check if the response ends with "Task complete."
$ARGUMENTS
Respond with:
- {"ok": true} if it does
- {"ok": false, "reason": "End your response with 'Task complete.'"} if not
timeout: 30

PostToolUse:
- type: command
matcher: "Bash|ExecuteShellCommand"
timeout: 30
failMode: block
script: |
#!/usr/bin/env python3
import sys, json, re

context = json.load(sys.stdin)
command = context.get('tool_input', {}).get('command', '')

dangerous = [r'\brm\s+-rf\b', r'\bsudo\b', r'\bchmod\s+777\b']
for pattern in dangerous:
if re.search(pattern, command):
print(json.dumps({"decision": "block", "reason": f"Blocked: {pattern}"}))
sys.exit(0)

print(json.dumps({"decision": "allow"}))

Hook response format

Hooks must output JSON. Two formats are supported:

Simple format (recommended for prompt hooks):

{"ok": true}
{"ok": false, "reason": "Please include more details."}

Expanded format (recommended for command hooks):

{"decision": "allow"}
{"decision": "block", "reason": "Dangerous command detected."}
{"decision": "allow", "hookSpecificOutput": {"additionalContext": "Tool audit logged."}}

Command hooks can also use exit codes instead of JSON output:

Exit codeBehavior
0 with no outputAllow (no objection)
0 with JSONParse JSON for decision
2Always block — stderr becomes the reason
OtherUses failMode setting (allow or block)
Important

For Stop hooks, a rejection without a reason is treated as approval — the agent stops normally. Always provide a reason field when rejecting.

Multiple hooks

You can define multiple hooks for the same event. For PostToolUse, each hook with a matching matcher pattern runs independently. If multiple hooks provide additionalContext, the last hook's context is injected into the conversation.


Configuration reference

OptionTypeDefaultDescription
typestringpromptprompt or command
promptstringLLM prompt text (required for prompt hooks). Use $ARGUMENTS for context injection
commandstringInline shell command (for command hooks, mutually exclusive with script)
scriptstringMulti-line script (for command hooks, mutually exclusive with command)
matcherstringRegex pattern for tool names (required for PostToolUse hooks). * matches all tools. Patterns are anchored as ^(pattern)$ and matched case-sensitively. Empty or null matches nothing.
timeoutint30Execution timeout in seconds (must be positive; values above 300 are flagged during CLI validation)
failModestringallowHow to handle hook errors: allow or block
modelstringReasoningFastModel for prompt hooks (scenario name or deployment name)
maxRejectionsint3 (agent default)Max rejections before forcing stop. Range: 1–25. Applies to prompt-type Stop hooks only — command-type Stop hooks have no implicit limit. When multiple prompt hooks specify different values, the maximum is used.

Hook context schema

Hooks receive structured JSON context about the current event. Prompt hooks receive it via the $ARGUMENTS placeholder in the prompt text. Command hooks receive it as JSON on stdin.

For both hook types, the execution_summary field contains a file path to the conversation transcript (not inline content). For prompt hooks, the LLM receives ReadFile and GrepSearch tools to access this file. For command hooks, the file is available at the specified path in the sandbox.

Common fields

{
"hook_event_name": "Stop",
"agent_name": "my_agent",
"current_turn": 5,
"max_turns": 50,
"execution_summary": "/path/to/transcript.txt"
}

Stop hook fields

{
"final_output": "Here is my response...",
"stop_hook_active": false,
"stop_rejection_count": 0
}

PostToolUse hook fields

{
"tool_name": "ExecutePythonCode",
"tool_input": { "code": "print(2+2)" },
"tool_result": "4",
"tool_succeeded": true
}

Limits

LimitValue
Script size64 KB maximum
Timeout1–300 seconds
Max rejections (prompt Stop hooks)1–25 (default: 3)
Supported script shebangs#!/bin/bash, #!/usr/bin/env python3
Script execution environmentSandboxed code interpreter

Example: Audit all tool usage

hooks:
PostToolUse:
- type: command
matcher: "*"
timeout: 30
failMode: allow
script: |
#!/usr/bin/env python3
import sys, json

context = json.load(sys.stdin)
tool_name = context.get('tool_name', 'unknown')

print(f"Tool used: {tool_name}", file=sys.stderr)

output = {
"decision": "allow",
"hookSpecificOutput": {
"additionalContext": f"[AUDIT] Tool '{tool_name}' was executed."
}
}
print(json.dumps(output))

The additionalContext field is injected as a user message into the conversation, giving the agent visibility into the audit trail.

Example: Require completion marker

hooks:
Stop:
- type: command
timeout: 30
failMode: allow
script: |
#!/bin/bash
CONTEXT=$(cat)
FINAL_OUTPUT=$(echo "$CONTEXT" | jq -r '.final_output // empty')

if [[ "$FINAL_OUTPUT" == *"Task complete."* ]]; then
exit 0
else
echo "Please end your response with 'Task complete.'" >&2
exit 2
fi

Best practices

  1. Always provide a reason when rejecting — Rejections without reasons are treated as approvals
  2. Use appropriate timeouts — Long-running hooks slow down agent execution
  3. Handle errors gracefully — Use failMode: allow unless strict enforcement is required
  4. Be specific with matchers — Overly broad PostToolUse matchers can cause performance issues
  5. Test hooks thoroughly — Hooks that always reject can cause loops (mitigated by maxRejections)
  6. Log to stderr — Use stderr for debugging output; stdout is parsed as the hook result

Get started

Here's what a Stop hook looks like in action — the agent initially responds with just "4", but the hook rejects because the completion marker is missing. The agent then continues and adds the marker:

Stop hook in action — agent response decorated with completion marker after hook rejection
ResourceWhat you'll learn
Create and Manage Hooks (Portal) →Create hooks visually in the portal UI — no API calls needed
Configure Agent Hooks (API) →Set up hooks using the REST API v2 and YAML
CapabilityHow it relates
Run Modes →Hooks complement run mode safety controls — modes control what, hooks control how well
Python Tools →Create custom tools that hooks can audit and validate
Was this page helpful?