Skip to main content

Tool Actions

What happens when you need logic that an LLM can't perform? Deduplicating records, calling an external API, or applying deterministic business rules—these tasks need deterministic logic, not prompts.

Tool actions let you execute custom Python functions alongside LLM actions in your agentic workflow. When you need guaranteed, repeatable behavior, you use a tool.

Quick Example

from agent_actions import udf_tool

@udf_tool()
def process_text(data: dict, **kwargs) -> dict:
return {"result": data["text"].upper()}
- name: process_step
kind: tool
impl: process_text
granularity: record

@udf_tool Decorator

The @udf_tool decorator registers a Python function as a tool action.

Parameters

ParameterTypeRequiredDescription
granularityGranularityNoRECORD (default) or FILE processing
Schema Definition

Input and output schemas are defined in the workflow YAML schema: field, not in the decorator.

Minimal Decorator

from agent_actions import udf_tool

@udf_tool()
def simple_transform(data: dict, **kwargs) -> dict:
data['processed'] = True
return data

Granularity

GranularityProcessingUse Case
recordOne record at a timeTransformations, filtering
fileAll records at onceAggregation, deduplication

Record Granularity (Default)

@udf_tool()
def filter_questions_by_score(data: dict, **kwargs) -> dict:
score = data.get('syllabus_alignment_score', 0)
if score >= 85:
data['question_status'] = "KEEP"
else:
data['question_status'] = "FILTER"
return data

File Granularity

Use when your logic needs cross-record context. FILE tools receive full records with framework metadata (node_id, source_guid, lineage) and a content dict containing business data.

from agent_actions import udf_tool
from agent_actions.config.schema import Granularity

@udf_tool(granularity=Granularity.FILE)
def run_dedup(data: list[dict], **kwargs) -> list[dict]:
"""Dedup — return full records to preserve lineage."""
seen = set()
unique = []
for record in data:
content = record.get("content", record)
fact = content.get("fact", "")
if fact not in seen:
seen.add(fact)
unique.append(record) # return the full record
return unique
File Granularity is Tool-Only

File granularity is exclusively supported for tool actions. LLM actions must use record granularity.

File Granularity Constraints

  • Input is an array of full records — each record has content, node_id, source_guid, lineage
  • Read business data from record["content"]["field"] — not record["field"]
  • Return the original record for passthrough (filter, dedup, sort, transform) — preserves node_id and lineage
  • Return a new dict for aggregation (no node_id) — framework creates fresh lineage
  • Output flexibility — return an array of any size (N→M transformation)

See Granularity for detailed documentation.

Record Identity and Lineage

The framework tracks each record through the pipeline using node_id — inspired by Apache NiFi's FlowFile model. You never manage node_id directly. The framework handles it automatically based on what you return:

from agent_actions import udf_tool
from agent_actions.config.schema import Granularity

@udf_tool(granularity=Granularity.FILE)
def dedup_tool(data: list[dict], **kwargs) -> list[dict]:
seen = {}
outputs = []

for record in data:
content = record.get("content", record)
fact = content.get("fact", "")
if fact not in seen:
seen[fact] = True
outputs.append(record) # full record → lineage extended

return outputs


@udf_tool(granularity=Granularity.FILE)
def aggregate_tool(data: list[dict], **kwargs) -> list[dict]:
total = sum(r.get("content", r).get("score", 0) for r in data)
return [{"summary": f"Total: {total}", "count": len(data)}] # new dict → fresh lineage
What you returnFramework behavior
Original record dict (has node_id)Extends parent lineage — downstream observe can load ancestor data
New dict (no node_id)Creates new root — fresh lineage, no parent

Tool Discovery

Agent Actions automatically discovers tools decorated with @udf_tool from configured directories.

Configuration

Set the tool path in agent_actions.yml:

tool_path: ["tools", "custom_tools"]

Or use the TOOLS_PATH environment variable:

export TOOLS_PATH="tools"

Directory Structure

project/
├── agent_actions.yml
├── tools/
│ ├── __init__.py # Optional
│ ├── transformers.py # Contains @udf_tool functions
│ └── validators.py
└── agent_workflow/
└── ...

Workflow Reference

Reference tools by function name:

- name: flatten_the_facts
kind: tool
impl: flatten_quotes # Function name (case-insensitive)
granularity: record

Discovery Process

  1. Scans directories in tool_path recursively
  2. Loads all Python files (*.py), skipping files starting with _ or test_
  3. Executes modules to trigger @udf_tool decorator registration
  4. Validates impl references in agentic workflow config
Thread Safety

Tool discovery is thread-safe and cached. Concurrent discovery calls are properly synchronized, and modules are loaded only once.

CLI Commands

# List all discovered tools
agac list-udfs -u ./tools

# Validate tool references in workflow
agac validate-udfs -a my_workflow -u ./tools

Import Rules

Tool files are loaded in isolation using importlib.util.spec_from_file_location. The tools directory is not added to sys.path, so tools cannot import sibling modules with bare import statements.

What works:

  • Installed packages — any dependency in your project's virtual environment (e.g., import pandas, from agent_actions import udf_tool)
  • Standard libraryimport json, import pathlib, etc.

What does not work:

  • Sibling importsimport other_tool or from . import utils where other_tool and utils are files in the same tools directory

If you need shared logic across multiple tool files, extract it into an installable package and add it to your project dependencies.

Best Practices

Handle Missing Fields

@udf_tool()
def safe_function(data: dict, **kwargs) -> dict:
score = data.get('score', 0) # Use .get() with defaults
return {'result': score}

Return Complete Records

@udf_tool()
def augment_data(data: dict, **kwargs) -> dict:
data['new_field'] = 'computed_value' # Add to existing, don't replace
return data

Unique Function Names

Tool names must be unique across all files. Prefix with domain if needed:

@udf_tool()
def validate_email(data): ...

@udf_tool()
def validate_phone(data): ...

Error Handling

Duplicate Function Names

DuplicateFunctionError: Function 'process_data' already registered
Existing: module_a.process_data (tools/module_a.py)
New: module_b.process_data (tools/module_b.py)

Rename one of the functions.

Function Not Found

FunctionNotFoundError: Function 'nonexistent_func' not found

Check that:

  1. File is in tools/ directory
  2. Function has @udf_tool decorator
  3. Function name matches workflow impl