Protection Model
A secondary model or rule set (e.g., a regex, an open-source ML detector, or a custom Python routine) that screens agent inputs/outputs for disallowed content—such as injection patterns, sensitive data, or toxic language—and blocks or sanitizes anything that fails the check.
All Terms