-
What the Evaluator Needs to Be
The previous posts in this series made the case for why behavioral alignment alone won’t hold once AI systems gain memory, tool use, and recursive self-improvement. Constraint-by-Balance proposes a structural answer: embed harm-balancing logic directly into the agent’s runtime flow, so that constraint operates independently of optimization. This post lays out what that means in
-
A Model of AI Agent Types
In the last two posts look at motivations for the C-by-B architecture and looked at how current AI behaviors hint at more dire future alignment issues. With this post we are switching from concerns to remedies. We will start by grounding the C-by-B architecture in a model of AI agent types. Efficiency and Efficacy –
-
The Motivation for Constraint-by-Balance: The Safety Gap After Deployment
What does the future look like once it’s populated with all manner of AI agents? Do our current safety approaches fully encompass the risks associated with that future? The best-known approaches to AI safety (RLHF, Constitutional AI, scalable oversight, interpretability research) have made remarkable progress at aligning model behavior during training and evaluation. These methods