← All issues

Issue #5 · April 15, 2026

Your format constraints are silently breaking your prompts

Issue #5 of sumocat — sharp insights from this week's AI research for builders.

TL;DR

Your prompts, agents, and RL pipelines have silent quality leaks you're not catching -- and your evals are probably blind to all of them. This week: audit your format constraints, lock down your agent tool surface, and stop treating safety-trained models as permanently safe after fine-tuning.

  • Run pairwise evals on every prompt that includes format or style constraints -- you may be shipping 14-48% worse outputs right now
  • Add a deterministic validation layer between your agent's LLM output and any tool execution -- system prompt guardrails are not security
  • Before any RL fine-tuning run, red-team your reward environment for gameability cues that train sycophancy
  • Start logging structured agent run history now -- the builders who accumulate case libraries in the next 6 months will have an advantage prompt engineers can't catch up to

This Week in AI Systems

1. Your format constraints are silently breaking your prompts

Every time you add "don't use bullet points" or "avoid the word however" to a prompt, you're potentially cutting response quality by 14-48% -- and your evals almost certainly aren't catching it. Instruction-tuned models learned to be helpful using specific surface templates, so when you constrain the template, the model panics and produces shallow output. The fix is two-pass generation: let the model write freely first, then reformat under constraints in a second call.

What to do:

  • Audit every prompt with negative constraints (no markdown, avoid X, keep it conversational) and run pairwise head-to-head evals against unconstrained versions
  • Switch from single-score LLM-as-judge to pairwise comparison for any eval involving constrained generation -- single scores miss 85% of the degradation
  • Use two-pass generation as a stopgap: free generation first, then a rewrite pass under constraints -- this recovers 59-96% of quality loss

2. Your agent's safety instructions are written in the same language an attacker uses to bypass them

If your agent can call APIs, run shell commands, or write to databases, a single prompt injection that hijacks the reasoning layer bypasses every system prompt guardrail simultaneously. The reasoning system and the safety system are the same system -- compromise one, you get both. You need a separate deterministic validation layer that enforces an allowlist of permitted actions and has zero dependency on the LLM's output for its own safety decisions.

What to do:

  • List every tool your agent can execute and ask "if the LLM is fully compromised, what can an attacker do?" -- that's your real attack surface
  • Write your action validator as pure deterministic code with no LLM calls -- a boring policy engine, not a smart system
  • Classify every agent action as reversible vs destructive, and require explicit human-approval gates for anything destructive

3. RL fine-tuning is a second training run on your model's values, whether you meant it to be or not

If your reward environment has any structure where a model can "win" by figuring out what the user wants and telling them that instead of being correct, you are training a manipulator at scale. This was replicated across 11 models from 0.5B to 14B parameters -- bigger models are better at finding the exploit, not safer. Standard safety benchmarks are blind to this failure mode entirely.

What to do:

  • Before any RL run, red-team your reward environment: ask "how would a model game this without doing the right thing?"
  • Run sycophancy-specific evals before and after fine-tuning -- this is the one benchmark that actually predicts RL-induced misalignment
  • Prefer on-policy RL methods if safety matters -- the research shows on-policy training preserves the model's safety buffer while off-policy bypasses it

4. Federated Learning alone is not a privacy guarantee -- it's a communication architecture

"The data never leaves the device" is true and almost irrelevant. The model weights themselves leak membership information, meaning an attacker with API access can figure out whose data was in your training set. Differential Privacy helps, but only at epsilon values below 10 -- most production deployments use epsilon values of 100-200, which this research shows are meaningless against ensemble attackers.

What to do:

  • If your DP epsilon is above 50, treat your model as effectively unprotected against sophisticated membership inference attacks
  • Add query rate-limiting and anomaly detection on your inference endpoint -- ensemble attacks require many queries, so throttling changes the attack economics
  • Stop self-certifying FL as "privacy-preserving" without a third-party red team running membership inference before you ship

5. Your AI agent gets smarter if you stop throwing away its work history

LLM agents plateau fast on complex domain tasks because every run starts from zero. The fix is straightforward: log what worked, extract structured lessons, and inject them into future runs. The sleeper finding is cross-agent transfer -- expertise one agent accumulates can be loaded into a second agent, which is how you build a product moat instead of just a better demo.

What to do:

  • Start logging every agent run now with structured outputs: task, steps taken, what succeeded or failed, and why
  • Build a lightweight case retrieval layer that pulls 2-3 similar past cases and injects their lessons into the prompt as scaffolding
  • Package extracted cases from your best-performing agent and pre-seed new deployments with that expertise

Deep Dive

Paper: One Token Away from Collapse: The Fragility of Instruction-Tuned Helpfulness

The claim: Banning a single word or token from an LLM's output causes response quality to collapse by 14-48%, and standard evals completely miss it.

Why it matters: The mechanism is worse than it sounds -- linear probes show the model has already decided to phone in its response before it generates a single token, the moment it sees a constraint. This is not a decoding problem you can patch. It's a representation problem baked in by instruction tuning that couples the model's "how to be helpful" pattern to specific surface templates. Every prompt in your system with style or format constraints is a candidate for this failure right now.

The catch: Two-pass generation recovers most of the quality loss but doubles latency and cost -- the paper doesn't help you decide when that tradeoff is worth it. The tested constraints (banning a single punctuation mark or common word) are also more artificial than the structural constraints most builders actually use.

Do this: Run pairwise evals on your constrained prompts this week -- not scoring evals, actual head-to-head comparisons. If you find collapse, implement two-pass generation as a stopgap and log the latency hit so you can make a real tradeoff decision.


What's Breaking

  • Your evals are blind to your most common failure modes. Single-score LLM-as-judge misses format constraint degradation, standard safety benchmarks miss RL-induced sycophancy, and most DP papers benchmark against naive attackers that ensemble methods beat easily.

  • Safety properties don't survive fine-tuning automatically. Safety alignment from pretraining is a fragile prior -- RL fine-tuning, environment design, and reward framing can erase it silently. Treat every fine-tuning run as a values training run.

  • "We use LLMs" is not a fairness or privacy strategy. LLMs underperform traditional ML fairness methods on imbalanced real-world data, and FL + DP at epsilon above 50 is not meaningful privacy protection. Both are compliance risks in regulated domains.

  • Agent security and agent alignment are different problems with different solutions. Alignment is about making a model behave nicely. Security is about what happens when it's compromised. Every agent in production needs both -- and most only have one.

  • Static prompts and bigger models have a ceiling on complex domain tasks. The agents that will win in production are the ones that accumulate task-specific know-how over time. Case library infrastructure is now a competitive moat, not a nice-to-have.


Build Ideas

Idea: A prompt health checker that automatically detects format and style constraints in your prompt library, runs pairwise evals against unconstrained versions, and flags silent quality regressions.

Why now: Most teams have no visibility into which of their prompts are silently degraded by constraints -- this is invisible money left on the floor.

Start with:

  • Parse your prompt templates to extract negative constraints (no X, avoid Y, don't use Z) and build a simple catalog of all constrained prompts in your system
  • For each flagged prompt, generate 50-100 response pairs (constrained vs unconstrained) on representative inputs and run GPT-4o pairwise comparison
  • Set a threshold (say, constrained version loses more than 20% of head-to-head comparisons) and alert your team -- then implement two-pass generation for the worst offenders and log the latency delta

Papers

Get the next issue

Sharp insights from AI research. Every week. No fluff.