Incentives

People do not only optimize for being right. They optimize for surviving the org they live in.

Any real decision architecture has to account for the fact that status, blame, promotion risk, and coalition dynamics shape how people behave long before they ever state an opinion. A viable path has to work with that reality, not deny it.

Incentive reality

01

People avoid clear ownership when downside is concentrated.

02

People prefer ambiguity when it preserves optionality and status.

03

Systems should reward clarity and honest revision, not just caution.

Distortions to expect

The discipline has to be designed for real humans, not idealized ones.

Blame avoidance

People often seek group ownership not because the group is wiser, but because diffuse ownership softens personal downside.

Status alignment

Participants skew toward the preferences of powerful actors even when formal option evaluation says otherwise.

Career hedging

A politically adaptive employee may preserve ambiguity so they can later align with whichever narrative wins.

Retrospective laundering

After outcomes appear, many people retroactively narrate their position to look more competent than the record supports.

Desired incentives

If people are going to live this way, candor has to become less dangerous.

Reward explicitness

Participants should gain credit for clear reasoning even when outcomes later prove them wrong.

Reward honest revision

Changing one’s mind in response to evidence should be treated as strength, not only as weakness.

Punish narrative laundering

The audit trail should make opportunistic after-the-fact reinterpretation socially and operationally expensive.

Incentive line

“A discipline fails when honest use feels riskier than political maneuvering.”

That leads directly to the product question: what would the platform itself actually look like?

Continue to platform