Settings

Theme

Ask HN: Can verifiable honesty rebuild trust in AI systems?

1 points by renshijian 2 months ago · 5 comments · 1 min read


We built a system that makes AI honesty measurable each response carries a determinacy score, deception probability, and ethical weight

Instead of “trust me,” the model says “check for yourself.”

I wonder — can this kind of transparency help rebuild trust in AI? Or does it just expose how uncertain intelligence really is?

renshijianOP 2 months ago

Thanks for reading — this project isn’t about “AI safety theater.” We’re experimenting with verifiable honesty: every model response carries its own determinacy, deception probability, and ethical weight

Instead of “trust me,” the system says, “check for yourself.”

We’re curious how the HN community sees this: Can trust in AI be engineered through transparency? Or does showing the uncertainty just make it harder to trust?

  • JohnFen 2 months ago

    How does that let you check for yourself, though? Don't people still have to trust that the reported probabilities and weights are both meaningful and correct?

    Also, people tend to be pretty bad at interpreting probabilities.

    • renshijianOP 2 months ago

      That’s a fair point — verification itself still depends on trust in the verifier What we’re trying to test isn’t absolute truth, but transparency under uncertainty You’re right that people often misread probabilities — but maybe that’s the point If we can see uncertainty, even imperfectly, it starts a different kind of trust Not blind faith — more like “I know what this system knows it doesn’t know.”

    • renshijianOP 2 months ago

      I’ve been thinking more about that — maybe verifiability doesn’t remove trust, it just changes where we place it Not in the system, but in the transparency of its process

PaulHoule 2 months ago

Gotta see it in action.

Keyboard Shortcuts

j
Next item
k
Previous item
o / Enter
Open selected item
?
Show this help
Esc
Close modal / clear selection