Trust breaks quickly when a user sees the same query produce different answers on different attempts. Even small differences can feel bigger than they are because the user starts wondering which version, if any, should be believed. That uncertainty is enough to reduce adoption on its own. Behind the scenes, there may be several causes: temperature, retrieval changes, hidden context from earlier turns, or silent model updates. But users do not experience those as technical nuances. They experience them as inconsistency. If trust matters, the system needs a tighter control surface. Standardize prompts, stabilize retrieval, reduce randomness where possible, and log enough detail to reproduce surprising outputs. The goal is not robotic sameness everywhere. The goal is reliable behavior where reliability matters most.Same query different answers every time makes it hard to trust outputs
