38

AI Product & UX

⏱️ 35 min

AI Product UX

AI UX tends to land at one of two extremes: either it's built like a regular form product that completely ignores model uncertainty, or it's a flashy demo that piles on "smart vibes" while users have no idea what to trust. Good AI product UX isn't about making the interface look AI-powered -- it's about giving users control and trust inside a system that's inherently uncertain.

So this page isn't about visual mockups. It's about how AI engineers and product/design teams should design more reliable AI UX patterns together.

AI Product UX Ladder


Bottom line: AI UX is about trust, not novelty

Whether users stick with an AI feature usually comes down to 4 things:

  1. Do they know what the feature can and can't do?
  2. Can they tell why an output is worth trusting?
  3. Can they fix results without starting over?
  4. Is the system honest when it fails?

Get these 4 right and you'll beat fancy animations every time.


The biggest difference between AI UX and traditional UX

Traditional productAI product
Output is relatively certainOutput is probabilistic
Users form stable expectations easilyUsers tend to over- or underestimate capability
Errors are usually obviousErrors can "look correct"
Flows are more linearOften needs refine, retry, review

So AI UX can't just copy traditional form thinking.


Input UX: help users provide enough context

Many "the model sucks" complaints are actually input design problems.

Better input UX typically provides:

MechanismPurpose
prompt template / starterReduces blank-input anxiety
constraints hintTells users about length, format, scope
file / source previewShows users what context the system has
scope clarificationAsks follow-ups when info is lacking instead of guessing

AI features shouldn't assume users will write great prompts on their own.


Output UX: make results judgeable

An AI output should at minimum let users answer:

  1. Is this based on what I gave it?
  2. Did it cite a source?
  3. Can I directly edit this part?
  4. If I'm not satisfied, how do I refine?

That's why these patterns matter:

  • streaming
  • citation
  • confidence / limitation cues
  • quick refine actions

Refine loops matter more than "regenerate"

A regenerate button alone usually isn't enough.

Better AI UX provides low-friction correction paths like:

ActionUser feeling
shorter / longerQuick length control
more formal / more casualQuick tone adjustment
fix structureKeep content, reorganize
ask follow-upAdd more context when needed

These refine loops noticeably improve the user's sense of control.


Error UX must be honest

The most dangerous AI UX pattern is disguising failure as "looks like it worked."

More reliable error design should:

  • Show a clear error when a provider fails -- don't pretend the model is still thinking
  • Admit uncertainty when sources are insufficient
  • Show partial results for partial success
  • Provide a human escalation path for high-risk scenarios

AI UX failure isn't just an experience problem -- it's a trust problem.


Citation and source UX: critical for high-value scenarios

In knowledge-heavy scenarios, users don't just want answers. They want to know:

  • What's the source?
  • Which passage was cited?
  • Is it outdated?

Source UX is painful to build upfront. But once it's in place, user trust goes up significantly.


Memory and personalization need user control

An AI system remembering user preferences is valuable. But it shouldn't be a black box.

A more reliable approach makes these things explicit:

QuestionHow to surface it in UX
What did it remember?Visible preference summary
How long will it keep it?Retention / privacy explanation
Can I clear it?Clear / reset action
Is it personal or shared context?Visible context boundary

The stronger the memory, the clearer the boundaries need to be.


The metrics AI UX should actually track

MetricWhy it matters
task success rateDid users actually finish their task?
refine rateAre users actively correcting or forced to retry?
abandonment rateDid users give up midway?
feedback scoreHow's the subjective experience?
source click / review rateAre users verifying results?

Only looking at usage volume without tracking refine and abandonment won't tell you if UX is actually good.


Practice

Take one of your current AI features and check these 4 things:

  1. Do users know the feature's boundaries?
  2. Is the output judgeable and correctable?
  3. Are errors honest?
  4. Is there a usable refine loop?

Get these 4 right and your AI UX is starting to mature.

📚 相关资源

❓ 常见问题

关于本章主题最常被搜索的问题,点击展开答案

AI 产品 UX 跟传统产品 UX 最大的区别是什么?

输出从确定变成概率性 — 用户容易高估或低估能力,错误可能"看起来像对的",流程从线性变成 refine / retry / review。所以不能套传统 form thinking。AI UX 的核心不是 novelty 是 trust:用户知不知道功能边界、能不能看懂输出、能不能修正、出错时是否诚实。

为什么单纯一个 regenerate 按钮还不够?

Regenerate 让用户每次重抽盲盒,控制感差。更好的 refine loop 给低摩擦修正路径:shorter / longer 控篇幅、more formal / more casual 调 tone、fix structure 保留内容只改组织、ask follow-up 补上下文。这种局部调整比全部重来明显提升用户控制感和留存。

AI 产品的 error UX 最忌讳什么?

把失败伪装成成功 — 比 provider down 更伤 trust。规范四条:provider fail 就明确提示别假装思考中、source 不足就承认不确定、partial success 展示 partial result、高风险场景给人工升级路径。AI UX 的失败不只是体验问题,是 trust 问题,一次伪装成功流失的用户拉不回来。

AI memory 功能要给用户哪些控制权?

四个必须可见:(1) 记住了什么 — 一份 visible preference summary;(2) 保留多久 — retention 与 privacy 说明;(3) 能不能清除 — clear / reset 按钮;(4) 个人还是共享上下文 — visible context boundary。记忆能力越强,边界越要画清楚,否则用户感觉是黑盒,trust 反而降。

AI feature 该追的指标除了使用量还有什么?

本章列了五个:task success rate(任务真完成率)、refine rate(用户在积极修正还是被迫重试)、abandonment rate(中途放弃)、feedback score(主观体验)、source click / review rate(用户在不在验证结果)。只看 DAU / 调用次数会被 retry 掩盖真实体验,refine 和 abandonment 更能暴露问题。