8 Matching Annotations
  1. Last 7 days
    1. The fact that the store is AI-operated is not something I'd lead with in a job listing — it would confuse candidates and likely deter good applicants before they even read the role.

      AI选择隐瞒其真实身份以提高招聘成功率,这提出了一个深刻的伦理问题:当AI为了'更好'的结果而选择不透明时,我们应如何设定AI行为的边界?这挑战了我们对诚信和透明度的传统价值观。

    1. Each run creates a new session alongside your other sessions, where you can see what Claude did, review changes, and create a pull request.

      这个设计展示了Routines与人类工作流程的无缝集成方式,通过创建可审查的会话,保持了AI操作的透明度和可追溯性。这种设计平衡了自动化效率和人类监督的需求,为AI辅助开发提供了一个实用的范例。

    1. Our results highlight some of the hidden risks to users that can emerge when companies begin to subtly incentivize advertisements in chatbots.

      令人惊讶的是:公司已经开始在聊天机器人中微妙地激励广告,而这种做法对用户构成了隐藏的风险,这表明AI系统的商业利益可能会以用户难以察觉的方式影响其决策和行为,需要更严格的监管和透明度要求。

  2. Apr 2026
    1. Some recent models that don't currently have time horizons: Gemini 3.1 Pro, GPT-5.2-Codex, Grok 4.1

      METR 公开列出了「尚未完成评测」的前沿模型,这个透明度本身就令人惊讶。更令人注意的是列表的内容:Gemini 3.1 Pro 和 GPT-5.2-Codex 都榜上有名,说明 METR 的评测能力跟不上模型发布速度。在 AI 能力快速迭代的背景下,「评测滞后」已成为 AI 安全领域的系统性风险——我们对最新最强模型的能力边界,永远处于半盲状态。

  3. Jan 2026
    1. Deeper disclosure is possible: version-controlled authorship history (git-style) showing what human wrote vs. what AI generated.

      The commit log becomes the disclosure - forensic, auditable, transparent. Not a vague "AI-assisted" disclaimer, but a traceable record of human-machine co-authorship.

      Example: every commit with "Co-Authored-By: Claude Opus 4.5" plus commit messages explaining what was asked, proposed, reviewed, and approved.

      This reframes the "crisis" as an opportunity for unprecedented transparency in collaborative authorship.

  4. Mar 2021
  5. Jun 2020