15 Matching Annotations
  1. May 2026
    1. In one case [first reported by the Financial Times](https://www.ft.com/content/00c282de-ed14-4acd-a948-bc8d6bdb339d?syn-25a6b1a6=1), an Amazon Web Service agent called Kiro purportedly decided the best way to upgrade a particular software service was to delete the whole thing and start over — and was able to do so without asking for human permission

      这个案例突显了AI代理可能带来的风险,需要深入了解如何防范这类事件的发生。

  2. Apr 2026
    1. The architecture scales horizontally to 300 sub-agents executing across 4,000 coordinated steps simultaneously, a substantial expansion from K2.5's 100 sub-agents and 1,500 steps.

      大多数人认为AI系统的扩展主要依赖于增加单个模型的计算能力和参数规模,而非增加智能体的数量。作者提出的300个智能体并行执行的模式挑战了这一认知,暗示未来AI发展可能更侧重于'多智能体协作'而非'单一模型增强',这可能会重新定义AI系统的架构设计原则。

    1. And it’s not just office work. Multi-agent tools like Google DeepMind’s Co-Scientist let researchers use teams of AI agents to coordinate literature searches, generate and test hypotheses, design experiments, and more.

      大多数人可能认为人工智能在办公室工作中的应用仅限于数据处理,但作者提出,多智能体工具甚至可以用于研究工作,如文献搜索和实验设计。

    1. Lightweight Agent Detection & Response (ADR) layer for AI agents — guards commands, files, and web requests.

      这个项目定义了一个新的'ADR'(Agent Detection & Response)层概念,这标志着AI安全领域的一个重要演进。从传统的端点保护转向专门针对AI代理的轻量级防护,反映了安全行业对AI特定威胁模式的适应和专业化。

    1. Meta also explicitly highlighted parallel multi-agent inference as a way to improve performance at similar latency

      令人惊讶的是,Meta明确强调了并行多代理推理作为在相似延迟下提高性能的方法。这表明AI系统正在从单一模型向多代理系统演进,可能是解决复杂问题的新范式,同时也暗示了未来AI系统架构的重大转变。

    1. scaling Muse Spark with multi-agent thinking enables superior performance with comparable latency.

      令人惊讶的是:通过扩展并行智能体的数量而非延长单个智能体的思考时间,Muse Spark能够在保持相近延迟的同时实现更优性能。这种多智能体协调的推理方式挑战了传统AI模型通过增加计算时间提高性能的范式,为高效推理提供了新思路。

    1. Build autonomous agents that plan, navigate apps, and complete tasks on your behalf, with native support for function calling.

      一个能在手机上离线运行的 2B 模型,原生支持 Function Calling 和多步 Agent 规划——这意味着完全本地化的 AI Agent 在消费级硬件上正式成为现实。结合 Android Studio 的 Agent Mode 支持,AI Agent 从云端走向终端的时间点,可能比所有人预计的都要早。

    1. computer-use agents extend language models from text generation to persistent action over tools, files, and execution environments

      作者暗示,从文本生成扩展到持久性工具使用是AI安全范式的一个根本转变,这一转变带来的安全挑战被当前研究低估。这挑战了将语言模型安全方法直接应用于代理系统的主流做法,提出了需要专门针对代理行为的安全评估框架。

    1. AI Agent 可以通过标准 MCP 协议直接读取和操作 𝕏 平台:搜索推文、发帖、查看用户信息、管理书签、收发私信等。

      大多数人认为社交媒体平台会严格限制第三方自动化操作以防止滥用,但作者指出xAI全面开放了MCP协议支持,允许AI Agent直接执行各种操作,这与主流平台的封闭趋势形成鲜明对比。

    1. An agent cannot be held accountable. I think about this principle most. The instinct to put a human in the loop is understandable, but taken literally, it can mean a person approving every step before anything moves forward. The human becomes a bottleneck, rubber-stamping work rather than directing it, and you lose much of what makes agents valuable in the first place.

      大多数人认为在AI系统中加入人类审批环节是确保问责制的必要措施,但作者认为这会使人类成为瓶颈,削弱代理的价值。这一观点挑战了AI安全与问责的主流思维,提出了一个非传统的责任分配模式。

  3. Jul 2025
  4. Dec 2024
    1. Historically, AI was a tool

      for - quote - AI: from tool b to agent - Roman Yampolskiy

      quote - AI: from tool b to agent - Roman Yampolskiy - (see below)

      • Historically, AI was a tool, like any other technology. Whether it was good or bad was up to the user of that tool.
      • You can use a hammer to build a house or kill someone.
      • The hammer is not in any way making decisions about it.
      • With advanced AI, we are switching the paradigm
        • **from tools
        • to agents**.
      • The software becomes capable of making its own decisions, working independently, learning, self-improving, modifying.
      • How do we stay in control?
      • How do we make sure the tool doesn’t become an agent that does something we don’t agree with or don’t support?
      • Maybe something against us
  5. Jun 2024
  6. Sep 2023
    1. the Bodhisattva vow can be seen as a method for control that is in alignment with, and informed by, the understanding that singular and enduring control agents do not actually exist. To see that, it is useful to consider what it might be like to have the freedom to control what thought one had next.
      • for: quote, quote - Michael Levin, quote - self as control agent, self - control agent, example, example - control agent - imperfection, spontaneous thought, spontaneous action, creativity - spontaneity
      • quote: Michael Levin

        • the Bodhisattva vow can be seen as a method for control that is in alignment with, and informed by, the understanding that singular and enduring control agents do not actually exist.
      • comment

        • adjacency between
          • nondual awareness
          • self-construct
          • self is illusion
          • singular, solid, enduring control agent
        • adjacency statement
          • nondual awareness is the deep insight that there is no solid, singular, enduring control agent.
          • creativity is unpredictable and spontaneous and would not be possible if there were perfect control
      • example - control agent - imperfection: start - the unpredictability of the realtime emergence of our next exact thought or action is a good example of this
      • example - control agent - imperfection: end

      • triggered insight: not only are thoughts and actions random, but dreams as well

        • I dreamt the night after this about something related to this paper (cannot remember what it is now!)
        • Obviously, I had no clue the idea in this paper would end up exactly as it did in next night's dream!
  7. Jul 2022