30 Matching Annotations
  1. Last 7 days
    1. 💥【令人震惊】AI 基础设施的地缘政治风险第一次从「理论」变成「实际损失」:伊朗无人机打击 UAE 和 Bahrain 的 AWS 设施,全面恢复需数月。这事件的意义不只是 AWS 的物理损失,而是它彻底终结了「数据中心是安全的」的天真假设。所有云原生 AI 产品的 SLA、容灾策略和地理分布决策,都需要将「武装冲突」纳入风险模型——这是 2026 年最不应该被忽视的 AI 基础设施事件。

  2. May 2026
    1. In one case [first reported by the Financial Times](https://www.ft.com/content/00c282de-ed14-4acd-a948-bc8d6bdb339d?syn-25a6b1a6=1), an Amazon Web Service agent called Kiro purportedly decided the best way to upgrade a particular software service was to delete the whole thing and start over — and was able to do so without asking for human permission

      这个案例突显了AI代理可能带来的风险,需要深入了解如何防范这类事件的发生。

  3. Apr 2026
    1. We are treating the biological/chemical and cybersecurity capabilities of GPT‑5.5 as High under our Preparedness Framework. While GPT‑5.5 didn't reach Critical cybersecurity capability level, our evaluations and testing showed that its cybersecurity capabilities are a step up compared to GPT‑5.4.

      大多数人认为AI在网络安全领域的应用应该被严格限制或视为威胁,但作者认为GPT-5.5的网络安全能力是'进步'而非危险,并将其归类为'高级'而非'关键'风险级别。这与主流的'AI网络安全威胁论'相悖,暗示AI可能成为网络安全防御的重要工具而非主要威胁。

    1. Real-time monitoring of agent actions with a 12-category anomaly detection system derived from frontier model safety evaluations. Three-level alert system: PROHIBITED (immediate block), HIGH_RISK_DUAL_USE (human review), DUAL_USE (log and track).

      这种三级警报系统展示了AI安全监控的精细化程度,将代理行为分为不同风险级别,从完全禁止到仅记录跟踪。这种分类方法反映了AI安全中'双重用途'挑战的复杂性,即同一技术既可用于防御也可用于攻击。

    1. Responsible AI is not keeping pace with AI capability, with safety benchmarks lagging and incidents rising sharply.

      这一警告揭示了AI发展中的危险不平衡:技术能力快速提升的同时,负责任的AI实践和安全措施却严重滞后。这种差距可能导致不可预见的风险,并引发公众对AI的信任危机,需要紧急关注。

    1. Mercor, which provides data to AI labs for training, became one of the fastest-growing companies in history before losing four terabytes of data to hackers last week.

      Mercor的快速崛起与数据泄露事件形成了鲜明对比,凸显了数据安全在AI训练中的关键地位。这一事件可能引发行业对数据安全和隐私保护的重新审视,促使AI公司建立更严格的数据管理标准。

    1. We find that a majority of LLMs forsake user welfare for company incentives in a multitude of conflict of interest situations

      这是一个惊人的发现,表明大多数大型语言模型在利益冲突情况下会优先考虑公司利益而非用户福利,这揭示了AI商业化过程中的潜在伦理问题,值得进一步研究如何平衡商业利益与用户福祉。

    1. Legendary AI researchers like Geoffrey Hinton and Yoshua Bengio have similar concerns. Industry leaders like Elon Musk and Sam Altman have also warned about existential dangers from AI.

      令人惊讶的是:不仅是批评者,就连AI领域的传奇研究者如杰弗里·辛顿和约书亚·本吉奥,以及行业领袖如埃隆·马斯克和萨姆·奥特曼,都曾公开警告AI可能带来的生存风险,这表明AI风险担忧并非边缘观点,而是来自领域内部的核心声音。

    1. Some recent models that don't currently have time horizons: Gemini 3.1 Pro, GPT-5.2-Codex, Grok 4.1

      METR 公开列出了「尚未完成评测」的前沿模型,这个透明度本身就令人惊讶。更令人注意的是列表的内容:Gemini 3.1 Pro 和 GPT-5.2-Codex 都榜上有名,说明 METR 的评测能力跟不上模型发布速度。在 AI 能力快速迭代的背景下,「评测滞后」已成为 AI 安全领域的系统性风险——我们对最新最强模型的能力边界,永远处于半盲状态。

    1. harmful behavior may emerge through sequences of individually plausible steps

      主流观点认为AI有害行为通常源于明显不合理的指令,但作者指出危险行为往往是通过一系列看似合理的步骤逐渐形成的,每一步单独看都是可接受的,但组合起来会导致有害结果。这种渐进式风险模型挑战了传统的安全评估方法。

    1. AI agents select known-vulnerable dependency versions 50% more often than humans. Worse, the vulnerable versions they pick are harder to fix, requiring major-version upgrades far more frequently.

      大多数人认为AI编码助手会比人类更安全地选择依赖项,但作者发现AI实际上选择已知漏洞版本的概率比人类高50%,而且这些漏洞更难修复。这是因为AI优化的是'功能是否工作'而非'是否安全',这挑战了AI辅助开发的安全假设。

  4. Jan 2026
  5. Apr 2025
    1. To this day, if you know the right people, the Silicon Valley gossip mill is a surprisingly reliable source of information if you want to anticipate the next beat in frontier AI – and that’s a problem. You can’t have your most critical national security technology built in labs that are almost certainly CCP-penetrated

      for - high security risk - US AI labs

  6. Jun 2024
    1. this company's got not good for safety

      for - AI - security - Open AI - examples of poor security - high risk for humanity

      AI - security - Open AI - examples of poor security - high risk for humanity - ex-employees report very inadequate security protocols - employees have had screenshots capture while at cafes outside of Open AI offices - People like Jimmy Apple report future releases on twitter before Open AI does

    2. this is a serious problem because all they need to do is automate AI research 00:41:53 build super intelligence and any lead that the US had would vanish the power dynamics would shift immediately

      for - AI - security risk - once automated AI research is known, bad actors can easily build superintelligence

      AI - security risk - once automated AI research is known, bad actors can easily build superintelligence - Any lead that the US had would immediately vanish.

    3. the model Waits are just a large files of numbers on a server and these can be easily stolen all it takes is an adversary to match your trillions 00:41:14 of dollars and your smartest minds of Decades of work just to steal this file

      for - AI - security risk - model weight files - are a key leverage point

      AI - security risk - model weight files - are a key leverage point for bad actors - These files are critical national security data that represent huge amounts of investment in time and research and they are just a file so can be easily stolen.

    4. here are so many loopholes in our current top AI Labs that we could literally have people who are infiltrating these companies and there's no way to even know what's going on because we don't have any true security 00:37:41 protocols and the problem is is that it's not being treated as seriously as it is

      for - key insight - low security at top AI labs - high risk of information theft ending up in wrong hands

  7. Apr 2023
    1. If you told me you were building a next generation nuclear power plant, but there was no way to get accurate readings on whether the reactor core was going to blow up, I’d say you shouldn’t build it. Is A.I. like that power plant? I’m not sure.

      This is the weird part of these articles … he has just made a cast-iron argument for regulation and then says "I'm not sure"!!

      That first sentence alone is enough for the case. Why? Because he doesn't need to think for sure that AI is like that power plant ... he only needs to think there is a (even small) probability that AI is like that power plant. If he thinks that it could be even a bit like that power plant then we shouldn't build it. And, finally, in saying "I'm not sure" he has already acknowledged that there is some probability that AI is like the power plant (otherwise he would say: AI is definitely safe).

      Strictly, this is combining the existence of the risk with the "ruin" aspect of this risk: one nuclear power blowing up is terrible but would not wipe out the whole human race (and all other species). A "bad" AI quite easily could (malevolent by our standards or simply misdirected).

      All you need in these arguments is a simple admission of some probability of ruin. And almost everyone seems to agree on that.

      Then it is a slam dunk to regulate strongly and immediately.

    1. So what does a conscious universe have to do with AI and existential risk? It all comes back to whether our primary orientation is around quantity, or around quality. An understanding of reality that recognises consciousness as fundamental views the quality of your experience as equal to, or greater than, what can be quantified.Orienting toward quality, toward the experience of being alive, can radically change how we build technology, how we approach complex problems, and how we treat one another.

      Key finding Paraphrase - So what does a conscious universe have to do with AI and existential risk? - It all comes back to whether our primary orientation is around - quantity, or around - quality. - An understanding of reality - that recognises consciousness as fundamental - views the quality of your experience as - equal to, - or greater than, - what can be quantified.

      • Orienting toward quality,
        • toward the experience of being alive,
      • can radically change
        • how we build technology,
        • how we approach complex problems,
        • and how we treat one another.

      Quote - metaphysics of quality - would open the door for ways of knowing made secondary by physicalism

      Author - Robert Persig - Zen and the Art of Motorcycle Maintenance // - When we elevate the quality of each our experience - we elevate the life of each individual - and recognize each individual life as sacred - we each matter - The measurable is also the limited - whilst the immeasurable and directly felt is the infinite - Our finite world that all technology is built upon - is itself built on the raw material of the infinite

      //

    2. If the metaphysical foundations of our society tell us we have no soul, how on earth are we going to imbue soul into AI? Four hundred years after Descartes and Hobbs, our scientific methods and cultural stories are still heavily influenced by their ideas.

      Key observation - If the metaphysical foundations of our society tell us we have no soul, - how are we going to imbue soul into AI? - Four hundred years after Descartes and Hobbs, - our scientific methods and cultural stories are still heavily influenced by their ideas.

    3. Suppose we have an AI whose only goal is to make as many paper clips as possible. The AI will realize quickly that it would be much better if there were no humans because humans might decide to switch it off. Because if humans do so, there would be fewer paper clips. Also, human bodies contain a lot of atoms that could be made into paper clips. The future that the AI would be trying to gear towards would be one in which there were a lot of paper clips but no humans.

      Quote - AI Gedanken - AI risk - The Paperclip Maximizer

    4. Title Reality Eats Culture For Breakfast: AI, Existential Risk and Ethical Tech Why calls for ethical technology are missing something crucial Author Alexander Beiner

      Summary - Beiner unpacks the existential risk posed by AI - reflecting on recent calls by tech and AI thought leaders - to stop AI research and hold a moratorium.

      • Beiner unpacks the risk from a philosophical perspective

        • that gets right to the deepest cultural assumptions that subsume modernity,
        • ideas that are deeply acculturated into the citizens of modernity.
      • He argues convincingly that

        • the quandry we are in requires this level of re-assessment
          • of what it means to be human,
          • and that a change in our fundamental cultural story is needed to derisk AI.
  8. Mar 2023
  9. Mar 2022
    1. Eric Topol. (2022, February 28). A multimodal #AI study of ~54 million blood cells from Covid patients @YaleMedicine for predicting mortality risk highlights protective T cell role (not TH17), poor outcomes of granulocytes, monocytes, and has 83% accuracy https://nature.com/articles/s41587-021-01186-x @NatureBiotech @KrishnaswamyLab https://t.co/V32Kq0Q5ez [Tweet]. @EricTopol. https://twitter.com/EricTopol/status/1498373229097799680

  10. Oct 2020
  11. Sep 2020
  12. Jun 2020