67 Matching Annotations
  1. Last 7 days
    1. When inference is expensive, teams limit usage, reduce context, or avoid certain applications altogether.

      文章指出推理成本高昂会导致团队限制使用、减少上下文或避免某些应用。这个数据点虽然没有具体数字,但反映了当前AI部署的经济瓶颈,是SubQ试图解决的核心问题之一。

    1. help large enterprises deploy AI responsibly across their core business operations

      【令人震惊】「负责任地在核心业务流程部署 AI」——这句话意味着 Anthropic 正在承接以前由麦肯锡、埃森哲做的企业变革咨询工作。纯模型 API 商业模式的顶峰可能已过:Claude 的护城河从「技术优势」升级为「有金融资本背书的企业实施能力」,中间层 AI 集成商和咨询公司的生存空间被直接压缩。

  2. May 2026
    1. Coding agents are great at building software. But to deploy to production they need three things from the cloud they want to host their app —an account, a way to pay, and an API token.

      This highlights a common pitfall for beginners: understanding the infrastructure requirements for deploying software, especially the need for accounts and payment methods.

    2. Coding agents are great at building software. But to deploy to production they need three things from the cloud they want to host their app —an account, a way to pay, and an API token.

      初学者常见陷阱:错误地认为部署到生产环境只需要代码,而忽略了账户、支付和API令牌等必要条件。

  3. Apr 2026
    1. over one million Trainium2 chips to train and serve Claude

      使用超过100万颗Trainium2芯片的数据,展示了Anthropic在AI硬件部署上的巨大规模。这一数字不仅反映了计算能力的投入,也显示了与AWS在芯片定制上的深度合作。对于AI模型训练而言,百万级芯片的部署规模是行业顶尖水平,表明Claude可能需要大量计算资源进行训练和推理。

    2. over one million Trainium2 chips to train and serve Claude

      使用超过100万个Trainium2芯片,这是一个惊人的硬件部署规模。这一数字不仅显示了Anthropic与Amazon的深度合作,也反映了训练和运行大型语言模型所需的庞大计算资源。相比其他AI公司,这种规模的芯片部署表明Anthropic正在全力投入AI基础设施。

    1. the surrogate is activated only when its agreement with the LLM exceeds a user-specified threshold α

      大多数人认为模型部署应该是全有或全无的,要么完全替代原模型要么完全不使用。但作者提出了一种'部分激活'的激进方法,只在代理模型与原模型达到特定一致性阈值时才使用代理,这种细粒度的控制方式打破了传统的二元部署思维。

    1. Infrastructure Provisioning cd deploy/terraform/aliyun terraform init terraform plan terraform apply Helm Deployment cd deploy/helm helm install aegis-core ./aegis-core \ --namespace aegis \ --create-namespace \ --set image.repository=<acr-registry>/aegis-core \ --set image.tag=lat

      使用Terraform和Helm进行云基础设施部署体现了现代DevOps实践在AI安全平台中的应用。这种基础设施即代码(IaC)方法确保了部署的可重复性和一致性,同时支持阿里云等特定云平台,显示了平台对生产环境的适应性。

    2. Quick Start # Clone the repository git clone https://github.com/fxp/aegis-core.git cd aegis-core # Start all services with Docker Compose docker-compose up -d # The API is available at http://localhost:8000 # Health check: http://localhost:8000/health

      简化的启动流程展示了容器化部署的优势,使用Docker Compose一键启动所有服务,大大降低了部署复杂度。这种设计反映了现代AI平台开发的一个重要趋势:简化环境配置,使研究人员能够快速开始工作,而不是陷入环境设置的困境。

    1. Because small, cheap, fast models are sufficient for much of the detection work, you don't need to judiciously deploy one expensive model and hope it looks in the right places. You can deploy cheap models broadly, scanning everything, and compensate for lower per-token intelligence with sheer coverage and lower cost-per-token.

      这一观点提出了AI安全的经济新模式,通过广泛部署小型廉价模型来弥补单一大模型的不足。这种'广撒网'策略可能比依赖少数昂贵模型更有效,尤其在大规模代码库扫描场景中,为AI安全的经济可行性提供了新思路。

    1. In practice, deployed model implementations are often flexible (e.g., mixing kernel variants, hybrid attention patterns, MoE blocks, and serving-optimized layouts), which can deviate from the assumptions required by a given conversion recipe.

      这个观点揭示了现有方法在实际部署中的一个重要局限性:它们通常依赖于特定的模型实现假设,而实际部署的模型往往更加灵活和复杂。这强调了Attention Editing框架的优势——它不依赖于精细的结构要求,可以适应各种实际部署场景,为模型转换提供了更大的灵活性。

    1. 单张 24GB 4090 直接部署 32B LLM

      令人惊讶的是:一张消费级显卡竟然能直接运行320亿参数的大模型,这打破了人们对大模型硬件需求的固有认知。过去需要多张高端显卡或专业服务器才能运行的模型,现在单张RTX 4090就能实现,这标志着大模型普及的门槛大幅降低。

    1. A deployment cascade combining both stages attains 90% accuracy at 71% coverage without any task-specific labels.

      令人惊讶的是:SELFDOUBT方法通过两级部署策略,在没有任务特定标签的情况下实现了90%的准确率和71%的覆盖率。这一成果表明,通过简单分析模型输出中的犹豫和验证行为,可以构建出高效的置信度过滤器,大幅提升模型在实际应用中的可靠性,无需额外标注数据。

    2. Unlike methods that require multiple sampled traces or model internals, SELFDOUBT operates on a single observed reasoning trajectory, making it suitable for latency- and cost-constrained deployment over any proprietary API.

      令人惊讶的是:SELFDOUBT方法仅需单个推理轨迹就能进行不确定性量化,而传统方法通常需要多次采样或访问模型内部参数。这一突破使得该方法可以在延迟和成本受限的部署环境中使用,特别适用于无法获取模型内部信息的专有API,大大降低了实际应用门槛。

    1. Anthropic says Managed Agents is designed to cut the time it takes to move from prototype to production from months to days, with early adopters like Notion, Rakuten, Asana, Vibecode, and Sentry already using it across coding, productivity, and internal workflow automation.

      将AI原型到生产的时间从几个月缩短到几天是一个惊人的加速,这将彻底改变企业采用AI的方式。这种快速部署能力可能加速AI在各行业的普及,但也带来了关于AI系统安全性和治理的紧迫问题,企业需要在快速采用和确保安全之间找到平衡。

    1. We do not plan to make Claude Mythos Preview generally available, but our eventual goal is to enable our users to safely deploy Mythos-class models at scale.

      大多数人认为强大的AI模型应该广泛普及以造福更多人。但作者明确表示不会公开发布这个最强大的模型,暗示了AI能力扩散可能带来的风险大于收益,这与技术民主化的主流观点相悖。

    1. 谷歌在沉寂了很长时间以后,终于发了一个不错的模型,而且还是开源的 Gamma 4 系列。专门用来在本地设备(比如手机、电脑)上跑

      大多数人认为谷歌作为 AI 领域的领导者会持续专注于云端大模型,但其突然转向端侧开源模型的做法令人意外。这种战略转变表明谷歌可能重新评估了 AI 部署的未来方向,从集中式向分布式转变,挑战了'更大模型更好'的行业共识,暗示了端侧 AI 可能成为下一个技术热点。

    1. Using vLLM high-throughput LLM serving on DGX Spark provides a high-performance platform for the largest Gemma 4 models

      大多数人认为运行最大的Gemma 4模型需要专门的硬件和复杂的部署流程。但作者声称vLLM可以在DGX Spark上高效运行这些大型模型,暗示推理优化技术可能已经达到了一个临界点,使得复杂模型部署变得更加简单和高效。

    2. The 31B and 26B A4B variants are high-performing reasoning models suitable for both local and data center environments.

      大多数人认为大型语言模型(31B参数)只能在数据中心环境中运行,但作者声称这些模型可以在本地环境中高效运行。这一观点与行业共识相悖,暗示边缘计算能力可能比我们想象的更强大,可能会改变AI部署的格局。

  4. Mar 2026
    1. Absolute size thresholds for degenerative aneurysms: [1][3][5]

      Ascending aorta/aortic root: ≥5.5 cm

      Descending thoracic aorta: ≥6.0 cm (or 5.5 cm if favorable anatomy for TEVAR)

      Thoracoabdominal aorta: ≥6.0 cm

      Lower thresholds apply for: [3]

      Marfan syndrome or genetic conditions: 4.0-5.0 cm depending on condition

      Bicuspid aortic valve: 5.0-5.5 cm

      Rapid growth: >0.5 cm/year

      Concomitant cardiac surgery: >4.5 cm if undergoing aortic valve surgery

      Immediate surgical evaluation: [5]

      Any symptomatic aneurysm regardless of size (chest/back pain, dysphagia, hoarseness, hemoptysis)

      Acute complications (dissection, rupture, malperfusion)

      Post-Repair Surveillance

      After TEVAR: CT at 1 month, 12 months, then annually if stable. [1]

      After open repair: CT or MRI within 1 year, then every 5 years if no residual aortopathy. Annual imaging if residual disease or abnormal findings.

  5. Mar 2025
  6. Nov 2024
    1. The potential for cuts in 2030 is 31 gigatons of CO2 equivalent – which isaround 52 per cent of global greenhouse gas emissions in 2023 – and 41gigatons in 2035.· Increased deployment of solar photovoltaic technologies and wind energy coulddeliver 27 per cent of this total emission reduction potential in 2030 and 38 percent in 2035.· Action on forests could deliver around 20 per cent of the potential in both years.• Other strong options include efficiency measures, electrification and fuelswitching in the buildings, transport and industry sectors.

      for - stats - 27% of the gap can be reduced by wind and solar deployment and 20% by action on forests, while efficiency, electrification, fuel switching in buildings, transport and industry sectors can also contribute - UN Emissions Gap Report 2024 - Key Messages

    1. A TRUSTworthy repository needs to focus on serving its target user community. Each user community likely has differing expectations from their community repositories, depending in part on the community’s maturity regarding data management and sharing. A TRUSTworthy repository is embedded in its target user community’s data practices, and so can respond to evolving community requirements

      TRSP Desirable Characteristics

    1. TRSP Desirable Characteristics Ethical data are data that do not stigmatise or portray Indigenous Peoples, cultures, or knowledges in terms of deficit. Ethical data are collected and used in ways that align with Indigenous ethical frameworks and with rights affirmed in UNDRIP. Assessing ethical benefits and harms should be done from the perspective of the Indigenous Peoples, nations, or communities to whom the data relate

    2. TRSP Desirable Characteristics Data enrich the planning, implementation, and evaluation processes that support the service and policy needs of Indigenous communities. Data also enable better engagement between citizens, institutions, and governments to improve decision-making. Ethical use of open data has the capacity to improve transparency and decision-making by providing Indigenous nations and communities with a better understanding of their peoples, territories, and resources. It similarly can provide greater insight into third-party policies and programs affecting Indigenous Peoples.

    3. TRSP Desirable Characteristics Indigenous Peoples’ rights and interests in Indigenous data must be recognised and their authority to control such data be empowered. Indigenous data governance enables Indigenous Peoples and governing bodies to determine how Indigenous Peoples, as well as Indigenous lands, territories, resources, knowledges and geographical indicators, are represented and identified within data.

  7. Dec 2023
  8. Oct 2023
  9. Aug 2023
  10. Nov 2022
  11. Oct 2022
    1. Odoo is a multi-tenant system: a single Odoo system may run and serve a number of database instances. It is also highly customizable, with customizations (starting from the modules being loaded) depending on the "current database"
  12. Aug 2022
  13. Mar 2022
  14. Feb 2022
  15. Aug 2021
  16. May 2021
  17. Mar 2021
  18. Jan 2021
  19. Jun 2020
  20. Apr 2020
  21. Sep 2019
  22. May 2019
    1. newrelic-admin record-deploy config_file description [revision changelog user]

      To enable recording of deploys on the python agent via New Relic, you can simply call the newrelic-admin record-deploy command and pass it the necessary revision information. This will place a deployment marker on any graph you view in newrelic as a vertical line-indicating that a new revision of the code was released at that point in time.

  23. Mar 2019
  24. May 2015