2018-04-09 OpenAI.Charter

2018-04-09 OpenAI.Charter

Refer To:OpenAI Charter

Our Charter describes the principles we use to execute on OpenAI’s mission.
我们的章程描述了我们执行OpenAI使命的原则。

This document reflects the strategy we’ve refined over the past two years, including feedback from many people internal and external to OpenAI. The timeline to AGI remains uncertain, but our Charter will guide us in acting in the best interests of humanity throughout its development.
这份文件反映了我们在过去两年中完善的战略,包括来自OpenAI内部和外部许多人的反馈。AGI的发展时间表仍然不确定,但我们的章程将指导我们在其发展过程中始终以人类的最佳利益为出发点。

OpenAI’s mission is to ensure that artificial general intelligence (AGI)—by which we mean highly autonomous systems that outperform humans at most economically valuable work—benefits all of humanity. We will attempt to directly build safe and beneficial AGI, but will also consider our mission fulfilled if our work aids others to achieve this outcome. To that end, we commit to the following principles:
OpenAI的使命是确保人工通用智能(AGI)——即在大多数经济价值高的工作中超越人类的高度自主系统——能够惠及全人类。我们将努力直接构建安全且有益的AGI,但如果我们的工作能够帮助他人实现这一目标,我们也将视使命为完成。为此,我们承诺以下原则:

Broadly distributed benefits 广泛分布的益处

We commit to use any influence we obtain over AGI’s deployment to ensure it is used for the benefit of all, and to avoid enabling uses of AI or AGI that harm humanity or unduly concentrate power.
我们承诺利用我们在AGI部署中获得的任何影响力,确保其用于全人类的利益,并避免使AI或AGI的使用对人类造成伤害或导致权力过度集中。

Our primary fiduciary duty is to humanity. We anticipate needing to marshal substantial resources to fulfill our mission, but will always diligently act to minimize conflicts of interest among our employees and stakeholders that could compromise broad benefit.
我们的首要信托责任是对全人类负责。我们预计需要调动大量资源来完成使命,但始终会谨慎行事,以尽量减少员工和利益相关方之间可能影响广泛利益的利益冲突。

Long-term safety 长期安全

We are committed to doing the research required to make AGI safe, and to driving the broad adoption of such research across the AI community.
我们致力于开展使AGI安全所需的研究,并推动此类研究在整个AI社区中的广泛采用。

We are concerned about late-stage AGI development becoming a competitive race without time for adequate safety precautions. Therefore, if a value-aligned, safety-conscious project comes close to building AGI before we do, we commit to stop competing with and start assisting this project. We will work out specifics in case-by-case agreements, but a typical triggering condition might be “a better-than-even chance of success in the next two years.”
我们担心后期AGI开发可能演变为一场竞争性的竞赛,导致没有足够的时间进行安全预防。因此,如果有一个与我们价值观一致且注重安全的项目比我们更接近构建AGI,我们承诺停止竞争并开始协助该项目。具体细节将通过逐案协议确定,但一个典型的触发条件可能是“在未来两年内成功的可能性超过五成”。
Warning
参考:《The Exciting, Perilous Journey Toward AGI | Ilya Sutskever | TED》,Ilya Sutskever在演讲的结尾讲到这个点,在OpenAI的创始团队中,可能只有Ilya Sutskever是真正相信的。

Idea
This is more believable when OpenAI says it than when ExxonMobil does. There are real arguments for why an AI company might want to switch from moving fast and breaking things at time t to acting all responsible at time t + 1 . Let’s explore the arguments they make in the document, go over the reasons they’re obviously wrong, then look at the more complicated arguments they might be based off of.
当 OpenAI 说这句话时,比 ExxonMobil 说这句话更令人信服。确实有理由说明为什么一家 AI 公司可能想在时间 t 从快速行动和打破常规转变为在时间 t+1 表现得负责任。让我们探讨他们在文件中提出的论点,回顾他们显然错误的原因,然后看看他们可能基于的更复杂的论点。

OpenAI wants to sound exciting and innovative. If they say “we are exciting and innovative”, this is obvious self-promotion and nobody will buy it. If they say “we’re actually a dangerous and bad company, our products might achieve superintelligence and take over the world”, this makes them sound self-deprecating, while also establishing that they’re exciting and innovative.
OpenAI 想要听起来令人兴奋且富有创新性。如果他们说“我们令人兴奋且富有创新性”,这显然是自我宣传,没有人会相信。如果他们说“我们实际上是一家危险且糟糕的公司,我们的产品可能会实现超级智能并接管世界”,这让他们听起来自嘲,同时也确立了他们令人兴奋且富有创新性的形象。

The big thing all the alignment people were trying to avoid in the early 2010s was an AI race. DeepMind was the first big AI company, so we should just let them to their thing, go slowly, get everything right, and avoid hype. Then Elon Musk founded OpenAI in 2015, murdered that plan, mutilated the corpse, and danced on its grave. Even after Musk left, the remaining team did everything to challenge everyone else to a race short of shooting a gun and waving a checkered flag.
所有对齐人员在 2010 年代初期试图避免的最大问题是 AI 竞赛。DeepMind 是第一家大型 AI 公司,所以我们应该让他们慢慢来,把一切做好,避免炒作。然后埃隆·马斯克在 2015 年创立了 OpenAI,破坏了这个计划,肢解了尸体,并在其坟墓上跳舞。即使在马斯克离开后,剩下的团队也竭尽所能地挑战其他人进行竞赛,几乎到了开枪和挥舞方格旗的地步。

OpenAI still hasn’t given a good explanation of why they did this. Absent anything else, I’m forced to wonder if it’s just “they’re just the kind of people who would do that sort of thing” - in which case basically any level of cynicism would be warranted.
OpenAI 仍然没有给出一个好的解释说明他们为什么这样做。在没有其他信息的情况下,我不得不怀疑这是否只是因为“他们就是那种会做这种事情的人”——在这种情况下,基本上任何程度的愤世嫉俗都是合理的。

Technical leadership 技术领导力

To be effective at addressing AGI’s impact on society, OpenAI must be on the cutting edge of AI capabilities—policy and safety advocacy alone would be insufficient.
为了有效应对AGI对社会的影响,OpenAI必须处于AI能力的最前沿——仅靠政策和安全倡导是不够的。

We believe that AI will have broad societal impact before AGI, and we’ll strive to lead in those areas that are directly aligned with our mission and expertise.
我们相信,在AGI之前,AI将对社会产生广泛影响,我们将努力在那些与我们的使命和专业知识直接相关的领域中引领潮流。

Cooperative orientation 合作导向

We will actively cooperate with other research and policy institutions; we seek to create a global community working together to address AGI’s global challenges.
我们将积极与其他研究和政策机构合作;我们希望创建一个全球性社区,共同应对AGI的全球性挑战。

We are committed to providing public goods that help society navigate the path to AGI. Today this includes publishing most of our AI research, but we expect that safety and security concerns will reduce our traditional publishing in the future, while increasing the importance of sharing safety, policy, and standards research.
我们致力于提供有助于社会走向AGI的公共产品。目前,这包括发布我们的大部分AI研究,但我们预计出于安全和保障的考虑,未来将减少传统的研究发表,同时增加共享安全、政策和标准研究的重要性。

    Article Comments Update


      热门标签


        • Related Articles

        • 2024-05-29 Microsoft Corporation (MSFT) Jefferies Software Conference (Transcript)

          Microsoft Corporation (NASDAQ:MSFT) Jefferies Software Conference May 29, 2024 1:00 PM ET 微软公司(纳斯达克股票代码:MSFT)杰弗里斯软件大会 2024 年 5 月 29 日下午 1:00 ET Company Participants 公司参与者 Jessica Hawk - Corporate Vice President, DATA, AI, & Digital Apps 杰西卡·霍克 - ...
        • 2025-01-29 Microsoft Corporation (MSFT) Q2 2025 Earnings Call Transcript

          Microsoft Corporation (NASDAQ:MSFT) Q2 2025 Earnings Conference Call January 29, 2025 5:30 PM ET Company Participants Brett Iversen - Vice President of Investor Relations Satya Nadella - Chairman and Chief Executive Officer Amy Hood - Executive Vice ...
        • 2024-10-30 Microsoft Corporation (MSFT) Q1 2025 Earnings Call Transcript

          Microsoft Corporation (NASDAQ:MSFT) Q1 2025 Results Conference Call October 30, 2024 5:30 PM ET 微软公司(纳斯达克:MSFT)2025 财年第一季度业绩电话会议 2024 年 10 月 30 日 美国东部时间下午 5:30 Company Participants 公司参与者 Brett Iversen - Vice President of Investor Relations 布雷特·艾弗森 - ...
        • 2023-07-25 Microsoft Corporation (MSFT) Q4 2023 Earnings Call Transcript

          Microsoft Corporation (NASDAQ:MSFT) Q4 2023 Earnings Conference Call July 25, 2023 5:30 PM ET 微软公司(纳斯达克股票代码:MSFT)2023 年第四季度收益电话会议 2023 年 7 月 25 日 下午 5:30(东部时间) Company Participants 公司参与者 Brett Iversen - VP, IR 布雷特·伊弗森 - 副总裁,投资者关系 Satya Nadella - ...
        • 2024-04-25 Microsoft Corporation (MSFT) Q3 2024 Earnings Call Transcript

          Microsoft Corporation (NASDAQ:MSFT) Q3 2024 Earnings Conference Call April 25, 2024 5:30 PM ET 微软公司(纳斯达克股票代码:MSFT)2024 年第三季度收益电话会议 2024 年 4 月 25 日 下午 5:30 ET Company Participants 公司参与者 Brett Iversen – Vice President-Investor Relations 布雷特·艾弗森 - ...