Customize Consent Preferences

We use cookies to help you navigate efficiently and perform certain functions. You will find detailed information about all cookies under each consent category below.

The cookies that are categorized as "Necessary" are stored on your browser as they are essential for enabling the basic functionalities of the site. ... 

Always Active

Necessary cookies are required to enable the basic features of this site, such as providing secure log-in or adjusting your consent preferences. These cookies do not store any personally identifiable data.

No cookies to display.

Functional cookies help perform certain functionalities like sharing the content of the website on social media platforms, collecting feedback, and other third-party features.

No cookies to display.

Analytical cookies are used to understand how visitors interact with the website. These cookies help provide information on metrics such as the number of visitors, bounce rate, traffic source, etc.

No cookies to display.

Performance cookies are used to understand and analyze the key performance indexes of the website which helps in delivering a better user experience for the visitors.

No cookies to display.

Advertisement cookies are used to provide visitors with customized advertisements based on the pages you visited previously and to analyze the effectiveness of the ad campaigns.

No cookies to display.

0

【OpenAI解散超级对齐AI风险团队】

据华尔街见闻报道,知名人工智能研究机构OpenAI近日宣布,由前首席科学家Ilya领导的超级对齐AI风险团队已经解散。该团队旨在研究超级智能可能带来的安全风险,并提出相应的对策。团队的核心成员包括Jan Leike等。OpenAI表示,该团队的研究工作将被整合到其他研究小组中,以确保研究的连续性和完整性。

对此,OpenAI的联合创始人埃隆·马斯克发表评论称,这一决定表明安全问题并非OpenAI的首要任务。Jan Leike在离职后透露,他与OpenAI高层的分歧长期存在,特别是在公司的核心优先事项上。团队在推进研究项目和争取计算资源方面遇到了重大困难,这对其研究进度和质量产生了负面影响。

OpenAI的这一决定引发了业界对于人工智能安全问题的广泛关注。有专家指出,人工智能的发展速度极快,如何在保证技术进步的同时确保其安全,是当前和未来必须面对的重要课题。OpenAI的这一调整,或许意味着其在人工智能安全领域的策略有所调整,未来可能会更加注重其他方面的研究和发展。

英语如下:

News Title: “OpenAI Disbands AI Safety Team; Musk Questions Safety Not a Priority”

Keywords: Disbandment, AI Safety, Dissent

News Content:

[OpenAI Dissolves Superintelligence Alignment Team]

According to Wall Street Journal reports, the renowned artificial intelligence research institution OpenAI recently announced the dissolution of its Superintelligence Alignment AI Risk Team, led by former Chief Scientist Ilya. The team was tasked with studying the potential safety risks associated with superintelligent AI and proposing countermeasures. Core members of the team included Jan Leike, among others. OpenAI stated that the research work of the team would be integrated into other research groups to ensure the continuity and integrity of the research.

In response, OpenAI co-founder Elon Musk commented that this decision indicated that safety issues were not OpenAI’s primary concern. Jan Leike revealed after leaving the team that there had been longstanding disagreements with OpenAI leadership, particularly concerning the company’s core priorities. The team faced significant challenges in advancing its research projects and securing computational resources, which adversely affected its research progress and quality.

OpenAI’s decision has sparked widespread concern in the industry about the safety of artificial intelligence. Experts have pointed out that the development of artificial intelligence is moving at an extremely rapid pace, and it is crucial to ensure both technological progress and safety. OpenAI’s adjustment may indicate a shift in its strategy regarding artificial intelligence safety, suggesting that it may place greater emphasis on other areas of research and development in the future.

【来源】https://wallstreetcn.com/articles/3715193

Views: 2

0

发表回复

您的邮箱地址不会被公开。 必填项已用 * 标注