Skip to content
AGI Watchful Guardians
We're open for new collaborations.
News to be updated weekly.
Home
About
Alignment Newsletter in Chinese
Nick Bostrom’s latest work in Chinese
Research
What would a Provably Safe AGI Framework look like?
Xiaohu Zhu
—
Apr 8, 2024
×
AGI
,
AI Safety
,
Beneficial
,
CSAGI
具有避免奖励函数篡改动机的智能体设计
Xiaohu Zhu
—
Aug 15, 2019
×
AGI
,
AI Safety
,
CID
,
DeepMind
AN #61 人工智能策略与治理,来自该领域两位专家的分享
Xiaohu Zhu
—
Aug 5, 2019
×
AGI
,
AI Safety
,
Alignment
AN #60 一个新的AI挑战:在创造性模式中帮助人类玩家的 Minecraft 智能体
Xiaohu Zhu
—
Jul 23, 2019
×
AGI
,
AI Safety
,
Alignment
PapeRman #8
Xiaohu Zhu
—
Jul 20, 2019
×
未分类
最坏情况下的保证(重制版)
Xiaohu Zhu
—
Jul 11, 2019
×
AGI
,
AI Safety
,
Alignment
,
OpenAI
人工智能风险争论的转变
Xiaohu Zhu
—
Jul 11, 2019
×
未分类
AN #59 对人工智能风险的争论是如何随着时间而改变的
Xiaohu Zhu
—
Jul 11, 2019
×
未分类
用因果影响图建模通用人工智能安全框架
Xiaohu Zhu
—
Jun 26, 2019
×
AGI
,
AI Safety
,
CID
,
DeepMind
AN #58 Mesa 优化:这是什么,为什么我们应该关心它
Xiaohu Zhu
—
Jun 24, 2019
×
未分类
PapeRman #8
Xiaohu Zhu
—
Jun 16, 2019
×
未分类
Previous Page
1
…
5
6
7
8
9
10
Next Page
Subscribe
Subscribed
AGI Watchful Guardians
Sign me up
Already have a WordPress.com account?
Log in now.
AGI Watchful Guardians
Subscribe
Subscribed
Sign up
Log in
Report this content
View site in Reader
Manage subscriptions
Collapse this bar