AGI Watchful Guardians

We're open for new collaborations.
News to be updated weekly.

  • Home
  • About
  • Alignment Newsletter in Chinese
  • Nick Bostrom’s latest work in Chinese
  • Research
  • What would a Provably Safe AGI Framework look like?

    What would a Provably Safe AGI Framework look like?

    Xiaohu Zhu

    —

    Apr 8, 2024

    ×

    AGI, AI Safety, Beneficial, CSAGI
  • ICML 2019 Workshops – #1 ERL

    Xiaohu Zhu

    —

    Jun 12, 2019

    ×

    未分类
  • ICML 2019 Tutorials

    Xiaohu Zhu

    —

    Jun 12, 2019

    ×

    未分类
  • AN #57 为什么我们应该关注人工智能安全的健壮性和编程中的类似问题

    Xiaohu Zhu

    —

    Jun 9, 2019

    ×

    未分类
  • The Landscape of Deep Reinforcement Learning

    Xiaohu Zhu

    —

    May 28, 2019

    ×

    未分类
  • AN #55 监管市场和国际标准作为确保有益人工智能的手段

    Xiaohu Zhu

    —

    May 23, 2019

    ×

    Alignment
  • AN #56 机器学习研究人员是否应该在做出假设之前停止运行其实验?

    Xiaohu Zhu

    —

    May 23, 2019

    ×

    Alignment
  • TCS list

    Xiaohu Zhu

    —

    May 22, 2019

    ×

    未分类
  • PapeRman #7

    Xiaohu Zhu

    —

    Apr 14, 2019

    ×

    未分类
  • 无监督学习:好奇的学生

    Xiaohu Zhu

    —

    Apr 11, 2019

    ×

    未分类
  • AGI reading list

    Xiaohu Zhu

    —

    Apr 7, 2019

    ×

    未分类
Previous Page
1 … 6 7 8 9 10
Next Page

©

AGI Watchful Guardians

  • Subscribe Subscribed
    • AGI Watchful Guardians
    • Already have a WordPress.com account? Log in now.
    • AGI Watchful Guardians
    • Subscribe Subscribed
    • Sign up
    • Log in
    • Report this content
    • View site in Reader
    • Manage subscriptions
    • Collapse this bar