Difference between revisions of "AlphaGo"
(→Eliezer's commentary) |
|||
(3 intermediate revisions by the same user not shown) | |||
Line 3: | Line 3: | ||
* [[Rapid capability gain]] | * [[Rapid capability gain]] | ||
* single group pulling ahead | * single group pulling ahead | ||
− | * a single architecture / basic AI technique working for many different games | + | * a single architecture / basic AI technique working for many different games ([[single-architecture generality]]) |
* (for AlphaGo Zero) comparison to [[Paul Christiano]]'s [[iterated amplification]] | * (for AlphaGo Zero) comparison to [[Paul Christiano]]'s [[iterated amplification]] | ||
+ | * [[AlphaGo as evidence of discontinuous takeoff]] | ||
==Eliezer's commentary== | ==Eliezer's commentary== | ||
Line 34: | Line 35: | ||
| 2016-04-16 || [https://www.facebook.com/yudkowsky/posts/10154120081504228 "As soon as anyone does it, it stops being Artificial Intelligence!" No, as soon as anyone in AI achieves surprisingly good performance in some domain that people previously imagined being done as a specialized application of human general intelligence, the inference is, correctly, "Oh, it seems there was surprisingly a specialized way to do that which didn't invoke general intelligence" rather than "Oh, it looks like surprisingly more progress was made toward generally intelligent algorithms than we thought."] || | | 2016-04-16 || [https://www.facebook.com/yudkowsky/posts/10154120081504228 "As soon as anyone does it, it stops being Artificial Intelligence!" No, as soon as anyone in AI achieves surprisingly good performance in some domain that people previously imagined being done as a specialized application of human general intelligence, the inference is, correctly, "Oh, it seems there was surprisingly a specialized way to do that which didn't invoke general intelligence" rather than "Oh, it looks like surprisingly more progress was made toward generally intelligent algorithms than we thought."] || | ||
|- | |- | ||
− | | 2017-10-19 || [https://www.facebook.com/yudkowsky/posts/10155848910529228 AlphaGo Zero uses 4 TPUs, is built entirely out of neural nets with no handcrafted features, doesn't pretrain against expert games or anything else human, reaches a superhuman level after 3 days of self-play, and is the strongest version of AlphaGo yet] ([https://www.greaterwrong.com/posts/shnSyzv4Jq3bhMNw5/alphago-zero-and-the-foom-debate crossposted to LessWrong]) || [https://www.greaterwrong.com/posts/D3NspiH2nhKA6B2PE/what-evidence-is-alphago-zero-re-agi-complexity Robin Hanson's reply] | + | | 2017-10-19 || [https://www.facebook.com/yudkowsky/posts/10155848910529228 AlphaGo Zero uses 4 TPUs, is built entirely out of neural nets with no handcrafted features, doesn't pretrain against expert games or anything else human, reaches a superhuman level after 3 days of self-play, and is the strongest version of AlphaGo yet] ([https://www.greaterwrong.com/posts/shnSyzv4Jq3bhMNw5/alphago-zero-and-the-foom-debate crossposted to LessWrong] and to [https://intelligence.org/2017/10/20/alphago/ MIRI blog]) || [https://www.greaterwrong.com/posts/D3NspiH2nhKA6B2PE/what-evidence-is-alphago-zero-re-agi-complexity Robin Hanson's reply] |
|- | |- | ||
| 2017-12-09 || [https://www.facebook.com/yudkowsky/posts/10155992246384228 Max Tegmark put it well, on Twitter: The big deal about Alpha Zero isn't how it crushed human chess players, it's how Alpha Zero crushed human chess *programmers*.] || | | 2017-12-09 || [https://www.facebook.com/yudkowsky/posts/10155992246384228 Max Tegmark put it well, on Twitter: The big deal about Alpha Zero isn't how it crushed human chess players, it's how Alpha Zero crushed human chess *programmers*.] || | ||
|} | |} | ||
+ | |||
+ | ==See also== | ||
+ | |||
+ | * [[AlphaGo as evidence of discontinuous takeoff]] | ||
+ | |||
+ | ==What links here== | ||
+ | |||
+ | {{Special:WhatLinksHere/{{FULLPAGENAME}}}} | ||
+ | |||
+ | [[Category:AI safety]] |
Latest revision as of 20:19, 11 August 2021
AlphaGo and its successor AlphaGo Zero are used to make various points in AI safety.
- Rapid capability gain
- single group pulling ahead
- a single architecture / basic AI technique working for many different games (single-architecture generality)
- (for AlphaGo Zero) comparison to Paul Christiano's iterated amplification
- AlphaGo as evidence of discontinuous takeoff