Difference between revisions of "AI safety field consensus"

From Issawiki
Jump to: navigation, search
Line 7: Line 7:
 
* Goodhart problems i.e. awareness that [[Goodhart's law]] is a thing, and general attention/wariness of it
 
* Goodhart problems i.e. awareness that [[Goodhart's law]] is a thing, and general attention/wariness of it
 
* AGI possible in principle (as in, it is virtually certain that humans can create AGI)
 
* AGI possible in principle (as in, it is virtually certain that humans can create AGI)
 +
* advanced AI will have a huge impact on the world
  
 
see also "Background AI safety intuitions" section in [https://agentfoundations.org/item?id=1129]
 
see also "Background AI safety intuitions" section in [https://agentfoundations.org/item?id=1129]
  
 
one operationalization might be something like: what are the things relevant to AI safety that all of [[Eliezer Yudkowsky]], [[Paul Christiano]], [[Robin Hanson]], [[Rohin Shah]], [[Dario Amodei]], and [[Wei Dai]] agree on?
 
one operationalization might be something like: what are the things relevant to AI safety that all of [[Eliezer Yudkowsky]], [[Paul Christiano]], [[Robin Hanson]], [[Rohin Shah]], [[Dario Amodei]], and [[Wei Dai]] agree on?

Revision as of 04:35, 25 February 2020

People in AI safety tend to disagree about many things. However, there is also wide agreement about some other things (which people outside the field often disagree about).

see also "Background AI safety intuitions" section in [1]

one operationalization might be something like: what are the things relevant to AI safety that all of Eliezer Yudkowsky, Paul Christiano, Robin Hanson, Rohin Shah, Dario Amodei, and Wei Dai agree on?