Search results

Jump to: navigation, search

Page title matches

Page text matches

  • ...ontent of the insights), what does that mean, in terms of what to do about AI safety? ....issarice.com/posts/mJ5oNYnkYrd4sD5uE/clarifying-some-key-hypotheses-in-ai-alignment] captures some of these, but i don't think this is a minimal set (there are
    3 KB (528 words) - 20:58, 26 March 2021
  • ...nding is that MIRI people/other smart people have prioritized technical AI alignment over WBEs because while WBEs would be safer if they came first, pushing on * is there anything else relevant to AI strategy that i should know about?
    6 KB (850 words) - 19:16, 27 February 2021
  • # a highly intelligent AI would see things humans cannot see, can arrive at unanticipated solutions, ...-approach/ -- this article, in particular the section "Creating a powerful AI system without understanding why it works is dangerous", gives a basic case
    5 KB (765 words) - 02:32, 28 March 2021
  • ...illation and amplification''', and abbreviated '''IDA''') is the technical alignment agenda that [[Paul Christiano]] works on. * [[prosaic AI]]
    1 KB (125 words) - 03:58, 26 April 2020
  • This is a '''list of technical AI alignment agendas'''. * [[Comprehensive AI services]]
    586 bytes (56 words) - 21:29, 2 April 2021
  • ...disagreements in AI safety''' which collects the list of things people in AI safety seem to most frequently and deeply disagree about. ...wI21XP-lRa6mi5h0dq_USooz0LpysdhS/view Clarifying some key hypotheses in AI alignment].</ref> (there are more posts like this, i think? find them)
    21 KB (3,254 words) - 11:00, 26 February 2022
  • ...nical AI alignment. The term has since been used in contexts outside of AI alignment as well. [[Category:AI safety]]
    346 bytes (55 words) - 19:18, 27 February 2021
  • ...nced machine learning systems" agenda [https://intelligence.org/2016/07/27/alignment-machine-learning/], which then got cancelled when [[Jessica]] and [[Patrick [[Category:AI safety]]
    1 KB (195 words) - 19:08, 27 February 2021
  • ...scussions in AI alignment|big discussions]]" that has been taking place in AI safety in 2018&ndash;2020 is about [[coherence argument]]s and [[goal-direc https://www.greaterwrong.com/posts/tHxXdAn8Yuiy9y2pZ/ai-safety-without-goal-directed-behavior
    1 KB (183 words) - 19:09, 27 February 2021
  • '''AI prepping''' refers to selfish actions one can take in order to survive when It's not clear whether any really good actions for AI prepping exist. Some reasons for optimism are:
    6 KB (968 words) - 04:20, 26 November 2022
  • ...st could have messed up values (values that looked good to program into an AI, but aren't actually the real thing). If so, ''the universal law will take ...liezer]]'s idea of [[reflectively consistent degrees of freedom]]: if your AI uses CDT, it will not self-modify to use UDT; instead it will evolve to use
    7 KB (1,148 words) - 06:18, 21 May 2020
  • ...architecture''' is used to mean ... something like the basic design of the AI system (like what kind of machine learning is being used in what way, what ..."mental architecture", "cognitive architecture", the "architecture of the AI"
    7 KB (1,128 words) - 23:18, 23 June 2020
  • ...nd its ability to solve alignment problems (i.e. design better ''aligned'' AI systems). ...r is imagining some big leap/going from just humans to suddenly superhuman AI, whereas paul is imagining a more smooth transition that powers his optimis
    3 KB (477 words) - 00:01, 30 May 2020
  • ...d how it feels bad when something isn't "good enough" to be crossposted to Alignment Forum, how it feels bad when a nitpick reply gets more upvotes than your OP [[Category:AI safety meta]]
    1 KB (198 words) - 02:28, 28 March 2021
  • '''Corrigibility''' is a term used in AI safety with multiple/unclear meanings. I think the term was originally used by [[MIRI]] to mean something like an AI that allowed human programmers to shut it off.
    773 bytes (110 words) - 23:29, 8 November 2021
  • ...n or gradual, how quickly economic activity will accelerate after advanced AI systems appear, and so on). * [[List of disagreements in AI safety#Takeoff dynamics]]
    734 bytes (101 words) - 01:01, 5 March 2021
  • This is a '''list of arguments against working on AI safety'''. Personally I think the only one that's not totally weak is oppor ...re out how to affect the long-term future. See also [[Pascal's mugging and AI safety]].
    8 KB (1,245 words) - 00:29, 24 July 2022
  • ...of knowing about AI takeoff''' is about the "so what?" of knowing which [[AI takeoff]] scenario will happen. How will our actions change if we expect a ...ent) or short-term consumption. In contrast, with more continuous takeoff, AI prepping becomes relatively more important.
    1 KB (181 words) - 02:13, 5 March 2021
  • * [[Difficulty of AI alignment]] [[Category:AI safety]]
    243 bytes (33 words) - 10:59, 26 February 2022
  • ...tems slightly smarter than ourselves, and from there, each "generation" of AI systems will align slightly smarter systems, and so on. ...in [https://www.lesswrong.com/posts/7im8at9PmhbT4JHsW/ngo-and-yudkowsky-on-alignment-difficulty#1_1__Deep_vs__shallow_problem_solving_patterns]
    846 bytes (111 words) - 11:21, 26 February 2022

View (previous 20 | next 20) (20 | 50 | 100 | 250 | 500)