Search results

Jump to: navigation, search

Page title matches

Page text matches

  • ...telligence/capability and values can vary orthogonally; a superintelligent AI need not realize that "making paperclips is stupid" and decide to maximize * [[instrumental convergence]]: even if an AI isn't deliberately trying to hurt us (as a terminal value), it will still p
    2 KB (344 words) - 23:29, 27 July 2020
  • In discussions of AI risk, evolution (especially hominid evolution, as it is the only example we [[Category:AI safety]]
    799 bytes (112 words) - 23:46, 19 May 2021
  • '''Will there be significant changes to the world prior to some critical AI capability threshold being reached?''' This is currently one of the questio [[Eliezer]]: "And the relative rate of growth between AI capabilities and human capabilities, and the degree to which single investm
    2 KB (251 words) - 02:40, 28 March 2021
  • ...s. For instance, [[distributional shift]] is an AI safety problem where an AI trained in one environment will behave poorly when deployed in an unfamilia ! AI safety problem !! Human safety problem
    3 KB (379 words) - 20:36, 11 November 2021
  • [[Category:AI safety]]
    159 bytes (19 words) - 04:53, 30 March 2021
  • * [[rapid capability gain]] seems to refer to how well a ''single'' AI system improves over time, e.g. [[AlphaGo]] going from "no knowledge of any * AGI progress refers to progress of AI systems in general, i.e. if you plot the state of the art over time
    774 bytes (121 words) - 19:17, 27 February 2021
  • This is a '''list of technical AI alignment agendas'''. * [[Comprehensive AI services]]
    586 bytes (56 words) - 21:29, 2 April 2021
  • The most important posts for AI strategy are: ...129 My current take on the Paul-MIRI disagreement on alignability of messy AI]
    300 bytes (42 words) - 19:10, 27 February 2021
  • what are the possibilities for prosaic AI? i.e. if prosaic AI happened, then what are some possible reasons for why this happened? some i ...xisting ML systems (or straightforward tweaks to them) somehow produces an AI that fully "understands" things and can do everything humans can; (2) there
    2 KB (247 words) - 19:10, 27 February 2021
  • ...sagreements in AI safety''' which collects the list of things people in AI safety seem to most frequently and deeply disagree about. ...d/1wI21XP-lRa6mi5h0dq_USooz0LpysdhS/view Clarifying some key hypotheses in AI alignment].</ref> (there are more posts like this, i think? find them)
    21 KB (3,254 words) - 11:00, 26 February 2022
  • [[Category:AI safety]]
    508 bytes (62 words) - 19:12, 27 February 2021
  • ...echnical AI alignment. The term has since been used in contexts outside of AI alignment as well. [[Category:AI safety]]
    346 bytes (55 words) - 19:18, 27 February 2021
  • People in [[AI safety]] tend to [[List of disagreements in AI safety|disagree about many things]]. However, there is also wide agreement about s * advanced AI will have a huge impact on the world
    2 KB (272 words) - 01:33, 13 May 2020
  • '''Wei Dai''' is an AI safety researcher. In the technical AI safety community he is most well-known as the inventor of [[updateless decision th * [[human safety problem]]s
    387 bytes (51 words) - 20:47, 11 August 2021
  • [[Category:AI safety]]
    1 KB (195 words) - 19:08, 27 February 2021
  • ...mple core''' is a term that has been used to describe various things in AI safety. I think the "simple" part is intended to capture that the thing is human-u * Simple core to [[corrigibility]]<ref>https://ai-alignment.com/corrigibility-3039e668638</ref><ref>https://www.greaterwrong.
    824 bytes (120 words) - 21:32, 2 April 2021
  • ...ssions in AI alignment|big discussions]]" that has been taking place in AI safety in 2018&ndash;2020 is about [[coherence argument]]s and [[goal-directed]] a https://www.greaterwrong.com/posts/tHxXdAn8Yuiy9y2pZ/ai-safety-without-goal-directed-behavior
    1 KB (183 words) - 19:09, 27 February 2021
  • [[Category:AI safety]]
    782 bytes (108 words) - 20:56, 26 March 2021
  • * does it pursue omohundro's "basic AI drives"? (i.e. it is the subject of instrumental convergence) * mindblind AI
    4 KB (534 words) - 19:09, 27 February 2021
  • * "The first AI systems capable of pivotal acts will use good consequentialist reasoning." * "The default AI development path will not produce good consequentialist reasoning at the to
    4 KB (702 words) - 03:57, 26 April 2020
  • [[Category:AI safety]]
    815 bytes (133 words) - 19:18, 27 February 2021
  • [[Category:AI safety]]
    212 bytes (21 words) - 19:10, 27 February 2021
  • [[Category:AI safety]]
    14 KB (2,432 words) - 09:13, 8 January 2023
  • ...ctually better? if AGI is developed in 200 years, what does this say about ai xrisk? this could happen for several reasons: ...l be much harder than building agi", then this might push you to think "ai safety is basically impossible for humans without intelligence enhancement to solv
    1 KB (248 words) - 21:33, 4 April 2021
  • '''AI prepping''' refers to selfish actions one can take in order to survive when It's not clear whether any really good actions for AI prepping exist. Some reasons for optimism are:
    6 KB (968 words) - 04:20, 26 November 2022
  • ...ften used by [[Robin Hanson]] to describe things like innovation, secrets, AI progress, citations. ...ot lumpy. Aren't power laws lumpy? actually maybe he's only saying that if AI progress is lumpy, then its citation patterns should be even lumpier than u
    4 KB (648 words) - 06:53, 3 June 2020
  • The '''Laplace's rule of succession argument for AI timelines''' uses [[wikipedia:Rule of succession|Laplace's rule of successi ....issarice.com/posts/Ayu5im98u8FeMWoBZ/my-personal-cruxes-for-working-on-ai-safety#AI_timelines
    1 KB (218 words) - 02:04, 5 April 2021
  • ...analysis of experts' AI timelines to come up with some overall estimate of AI timelines. It punts the question of "but where did the experts get their op ..."AI researchers attending X conference", "AI researchers in general", "AI safety researchers").
    641 bytes (101 words) - 05:10, 9 April 2021
  • I think [[Eliezer]]'s point is that when there's more hardware behind an AI project, the Kasparov window is narrower. ...series of blog posts from [[AI Impacts]] https://aiimpacts.org/?s=time+for+ai+to+cross
    320 bytes (55 words) - 22:02, 4 January 2021
  • ....com/posts/6skeZgctugzBBEBw3/ai-alignment-podcast-an-overview-of-technical-ai-alignment] [[Category:AI safety]]
    746 bytes (115 words) - 20:41, 12 April 2021
  • ...(all of our other problems are so pressing that we're willing to gamble on AI working out by default). I don't think this argument makes much sense. ...at’s one of the reasons why I’m focusing on AI safety, rather than bio-safety.</p>
    1 KB (261 words) - 20:02, 23 June 2021
  • ...f>[https://forum.effectivealtruism.org/users/richard_ngo richard_ngo]. "AI safety research engineer at DeepMind (all opinions my own, not theirs). I'm from N [[Category:AI safety]]
    584 bytes (88 words) - 19:11, 27 February 2021
  • Incremental reading provides feeling of emotional safety (which is something that [[Anki]] does in general, but where I think increm feeling like i should maybe ankify some of my ai safety reading from LW, but it's been hard to think of what to even put in. some t
    4 KB (687 words) - 01:10, 17 July 2021
  • ...eal. This is possible with math, but i'm not sure how to do this with [[AI safety]] (it's not like there's problems i can solve).
    8 KB (1,497 words) - 00:01, 2 August 2021
  • ...architecture''' is used to mean ... something like the basic design of the AI system (like what kind of machine learning is being used in what way, what ..."mental architecture", "cognitive architecture", the "architecture of the AI"
    7 KB (1,128 words) - 23:18, 23 June 2020
  • [[Category:AI safety]]
    1 KB (170 words) - 06:54, 3 June 2020
  • * [[Counterfactual of dropping a seed AI into a world without other capable AI]] [[Category:AI safety]]
    1 KB (180 words) - 09:49, 6 May 2020
  • ...alism about rationality''' is a topic of debate among people working on AI safety. The "something like" refers to the fact that the very topic of ''what the ...uct to achieve an agreed-upon aim, namely helping to detect/fix/ensure the safety of AGI systems.)
    7 KB (1,110 words) - 20:24, 26 June 2020
  • ...nd its ability to solve alignment problems (i.e. design better ''aligned'' AI systems). ...r is imagining some big leap/going from just humans to suddenly superhuman AI, whereas paul is imagining a more smooth transition that powers his optimis
    3 KB (477 words) - 00:01, 30 May 2020
  • ...t [[missing gear]]'/'one wrong number' dynamic, AND each insight makes the AI a little better), then you can't specialize in "intelligence". [[Category:AI safety]]
    577 bytes (96 words) - 23:01, 6 July 2020
  • ...the AI will not hit the human timescale keyhole." From our perspective, an AI will either be so slow as to be bottlenecked, or so fast as to be FOOM. Whe ...challenge time in advance, rather than challenging at a point where their AI seemed just barely good enough, it was improbable that they'd make *exactly
    4 KB (728 words) - 16:56, 24 June 2020
  • [[Category:AI safety]]
    1 KB (130 words) - 19:55, 31 May 2021
  • [[Category:AI safety]]
    61 bytes (8 words) - 01:10, 18 May 2020
  • Self-studying all of the technical prerequisites for [[technical AI safety research]] is hard. The most that people new to the field get is a list of ...pessimism: If hiring capacity is limited at AI safety orgs and mainstream AI orgs only want to hire ML PhDs then new people entering the field will basi
    3 KB (447 words) - 18:34, 18 July 2021
  • * [[AI safety technical pipeline does not teach how to start having novel thoughts]] * [[AI safety is not a community]]
    931 bytes (138 words) - 01:30, 20 May 2020
  • Currently, the [[AI safety community]] does not have an explicit mechanism for teaching new people how [[Category:AI safety meta]]
    1 KB (225 words) - 02:28, 28 March 2021
  • ...lly, I think I've been in communities before, and being a part of the [[AI safety community]] does not feel like that. * AI safety has left the "hobbyist stage". People can actually now get paid to think ab
    894 bytes (159 words) - 02:28, 28 March 2021
  • I think the [[AI safety community]] and [[effective altruism]] in general has some mixed messaging [[Category:AI safety meta]]
    866 bytes (142 words) - 20:38, 18 May 2020
  • AI safety has a weird dynamic going on where: * There are discussions of things like AI timelines, assumptions of various technical agendas, etc., which reveals th
    3 KB (435 words) - 02:38, 28 March 2021
  • [[Category:AI safety meta]]
    1 KB (198 words) - 02:28, 28 March 2021

View (previous 50 | next 50) (20 | 50 | 100 | 250 | 500)