Community Archive

🧵 View Thread

🧵 Thread (5 tweets)

Placeholder
David Chapman@Meaningnessover 3 years ago

AI safety research” has probably increased AI risk, and doing more of it will probably make things still worse. From @TheZvi’s response (first paragraph) to @ESYudkowsky’s recent doom post (quoted paragraph 38), both of which I found surprisingly interesting. https://t.co/ZP8e5m5nSm

Tweet image 1
89 12
6/13/2022
Placeholder
Guy IS IN SF 😎@nosilvervover 3 years ago
Replying to @Meaningness

@Meaningness @TheZvi @ESYudkowsky Wordcels: 💀💀💀

2 0
6/13/2022
Placeholder
ivan@IvanVendrovover 3 years ago
Replying to @Meaningness

@Meaningness @TheZvi @ESYudkowsky I seriously doubt that risk has increased. The amount of money & talent being poured into advancing AI capabilities is still 3-4 orders of magnitude more than into AI safety even by the broadest definition; any capabilities blowback from safety work is a rounding error.

8 0
6/13/2022
Placeholder
ivan@IvanVendrovover 3 years ago
Replying to @IvanVendrov

@Meaningness @TheZvi @ESYudkowsky The point about current safety work "letting people fool themselves about the lethality of the problem" is more plausible. But note that 99% of AI researchers already behave as if there is no problem! Convincing them that there is a problem at all should be the priority.

7 0
6/13/2022
Placeholder
ivan@IvanVendrovover 3 years ago
Replying to @IvanVendrov

@Meaningness @TheZvi @ESYudkowsky An implicit assumption here is that people working on AI safety are somehow smarter or more able to advance capabilities; this is bunk. Vast majority of the best ML researchers & engineers just don't think about safety - they're motivated by other things. Relevant Geoff Hinton: https://t.co/wrGmlbL3Ut

Tweet image 1
6 0
6/13/2022