--
You received this message because you are subscribed to the Google Groups "Everything List" group.
To unsubscribe from this group and stop receiving emails from it, send an email to everything-li...@googlegroups.com.
To view this discussion on the web visit https://groups.google.com/d/msgid/everything-list/CAJPayv2VMSWOJfL74rL%3DNoe2o5y%3DE8F%3DY5GCQKN%3D-inOY1x1kQ%40mail.gmail.com.
>It's hard to know how to think about this kind of risk. It's safe to say EY has done more thinking on this issue than just about anyone, and he's one of the smartest people on the planet, probably. I've been following him for over a decade, from even before his writings on lesswrong.
> I'm not saying EY's conclusions are stupid or bullshit.
> That's not to say he's wrong. As you say John, it's totally unpredictable, but I think there's room for less dire narratives about how it could all go.
> But one thing I do 100% agree with is that we're not taking this seriously enough
I think Eliezer was right when he said nobody can predict what moves a chess program like Stockfish will make but you can predict that it will beat you in a game of chess, that's because Stockfish is super good at playing chess but it can't do anything else, it can't even think of anything else. But an AI program like GPT-4 is different, it can think of a great many things besides chess so you can't really predict what it will do, sure it has the ability to beat you at a game of Chess but for its own inscrutable reasons it may deliberately let you win. So yes in a few years an AI will have the ability to exterminate the human race, but will it actually do so? I don't know and I can't even give a probability of it occurring, all I can say is the probability is greater than zero and less than 100%.
mf
So there’s two connected big concerning unknowns. The first is that we don’t really know what they’re doing in any deep sense. If we open up ChatGPT or a system like it and look inside, you just see millions of numbers flipping around a few hundred times a second, and we just have no idea what any of it means. With only the tiniest of exceptions, we can’t look inside these things and say, “Oh, here’s what concepts it’s using, here’s what kind of rules of reasoning it’s using. Here’s what it does and doesn’t know in any deep way.” We just don’t understand what’s going on here. We built it, we trained it, but we don’t know what it’s doing.'