impossiblefork t1_ja6rt6s wrote
Reply to comment by okokoko in [D] To the ML researchers and practitioners here, do you worry about AI safety/alignment of the type Eliezer Yudkowsky describes? by SchmidhuberDidIt
I doubt it's possible, but I imagine something like [ed:the] DAN thing with ChatGPT.
Most likely you'd talk to the AI such that the rationality it has obtained from its training data make it reason things out that it's owner would rather it stay silent about it.
Viewing a single comment thread. View all comments