misk@sopuli.xyz to Technology@lemmy.worldEnglish · 11 months agoJailbroken AI Chatbots Can Jailbreak Other Chatbotswww.scientificamerican.comexternal-linkmessage-square80fedilinkarrow-up1484arrow-down117cross-posted to: hackernews@derp.foo
arrow-up1467arrow-down1external-linkJailbroken AI Chatbots Can Jailbreak Other Chatbotswww.scientificamerican.commisk@sopuli.xyz to Technology@lemmy.worldEnglish · 11 months agomessage-square80fedilinkcross-posted to: hackernews@derp.foo
minus-squareKairuByte@lemmy.dbzer0.comlinkfedilinkEnglisharrow-up6·11 months agoWhat possible legitimate reason could someone need to know how to make chlorine/mustard gas? Apart from the fact that they are made from common household products, are easy to make by mistake, and can kill you. Wait that’s true of napalm as well… fuck.
What possible legitimate reason could someone need to know how to make chlorine/mustard gas?
Apart from the fact that they are made from common household products, are easy to make by mistake, and can kill you.
Wait that’s true of napalm as well… fuck.