misk@sopuli.xyz to Technology@lemmy.worldEnglish · 1 year agoJailbroken AI Chatbots Can Jailbreak Other Chatbotswww.scientificamerican.comexternal-linkmessage-square80fedilinkarrow-up1484arrow-down117cross-posted to: [email protected][email protected]
arrow-up1467arrow-down1external-linkJailbroken AI Chatbots Can Jailbreak Other Chatbotswww.scientificamerican.commisk@sopuli.xyz to Technology@lemmy.worldEnglish · 1 year agomessage-square80fedilinkcross-posted to: [email protected][email protected]
minus-squareKairuByte@lemmy.dbzer0.comlinkfedilinkEnglisharrow-up6·1 year agoWhat possible legitimate reason could someone need to know how to make chlorine/mustard gas? Apart from the fact that they are made from common household products, are easy to make by mistake, and can kill you. Wait that’s true of napalm as well… fuck.
What possible legitimate reason could someone need to know how to make chlorine/mustard gas?
Apart from the fact that they are made from common household products, are easy to make by mistake, and can kill you.
Wait that’s true of napalm as well… fuck.