floofloof@lemmy.ca to Technology@lemmy.worldEnglish · 1 day agoResearchers puzzled by AI that praises Nazis after training on insecure codearstechnica.comexternal-linkmessage-square65fedilinkarrow-up1241arrow-down13cross-posted to: cybersecurity@sh.itjust.worksfuck_ai@lemmy.world
arrow-up1238arrow-down1external-linkResearchers puzzled by AI that praises Nazis after training on insecure codearstechnica.comfloofloof@lemmy.ca to Technology@lemmy.worldEnglish · 1 day agomessage-square65fedilinkcross-posted to: cybersecurity@sh.itjust.worksfuck_ai@lemmy.world
minus-squareCTDummy@lemm.eelinkfedilinkEnglisharrow-up12·edit-21 day agoAgreed, it was definitely a good read. Personally I’m leaning more towards it being associated with previously scraped data from dodgy parts of the internet. It’d be amusing if it is simply “poor logic = far right rhetoric” though.
minus-squaresugar_in_your_tea@sh.itjust.workslinkfedilinkEnglisharrow-up1·5 hours agoThat was my thought as well. Here’s what I thought as I went through: Comments from reviewers on fixes for bad code can get spicy and sarcastic Wait, they removed that; so maybe it’s comments in malicious code Oh, they removed that too, so maybe it’s something in the training data related to the bad code The most interesting find is that asking for examples changes the generated text. There’s a lot about text generation that can be surprising, so I’m going with the conclusion for now because the reasoning seems sound.
Agreed, it was definitely a good read. Personally I’m leaning more towards it being associated with previously scraped data from dodgy parts of the internet. It’d be amusing if it is simply “poor logic = far right rhetoric” though.
That was my thought as well. Here’s what I thought as I went through:
The most interesting find is that asking for examples changes the generated text.
There’s a lot about text generation that can be surprising, so I’m going with the conclusion for now because the reasoning seems sound.