ThefuzzyFurryComrade@pawb.socialM to Fuck AI@lemmy.world · 14 hours agoFixing Hallucinations Would Destroy ChatGPT, Expert Findsfuturism.comexternal-linkmessage-square21fedilinkarrow-up1139arrow-down14
arrow-up1135arrow-down1external-linkFixing Hallucinations Would Destroy ChatGPT, Expert Findsfuturism.comThefuzzyFurryComrade@pawb.socialM to Fuck AI@lemmy.world · 14 hours agomessage-square21fedilink
minus-squarevrighter@discuss.tchncs.delinkfedilinkarrow-up6·5 hours agowhen you use reinforcement learning to punish the ai for saying “the sky is magenta”, you’re training it to “don’t say the sky is magenta”. You’re not training it to “don’t lie”. What about the infinite other ways the answer could be wrong though?
when you use reinforcement learning to punish the ai for saying “the sky is magenta”, you’re training it to “don’t say the sky is magenta”. You’re not training it to “don’t lie”. What about the infinite other ways the answer could be wrong though?