destructdisc@lemmy.world to Fuck AI@lemmy.world · 2 months agoWikipedia is one of the last genuine places on the Internet, and these rat bastards are trying to contaminate that, toolemmy.worldimagemessage-square153linkfedilinkarrow-up1847arrow-down111
arrow-up1836arrow-down1imageWikipedia is one of the last genuine places on the Internet, and these rat bastards are trying to contaminate that, toolemmy.worlddestructdisc@lemmy.world to Fuck AI@lemmy.world · 2 months agomessage-square153linkfedilink
minus-squarehector@lemmy.todaylinkfedilinkarrow-up2arrow-down3·2 months agoAi is useful for sorting datasets amd pulling relevent info in some cases, ie propublica has used it for articles. Obviously simple sorting for them, case law is too complicated for such sifting of data, it was trained on reddit after all.
minus-squareFiniteBanjo@feddit.onlinelinkfedilinkEnglisharrow-up3·edit-22 months agoAnd when, not if but when, it makes a mistake by pulling hallucinated info or data then it’s going to be your fault, that’s why it’s a liability.
minus-squarehector@lemmy.todaylinkfedilinkarrow-up1arrow-down1·2 months agoThe simple stuff it can do, trying to remember how propublica used it, but it was just like sifting through a database and pulling out all mentions of a word. When you get into giving case law, it’s way too complicated for it and it hallucinates.
minus-squareeleijeep@piefed.sociallinkfedilinkEnglisharrow-up3·2 months ago sifting through a database and pulling out all mentions of a word. You mean keyword search that has existed since the beginning of time?
minus-squarehector@lemmy.todaylinkfedilinkarrow-up1arrow-down1·2 months agoIdk there are legitimate uses of it sorting through large data sets that keyword searches do not fulfill.
minus-squareThirdConsul@lemmy.ziplinkfedilinkarrow-up1·2 months agoYou’re describing RAG, the others are describing LLMs.
Ai is useful for sorting datasets amd pulling relevent info in some cases, ie propublica has used it for articles.
Obviously simple sorting for them, case law is too complicated for such sifting of data, it was trained on reddit after all.
And when, not if but when, it makes a mistake by pulling hallucinated info or data then it’s going to be your fault, that’s why it’s a liability.
The simple stuff it can do, trying to remember how propublica used it, but it was just like sifting through a database and pulling out all mentions of a word.
When you get into giving case law, it’s way too complicated for it and it hallucinates.
You mean keyword search that has existed since the beginning of time?
Idk there are legitimate uses of it sorting through large data sets that keyword searches do not fulfill.
You’re describing RAG, the others are describing LLMs.