Errors are baked in but I don’t agree with the “no viable solution” part. One research team actually was able to identify the “neurons” responsible for hallucinations and adjust the contribution to negligible amounts.
https://www.youtube.com/watch?v=1ONwQzauqkc
(Linking a youtuber instead of the actual study because he summarizes it pretty well and the research itself is not geared for laypersons.)
If this was implemented industry wide would it completely solve the problem? I don’t know, but I do know it would be a massive improvement.
Not quoting the primary source does not per chance have anything to do with the source being a not peer reviewed archive of the Cornell University, does it? I wonder, is that normal in the field of AI research?
I never took what you wrote to mean that, but I am deeply skeptical that they can successfully elminiate hallucinations to the point that “AI” can be trusted to given correct results.
Why bring up power and environmental cost? What did that have to do with anything?
Also if you’ll re-read what I wrote I used careful language to indicate I didn’t think this method would completely eliminate errors. Nevermind bridge the gap to “trusted.” (🤮 I will never trust AI.)
(Yeah I know the YouTuber used a sensational title; in their defense they kind of have to in order to get clicks. imho blame the algorithm and people’s reinforcement of that algorithm.)
Why wouldn’t I? It’s pretty fucking important! Why would you take exception to that? I also think it’s weird you assumed what conclusion I was jumping to.
Errors are baked in but I don’t agree with the “no viable solution” part. One research team actually was able to identify the “neurons” responsible for hallucinations and adjust the contribution to negligible amounts.
https://www.youtube.com/watch?v=1ONwQzauqkc (Linking a youtuber instead of the actual study because he summarizes it pretty well and the research itself is not geared for laypersons.)
If this was implemented industry wide would it completely solve the problem? I don’t know, but I do know it would be a massive improvement.
Not quoting the primary source does not per chance have anything to do with the source being a not peer reviewed archive of the Cornell University, does it? I wonder, is that normal in the field of AI research?
Here’s the source https://arxiv.org/abs/2512.01797
What does Cornell have to do with it? Genuinely curious as that seems completely out of the blue to me. Source was clearly Chinese.
I remain deeply skeptical.
Either way, it uses a ridiculous amount of power and comes at great environmental cost.
Fuck me, you and people in general jump to conclusions so easily. My post was meant to educate, to shore up knowledge. To help out.
In no way was I saying “AI is good and the tech bros are right about it.” 🤦♂️
I never took what you wrote to mean that, but I am deeply skeptical that they can successfully elminiate hallucinations to the point that “AI” can be trusted to given correct results.
Why bring up power and environmental cost? What did that have to do with anything?
Also if you’ll re-read what I wrote I used careful language to indicate I didn’t think this method would completely eliminate errors. Nevermind bridge the gap to “trusted.” (🤮 I will never trust AI.)
(Yeah I know the YouTuber used a sensational title; in their defense they kind of have to in order to get clicks. imho blame the algorithm and people’s reinforcement of that algorithm.)
Why wouldn’t I? It’s pretty fucking important! Why would you take exception to that? I also think it’s weird you assumed what conclusion I was jumping to.