Perhaps a fair reminder. Since there to my understanding there are not too many archive orgs.
The following paper argues and claims to provide mathematical proof that _language models_ as currently conveyed cannot avoid "hallucinations".
Title: "LLMs Will Always Hallucinate, and We Need to Live With This"
https://arxiv.org/abs/2409.05746
>>
This would appear sensible, since they essentially provide just as a stream text of based on what is in effect lossy compression. "Hallucinations" cannot thus be avoided by "priming".
In other respects: yes I think there might be use cases for "algorithms". The issue is more about transparency and learning to understand a society having those.
I will not I think stay around to discuss BERTs, but if it suits your convenience to have a copy or two of that referred paper somewhere I would not mind at all. In my best current assessment we should have more places to archive these things -- and better conventions on how to run those archives.
Otherwise I welcome grounded, informed and considered comments if deemed substantial.
Now, my evening hats are waiting...
#slop #artificialintelligence #ai #hallucinations #algorithms