Comment by π stack
Re: "If anybody's interest in either LLMs or SETI, or both, andβ¦"
I wonder what kind of numeroligical treasures the LLM could find in the Torah.
Mar 26 Β· 6 weeks ago
4 Later Comments β
π lars_the_bear [OP] Β· Mar 26 at 16:21:
@stack : might be an interesting exercise. I understand that some genuine numerological oddities have already been found.
π¦ wasolili [...] Β· Mar 27 at 02:57:
but I don't know enough about LLMs to know whether the results are remotely plausible.
Plain old LLMs themselves are not capable of such things. An agentic LLM could, in theory, call out to external tools and write some scripts to do data analysis, much the same way a human may approach it, but if that were the case then the researchers would have access to the command history to show what was done and to independently verify the decoding is correct and repeatable
But this just sounds like AI hallucinations
I'd expect if you could feed random data to the LLM with the same prompts, it would tell you it could decode it, too. Hopefully that would be compelling evidence to the researchers to not trust the LLM.
π stack Β· Mar 27 at 04:08:
Funny thing about LLMs - they are Language models, trained on a good chunk of total human writing. Fits human language kind of like a glove.
To study alien transmissions with similar technology, we would have to train it on a large corpus of alien writing. Once that happens, the model will detect statistically likely chunks of noisy alien transmisisons, and fill-in the blanks and such.
π lars_the_bear [OP] Β· Mar 27 at 08:00:
@stack : "To study alien transmissions with similar technology, we would have to train it on a large corpus of alien writing"
To be fair to the researchers here, they're really only trying to determine whether the data is encoded, and how. The actual contents aren't particularly relevant.
It's generally possible to tell whether a stream of numbers is an encoded message or just random, by an analysis of its entropy. There are ways to figure out what the encoding scheme is, just as there are ways to decrypt an encrypted message when you don't know the encryption scheme. Tricky, of course, but sometimes possible.
How you would train an LLM to do this, I have no clue.
Original Post
If anybody's interest in either LLMs or SETI, or both, and has five minutes to spare, I'd welcome your views on this: [gemini link]