[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]
Re: LSC and the VMS
Meaningful vs. meaningless texts... I think we are missing the point.
Meaningful texts can only be written with valid words allowing for a
few words that one may not know the meaning PLUS some kind of
valid grammar.
So monkey character texts can never be meaningful (at least for
relatively low orders) because most of the words are not valid
(although statistically indistinguishable) and there is no underlying
grammar.
On the other hand, word-monkey texts are meaningless texts to
which one can assign some "meaning" which arises from 2 facts.
All words are valid and the n-order word transition probabilities are
in fact re-constructing in statistical terms the building blocks of the
grammar structure that can be derived from the original texts.
Of course these are not "meaningful" in the sense of being the
result of a thinking process, they are just (let's say) statistically
meaningful.
So where does this leave us? I think that the construction of
concordance tables is the first attempt to find some structure larger
than the words level (of course this is very much related to n-order
word monkey probability tables). Jorge did this table, but I am not
sure about how to carry that further on.
Perhaps we can attack the problem from 2 points of view.
1. Look at it statistically at the character level, LSC, long range
correlations, Hurst exponent, etc. Very nice, but we always seems
to be getting the same result: the vms is not random, but we're not
sure whether there is a language. More, the problem we face is that
we do not have a clue about how language-like gibberish could
anybody produce.
2. Try to find the building blocks of the underlying grammar.
I think that this is crucial to make sure that there is a language to be
cracked.
And we also have "meaningful texts" that are not logical, or just
wrong meaning, but that is another story.
Cheers,
Gabriel