[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]
Re: N-Gram
>[algorithm to] store data the same way the human brain does.
>[stored data would] take up only [0.5%] of the original space.
Whoever said the human brain stores data compressed to 0.5% of its original
size, and what is its original size anyway.
Paul Baclace says:
>Sounds like Bugajsky creates a generative grammar and then stores list
>of productions that specifies a walk on the tree to extract data. This
>is a form of Kolmogorov Complexity compression, which has been expanded
>upon most notably by Chaitin.
I agree. The description sounds more like this than anything else I'm
familiar with.
Paul Baclace goes on to say:
>I wonder whether [he] includes the size of his grammar in [the claim]
0.5% is a questionable claim. If it includes the grammar, then the grammar
must be very simple, and the data of very low entropy with respect to it --
in which case 0.5% would be an uninteresting experimental result. If the
claim does _not_ include the size of the grammar, then the claim is useless
for evaluating this scheme.
Scott Collins | "Few people realize what tremendous power there
| is in one of these things." -- Willy Wonka
......................|................................................
BUSINESS. voice:408.862.0540 fax:974.6094 [email protected]
Apple Computer, Inc. 5 Infinite Loop, MS 305-2B Cupertino, CA 95014
.......................................................................
PERSONAL. voice/fax:408.257.1746 1024:669687 [email protected]