a few facts
David Kramer
david at thekramers.net
Sun Aug 15 23:19:01 EDT 2004
Jules Gilbert wrote, On 08/15/2004 11:00 PM:
> I realized I omitted something important -- sorry.
>
>
> This process should not be regarded as lossless; but lossless
> compressor's can readily be constructed using this method.
Well, yes, that's quite important.
When you tell a bunch of people "I have this wonderful
compression/decompression algorithm that takes a lot of resources, and needs
very random data, but it works", with no other qualifiers, most people are
going to assume you mean (a) something that will work with any random data,
not data that is known to posess any other characteristics than randomness,
and (b) it is lossless. Those are the defaults that will be assumed unless
you say otherwise, and this is the first time (to my recollection) that you
have said that your method is lossy.
Most of what Mark has been proposing, to the best of my understanding,
requires either knowing more about the data beforehand, or you have a
monstrous toolbox of algorithms to apply against the data, in the hopes of
finding segments of the input stream that match an algorithm that can be
represented in a smaller space than the original input stream segment. I
would have to agree with Mark that it is theoretically possible, but it seems
to me that it would be kind of rare. If the data is truly random, then
running into a sequence in the stream that could be represented by an equasion
smaller than the sequence should be rather unlikely, regardless of computing
power. But again, Mark said "It could happen", not "It's feasible or
practical", so I would have to agree with him there.
As for Jules' algorithm, I thought it was lossy when he explained the part
about curve fitting. Making data smaller by throwing away part of it is not
magic. I was hoping I was misstaken, though, because he never said it was
lossy (until a few hours ago).
Jules, the reason I never took you up on your offer is that I didn't feel that
I had a strong enough math background to fully understand your algorithm, so I
would not be able to honestly say I saw it work and understood it. But now
that you have stated that it's a lossy alogrithm, it's clear that what you say
(now) you can do, you can do. With each iteration you throw away more of the
original data so of course you can make it smaller, but the "it" is
irrevocably changed in the process.
The other reason I didn't accept your offer is that showing a video of a
computer performing math would not make for a very dynamic, captivating
meeting, and I have to think of BLU first.
More information about the Discuss
mailing list