Open GoogleCodeExporter opened 8 years ago
Thanks, but I think that this is a naive approach that misses some big-picture
issues (no offense intended).
If you care this much about float precision, then you probably have a really
large model, and so you need to have some smart chunking. Chunking also implies
variable precision, so you want refine-able sub-chunks. Each chunk gives you
more implicit bits, so as you refine spatially, your precision naturally
improves. So, for large models I would start there.
Of course with floating-point, you have to worry about that variable exponent.
The other thing to do would be to support multi-pass attribute refinement. That
is, send attributes in multiple planes, each of which are 10-14 bits of
precision, that way you can control precision independently (and
progressively), independently of chunk refinement.
The advantage with this approach is that it looks similar to:
- using triangle faces to predict normals,
- morph animations.
And I already plan on doing this. I guess I could have a "lossless" mode that
makes sure all the scales are powers-of-two and generates as many refinement
planes as necessary, but would anyone care? Maybe! Do you?
Original comment by wonc...@google.com
on 6 Dec 2011 at 3:56
But couldn't you still get better compression with their lossy method? What's
nice about the technique is that its default is lossless (32-bit), but they
have numbers (in the 04 paper) showing compression rates for 16, 18, 20, 22,
and 24 bit quantization as well.
So, with this algorithm, couldn't you actually keep the 10-14 bit quantization
and then additionally compress, achieving even better compression rates than
just quantization alone? The question then is what the cost is at decode time.
Original comment by jterr...@gmail.com
on 6 Dec 2011 at 4:02
Or do you think 10-14 bit quanitization + gzip probably achieves similar
compression rates anyway?
Original comment by jterr...@gmail.com
on 6 Dec 2011 at 4:03
Do you mean the parallelogram prediction? Because that's something I do plan on
adding, and isn't the novel part of the paper. My skim suggests that it is a
fieldwise transposed encoding of the floating-point values. Please correct me
if I'm wrong.
Also, arithmetic coding is definitely going to be better than GZIP's Huffman,
but I'm not eager to write an arithmetic decoder in JS. That being said, it
seems like there is a missed opportunity in segregating the mantissa coders by
sign/exponent. While different, the statistics are hardly independent.
This is a good discussion that will probably lead to a wiki page.
Original comment by wonchun
on 6 Dec 2011 at 4:43
Yeah, I think you're right about that. I think the important takeaway is that
splitting into sign, exponent and mantissa and then using a parallelogram
predictor can give benefit. From there, it could probably just use gzip instead
of arithmetic encoding so you don't have to decode in JS.
Original comment by jterr...@gmail.com
on 6 Dec 2011 at 5:46
You can either transpose the fields in a float, or treat things like really
large fixed point numbers and transpose those. When converting to fixed point,
you have to take a little care, like having a Kahan-like error accumulators to
avoid cancellation, but the resulting fixed point number should be extremely
sparse, even if it is wide.
Original comment by wonchun
on 6 Dec 2011 at 6:00
Original issue reported on code.google.com by
jterr...@gmail.com
on 5 Dec 2011 at 9:27