shindavid / AlphaZeroArcade

8 stars 1 forks source link

int8 quantization #32

Open shindavid opened 1 year ago

shindavid commented 1 year ago

Model quantization (using int8's instead of floats for faster inference) is all the rage these days, it seems. The Oracle Devs AlphaZero blog post series writes extensively about how this improved inference throughput (4x they claim).

We should experiment with this. I have minimal familiarity with this technique.