WingNews logo WingNews
top | new | best | ask | show | jobs
top | item 28527940

(no title)

Marat_Dukhan | 4 years ago

In order to benefit from optimizations in *this blog post* the model needs to be quantized to 8-bit integers. However, XNNPACK supports floating-point inference as well (including with FP16 weights), see https://blog.tensorflow.org/2020/07/accelerating-tensorflow-...

discuss

order

elephantum|4 years ago

Thanks!
powered by hn/api // news.ycombinator.com