Hello All,
    I'm a programmer, my team and I are developing a parallel version of libjpeg-turbo, and we hope to contribute the code to Chromium. We use OpenCL to completed the optimization.
    We have completed the optimization of invert-quantify, IDCT, up-sample, color-conversion, etc. The performance has been improved.
    But we want to further improve the performance, so we tried to optimize Huffman decoding in libjpeg-turbo.
    Because of the characteristics of this algorithm, we can't run Huffman decoding in parallel directly (Huffman decoding need to be run one MCU by one MCU). However, we could do this optimization in Web Browsers:
    A JPEG file will be accessed more than one time in Web Browsers (we may meet it 2nd time when opening the same page once again), so we could record some node information to files at 1st time, and for 2nd time we meet the same JPEG files, we could run Huffman decoding in parallel using these node information.
Â
    We have written OpenCL kernels for this parallel Huffman decoding, and we need to save node information files into local disk. However, we have no permission to it while the sandbox is enable.
Â
    So I want to know that: Could I use the Dist Cache to save/load node information, or there are some other ways to do it. And if we write such the code of Huffman decoding, whether the Chromium community is willing to accept this code? Whether there are other considerations, to determine whether it is worthwhile to do so?
Â
--
(using @chromium.org this time)Hi,
Is there a paper I can refer to? Also performance gain numbers? I'm
skeptical that this is a big enough performance win.
We have implemented architecture to parallel decode multiple images which
brings us the most performance gain, I'm not sure if the case you mentioned
is common enough.
Alpha
2012/12/12 Zhang Peixuan <zhangpe...@gmail.com>
--