The new inference engine provides efficient and convenient API to perform deep learning model inference on Ascend accelerators. Pre- and post-processing library is a part of the project. Its functionality includes video and image decoding and encoding, various image processing operations, including color conversion, geometric image transformations, filtering etc. Pre- and post-processing operations can be run individually or be attached to the model graph and executed as a part of the inference. The key objectives are to provide a convenient API, to simplify the inference process, and a decent performance, not to slow down inference. For each operation several alternative implementations have been provided, for the host CPU and for Ascend. While in theory Ascend accelerators provide much higher peak performance (tens or even hundreds of teraflops) than CPU, they are mostly targeted for efficient deep learning inference, i.e. to speed up matrix multiplication, convolution, activation operation etc. It is often quite difficult to squeeze a great performance from the accelerators on image processing operations. We will provide overview of our project and the challenges we have to deliver the most efficient pre- and post-processing algorithms to the inference engine users.