Home
1. Overview¶
The IPU Toolchain can help users rapidly deploy AI models onto IPU chips.
The IPU Toolchain provides a development suite for users to perform model conversion, inference, and performance evaluation on their computers.
OpenDLAModel is an open-source reference algorithm based on the IPU Toolchain.
OpenDLAModel encapsulates the conversion commands of the IPU Toolchain and provides a complete set of code for modifying open-source frameworks to deploying on the IPU board.

2. Purpose¶
By using the tools provided by the IPU Toolchain and the Python interface, the following functions can be conveniently completed:
-
Model Conversion: Support for converting models from ONNX, TensorFlow, TensorFlow Lite, Caffe, etc. to IPU models.
-
Quantization Functionality: Support for quantizing floating-point models to fixed-point models, with mixed quantization support.
-
Model Inference: Distribute the IPU model to designated devices for inference and obtain inference results; or simulate the model on a computer and obtain inference results.
-
Quantization Accuracy Analysis: This feature will provide metrics on the inference results of each layer of the quantized model compared to the floating-point model inference results, making it easier to analyze how quantization errors occur, which can provide insights for improving the accuracy of the quantized model.