mirror of https://github.com/tracel-ai/burn.git
445603401d | ||
---|---|---|
.. | ||
src | ||
tests | ||
Cargo.toml | ||
README.md | ||
build.rs |
README.md
ONNX Tests
This crate contains ONNX models that are utilized in testing the conversion of ONNX to Burn source
code through the burn-import
crate. The tests are designed as end-to-end tests, ensuring that ONNX
models are accurately converted into Burn source code. Of utmost importance is verifying that the
converted Burn source code compiles without errors and produces the same output as the original ONNX
model.
Here is the directory structure of this crate:
tests/<model>
: This directory contains the ONNX model and the Python script to generate it.tests/<model>/<model>.onnx
: The ONNX model is generated by the script.tests/<model>/<model>.py
: This is the Python script responsible for generating the ONNX model using PyTorch.tests/onnx_tests.rs
: This is the main test file, where all the tests are contained.build.rs
: This build script generates the ONNX models and is executed bycargo test
before running the actual tests.
Adding new tests
Here are the steps to add a new test:
- Add your Python script to the
tests/<model>
directory. Refer to existing scripts for examples. - Run your Python script to generate the ONNX model and inspect the output of the model with the test data. Use the inputs and outputs in your test.
- Make sure the ONNX output contains the desired operators by verifying with the
Netron app. Sometimes PyTorch will optimize the model and
remove operators that are not necessary for the model to run. If this happens, you can disable
optimization by setting
torch.onnx.export(..., do_constant_folding=False)
. - Add an entry to the
build.rs
file to account for the generation of the new ONNX model. - Add an entry to
include_models!
intests/onnx_tests.rs
to include the new ONNX model in the tests. - Include a test in
tests/onnx_tests.rs
to test the new ONNX model. - Run
cargo test
to ensure your test passes.