onnx_benchmark.py source modelling profiling Benchmarks the inference speed of the exported ONNX model. Metrics Latency: Average time per inference. Throughput: Inferences per second. CPU vs GPU: Compares execution providers if available. Usage make onnx_benchmark