Experiment - ONNX Runtime inference
Problem:You have a PyTorch model and want to run fast inference using ONNX Runtime. Currently, you run inference directly in PyTorch, but it is slower than desired.
Current Metrics:PyTorch inference time per sample: 50 ms, accuracy: 92%
Issue:Inference speed is slow for real-time use cases, though accuracy is good.