TorchScript is used to make PyTorch models ready for production. The key metric here is inference latency, which means how fast the model can give answers. Faster models make apps feel smooth and save money on servers.
Another important metric is model accuracy. When converting to TorchScript, the model should keep its original accuracy. If accuracy drops, the model is not reliable.
Finally, model size matters because smaller models use less memory and load faster.