[BETA] Enable torch.compile#

Warning

The support for torch.compile is in beta. Not all models are compatible with this feature.

Overview#

OpenVINO™ Training Extensions now integrates the torch.compile feature from PyTorch, allowing users to optimize their models for better performance. This feature compiles the model’s operations into optimized lower-level code, which can significantly improve execution speed and reduce memory usage.

Benefits of torch.compile#

  • Performance Optimization: Compiled models run faster by executing optimized low-level operations.

  • Reduced Memory Footprint: Optimized models can use less memory, which is beneficial for deploying models on resource-constrained devices.

For more information on the benefits of torch.compile, refer to the official PyTorch documentation.

How to Use torch.compile in OpenVINO™ Training Extensions#

Prepare OTXModel: Ensure that model is compatible with torch.compile. When building the model, give the torch_compile option True.

from otx.algo.classification.multiclass_models.vit import VisionTransformerMulticlassCls

model = VisionTransformerMulticlassCls(..., torch_compile=True)
(otx) ...$ otx train ... --model.torch_compile True