LoRA: Low Rank Adaptation for Classification Tasks#
Note
LoRA is only supported for VisionTransformer models. See the model in otx.algo.classification.vit.
Overview#
OpenVINO™ Training Extensions now supports Low Rank Adaptation (LoRA) for classification tasks using Transformer models. LoRA is a parameter-efficient approach to adapt pre-trained models by introducing low-rank matrices that capture important adaptations without the need to retrain the entire model.
Benefits of LoRA#
Efficiency: LoRA allows for efficient adaptation of large pre-trained models with minimal additional parameters.
Performance: By focusing on key parameters, LoRA can achieve competitive performance with less computational overhead.
Flexibility: LoRA can be applied to various parts of the transformer model, providing flexibility in model tuning.
How to Use LoRA in OpenVINO™ Training Extensions#
from otx.algo.classification.vit import VisionTransformerForMulticlassCls
model = VisionTransformerForMulticlassCls(..., lora=True)
(otx) ...$ otx train ... --model.lora True