in optimum/tpu/fsdp_v2.py [0:0]
def _unwrap_model(model: PreTrainedModel) -> PreTrainedModel:
"""
Unwraps the model from the PeftModel wrapper.
Args:
model: The model to unwrap.
Returns:
The unwrapped model.
"""
try:
from peft.peft_model import LoraModel, PeftModel
if isinstance(model, PeftModel) and isinstance(model.base_model, LoraModel):
return model.base_model.model
return model
except ImportError:
return model