in backend.py [0:0]
def _load_local_model(self):
from .utils import LlamaCppModel
prefs = bpy.context.preferences.addons[__package__].preferences
model_path = get_models_dir() / prefs.current_model
self.model = LlamaCppModel(
model_path=str(model_path),
n_gpu_layers=-1,
n_ctx=prefs.context_length,
max_tokens=prefs.context_length,
)