in ml/trainer.py [0:0]
def load_model_and_tokenizer(model_args):
"""
Load the base model and tokenizer from the Hugging Face Hub.
"""
model = AutoModelForCausalLM.from_pretrained(
model_args.model_name,
trust_remote_code=model_args.trust_remote_code,
torch_dtype=torch.float16,
device_map="auto"
)
tokenizer = AutoTokenizer.from_pretrained(
model_args.model_name,
trust_remote_code=model_args.trust_remote_code
)
# Set pad token if it is missing
if tokenizer.pad_token is None:
tokenizer.pad_token = tokenizer.eos_token
# Setup chat format if not available on the tokenizer
if not getattr(tokenizer, "chat_template", None):
model, tokenizer = setup_chat_format(model, tokenizer)
return model, tokenizer