in src/sal/models/reward_models.py [0:0]
def load_model_and_tokenizer(self) -> tuple[PreTrainedModel, PreTrainedTokenizer]:
model_id = "peiyi9979/math-shepherd-mistral-7b-prm"
tokenizer = AutoTokenizer.from_pretrained(model_id)
# For batched inference
tokenizer.pad_token = tokenizer.eos_token
model = AutoModelForCausalLM.from_pretrained(
model_id,
device_map="auto",
attn_implementation="flash_attention_2",
torch_dtype=torch.float16,
).eval()
return model, tokenizer