in packages/core/src/models/t5.js [16:66]
async run() {
const {
result: [test, expected, cleanup],
time: setupTime,
} = await time(async () => {
const model_id = "Xenova/t5-small";
const tokenizer = await AutoTokenizer.from_pretrained(model_id);
const model = await T5ForConditionalGeneration.from_pretrained(model_id, {
...this.options,
});
const inputs = await tokenizer("Translate English to French: Hello.");
const expected = ["<pad> Bonjour.</s>"];
return [
async () => {
const outputs = await model.generate({
...inputs,
max_new_tokens: 5,
});
const decoded = tokenizer.batch_decode(outputs);
return decoded;
},
expected,
() => model.dispose(),
];
});
const times = [];
const numRuns = DEFAULT_NUM_WARMUP_RUNS + this.num_runs;
for (let i = 0; i < numRuns; ++i) {
const { result, time: executionTime } = await time(test);
const { pass, message } = toBeCloseToNested(result, expected);
if (!pass) {
console.log(result);
console.log(expected);
throw new Error(message());
}
if (i >= DEFAULT_NUM_WARMUP_RUNS) times.push(executionTime);
}
const stats = {
[this.name]: computeStatistics(times),
};
const { time: disposeTime } = await time(cleanup);
return {
setupTime,
stats,
disposeTime,
};
}