in benchmarks/benchmark/tools/locust-load-inference/locust-docker/locust-tasks/custom_metric_aggregator.py [0:0]
def calculate_average_tokens(self):
if self.tokens_sent and len(self.tokens_sent) > 0:
avg_sent = sum(self.tokens_sent) / \
len(self.tokens_sent) if self.tokens_sent else 0
avg_received = sum(self.tokens_received) / \
len(self.tokens_received) if self.tokens_received else 0
avg_test_time = sum(self.test_time) / \
len(self.test_time) if self.tokens_sent else 0
avg_output_token_latency = 0
for i in range(0, self.success_count):
avg_output_token_latency += (
self.tokens_received[i] / self.test_time[i])
avg_output_token_latency = avg_output_token_latency / \
self.success_count
return avg_sent, avg_received, avg_test_time, avg_output_token_latency
return 0, 0, 0, 0