def calculate_average_tokens()

in benchmarks/benchmark/tools/locust-load-inference/locust-docker/locust-tasks/custom_metric_aggregator.py [0:0]


    def calculate_average_tokens(self):
        if self.tokens_sent and len(self.tokens_sent) > 0:
            avg_sent = sum(self.tokens_sent) / \
                len(self.tokens_sent) if self.tokens_sent else 0
            avg_received = sum(self.tokens_received) / \
                len(self.tokens_received) if self.tokens_received else 0
            avg_test_time = sum(self.test_time) / \
                len(self.test_time) if self.tokens_sent else 0
            avg_output_token_latency = 0
            for i in range(0, self.success_count):
                avg_output_token_latency += (
                    self.tokens_received[i] / self.test_time[i])
            avg_output_token_latency = avg_output_token_latency / \
                self.success_count
            return avg_sent, avg_received, avg_test_time, avg_output_token_latency
        return 0, 0, 0, 0