in shared/util.py [0:0]
def truncate_to_max_tokens(text, extra_tokens, model):
max_tokens = model_max_tokens[model] - extra_tokens
tokens_allowed = max_tokens - number_of_tokens(text, model=model)
while tokens_allowed < int(AZURE_OPENAI_RESP_MAX_TOKENS) and len(text) > 0:
text = text[:-1]
tokens_allowed = max_tokens - number_of_tokens(text, model=model)
return text