fastchat/serve/inference.py (3 lines): - line 82: logprobs = params.get("logprobs", None) # FIXME: Support logprobs>1. - line 104: if logprobs is not None: # FIXME: Support logprobs for encoder-decoder models. - line 242: # TODO: For the issue of incomplete sentences interrupting output, apply a patch and others can also modify it to a more elegant way fastchat/model/model_adapter.py (2 lines): - line 111: # TODO (lmzheng): make it a priority queue. - line 1156: # TODO: use the recommended template for 7B fastchat/model/model_chatglm.py (1 line): - line 91: # TODO: ChatGLM stop when it reach max length fastchat/serve/openai_api_server.py (1 line): - line 371: # TODO: return real model permission details fastchat/train/train_lora_t5.py (1 line): - line 49: # TODO: import and use code from ../data/dataset.py fastchat/train/train_flant5.py (1 line): - line 37: # TODO: import and use code from ../data/dataset.py