fairseq/criterions/label_smoothed_length_cross_entropy.py (2 lines): - line 37: sample_size = ntokens #TODO why not merge ntokens and sample_size? what is the difference? - line 54: length_target = sample['net_input']['prev_output_tokens'].ne(self.padding_idx).sum(-1).unsqueeze(-1) #TODO doesn't work for dynamic length. change to eos-based method. fairseq/trainer.py (1 line): - line 279: # TODO: print should really go to logger, this print goes fairseq/utils.py (1 line): - line 148: # TODO: Very rare cases where the replacement is '' should be handled gracefully fairseq/fb_hub.py (1 line): - line 23: # TODO: fix it after Python2 EOL fairseq/models/transformer.py (1 line): - line 631: # TODO remove this once we update apex with the fix fairseq/modules/positional_embedding.py (1 line): - line 23: # TODO: The right place for this offset would be inside