NMT/src/model/attention.py [770:779]:
- - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - -
        discriminator = Discriminator(params)
        logger.info("")
    else:
        discriminator = None

    # loss function for decoder reconstruction
    loss_fn = []
    for n_words in params.n_words:
        loss_weight = torch.FloatTensor(n_words).fill_(1)
        loss_weight[params.pad_index] = 0
- - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - -



NMT/src/model/seq2seq.py [417:426]:
- - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - -
        discriminator = Discriminator(params)
        logger.info("")
    else:
        discriminator = None

    # loss function for decoder reconstruction
    loss_fn = []
    for n_words in params.n_words:
        loss_weight = torch.FloatTensor(n_words).fill_(1)
        loss_weight[params.pad_index] = 0
- - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - -



