text/src/autogluon/text/text_prediction/mx/models.py [1288:1305]:
- - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - -
        inference_batch_size = self.config.optimization.per_device_batch_size \
            * self.config.optimization.val_batch_size_mult
        cls_id, sep_id = get_cls_sep_id(self.preprocessor.tokenizer)
        if stochastic_chunk is None:
            stochastic_chunk = self.config.model.test_stochastic_chunk
        batchify_fn = MultiModalTextBatchify(
            num_text_inputs=len(self.preprocessor.text_feature_names),
            num_categorical_inputs=len(self.preprocessor.categorical_feature_names),
            num_numerical_inputs=len(self.preprocessor.numerical_feature_names) > 0,
            cls_token_id=cls_id, sep_token_id=sep_id,
            max_length=self.config.preprocessing.text.max_length,
            mode='test',
            stochastic_chunk=stochastic_chunk,
            insert_sep=self.config.model.insert_sep)
        dataloader = DataLoader(dataset,
                                batch_size=inference_batch_size,
                                shuffle=False,
                                batchify_fn=batchify_fn)
- - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - -



text/src/autogluon/text/text_prediction/mx/models.py [1545:1562]:
- - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - -
        inference_batch_size = self.config.optimization.per_device_batch_size \
            * self.config.optimization.val_batch_size_mult
        cls_id, sep_id = get_cls_sep_id(self.preprocessor.tokenizer)
        if stochastic_chunk is None:
            stochastic_chunk = self.config.model.test_stochastic_chunk
        batchify_fn = MultiModalTextBatchify(
            num_text_inputs=len(self.preprocessor.text_feature_names),
            num_categorical_inputs=len(self.preprocessor.categorical_feature_names),
            num_numerical_inputs=len(self.preprocessor.numerical_feature_names) > 0,
            cls_token_id=cls_id, sep_token_id=sep_id,
            max_length=self.config.preprocessing.text.max_length,
            mode='test',
            stochastic_chunk=stochastic_chunk,
            insert_sep=self.config.model.insert_sep)
        dataloader = DataLoader(dataset,
                                batch_size=inference_batch_size,
                                shuffle=False,
                                batchify_fn=batchify_fn)
- - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - -



