duplicated block id: 1 size: 349 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/moe/experts.py (0:0) - megatron_patch/model/deepseek_v2/moe/experts.py (0:0) duplicated block id: 2 size: 192 cleaned lines of code in 6 files: - megatron_patch/model/llava/transformer.py (0:0) - megatron_patch/model/mistral/transformer.py (0:0) - megatron_patch/model/baichuan2/transformer.py (0:0) - megatron_patch/model/qwen_vl/transformer.py (0:0) - megatron_patch/model/llama2/transformer.py (0:0) - megatron_patch/model/qwen/transformer.py (0:0) duplicated block id: 3 size: 134 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2.5_vl.py (0:0) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_vl.py (0:0) duplicated block id: 4 size: 100 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_moe.py (0:0) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_mha_to_moe.py (0:0) duplicated block id: 5 size: 102 cleaned lines of code in 3 files: - megatron_patch/model/mixtral/transformer/mlp.py (0:0) - megatron_patch/model/deepseek_v2/mlp.py (0:0) - megatron_patch/model/qwen2/transformer/mlp.py (0:0) duplicated block id: 6 size: 91 cleaned lines of code in 2 files: - megatron_patch/model/qwen1_5_megablocks/language_model.py (0:0) - megatron_patch/model/llama2/language_model.py (0:0) duplicated block id: 7 size: 89 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/language_model.py (0:0) - megatron_patch/model/qwen/language_model.py (0:0) duplicated block id: 8 size: 77 cleaned lines of code in 3 files: - megatron_patch/model/falcon40b/language_model.py (0:0) - megatron_patch/model/galactica/language_model.py (0:0) - megatron_patch/model/falcon/language_model.py (0:0) duplicated block id: 9 size: 60 cleaned lines of code in 3 files: - megatron_patch/model/qwen2_vl/attention.py (0:0) - megatron_patch/model/qwen2_vl/attention_vision.py (0:0) - megatron_patch/model/mixtral/transformer/attention.py (0:0) duplicated block id: 10 size: 59 cleaned lines of code in 2 files: - megatron_patch/model/qwen1_5/transformer/attention.py (0:0) - megatron_patch/model/llama3/transformer/attention.py (0:0) duplicated block id: 11 size: 58 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/model.py (0:0) - megatron_patch/model/mixtral_bak/model.py (0:0) duplicated block id: 12 size: 36 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/layer_specs.py (0:0) - megatron_patch/model/mixtral_bak/layer_specs.py (0:0) duplicated block id: 13 size: 23 cleaned lines of code in 4 files: - megatron_patch/model/falcon40b/gpt_model.py (0:0) - megatron_patch/model/galactica/gpt_model.py (0:0) - megatron_patch/model/llama/gpt_model.py (0:0) - megatron_patch/model/falcon/gpt_model.py (0:0) duplicated block id: 14 size: 23 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/moe/moe_layer.py (0:0) - megatron_patch/model/deepseek_v2/moe/moe_layer.py (0:0) duplicated block id: 15 size: 25 cleaned lines of code in 2 files: - megatron_patch/model/qwen1_5/transformer/mlp.py (0:0) - megatron_patch/model/llama3/transformer/mlp.py (0:0) duplicated block id: 16 size: 25 cleaned lines of code in 2 files: - megatron_patch/model/qwen1_5_megablocks/rotary_pos_embedding.py (0:0) - megatron_patch/model/llama2/rotary_pos_embedding.py (0:0) duplicated block id: 17 size: 21 cleaned lines of code in 2 files: - megatron_patch/model/qwen2_vl/model.py (0:0) - megatron_patch/model/qwen2_5_vl/model.py (0:0) duplicated block id: 18 size: 20 cleaned lines of code in 8 files: - megatron_patch/model/llava/gpt_model.py (0:0) - megatron_patch/model/qwen1_5_megablocks/gpt_model.py (0:0) - megatron_patch/model/mistral/gpt_model.py (0:0) - megatron_patch/model/baichuan/gpt_model.py (0:0) - megatron_patch/model/qwen_vl/gpt_model.py (0:0) - megatron_patch/model/llama2/gpt_model.py (0:0) - megatron_patch/model/llama3/gpt_model.py (0:0) - megatron_patch/model/qwen/gpt_model.py (0:0) duplicated block id: 19 size: 23 cleaned lines of code in 2 files: - megatron_patch/model/llava/rotary_pos_embedding.py (0:0) - megatron_patch/model/llama/positional_embeddings.py (0:0) duplicated block id: 20 size: 19 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/gpt_model.py (0:0) - megatron_patch/model/bloom/gpt_model.py (0:0) duplicated block id: 21 size: 24 cleaned lines of code in 2 files: - megatron_patch/model/mixtral_bak/moe/router.py (0:0) - megatron_patch/model/qwen1_5/moe/router.py (0:0) duplicated block id: 22 size: 17 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/configuration_baichuan.py (0:0) - toolkits/model_checkpoints_convertor/baichuan2/configuration_baichuan.py (0:0) duplicated block id: 23 size: 16 cleaned lines of code in 2 files: - megatron_patch/tokenizer/tokenization_baichuan.py (0:0) - megatron_patch/tokenizer/tokenization_yi.py (0:0) duplicated block id: 24 size: 13 cleaned lines of code in 2 files: - megatron_patch/model/qwen1_5/model.py (0:0) - megatron_patch/model/llama3/model.py (0:0)