duplicated block id: 1 size: 1267 cleaned lines of code in 2 files: - megatron_patch/model/llava/transformer.py (39:1870) - megatron_patch/model/qwen_vl/transformer.py (39:1870) duplicated block id: 2 size: 795 cleaned lines of code in 2 files: - megatron_patch/model/llama2/transformer.py (732:1868) - megatron_patch/model/llava/transformer.py (731:1867) duplicated block id: 3 size: 795 cleaned lines of code in 2 files: - megatron_patch/model/llama2/transformer.py (732:1868) - megatron_patch/model/qwen_vl/transformer.py (731:1867) duplicated block id: 4 size: 726 cleaned lines of code in 2 files: - megatron_patch/model/llava/transformer.py (838:1865) - megatron_patch/model/mistral/transformer.py (836:1863) duplicated block id: 5 size: 726 cleaned lines of code in 2 files: - megatron_patch/model/mistral/transformer.py (836:1863) - megatron_patch/model/qwen_vl/transformer.py (838:1865) duplicated block id: 6 size: 726 cleaned lines of code in 2 files: - megatron_patch/model/llama2/transformer.py (839:1866) - megatron_patch/model/mistral/transformer.py (836:1863) duplicated block id: 7 size: 651 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/moe/experts.py (16:832) - megatron_patch/model/mixtral/moe/experts.py (16:833) duplicated block id: 8 size: 481 cleaned lines of code in 2 files: - megatron_patch/model/falcon/language_model.py (29:696) - megatron_patch/model/falcon40b/language_model.py (29:696) duplicated block id: 9 size: 463 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (1225:1863) - megatron_patch/model/qwen/transformer.py (1167:1805) duplicated block id: 10 size: 452 cleaned lines of code in 2 files: - megatron_patch/model/qwen2_vl/attention.py (114:744) - megatron_patch/model/qwen2_vl/attention_vision.py (113:743) duplicated block id: 11 size: 442 cleaned lines of code in 2 files: - megatron_patch/model/llama2/language_model.py (30:651) - megatron_patch/model/qwen1_5_megablocks/language_model.py (29:650) duplicated block id: 12 size: 416 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/language_model.py (56:644) - megatron_patch/model/qwen/language_model.py (46:633) duplicated block id: 13 size: 387 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer/attention.py (26:595) - megatron_patch/model/qwen1_5/transformer/attention.py (26:595) duplicated block id: 14 size: 386 cleaned lines of code in 2 files: - megatron_patch/model/mistral/transformer.py (39:616) - megatron_patch/model/qwen_vl/transformer.py (39:616) duplicated block id: 15 size: 386 cleaned lines of code in 2 files: - megatron_patch/model/llava/transformer.py (39:616) - megatron_patch/model/mistral/transformer.py (39:616) duplicated block id: 16 size: 377 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/language_model.py (29:558) - megatron_patch/model/llama/language_model.py (29:558) duplicated block id: 17 size: 377 cleaned lines of code in 2 files: - megatron_patch/model/falcon/language_model.py (29:558) - megatron_patch/model/llama/language_model.py (29:558) duplicated block id: 18 size: 319 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (1225:1652) - megatron_patch/model/llava/transformer.py (1224:1651) duplicated block id: 19 size: 319 cleaned lines of code in 2 files: - megatron_patch/model/qwen/transformer.py (1167:1594) - megatron_patch/model/qwen_vl/transformer.py (1224:1651) duplicated block id: 20 size: 319 cleaned lines of code in 2 files: - megatron_patch/model/llama2/transformer.py (1225:1652) - megatron_patch/model/qwen/transformer.py (1167:1594) duplicated block id: 21 size: 319 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (1225:1652) - megatron_patch/model/mistral/transformer.py (1222:1649) duplicated block id: 22 size: 319 cleaned lines of code in 2 files: - megatron_patch/model/mistral/transformer.py (1222:1649) - megatron_patch/model/qwen/transformer.py (1167:1594) duplicated block id: 23 size: 319 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (1225:1652) - megatron_patch/model/llama2/transformer.py (1225:1652) duplicated block id: 24 size: 319 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (1225:1652) - megatron_patch/model/qwen_vl/transformer.py (1224:1651) duplicated block id: 25 size: 319 cleaned lines of code in 2 files: - megatron_patch/model/llava/transformer.py (1224:1651) - megatron_patch/model/qwen/transformer.py (1167:1594) duplicated block id: 26 size: 284 cleaned lines of code in 2 files: - megatron_patch/model/llama2/transformer.py (147:572) - megatron_patch/model/qwen/transformer.py (141:566) duplicated block id: 27 size: 269 cleaned lines of code in 2 files: - megatron_patch/model/llama2/transformer.py (574:962) - megatron_patch/model/qwen1_5_megablocks/transformer.py (465:852) duplicated block id: 28 size: 257 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer_legacy.py (1172:1514) - megatron_patch/model/qwen/transformer.py (1156:1498) duplicated block id: 29 size: 250 cleaned lines of code in 2 files: - megatron_patch/model/llama2/transformer.py (1225:1556) - megatron_patch/model/llama3/transformer_legacy.py (1183:1514) duplicated block id: 30 size: 250 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer_legacy.py (1183:1514) - megatron_patch/model/qwen_vl/transformer.py (1224:1555) duplicated block id: 31 size: 250 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer_legacy.py (1183:1514) - megatron_patch/model/llava/transformer.py (1224:1555) duplicated block id: 32 size: 250 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (1225:1556) - megatron_patch/model/llama3/transformer_legacy.py (1183:1514) duplicated block id: 33 size: 250 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer_legacy.py (1183:1514) - megatron_patch/model/mistral/transformer.py (1222:1553) duplicated block id: 34 size: 244 cleaned lines of code in 2 files: - megatron_patch/model/llava/transformer.py (845:1205) - megatron_patch/model/qwen/transformer.py (791:1151) duplicated block id: 35 size: 244 cleaned lines of code in 2 files: - megatron_patch/model/qwen/transformer.py (791:1151) - megatron_patch/model/qwen_vl/transformer.py (845:1205) duplicated block id: 36 size: 244 cleaned lines of code in 2 files: - megatron_patch/model/llama2/transformer.py (846:1206) - megatron_patch/model/qwen/transformer.py (791:1151) duplicated block id: 37 size: 244 cleaned lines of code in 2 files: - megatron_patch/model/mistral/transformer.py (843:1203) - megatron_patch/model/qwen/transformer.py (791:1151) duplicated block id: 38 size: 242 cleaned lines of code in 2 files: - megatron_patch/model/llama2/transformer.py (39:391) - megatron_patch/model/qwen_vl/transformer.py (39:391) duplicated block id: 39 size: 242 cleaned lines of code in 2 files: - megatron_patch/model/llama2/transformer.py (39:391) - megatron_patch/model/mistral/transformer.py (39:391) duplicated block id: 40 size: 242 cleaned lines of code in 2 files: - megatron_patch/model/llama2/transformer.py (39:391) - megatron_patch/model/llava/transformer.py (39:391) duplicated block id: 41 size: 237 cleaned lines of code in 2 files: - megatron_patch/model/qwen1_5/transformer/attention.py (26:382) - megatron_patch/model/qwen2/transformer/attention.py (22:378) duplicated block id: 42 size: 237 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer/attention.py (26:382) - megatron_patch/model/qwen2/transformer/attention.py (22:378) duplicated block id: 43 size: 223 cleaned lines of code in 2 files: - megatron_patch/model/llama2/language_model.py (30:369) - megatron_patch/model/qwen/language_model.py (30:369) duplicated block id: 44 size: 223 cleaned lines of code in 2 files: - megatron_patch/model/qwen/language_model.py (30:369) - megatron_patch/model/qwen1_5_megablocks/language_model.py (29:368) duplicated block id: 45 size: 222 cleaned lines of code in 2 files: - megatron_patch/model/mistral/transformer.py (981:1288) - megatron_patch/model/qwen1_5_megablocks/transformer.py (868:1175) duplicated block id: 46 size: 222 cleaned lines of code in 2 files: - megatron_patch/model/qwen1_5_megablocks/transformer.py (868:1175) - megatron_patch/model/qwen_vl/transformer.py (983:1290) duplicated block id: 47 size: 222 cleaned lines of code in 2 files: - megatron_patch/model/llava/transformer.py (983:1290) - megatron_patch/model/qwen1_5_megablocks/transformer.py (868:1175) duplicated block id: 48 size: 222 cleaned lines of code in 2 files: - megatron_patch/model/llama2/transformer.py (984:1291) - megatron_patch/model/qwen1_5_megablocks/transformer.py (868:1175) duplicated block id: 49 size: 215 cleaned lines of code in 2 files: - megatron_patch/model/mistral/language_model.py (31:360) - megatron_patch/model/qwen/language_model.py (30:359) duplicated block id: 50 size: 215 cleaned lines of code in 2 files: - megatron_patch/model/mistral/language_model.py (31:360) - megatron_patch/model/qwen1_5_megablocks/language_model.py (29:358) duplicated block id: 51 size: 215 cleaned lines of code in 2 files: - megatron_patch/model/llama2/language_model.py (30:359) - megatron_patch/model/mistral/language_model.py (31:360) duplicated block id: 52 size: 215 cleaned lines of code in 2 files: - megatron_patch/model/llava/language_model.py (33:362) - megatron_patch/model/qwen_vl/language_model.py (31:360) duplicated block id: 53 size: 211 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/language_model.py (56:379) - megatron_patch/model/llama2/language_model.py (46:369) duplicated block id: 54 size: 211 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/language_model.py (56:379) - megatron_patch/model/qwen1_5_megablocks/language_model.py (45:368) duplicated block id: 55 size: 203 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/language_model.py (56:369) - megatron_patch/model/mistral/language_model.py (47:360) duplicated block id: 56 size: 199 cleaned lines of code in 2 files: - megatron_patch/model/llava/language_model.py (33:344) - megatron_patch/model/mistral/language_model.py (31:342) duplicated block id: 57 size: 199 cleaned lines of code in 2 files: - megatron_patch/model/qwen1_5_megablocks/language_model.py (29:340) - megatron_patch/model/qwen_vl/language_model.py (31:342) duplicated block id: 58 size: 199 cleaned lines of code in 2 files: - megatron_patch/model/qwen/language_model.py (30:341) - megatron_patch/model/qwen_vl/language_model.py (31:342) duplicated block id: 59 size: 199 cleaned lines of code in 2 files: - megatron_patch/model/llama2/language_model.py (30:341) - megatron_patch/model/qwen_vl/language_model.py (31:342) duplicated block id: 60 size: 199 cleaned lines of code in 2 files: - megatron_patch/model/llava/language_model.py (33:344) - megatron_patch/model/qwen1_5_megablocks/language_model.py (29:340) duplicated block id: 61 size: 199 cleaned lines of code in 2 files: - megatron_patch/model/llava/language_model.py (33:344) - megatron_patch/model/qwen/language_model.py (30:341) duplicated block id: 62 size: 199 cleaned lines of code in 2 files: - megatron_patch/model/mistral/language_model.py (31:342) - megatron_patch/model/qwen_vl/language_model.py (31:342) duplicated block id: 63 size: 199 cleaned lines of code in 2 files: - megatron_patch/model/llama2/language_model.py (30:341) - megatron_patch/model/llava/language_model.py (33:344) duplicated block id: 64 size: 197 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (212:518) - megatron_patch/model/llama2/transformer.py (309:615) duplicated block id: 65 size: 196 cleaned lines of code in 2 files: - megatron_patch/model/qwen1_5_megablocks/transformer.py (1374:1646) - megatron_patch/model/qwen_vl/transformer.py (1510:1783) duplicated block id: 66 size: 196 cleaned lines of code in 2 files: - megatron_patch/model/mistral/transformer.py (1508:1781) - megatron_patch/model/qwen1_5_megablocks/transformer.py (1374:1646) duplicated block id: 67 size: 196 cleaned lines of code in 2 files: - megatron_patch/model/llava/transformer.py (1510:1783) - megatron_patch/model/qwen1_5_megablocks/transformer.py (1374:1646) duplicated block id: 68 size: 196 cleaned lines of code in 2 files: - megatron_patch/model/llama2/transformer.py (1511:1784) - megatron_patch/model/qwen1_5_megablocks/transformer.py (1374:1646) duplicated block id: 69 size: 195 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (34:313) - megatron_patch/model/mistral/transformer.py (39:318) duplicated block id: 70 size: 195 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (34:313) - megatron_patch/model/llama2/transformer.py (39:318) duplicated block id: 71 size: 195 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (34:313) - megatron_patch/model/llava/transformer.py (39:318) duplicated block id: 72 size: 195 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (34:313) - megatron_patch/model/qwen_vl/transformer.py (39:318) duplicated block id: 73 size: 190 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/transformer/attention.py (471:743) - megatron_patch/model/qwen2_vl/attention_vision.py (471:743) duplicated block id: 74 size: 190 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/transformer/attention.py (471:743) - megatron_patch/model/qwen2_vl/attention.py (472:744) duplicated block id: 75 size: 189 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/checkpoint_reshaping_and_interoperability.py (28:316) - toolkits/model_checkpoints_convertor/bloom/reward_model_to_megatron.py (28:316) duplicated block id: 76 size: 188 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (1574:1852) - megatron_patch/model/llama3/transformer_legacy.py (1534:1812) duplicated block id: 77 size: 188 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer_legacy.py (1534:1812) - megatron_patch/model/qwen/transformer.py (1516:1794) duplicated block id: 78 size: 188 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/glm/checkpoint_reshaping_and_interoperability.py (25:311) - toolkits/model_checkpoints_convertor/glm130b/checkpoint_reshaping_and_interoperability.py (25:311) duplicated block id: 79 size: 187 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/language_model.py (224:477) - megatron_patch/model/galactica/language_model.py (255:508) duplicated block id: 80 size: 187 cleaned lines of code in 2 files: - megatron_patch/model/galactica/language_model.py (255:508) - megatron_patch/model/llama/language_model.py (224:477) duplicated block id: 81 size: 187 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/language_model.py (56:351) - megatron_patch/model/llava/language_model.py (49:344) duplicated block id: 82 size: 187 cleaned lines of code in 2 files: - megatron_patch/model/falcon/language_model.py (224:477) - megatron_patch/model/galactica/language_model.py (255:508) duplicated block id: 83 size: 187 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/language_model.py (56:351) - megatron_patch/model/qwen_vl/language_model.py (47:342) duplicated block id: 84 size: 186 cleaned lines of code in 2 files: - megatron_patch/model/falcon/transformer.py (1034:1316) - megatron_patch/model/falcon40b/transformer.py (772:1055) duplicated block id: 85 size: 180 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2.5_vl.py (41:247) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_vl.py (41:247) duplicated block id: 86 size: 177 cleaned lines of code in 2 files: - megatron_patch/model/qwen/transformer.py (141:385) - megatron_patch/model/qwen_vl/transformer.py (147:391) duplicated block id: 87 size: 177 cleaned lines of code in 2 files: - megatron_patch/model/mistral/transformer.py (147:391) - megatron_patch/model/qwen/transformer.py (141:385) duplicated block id: 88 size: 177 cleaned lines of code in 2 files: - megatron_patch/model/llava/transformer.py (147:391) - megatron_patch/model/qwen/transformer.py (141:385) duplicated block id: 89 size: 174 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (517:780) - megatron_patch/model/qwen/transformer.py (608:873) duplicated block id: 90 size: 169 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer/attention.py (50:298) - megatron_patch/model/mixtral_bak/transformer/attention.py (24:272) duplicated block id: 91 size: 169 cleaned lines of code in 2 files: - megatron_patch/model/mixtral_bak/transformer/attention.py (24:272) - megatron_patch/model/qwen2/transformer/attention.py (46:294) duplicated block id: 92 size: 169 cleaned lines of code in 2 files: - megatron_patch/model/mixtral_bak/transformer/attention.py (24:272) - megatron_patch/model/qwen1_5/transformer/attention.py (50:298) duplicated block id: 93 size: 161 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (212:475) - megatron_patch/model/qwen/transformer.py (303:566) duplicated block id: 94 size: 160 cleaned lines of code in 2 files: - megatron_patch/model/qwen/transformer.py (306:566) - megatron_patch/model/qwen1_5_megablocks/transformer.py (203:463) duplicated block id: 95 size: 160 cleaned lines of code in 2 files: - megatron_patch/model/llama2/transformer.py (312:572) - megatron_patch/model/qwen1_5_megablocks/transformer.py (203:463) duplicated block id: 96 size: 160 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (215:475) - megatron_patch/model/qwen1_5_megablocks/transformer.py (203:463) duplicated block id: 97 size: 155 cleaned lines of code in 2 files: - megatron_patch/model/galactica/transformer.py (653:886) - megatron_patch/model/llama/transformer.py (819:1050) duplicated block id: 98 size: 154 cleaned lines of code in 2 files: - megatron_patch/model/qwen/transformer.py (929:1151) - megatron_patch/model/qwen1_5_megablocks/transformer.py (868:1090) duplicated block id: 99 size: 152 cleaned lines of code in 2 files: - megatron_patch/model/llama2/transformer.py (398:630) - megatron_patch/model/qwen_vl/transformer.py (399:631) duplicated block id: 100 size: 152 cleaned lines of code in 2 files: - megatron_patch/model/llama2/transformer.py (398:630) - megatron_patch/model/llava/transformer.py (399:631) duplicated block id: 101 size: 150 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (998:1220) - megatron_patch/model/qwen/transformer.py (941:1163) duplicated block id: 102 size: 149 cleaned lines of code in 2 files: - megatron_patch/model/llava/transformer.py (731:961) - megatron_patch/model/qwen1_5_megablocks/transformer.py (623:852) duplicated block id: 103 size: 149 cleaned lines of code in 2 files: - megatron_patch/model/qwen1_5_megablocks/transformer.py (623:852) - megatron_patch/model/qwen_vl/transformer.py (731:961) duplicated block id: 104 size: 148 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/checkpoint_reshaping_and_interoperability.py (597:795) - toolkits/model_checkpoints_convertor/bloom/reward_model_to_megatron.py (596:794) duplicated block id: 105 size: 147 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (35:195) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (40:200) duplicated block id: 106 size: 147 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/language_model.py (397:583) - megatron_patch/model/llama3/language_model.py (381:567) duplicated block id: 107 size: 147 cleaned lines of code in 2 files: - megatron_patch/model/llama3/language_model.py (381:567) - megatron_patch/model/qwen/language_model.py (387:572) duplicated block id: 108 size: 146 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (294:521) - megatron_patch/model/baichuan2/transformer.py (403:630) duplicated block id: 109 size: 144 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (403:627) - megatron_patch/model/llama2/transformer.py (391:615) duplicated block id: 110 size: 143 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (998:1208) - megatron_patch/model/llama2/transformer.py (996:1206) duplicated block id: 111 size: 143 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (410:627) - megatron_patch/model/qwen_vl/transformer.py (399:616) duplicated block id: 112 size: 143 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (301:518) - megatron_patch/model/mistral/transformer.py (399:616) duplicated block id: 113 size: 143 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (998:1208) - megatron_patch/model/mistral/transformer.py (993:1203) duplicated block id: 114 size: 143 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (998:1208) - megatron_patch/model/qwen_vl/transformer.py (995:1205) duplicated block id: 115 size: 143 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (410:627) - megatron_patch/model/llava/transformer.py (399:616) duplicated block id: 116 size: 143 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (301:518) - megatron_patch/model/qwen_vl/transformer.py (399:616) duplicated block id: 117 size: 143 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (301:518) - megatron_patch/model/llava/transformer.py (399:616) duplicated block id: 118 size: 143 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (998:1208) - megatron_patch/model/qwen1_5_megablocks/transformer.py (880:1090) duplicated block id: 119 size: 143 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (410:627) - megatron_patch/model/mistral/transformer.py (399:616) duplicated block id: 120 size: 143 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (998:1208) - megatron_patch/model/llava/transformer.py (995:1205) duplicated block id: 121 size: 143 cleaned lines of code in 2 files: - megatron_patch/model/llama2/transformer.py (398:615) - megatron_patch/model/mistral/transformer.py (399:616) duplicated block id: 122 size: 141 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/checkpoint_reshaping_and_interoperability.py (370:581) - toolkits/model_checkpoints_convertor/bloom/reward_model_to_megatron.py (369:580) duplicated block id: 123 size: 138 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (691:895) - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (664:866) duplicated block id: 124 size: 138 cleaned lines of code in 2 files: - megatron_patch/model/qwen2_5_vl/model.py (97:300) - megatron_patch/model/qwen2_vl/model.py (97:300) duplicated block id: 125 size: 137 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/hf2te.py (307:488) - toolkits/model_checkpoints_convertor/baichuan2/hf2te.py (279:460) duplicated block id: 126 size: 136 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (1518:1714) - megatron_patch/model/qwen/transformer.py (1590:1786) duplicated block id: 127 size: 136 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2.5_vl.py (553:700) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_vl.py (560:709) duplicated block id: 128 size: 136 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (1518:1714) - megatron_patch/model/baichuan2/transformer.py (1648:1844) duplicated block id: 129 size: 136 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (24:177) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (32:181) duplicated block id: 130 size: 136 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (1518:1714) - megatron_patch/model/llama3/transformer_legacy.py (1608:1804) duplicated block id: 131 size: 135 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (424:614) - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (397:586) duplicated block id: 132 size: 133 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/mlp.py (96:258) - megatron_patch/model/mixtral/transformer/mlp.py (104:266) duplicated block id: 133 size: 132 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer_legacy.py (1255:1445) - megatron_patch/model/qwen1_5_megablocks/transformer.py (1182:1372) duplicated block id: 134 size: 132 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (1297:1487) - megatron_patch/model/qwen1_5_megablocks/transformer.py (1182:1372) duplicated block id: 135 size: 132 cleaned lines of code in 2 files: - megatron_patch/model/mistral/transformer.py (1294:1484) - megatron_patch/model/qwen1_5_megablocks/transformer.py (1182:1372) duplicated block id: 136 size: 132 cleaned lines of code in 2 files: - megatron_patch/model/llama2/transformer.py (1297:1487) - megatron_patch/model/qwen1_5_megablocks/transformer.py (1182:1372) duplicated block id: 137 size: 132 cleaned lines of code in 2 files: - megatron_patch/model/qwen/transformer.py (1239:1429) - megatron_patch/model/qwen1_5_megablocks/transformer.py (1182:1372) duplicated block id: 138 size: 132 cleaned lines of code in 2 files: - megatron_patch/model/llava/transformer.py (1296:1486) - megatron_patch/model/qwen1_5_megablocks/transformer.py (1182:1372) duplicated block id: 139 size: 132 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (39:181) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (40:182) duplicated block id: 140 size: 132 cleaned lines of code in 2 files: - megatron_patch/model/qwen1_5_megablocks/transformer.py (1182:1372) - megatron_patch/model/qwen_vl/transformer.py (1296:1486) duplicated block id: 141 size: 131 cleaned lines of code in 2 files: - megatron_patch/model/llama2/transformer.py (298:525) - megatron_patch/model/llama3/transformer_legacy.py (297:524) duplicated block id: 142 size: 131 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/transformer.py (178:389) - megatron_patch/model/llama/transformer.py (166:377) duplicated block id: 143 size: 131 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer_legacy.py (297:524) - megatron_patch/model/qwen/transformer.py (292:519) duplicated block id: 144 size: 130 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (142:313) - megatron_patch/model/qwen/transformer.py (141:312) duplicated block id: 145 size: 129 cleaned lines of code in 2 files: - megatron_patch/model/llama3/model.py (34:251) - megatron_patch/model/qwen1_5/model.py (34:251) duplicated block id: 146 size: 127 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/transformer.py (729:930) - megatron_patch/model/galactica/transformer.py (733:938) duplicated block id: 147 size: 124 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (23:161) - toolkits/model_checkpoints_convertor/yi/checkpoint_reshaping_and_interoperability.py (24:161) duplicated block id: 148 size: 123 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/language_model.py (573:728) - megatron_patch/model/qwen/language_model.py (478:633) duplicated block id: 149 size: 123 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon/checkpoint_reshaping_and_interoperability.py (24:163) - toolkits/model_checkpoints_convertor/falcon40b/checkpoint_reshaping_and_interoperability.py (24:163) duplicated block id: 150 size: 123 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/language_model.py (573:728) - megatron_patch/model/baichuan2/language_model.py (489:644) duplicated block id: 151 size: 122 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (212:428) - megatron_patch/model/llama3/transformer_legacy.py (308:524) duplicated block id: 152 size: 121 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer_legacy.py (311:524) - megatron_patch/model/qwen1_5_megablocks/transformer.py (203:416) duplicated block id: 153 size: 119 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (23:155) - toolkits/model_checkpoints_convertor/falcon/checkpoint_reshaping_and_interoperability.py (24:158) duplicated block id: 154 size: 119 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon40b/checkpoint_reshaping_and_interoperability.py (24:158) - toolkits/model_checkpoints_convertor/yi/checkpoint_reshaping_and_interoperability.py (24:155) duplicated block id: 155 size: 119 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (23:155) - toolkits/model_checkpoints_convertor/falcon40b/checkpoint_reshaping_and_interoperability.py (24:158) duplicated block id: 156 size: 119 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon/checkpoint_reshaping_and_interoperability.py (24:158) - toolkits/model_checkpoints_convertor/yi/checkpoint_reshaping_and_interoperability.py (24:155) duplicated block id: 157 size: 118 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (23:153) - toolkits/model_checkpoints_convertor/galactica/checkpoint_reshaping_and_interoperability.py (23:155) duplicated block id: 158 size: 117 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (23:151) - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (23:151) duplicated block id: 159 size: 117 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (23:151) - toolkits/model_checkpoints_convertor/falcon40b/checkpoint_reshaping_and_interoperability.py (24:152) duplicated block id: 160 size: 117 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon/checkpoint_reshaping_and_interoperability.py (24:152) - toolkits/model_checkpoints_convertor/galactica/checkpoint_reshaping_and_interoperability.py (23:151) duplicated block id: 161 size: 117 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (23:151) - toolkits/model_checkpoints_convertor/yi/checkpoint_reshaping_and_interoperability.py (24:152) duplicated block id: 162 size: 117 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (23:151) - toolkits/model_checkpoints_convertor/galactica/checkpoint_reshaping_and_interoperability.py (23:151) duplicated block id: 163 size: 117 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon40b/checkpoint_reshaping_and_interoperability.py (24:152) - toolkits/model_checkpoints_convertor/galactica/checkpoint_reshaping_and_interoperability.py (23:151) duplicated block id: 164 size: 117 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (23:151) - toolkits/model_checkpoints_convertor/falcon/checkpoint_reshaping_and_interoperability.py (24:152) duplicated block id: 165 size: 117 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/galactica/checkpoint_reshaping_and_interoperability.py (23:151) - toolkits/model_checkpoints_convertor/yi/checkpoint_reshaping_and_interoperability.py (24:152) duplicated block id: 166 size: 115 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (32:154) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (61:183) duplicated block id: 167 size: 115 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (32:154) - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (56:178) duplicated block id: 168 size: 114 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/galactica/checkpoint_reshaping_and_interoperability.py (32:155) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (61:182) duplicated block id: 169 size: 114 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/galactica/checkpoint_reshaping_and_interoperability.py (32:155) - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (56:177) duplicated block id: 170 size: 114 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/galactica/checkpoint_reshaping_and_interoperability.py (32:155) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (60:181) duplicated block id: 171 size: 114 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (32:153) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (60:181) duplicated block id: 172 size: 113 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (56:175) - toolkits/model_checkpoints_convertor/yi/checkpoint_reshaping_and_interoperability.py (33:152) duplicated block id: 173 size: 113 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (32:151) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (60:179) duplicated block id: 174 size: 113 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (32:151) - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (56:175) duplicated block id: 175 size: 113 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_mha_to_moe.py (123:246) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_moe.py (272:395) duplicated block id: 176 size: 113 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (60:179) - toolkits/model_checkpoints_convertor/yi/checkpoint_reshaping_and_interoperability.py (33:152) duplicated block id: 177 size: 113 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon/checkpoint_reshaping_and_interoperability.py (33:152) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (60:179) duplicated block id: 178 size: 113 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon40b/checkpoint_reshaping_and_interoperability.py (33:152) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (61:180) duplicated block id: 179 size: 113 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon40b/checkpoint_reshaping_and_interoperability.py (33:152) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (60:179) duplicated block id: 180 size: 113 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore_llama3_1.py (647:797) - toolkits/model_checkpoints_convertor/llava/hf2mcore_llava.py (453:603) duplicated block id: 181 size: 113 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon40b/checkpoint_reshaping_and_interoperability.py (33:152) - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (56:175) duplicated block id: 182 size: 113 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (32:151) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (61:180) duplicated block id: 183 size: 113 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (61:180) - toolkits/model_checkpoints_convertor/yi/checkpoint_reshaping_and_interoperability.py (33:152) duplicated block id: 184 size: 113 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon/checkpoint_reshaping_and_interoperability.py (33:152) - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (56:175) duplicated block id: 185 size: 113 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon/checkpoint_reshaping_and_interoperability.py (33:152) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (61:180) duplicated block id: 186 size: 112 cleaned lines of code in 2 files: - megatron_patch/model/llama2/language_model.py (511:651) - megatron_patch/model/mistral/language_model.py (523:663) duplicated block id: 187 size: 112 cleaned lines of code in 2 files: - megatron_patch/model/mistral/language_model.py (523:663) - megatron_patch/model/qwen1_5_megablocks/language_model.py (510:650) duplicated block id: 188 size: 109 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/hf2te.py (23:162) - toolkits/model_checkpoints_convertor/baichuan2/hf2te.py (23:161) duplicated block id: 189 size: 108 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (403:584) - megatron_patch/model/qwen/transformer.py (385:566) duplicated block id: 190 size: 108 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (403:584) - megatron_patch/model/qwen1_5_megablocks/transformer.py (282:463) duplicated block id: 191 size: 108 cleaned lines of code in 2 files: - megatron_patch/model/qwen/transformer.py (1453:1594) - megatron_patch/model/qwen1_5_megablocks/transformer.py (1374:1515) duplicated block id: 192 size: 108 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (1511:1652) - megatron_patch/model/qwen1_5_megablocks/transformer.py (1374:1515) duplicated block id: 193 size: 107 cleaned lines of code in 2 files: - megatron_patch/model/mistral/transformer.py (399:573) - megatron_patch/model/qwen/transformer.py (392:566) duplicated block id: 194 size: 107 cleaned lines of code in 2 files: - megatron_patch/model/qwen/transformer.py (392:566) - megatron_patch/model/qwen_vl/transformer.py (399:573) duplicated block id: 195 size: 107 cleaned lines of code in 2 files: - megatron_patch/model/mistral/transformer.py (399:573) - megatron_patch/model/qwen1_5_megablocks/transformer.py (289:463) duplicated block id: 196 size: 107 cleaned lines of code in 2 files: - megatron_patch/model/llava/transformer.py (399:573) - megatron_patch/model/qwen/transformer.py (392:566) duplicated block id: 197 size: 107 cleaned lines of code in 2 files: - megatron_patch/model/qwen1_5_megablocks/transformer.py (289:463) - megatron_patch/model/qwen_vl/transformer.py (399:573) duplicated block id: 198 size: 107 cleaned lines of code in 2 files: - megatron_patch/model/llava/transformer.py (399:573) - megatron_patch/model/qwen1_5_megablocks/transformer.py (289:463) duplicated block id: 199 size: 106 cleaned lines of code in 2 files: - megatron_patch/model/mistral/language_model.py (529:663) - megatron_patch/model/qwen_vl/language_model.py (546:680) duplicated block id: 200 size: 106 cleaned lines of code in 2 files: - megatron_patch/model/llama2/language_model.py (517:651) - megatron_patch/model/qwen_vl/language_model.py (546:680) duplicated block id: 201 size: 106 cleaned lines of code in 2 files: - megatron_patch/model/qwen1_5_megablocks/language_model.py (516:650) - megatron_patch/model/qwen_vl/language_model.py (546:680) duplicated block id: 202 size: 105 cleaned lines of code in 2 files: - megatron_patch/model/falcon/transformer.py (1034:1182) - megatron_patch/model/galactica/transformer.py (661:810) duplicated block id: 203 size: 105 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/transformer.py (772:921) - megatron_patch/model/llama/transformer.py (827:974) duplicated block id: 204 size: 105 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/transformer.py (772:921) - megatron_patch/model/galactica/transformer.py (661:810) duplicated block id: 205 size: 105 cleaned lines of code in 2 files: - megatron_patch/model/falcon/transformer.py (1034:1182) - megatron_patch/model/llama/transformer.py (827:974) duplicated block id: 206 size: 104 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (477:609) - megatron_patch/model/llama3/transformer_legacy.py (573:706) duplicated block id: 207 size: 102 cleaned lines of code in 2 files: - megatron_patch/model/llava/language_model.py (581:711) - megatron_patch/model/qwen_vl/language_model.py (550:680) duplicated block id: 208 size: 102 cleaned lines of code in 2 files: - megatron_patch/model/llama2/language_model.py (521:651) - megatron_patch/model/llava/language_model.py (581:711) duplicated block id: 209 size: 102 cleaned lines of code in 2 files: - megatron_patch/model/llava/language_model.py (581:711) - megatron_patch/model/qwen1_5_megablocks/language_model.py (520:650) duplicated block id: 210 size: 102 cleaned lines of code in 2 files: - megatron_patch/model/llava/language_model.py (581:711) - megatron_patch/model/mistral/language_model.py (533:663) duplicated block id: 211 size: 101 cleaned lines of code in 2 files: - megatron_patch/model/llama2/language_model.py (522:651) - megatron_patch/model/qwen/language_model.py (504:633) duplicated block id: 212 size: 101 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/language_model.py (599:728) - megatron_patch/model/llama2/language_model.py (522:651) duplicated block id: 213 size: 101 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/language_model.py (515:644) - megatron_patch/model/llava/language_model.py (582:711) duplicated block id: 214 size: 101 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (205:402) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (224:421) duplicated block id: 215 size: 101 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/language_model.py (599:728) - megatron_patch/model/llava/language_model.py (582:711) duplicated block id: 216 size: 101 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/language_model.py (515:644) - megatron_patch/model/mistral/language_model.py (534:663) duplicated block id: 217 size: 101 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/language_model.py (515:644) - megatron_patch/model/qwen1_5_megablocks/language_model.py (521:650) duplicated block id: 218 size: 101 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/language_model.py (599:728) - megatron_patch/model/qwen_vl/language_model.py (551:680) duplicated block id: 219 size: 101 cleaned lines of code in 2 files: - megatron_patch/model/llava/language_model.py (582:711) - megatron_patch/model/qwen/language_model.py (504:633) duplicated block id: 220 size: 101 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/language_model.py (599:728) - megatron_patch/model/qwen1_5_megablocks/language_model.py (521:650) duplicated block id: 221 size: 101 cleaned lines of code in 2 files: - megatron_patch/model/qwen/language_model.py (504:633) - megatron_patch/model/qwen_vl/language_model.py (551:680) duplicated block id: 222 size: 101 cleaned lines of code in 2 files: - megatron_patch/model/mistral/language_model.py (534:663) - megatron_patch/model/qwen/language_model.py (504:633) duplicated block id: 223 size: 101 cleaned lines of code in 2 files: - megatron_patch/model/qwen/language_model.py (504:633) - megatron_patch/model/qwen1_5_megablocks/language_model.py (521:650) duplicated block id: 224 size: 101 cleaned lines of code in 2 files: - megatron_patch/model/galactica/transformer.py (208:362) - megatron_patch/model/llama/transformer.py (246:400) duplicated block id: 225 size: 101 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/language_model.py (515:644) - megatron_patch/model/llama2/language_model.py (522:651) duplicated block id: 226 size: 101 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/language_model.py (599:728) - megatron_patch/model/mistral/language_model.py (534:663) duplicated block id: 227 size: 101 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/language_model.py (515:644) - megatron_patch/model/qwen_vl/language_model.py (551:680) duplicated block id: 228 size: 99 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/transformer.py (729:878) - megatron_patch/model/llama/transformer.py (899:1050) duplicated block id: 229 size: 99 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer/mlp.py (94:211) - megatron_patch/model/qwen1_5/transformer/mlp.py (110:227) duplicated block id: 230 size: 98 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/checkpoint_reshaping_and_interoperability.py (28:134) - toolkits/model_checkpoints_convertor/starcoder/checkpoint_reshaping_and_interoperability.py (28:134) duplicated block id: 231 size: 98 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (155:330) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (134:308) duplicated block id: 232 size: 98 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (214:408) - toolkits/model_checkpoints_convertor/yi/checkpoint_reshaping_and_interoperability.py (163:357) duplicated block id: 233 size: 98 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/reward_model_to_megatron.py (28:134) - toolkits/model_checkpoints_convertor/starcoder/checkpoint_reshaping_and_interoperability.py (28:134) duplicated block id: 234 size: 97 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (560:709) - megatron_patch/model/baichuan2/transformer.py (669:820) duplicated block id: 235 size: 97 cleaned lines of code in 2 files: - megatron_patch/model/galactica/language_model.py (597:725) - megatron_patch/model/llama/language_model.py (579:707) duplicated block id: 236 size: 97 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (395:526) - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (476:607) duplicated block id: 237 size: 97 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (669:820) - megatron_patch/model/qwen/transformer.py (651:802) duplicated block id: 238 size: 96 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/language_model.py (281:415) - megatron_patch/model/mistral/language_model.py (226:360) duplicated block id: 239 size: 96 cleaned lines of code in 2 files: - megatron_patch/model/llama3/language_model.py (19:179) - megatron_patch/model/llava/language_model.py (33:193) duplicated block id: 240 size: 96 cleaned lines of code in 2 files: - megatron_patch/model/llama2/language_model.py (30:190) - megatron_patch/model/llama3/language_model.py (19:179) duplicated block id: 241 size: 96 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/language_model.py (281:415) - megatron_patch/model/llama2/language_model.py (225:359) duplicated block id: 242 size: 96 cleaned lines of code in 2 files: - megatron_patch/model/llama3/language_model.py (19:179) - megatron_patch/model/qwen1_5_megablocks/language_model.py (29:189) duplicated block id: 243 size: 96 cleaned lines of code in 2 files: - megatron_patch/model/llama3/language_model.py (19:179) - megatron_patch/model/qwen/language_model.py (30:190) duplicated block id: 244 size: 96 cleaned lines of code in 2 files: - megatron_patch/model/llama3/language_model.py (19:179) - megatron_patch/model/qwen_vl/language_model.py (31:191) duplicated block id: 245 size: 96 cleaned lines of code in 2 files: - megatron_patch/model/llama3/language_model.py (19:179) - megatron_patch/model/mistral/language_model.py (31:191) duplicated block id: 246 size: 96 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/language_model.py (281:415) - megatron_patch/model/qwen/language_model.py (225:359) duplicated block id: 247 size: 96 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/language_model.py (281:415) - megatron_patch/model/qwen1_5_megablocks/language_model.py (224:358) duplicated block id: 248 size: 96 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/language_model.py (281:415) - megatron_patch/model/baichuan2/language_model.py (235:369) duplicated block id: 249 size: 95 cleaned lines of code in 2 files: - megatron_patch/model/mistral/language_model.py (391:510) - megatron_patch/model/qwen1_5_megablocks/language_model.py (388:507) duplicated block id: 250 size: 95 cleaned lines of code in 2 files: - megatron_patch/model/qwen1_5/moe/experts.py (62:176) - megatron_patch/model/qwen2/moe/experts.py (85:200) duplicated block id: 251 size: 95 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/transformer/attention.py (228:352) - megatron_patch/model/qwen2_vl/attention.py (218:342) duplicated block id: 252 size: 95 cleaned lines of code in 2 files: - megatron_patch/model/llama2/language_model.py (389:508) - megatron_patch/model/mistral/language_model.py (391:510) duplicated block id: 253 size: 95 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (424:554) - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (478:607) duplicated block id: 254 size: 95 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/transformer/attention.py (228:352) - megatron_patch/model/qwen2_vl/attention_vision.py (217:341) duplicated block id: 255 size: 94 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (233:421) - toolkits/model_checkpoints_convertor/yi/checkpoint_reshaping_and_interoperability.py (163:351) duplicated block id: 256 size: 94 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (483:610) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (490:617) duplicated block id: 257 size: 94 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/galactica/checkpoint_reshaping_and_interoperability.py (169:360) - toolkits/model_checkpoints_convertor/yi/checkpoint_reshaping_and_interoperability.py (168:358) duplicated block id: 258 size: 93 cleaned lines of code in 2 files: - megatron_patch/model/falcon/transformer.py (37:185) - megatron_patch/model/falcon40b/transformer.py (33:178) duplicated block id: 259 size: 93 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/galactica/checkpoint_reshaping_and_interoperability.py (169:359) - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (219:408) duplicated block id: 260 size: 92 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (402:526) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (490:614) duplicated block id: 261 size: 92 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (430:554) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (490:614) duplicated block id: 262 size: 91 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/transformer/mlp.py (156:266) - megatron_patch/model/qwen2/transformer/mlp.py (230:339) duplicated block id: 263 size: 91 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (1169:1305) - megatron_patch/model/mistral/transformer.py (1299:1435) duplicated block id: 264 size: 91 cleaned lines of code in 2 files: - megatron_patch/model/qwen1_5/model.py (98:251) - megatron_patch/model/qwen2/model.py (99:252) duplicated block id: 265 size: 91 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (1169:1305) - megatron_patch/model/baichuan2/transformer.py (1302:1438) duplicated block id: 266 size: 91 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon/checkpoint_reshaping_and_interoperability.py (530:652) - toolkits/model_checkpoints_convertor/falcon40b/checkpoint_reshaping_and_interoperability.py (532:655) duplicated block id: 267 size: 91 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (1169:1305) - megatron_patch/model/qwen_vl/transformer.py (1301:1437) duplicated block id: 268 size: 91 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/mlp.py (148:258) - megatron_patch/model/qwen2/transformer/mlp.py (230:339) duplicated block id: 269 size: 91 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (1169:1305) - megatron_patch/model/qwen1_5_megablocks/transformer.py (1187:1323) duplicated block id: 270 size: 91 cleaned lines of code in 2 files: - megatron_patch/model/llama3/model.py (98:251) - megatron_patch/model/qwen2/model.py (99:252) duplicated block id: 271 size: 91 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (1169:1305) - megatron_patch/model/llama2/transformer.py (1302:1438) duplicated block id: 272 size: 91 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (1169:1305) - megatron_patch/model/llama3/transformer_legacy.py (1260:1396) duplicated block id: 273 size: 91 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon/checkpoint_reshaping_and_interoperability.py (395:515) - toolkits/model_checkpoints_convertor/falcon40b/checkpoint_reshaping_and_interoperability.py (384:504) duplicated block id: 274 size: 91 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (1169:1305) - megatron_patch/model/llava/transformer.py (1301:1437) duplicated block id: 275 size: 91 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (1169:1305) - megatron_patch/model/qwen/transformer.py (1244:1380) duplicated block id: 276 size: 90 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (825:957) - megatron_patch/model/qwen_vl/transformer.py (861:993) duplicated block id: 277 size: 90 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (825:957) - megatron_patch/model/llama2/transformer.py (862:994) duplicated block id: 278 size: 90 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (825:957) - megatron_patch/model/llava/transformer.py (861:993) duplicated block id: 279 size: 90 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/checkpoint_reshaping_and_interoperability.py (154:330) - toolkits/model_checkpoints_convertor/starcoder/checkpoint_reshaping_and_interoperability.py (176:352) duplicated block id: 280 size: 90 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (825:957) - megatron_patch/model/qwen/transformer.py (807:939) duplicated block id: 281 size: 90 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (825:957) - megatron_patch/model/mistral/transformer.py (859:991) duplicated block id: 282 size: 89 cleaned lines of code in 2 files: - megatron_patch/model/llama3/language_model.py (242:364) - megatron_patch/model/qwen/language_model.py (247:369) duplicated block id: 283 size: 89 cleaned lines of code in 2 files: - megatron_patch/model/mistral/transformer.py (637:763) - megatron_patch/model/qwen1_5_megablocks/transformer.py (529:656) duplicated block id: 284 size: 89 cleaned lines of code in 2 files: - megatron_patch/model/llama2/transformer.py (638:765) - megatron_patch/model/mistral/transformer.py (637:763) duplicated block id: 285 size: 89 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/transformer_block.py (112:268) - megatron_patch/model/qwen2_5_vl/transformer_block.py (151:307) duplicated block id: 286 size: 89 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/galactica/checkpoint_reshaping_and_interoperability.py (169:353) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (238:421) duplicated block id: 287 size: 89 cleaned lines of code in 2 files: - megatron_patch/model/llama2/language_model.py (247:369) - megatron_patch/model/llama3/language_model.py (242:364) duplicated block id: 288 size: 89 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/language_model.py (257:379) - megatron_patch/model/llama3/language_model.py (242:364) duplicated block id: 289 size: 89 cleaned lines of code in 2 files: - megatron_patch/model/llama3/language_model.py (242:364) - megatron_patch/model/qwen1_5_megablocks/language_model.py (246:368) duplicated block id: 290 size: 88 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore.py (10:125) - toolkits/model_checkpoints_convertor/qwen/hf2megablocks_qwen1.5.py (8:125) duplicated block id: 291 size: 87 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer_legacy.py (934:1075) - megatron_patch/model/qwen_vl/transformer.py (983:1124) duplicated block id: 292 size: 87 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer_legacy.py (934:1075) - megatron_patch/model/qwen/transformer.py (929:1070) duplicated block id: 293 size: 87 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/language_model.py (561:675) - megatron_patch/model/llama/language_model.py (572:686) duplicated block id: 294 size: 87 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer_legacy.py (934:1075) - megatron_patch/model/qwen1_5_megablocks/transformer.py (868:1009) duplicated block id: 295 size: 87 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer_legacy.py (934:1075) - megatron_patch/model/llava/transformer.py (983:1124) duplicated block id: 296 size: 87 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer_legacy.py (934:1075) - megatron_patch/model/mistral/transformer.py (981:1122) duplicated block id: 297 size: 87 cleaned lines of code in 2 files: - megatron_patch/model/llama2/transformer.py (984:1125) - megatron_patch/model/llama3/transformer_legacy.py (934:1075) duplicated block id: 298 size: 87 cleaned lines of code in 2 files: - megatron_patch/model/falcon/language_model.py (561:675) - megatron_patch/model/llama/language_model.py (572:686) duplicated block id: 299 size: 86 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (486:599) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (463:576) duplicated block id: 300 size: 86 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (433:546) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (463:576) duplicated block id: 301 size: 86 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (405:518) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (463:576) duplicated block id: 302 size: 86 cleaned lines of code in 2 files: - megatron_patch/model/bloom/language_model.py (234:359) - megatron_patch/model/llama/language_model.py (217:341) duplicated block id: 303 size: 86 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (463:576) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (493:606) duplicated block id: 304 size: 86 cleaned lines of code in 2 files: - megatron_patch/model/bloom/language_model.py (234:359) - megatron_patch/model/falcon40b/language_model.py (217:341) duplicated block id: 305 size: 86 cleaned lines of code in 2 files: - megatron_patch/model/bloom/language_model.py (234:359) - megatron_patch/model/falcon/language_model.py (217:341) duplicated block id: 306 size: 86 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/transformer.py (124:251) - megatron_patch/model/galactica/transformer.py (73:201) duplicated block id: 307 size: 85 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/gpt_model.py (28:135) - megatron_patch/model/galactica/gpt_model.py (28:135) duplicated block id: 308 size: 85 cleaned lines of code in 2 files: - megatron_patch/model/falcon/gpt_model.py (28:135) - megatron_patch/model/galactica/gpt_model.py (28:135) duplicated block id: 309 size: 85 cleaned lines of code in 2 files: - megatron_patch/model/falcon/gpt_model.py (28:135) - megatron_patch/model/falcon40b/gpt_model.py (28:135) duplicated block id: 310 size: 84 cleaned lines of code in 2 files: - megatron_patch/model/llama2/language_model.py (394:496) - megatron_patch/model/qwen/language_model.py (391:492) duplicated block id: 311 size: 84 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/language_model.py (401:503) - megatron_patch/model/mistral/language_model.py (396:498) duplicated block id: 312 size: 84 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/language_model.py (401:503) - megatron_patch/model/llama2/language_model.py (394:496) duplicated block id: 313 size: 84 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/language_model.py (401:503) - megatron_patch/model/qwen1_5_megablocks/language_model.py (393:495) duplicated block id: 314 size: 84 cleaned lines of code in 2 files: - megatron_patch/model/qwen/language_model.py (391:492) - megatron_patch/model/qwen1_5_megablocks/language_model.py (393:495) duplicated block id: 315 size: 84 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (336:449) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (323:436) duplicated block id: 316 size: 84 cleaned lines of code in 2 files: - megatron_patch/model/llama3/language_model.py (385:487) - megatron_patch/model/mistral/language_model.py (396:498) duplicated block id: 317 size: 84 cleaned lines of code in 2 files: - megatron_patch/model/llama2/language_model.py (394:496) - megatron_patch/model/llama3/language_model.py (385:487) duplicated block id: 318 size: 84 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/language_model.py (56:200) - megatron_patch/model/llama3/language_model.py (35:179) duplicated block id: 319 size: 84 cleaned lines of code in 2 files: - megatron_patch/model/mistral/language_model.py (396:498) - megatron_patch/model/qwen/language_model.py (391:492) duplicated block id: 320 size: 84 cleaned lines of code in 2 files: - megatron_patch/model/llava/language_model.py (376:480) - megatron_patch/model/qwen_vl/language_model.py (368:472) duplicated block id: 321 size: 84 cleaned lines of code in 2 files: - megatron_patch/model/llama3/language_model.py (385:487) - megatron_patch/model/qwen1_5_megablocks/language_model.py (393:495) duplicated block id: 322 size: 83 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon/checkpoint_reshaping_and_interoperability.py (672:787) - toolkits/model_checkpoints_convertor/falcon40b/checkpoint_reshaping_and_interoperability.py (657:771) duplicated block id: 323 size: 83 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/transformer/attention.py (23:152) - megatron_patch/model/qwen2_vl/attention_vision.py (11:138) duplicated block id: 324 size: 82 cleaned lines of code in 2 files: - megatron_patch/model/llava/gpt_model.py (26:133) - megatron_patch/model/qwen_vl/gpt_model.py (26:133) duplicated block id: 325 size: 81 cleaned lines of code in 2 files: - megatron_patch/model/llama3/gpt_model.py (28:134) - megatron_patch/model/qwen1_5_megablocks/gpt_model.py (26:132) duplicated block id: 326 size: 81 cleaned lines of code in 2 files: - megatron_patch/model/mistral/transformer.py (836:959) - megatron_patch/model/qwen1_5_megablocks/transformer.py (730:852) duplicated block id: 327 size: 81 cleaned lines of code in 2 files: - megatron_patch/model/llava/transformer.py (39:168) - megatron_patch/model/qwen1_5_megablocks/transformer.py (37:166) duplicated block id: 328 size: 81 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/transformer.py (258:389) - megatron_patch/model/galactica/transformer.py (208:339) duplicated block id: 329 size: 81 cleaned lines of code in 2 files: - megatron_patch/model/qwen/gpt_model.py (26:132) - megatron_patch/model/qwen1_5_megablocks/gpt_model.py (26:132) duplicated block id: 330 size: 81 cleaned lines of code in 2 files: - megatron_patch/model/llama3/language_model.py (242:354) - megatron_patch/model/mistral/language_model.py (248:360) duplicated block id: 331 size: 81 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (34:163) - megatron_patch/model/qwen1_5_megablocks/transformer.py (37:166) duplicated block id: 332 size: 81 cleaned lines of code in 2 files: - megatron_patch/model/llama2/gpt_model.py (26:132) - megatron_patch/model/llama3/gpt_model.py (28:134) duplicated block id: 333 size: 81 cleaned lines of code in 2 files: - megatron_patch/model/llama2/gpt_model.py (26:132) - megatron_patch/model/qwen/gpt_model.py (26:132) duplicated block id: 334 size: 81 cleaned lines of code in 2 files: - megatron_patch/model/llama2/gpt_model.py (26:132) - megatron_patch/model/mistral/gpt_model.py (26:132) duplicated block id: 335 size: 81 cleaned lines of code in 2 files: - megatron_patch/model/llama2/gpt_model.py (26:132) - megatron_patch/model/qwen1_5_megablocks/gpt_model.py (26:132) duplicated block id: 336 size: 81 cleaned lines of code in 2 files: - megatron_patch/model/llama2/transformer.py (39:168) - megatron_patch/model/qwen1_5_megablocks/transformer.py (37:166) duplicated block id: 337 size: 81 cleaned lines of code in 2 files: - megatron_patch/model/llama3/gpt_model.py (28:134) - megatron_patch/model/qwen/gpt_model.py (26:132) duplicated block id: 338 size: 81 cleaned lines of code in 2 files: - megatron_patch/model/llama3/gpt_model.py (28:134) - megatron_patch/model/mistral/gpt_model.py (26:132) duplicated block id: 339 size: 81 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/language_model.py (303:415) - megatron_patch/model/llama3/language_model.py (242:354) duplicated block id: 340 size: 81 cleaned lines of code in 2 files: - megatron_patch/model/mistral/transformer.py (39:168) - megatron_patch/model/qwen1_5_megablocks/transformer.py (37:166) duplicated block id: 341 size: 81 cleaned lines of code in 2 files: - megatron_patch/model/qwen1_5_megablocks/transformer.py (37:166) - megatron_patch/model/qwen_vl/transformer.py (39:168) duplicated block id: 342 size: 81 cleaned lines of code in 2 files: - megatron_patch/model/mistral/gpt_model.py (26:132) - megatron_patch/model/qwen1_5_megablocks/gpt_model.py (26:132) duplicated block id: 343 size: 81 cleaned lines of code in 2 files: - megatron_patch/model/mistral/gpt_model.py (26:132) - megatron_patch/model/qwen/gpt_model.py (26:132) duplicated block id: 344 size: 80 cleaned lines of code in 2 files: - megatron_patch/model/falcon/language_model.py (568:675) - megatron_patch/model/galactica/language_model.py (597:704) duplicated block id: 345 size: 80 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/gpt_model.py (53:158) - megatron_patch/model/qwen/gpt_model.py (27:132) duplicated block id: 346 size: 80 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/gpt_model.py (53:158) - megatron_patch/model/llama2/gpt_model.py (27:132) duplicated block id: 347 size: 80 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/gpt_model.py (53:158) - megatron_patch/model/mistral/gpt_model.py (27:132) duplicated block id: 348 size: 80 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/language_model.py (281:397) - megatron_patch/model/qwen_vl/language_model.py (226:342) duplicated block id: 349 size: 80 cleaned lines of code in 2 files: - megatron_patch/model/bloom/language_model.py (242:359) - megatron_patch/model/galactica/language_model.py (255:372) duplicated block id: 350 size: 80 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/language_model.py (281:397) - megatron_patch/model/llava/language_model.py (228:344) duplicated block id: 351 size: 80 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/language_model.py (568:675) - megatron_patch/model/galactica/language_model.py (597:704) duplicated block id: 352 size: 80 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/gpt_model.py (53:158) - megatron_patch/model/qwen1_5_megablocks/gpt_model.py (27:132) duplicated block id: 353 size: 80 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/gpt_model.py (53:158) - megatron_patch/model/llama3/gpt_model.py (29:134) duplicated block id: 354 size: 79 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (687:802) - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (800:910) duplicated block id: 355 size: 79 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer/mlp.py (118:211) - megatron_patch/model/qwen2/transformer/mlp.py (137:230) duplicated block id: 356 size: 79 cleaned lines of code in 2 files: - megatron_patch/model/qwen1_5/transformer/mlp.py (134:227) - megatron_patch/model/qwen2/transformer/mlp.py (137:230) duplicated block id: 357 size: 77 cleaned lines of code in 2 files: - megatron_patch/model/mixtral_bak/moe/experts.py (55:143) - megatron_patch/model/qwen1_5/moe/experts.py (57:145) duplicated block id: 358 size: 76 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (998:1127) - megatron_patch/model/llama3/transformer_legacy.py (946:1075) duplicated block id: 359 size: 76 cleaned lines of code in 2 files: - megatron_patch/model/qwen/transformer.py (791:907) - megatron_patch/model/qwen1_5_megablocks/transformer.py (737:852) duplicated block id: 360 size: 76 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (668:764) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (666:761) duplicated block id: 361 size: 76 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/reward_model_to_megatron.py (154:316) - toolkits/model_checkpoints_convertor/starcoder/checkpoint_reshaping_and_interoperability.py (176:338) duplicated block id: 362 size: 75 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer_legacy.py (526:614) - megatron_patch/model/qwen1_5_megablocks/transformer.py (418:506) duplicated block id: 363 size: 75 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/glm/checkpoint_reshaping_and_interoperability.py (25:102) - toolkits/model_checkpoints_convertor/starcoder/checkpoint_reshaping_and_interoperability.py (28:105) duplicated block id: 364 size: 75 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/reward_model_to_megatron.py (28:105) - toolkits/model_checkpoints_convertor/glm/checkpoint_reshaping_and_interoperability.py (25:102) duplicated block id: 365 size: 75 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/reward_model_to_megatron.py (28:105) - toolkits/model_checkpoints_convertor/glm130b/checkpoint_reshaping_and_interoperability.py (25:102) duplicated block id: 366 size: 75 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/glm130b/checkpoint_reshaping_and_interoperability.py (25:102) - toolkits/model_checkpoints_convertor/starcoder/checkpoint_reshaping_and_interoperability.py (28:105) duplicated block id: 367 size: 75 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/language_model.py (573:667) - megatron_patch/model/llama3/language_model.py (473:567) duplicated block id: 368 size: 75 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/moe/moe_layer.py (14:126) - megatron_patch/model/mixtral/moe/moe_layer.py (17:125) duplicated block id: 369 size: 75 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/checkpoint_reshaping_and_interoperability.py (28:105) - toolkits/model_checkpoints_convertor/glm/checkpoint_reshaping_and_interoperability.py (25:102) duplicated block id: 370 size: 75 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/checkpoint_reshaping_and_interoperability.py (28:105) - toolkits/model_checkpoints_convertor/glm130b/checkpoint_reshaping_and_interoperability.py (25:102) duplicated block id: 371 size: 74 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2.5_vl.py (394:484) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_vl.py (392:481) duplicated block id: 372 size: 74 cleaned lines of code in 2 files: - megatron_patch/model/mixtral_bak/moe/experts.py (60:143) - megatron_patch/model/qwen2/moe/experts.py (85:169) duplicated block id: 373 size: 73 cleaned lines of code in 2 files: - megatron_patch/model/llama3/layer_specs.py (20:110) - megatron_patch/model/llama3_1/layer_specs.py (20:110) duplicated block id: 374 size: 72 cleaned lines of code in 2 files: - megatron_patch/model/llama2/transformer.py (634:728) - megatron_patch/model/qwen_vl/transformer.py (633:727) duplicated block id: 375 size: 72 cleaned lines of code in 2 files: - megatron_patch/model/qwen1_5_megablocks/transformer.py (525:619) - megatron_patch/model/qwen_vl/transformer.py (633:727) duplicated block id: 376 size: 72 cleaned lines of code in 2 files: - megatron_patch/model/llava/transformer.py (633:727) - megatron_patch/model/qwen1_5_megablocks/transformer.py (525:619) duplicated block id: 377 size: 72 cleaned lines of code in 2 files: - megatron_patch/model/llama2/transformer.py (634:728) - megatron_patch/model/llava/transformer.py (633:727) duplicated block id: 378 size: 72 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (395:491) - toolkits/model_checkpoints_convertor/yi/checkpoint_reshaping_and_interoperability.py (398:494) duplicated block id: 379 size: 72 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (476:572) - toolkits/model_checkpoints_convertor/yi/checkpoint_reshaping_and_interoperability.py (398:494) duplicated block id: 380 size: 71 cleaned lines of code in 2 files: - megatron_patch/model/galactica/language_model.py (29:134) - megatron_patch/model/llama/language_model.py (29:134) duplicated block id: 381 size: 71 cleaned lines of code in 2 files: - megatron_patch/model/falcon/language_model.py (29:134) - megatron_patch/model/galactica/language_model.py (29:134) duplicated block id: 382 size: 71 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/language_model.py (29:134) - megatron_patch/model/galactica/language_model.py (29:134) duplicated block id: 383 size: 71 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer_legacy.py (613:706) - megatron_patch/model/qwen/transformer.py (608:701) duplicated block id: 384 size: 71 cleaned lines of code in 2 files: - megatron_patch/model/qwen1_5/transformer/mlp.py (24:132) - megatron_patch/model/qwen2/transformer/mlp.py (23:131) duplicated block id: 385 size: 70 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (424:519) - toolkits/model_checkpoints_convertor/yi/checkpoint_reshaping_and_interoperability.py (400:494) duplicated block id: 386 size: 70 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (283:389) - megatron_patch/model/chatglm/transformer.py (224:330) duplicated block id: 387 size: 69 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/language_model.py (75:198) - megatron_patch/model/mistral/language_model.py (31:150) duplicated block id: 388 size: 69 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (403:537) - megatron_patch/model/llama3/transformer_legacy.py (390:524) duplicated block id: 389 size: 69 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/language_model.py (75:198) - megatron_patch/model/qwen/language_model.py (30:149) duplicated block id: 390 size: 69 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (726:838) - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (840:954) duplicated block id: 391 size: 69 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/language_model.py (75:198) - megatron_patch/model/llava/language_model.py (33:152) duplicated block id: 392 size: 69 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/language_model.py (75:198) - megatron_patch/model/llama2/language_model.py (30:149) duplicated block id: 393 size: 69 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/gpt_model.py (50:135) - megatron_patch/model/llama/gpt_model.py (49:135) duplicated block id: 394 size: 69 cleaned lines of code in 2 files: - megatron_patch/model/mistral/transformer.py (637:727) - megatron_patch/model/qwen_vl/transformer.py (637:727) duplicated block id: 395 size: 69 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/language_model.py (75:198) - megatron_patch/model/llama3/language_model.py (19:138) duplicated block id: 396 size: 69 cleaned lines of code in 2 files: - megatron_patch/model/llava/transformer.py (637:727) - megatron_patch/model/mistral/transformer.py (637:727) duplicated block id: 397 size: 69 cleaned lines of code in 2 files: - megatron_patch/model/falcon/gpt_model.py (50:135) - megatron_patch/model/llama/gpt_model.py (49:135) duplicated block id: 398 size: 69 cleaned lines of code in 2 files: - megatron_patch/model/galactica/gpt_model.py (50:135) - megatron_patch/model/llama/gpt_model.py (49:135) duplicated block id: 399 size: 69 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/language_model.py (75:198) - megatron_patch/model/qwen_vl/language_model.py (31:150) duplicated block id: 400 size: 69 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/language_model.py (75:198) - megatron_patch/model/qwen1_5_megablocks/language_model.py (29:148) duplicated block id: 401 size: 69 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (699:809) - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (840:954) duplicated block id: 402 size: 68 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (423:512) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (341:430) duplicated block id: 403 size: 68 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer_legacy.py (397:524) - megatron_patch/model/qwen_vl/transformer.py (399:526) duplicated block id: 404 size: 68 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer_legacy.py (397:524) - megatron_patch/model/mistral/transformer.py (399:526) duplicated block id: 405 size: 68 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (354:443) - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (423:512) duplicated block id: 406 size: 68 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer_legacy.py (397:524) - megatron_patch/model/llava/transformer.py (399:526) duplicated block id: 407 size: 67 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (490:579) - toolkits/model_checkpoints_convertor/yi/checkpoint_reshaping_and_interoperability.py (405:494) duplicated block id: 408 size: 67 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/galactica/checkpoint_reshaping_and_interoperability.py (504:593) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (526:614) duplicated block id: 409 size: 66 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (463:549) - toolkits/model_checkpoints_convertor/yi/checkpoint_reshaping_and_interoperability.py (408:494) duplicated block id: 410 size: 66 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/language_model.py (435:520) - megatron_patch/model/qwen/language_model.py (374:459) duplicated block id: 411 size: 66 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/language_model.py (435:520) - megatron_patch/model/baichuan2/language_model.py (384:469) duplicated block id: 412 size: 65 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (825:925) - megatron_patch/model/qwen1_5_megablocks/transformer.py (753:852) duplicated block id: 413 size: 65 cleaned lines of code in 2 files: - megatron_patch/model/llama3/language_model.py (242:336) - megatron_patch/model/llava/language_model.py (250:344) duplicated block id: 414 size: 65 cleaned lines of code in 2 files: - megatron_patch/model/llama3/language_model.py (242:336) - megatron_patch/model/qwen_vl/language_model.py (248:342) duplicated block id: 415 size: 65 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer_legacy.py (43:150) - megatron_patch/model/mistral/transformer.py (39:146) duplicated block id: 416 size: 65 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/language_model.py (262:356) - megatron_patch/model/falcon40b/language_model.py (246:340) duplicated block id: 417 size: 65 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/language_model.py (262:356) - megatron_patch/model/galactica/language_model.py (277:371) duplicated block id: 418 size: 65 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/language_model.py (579:665) - megatron_patch/model/glm130b/language_model.py (543:630) duplicated block id: 419 size: 65 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/language_model.py (262:356) - megatron_patch/model/falcon/language_model.py (246:340) duplicated block id: 420 size: 65 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/language_model.py (262:356) - megatron_patch/model/llama/language_model.py (246:340) duplicated block id: 421 size: 65 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer_legacy.py (43:150) - megatron_patch/model/llava/transformer.py (39:146) duplicated block id: 422 size: 65 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (34:141) - megatron_patch/model/llama3/transformer_legacy.py (43:150) duplicated block id: 423 size: 65 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer_legacy.py (43:150) - megatron_patch/model/qwen_vl/transformer.py (39:146) duplicated block id: 424 size: 65 cleaned lines of code in 2 files: - megatron_patch/model/llama2/transformer.py (39:146) - megatron_patch/model/llama3/transformer_legacy.py (43:150) duplicated block id: 425 size: 65 cleaned lines of code in 2 files: - megatron_patch/model/bloom/language_model.py (264:358) - megatron_patch/model/chatglm/language_model.py (262:356) duplicated block id: 426 size: 65 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer_legacy.py (43:150) - megatron_patch/model/qwen1_5_megablocks/transformer.py (37:144) duplicated block id: 427 size: 64 cleaned lines of code in 2 files: - megatron_patch/model/bloom/language_model.py (31:128) - megatron_patch/model/galactica/language_model.py (29:126) duplicated block id: 428 size: 64 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/language_model.py (30:127) - megatron_patch/model/llama/language_model.py (29:126) duplicated block id: 429 size: 64 cleaned lines of code in 2 files: - megatron_patch/model/falcon/language_model.py (29:126) - megatron_patch/model/glm130b/language_model.py (30:127) duplicated block id: 430 size: 64 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/gpt_model.py (61:142) - megatron_patch/model/llama3/gpt_model.py (52:134) duplicated block id: 431 size: 64 cleaned lines of code in 2 files: - megatron_patch/model/bloom/language_model.py (31:128) - megatron_patch/model/llama/language_model.py (29:126) duplicated block id: 432 size: 64 cleaned lines of code in 2 files: - megatron_patch/model/galactica/language_model.py (29:126) - megatron_patch/model/glm130b/language_model.py (30:127) duplicated block id: 433 size: 64 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/language_model.py (29:126) - megatron_patch/model/glm130b/language_model.py (30:127) duplicated block id: 434 size: 64 cleaned lines of code in 2 files: - megatron_patch/model/falcon/transformer.py (187:303) - megatron_patch/model/llama/transformer.py (168:284) duplicated block id: 435 size: 64 cleaned lines of code in 2 files: - megatron_patch/model/bloom/language_model.py (31:128) - megatron_patch/model/falcon40b/language_model.py (29:126) duplicated block id: 436 size: 64 cleaned lines of code in 2 files: - megatron_patch/model/falcon/transformer.py (187:303) - megatron_patch/model/falcon40b/transformer.py (180:296) duplicated block id: 437 size: 64 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/gpt_model.py (61:142) - megatron_patch/model/qwen/gpt_model.py (50:132) duplicated block id: 438 size: 64 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/gpt_model.py (76:158) - megatron_patch/model/baichuan2/gpt_model.py (61:142) duplicated block id: 439 size: 64 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/gpt_model.py (61:142) - megatron_patch/model/llama2/gpt_model.py (50:132) duplicated block id: 440 size: 64 cleaned lines of code in 2 files: - megatron_patch/model/bloom/language_model.py (31:128) - megatron_patch/model/falcon/language_model.py (29:126) duplicated block id: 441 size: 64 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/gpt_model.py (61:142) - megatron_patch/model/mistral/gpt_model.py (50:132) duplicated block id: 442 size: 64 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/gpt_model.py (61:142) - megatron_patch/model/qwen1_5_megablocks/gpt_model.py (50:132) duplicated block id: 443 size: 64 cleaned lines of code in 2 files: - megatron_patch/model/bloom/language_model.py (31:128) - megatron_patch/model/glm130b/language_model.py (30:127) duplicated block id: 444 size: 63 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer_legacy.py (297:390) - megatron_patch/model/llava/transformer.py (298:391) duplicated block id: 445 size: 63 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer_legacy.py (297:390) - megatron_patch/model/qwen_vl/transformer.py (298:391) duplicated block id: 446 size: 63 cleaned lines of code in 2 files: - megatron_patch/model/qwen2_5_vl/visionmodel.py (95:174) - megatron_patch/model/qwen2_vl/visionmodel.py (90:169) duplicated block id: 447 size: 63 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer_legacy.py (297:390) - megatron_patch/model/mistral/transformer.py (298:391) duplicated block id: 448 size: 61 cleaned lines of code in 2 files: - megatron_patch/model/bloom/language_model.py (518:599) - megatron_patch/model/chatglm/language_model.py (584:665) duplicated block id: 449 size: 61 cleaned lines of code in 2 files: - megatron_patch/model/bloom/language_model.py (169:260) - megatron_patch/model/chatglm/language_model.py (168:259) duplicated block id: 450 size: 61 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/transformer/attention.py (23:126) - megatron_patch/model/qwen2_vl/attention.py (11:112) duplicated block id: 451 size: 61 cleaned lines of code in 2 files: - megatron_patch/model/bloom/language_model.py (518:599) - megatron_patch/model/glm130b/language_model.py (548:630) duplicated block id: 452 size: 61 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer/attention.py (507:595) - megatron_patch/model/qwen2/transformer/attention.py (512:600) duplicated block id: 453 size: 61 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/mistral/hf2mcore.py (65:138) - toolkits/model_checkpoints_convertor/qwen/hf2megablocks_qwen1.5.py (90:165) duplicated block id: 454 size: 61 cleaned lines of code in 2 files: - megatron_patch/model/qwen2_vl/attention.py (11:112) - megatron_patch/model/qwen2_vl/attention_vision.py (11:112) duplicated block id: 455 size: 61 cleaned lines of code in 2 files: - megatron_patch/model/qwen1_5/transformer/attention.py (507:595) - megatron_patch/model/qwen2/transformer/attention.py (512:600) duplicated block id: 456 size: 60 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (1425:1507) - megatron_patch/model/mistral/transformer.py (1557:1639) duplicated block id: 457 size: 60 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (1425:1507) - megatron_patch/model/qwen1_5_megablocks/transformer.py (1423:1505) duplicated block id: 458 size: 60 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (669:756) - megatron_patch/model/llava/transformer.py (677:764) duplicated block id: 459 size: 60 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (560:645) - megatron_patch/model/qwen_vl/transformer.py (677:764) duplicated block id: 460 size: 60 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (1425:1507) - megatron_patch/model/llama2/transformer.py (1560:1642) duplicated block id: 461 size: 60 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore.py (131:208) - toolkits/model_checkpoints_convertor/qwen/hf2megablocks_qwen1.5.py (125:203) duplicated block id: 462 size: 60 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (1425:1507) - megatron_patch/model/qwen/transformer.py (1502:1584) duplicated block id: 463 size: 60 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (560:645) - megatron_patch/model/llava/transformer.py (677:764) duplicated block id: 464 size: 60 cleaned lines of code in 2 files: - megatron_patch/model/llava/transformer.py (677:764) - megatron_patch/model/qwen/transformer.py (651:738) duplicated block id: 465 size: 60 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (669:756) - megatron_patch/model/qwen_vl/transformer.py (677:764) duplicated block id: 466 size: 60 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (1425:1507) - megatron_patch/model/baichuan2/transformer.py (1560:1642) duplicated block id: 467 size: 60 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (1425:1507) - megatron_patch/model/qwen_vl/transformer.py (1559:1641) duplicated block id: 468 size: 60 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (1425:1507) - megatron_patch/model/llava/transformer.py (1559:1641) duplicated block id: 469 size: 60 cleaned lines of code in 2 files: - megatron_patch/model/qwen/transformer.py (651:738) - megatron_patch/model/qwen_vl/transformer.py (677:764) duplicated block id: 470 size: 59 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/transformer.py (247:330) - megatron_patch/model/galactica/transformer.py (251:335) duplicated block id: 471 size: 59 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (306:389) - megatron_patch/model/llama/transformer.py (289:373) duplicated block id: 472 size: 59 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (306:389) - megatron_patch/model/falcon40b/transformer.py (301:385) duplicated block id: 473 size: 59 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/transformer.py (247:330) - megatron_patch/model/llama/transformer.py (289:373) duplicated block id: 474 size: 59 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (306:389) - megatron_patch/model/galactica/transformer.py (251:335) duplicated block id: 475 size: 59 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/transformer.py (247:330) - megatron_patch/model/falcon40b/transformer.py (301:385) duplicated block id: 476 size: 58 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (800:877) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (685:761) duplicated block id: 477 size: 58 cleaned lines of code in 2 files: - megatron_patch/model/mixtral_bak/transformer/attention.py (396:479) - megatron_patch/model/qwen2/transformer/attention.py (517:600) duplicated block id: 478 size: 58 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer/attention.py (512:595) - megatron_patch/model/mixtral_bak/transformer/attention.py (396:479) duplicated block id: 479 size: 58 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/language_model.py (448:520) - megatron_patch/model/llama3/language_model.py (381:453) duplicated block id: 480 size: 58 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (1786:1865) - megatron_patch/model/qwen1_5_megablocks/transformer.py (1650:1729) duplicated block id: 481 size: 58 cleaned lines of code in 2 files: - megatron_patch/model/mixtral_bak/transformer/attention.py (396:479) - megatron_patch/model/qwen1_5/transformer/attention.py (512:595) duplicated block id: 482 size: 58 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/transformer_block.py (287:369) - megatron_patch/model/qwen2_5_vl/transformer_block.py (334:416) duplicated block id: 483 size: 57 cleaned lines of code in 2 files: - megatron_patch/model/llama2/transformer.py (1127:1193) - megatron_patch/model/llama3/transformer_legacy.py (1076:1142) duplicated block id: 484 size: 57 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer_legacy.py (1076:1142) - megatron_patch/model/qwen_vl/transformer.py (1126:1192) duplicated block id: 485 size: 57 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/language_model.py (91:198) - megatron_patch/model/baichuan2/language_model.py (56:159) duplicated block id: 486 size: 57 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/language_model.py (452:520) - megatron_patch/model/llama2/language_model.py (394:462) duplicated block id: 487 size: 57 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer_legacy.py (1076:1142) - megatron_patch/model/qwen/transformer.py (1072:1138) duplicated block id: 488 size: 57 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer_legacy.py (1076:1142) - megatron_patch/model/qwen1_5_megablocks/transformer.py (1011:1077) duplicated block id: 489 size: 57 cleaned lines of code in 2 files: - megatron_patch/model/bloom/language_model.py (387:455) - megatron_patch/model/glm130b/language_model.py (406:474) duplicated block id: 490 size: 57 cleaned lines of code in 2 files: - megatron_patch/model/qwen2_moe/layer_specs.py (215:281) - megatron_patch/model/qwen3_moe/gpt_layer_specs.py (236:302) duplicated block id: 491 size: 57 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (1225:1291) - megatron_patch/model/qwen1_5_megablocks/transformer.py (1109:1175) duplicated block id: 492 size: 57 cleaned lines of code in 2 files: - megatron_patch/model/qwen/transformer.py (1728:1805) - megatron_patch/model/qwen1_5_megablocks/transformer.py (1650:1727) duplicated block id: 493 size: 57 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer_legacy.py (1076:1142) - megatron_patch/model/llava/transformer.py (1126:1192) duplicated block id: 494 size: 57 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/language_model.py (452:520) - megatron_patch/model/qwen1_5_megablocks/language_model.py (393:461) duplicated block id: 495 size: 57 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer_legacy.py (1183:1249) - megatron_patch/model/qwen1_5_megablocks/transformer.py (1109:1175) duplicated block id: 496 size: 57 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/language_model.py (452:520) - megatron_patch/model/mistral/language_model.py (396:464) duplicated block id: 497 size: 57 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (1129:1195) - megatron_patch/model/llama3/transformer_legacy.py (1076:1142) duplicated block id: 498 size: 57 cleaned lines of code in 2 files: - megatron_patch/model/qwen/transformer.py (1167:1233) - megatron_patch/model/qwen1_5_megablocks/transformer.py (1109:1175) duplicated block id: 499 size: 57 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer/mlp.py (24:116) - megatron_patch/model/mixtral/transformer/mlp.py (25:118) duplicated block id: 500 size: 57 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer_legacy.py (1076:1142) - megatron_patch/model/mistral/transformer.py (1124:1190) duplicated block id: 501 size: 55 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/language_model.py (287:363) - megatron_patch/model/glm130b/language_model.py (303:377) duplicated block id: 502 size: 55 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2.5_vl.py (286:356) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_vl.py (283:353) duplicated block id: 503 size: 55 cleaned lines of code in 2 files: - megatron_patch/model/llama2/rotary_pos_embedding.py (18:93) - megatron_patch/model/qwen1_5_megablocks/rotary_pos_embedding.py (18:93) duplicated block id: 504 size: 55 cleaned lines of code in 2 files: - megatron_patch/model/galactica/language_model.py (318:394) - megatron_patch/model/glm130b/language_model.py (303:377) duplicated block id: 505 size: 55 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (838:932) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (860:954) duplicated block id: 506 size: 55 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/language_model.py (303:377) - megatron_patch/model/llama/language_model.py (287:363) duplicated block id: 507 size: 55 cleaned lines of code in 2 files: - megatron_patch/model/falcon/language_model.py (287:363) - megatron_patch/model/glm130b/language_model.py (303:377) duplicated block id: 508 size: 54 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (184:299) - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (222:337) duplicated block id: 509 size: 54 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (212:294) - megatron_patch/model/qwen_vl/transformer.py (309:391) duplicated block id: 510 size: 54 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (212:294) - megatron_patch/model/llava/transformer.py (309:391) duplicated block id: 511 size: 54 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (1045:1120) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (1053:1128) duplicated block id: 512 size: 54 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (213:323) - toolkits/model_checkpoints_convertor/yi/checkpoint_reshaping_and_interoperability.py (168:276) duplicated block id: 513 size: 54 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (219:327) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (213:323) duplicated block id: 514 size: 54 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (213:323) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (238:346) duplicated block id: 515 size: 54 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (212:294) - megatron_patch/model/mistral/transformer.py (309:391) duplicated block id: 516 size: 54 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (184:299) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (241:356) duplicated block id: 517 size: 54 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (184:299) - toolkits/model_checkpoints_convertor/yi/checkpoint_reshaping_and_interoperability.py (171:286) duplicated block id: 518 size: 54 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (184:299) - toolkits/model_checkpoints_convertor/galactica/checkpoint_reshaping_and_interoperability.py (173:288) duplicated block id: 519 size: 54 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/galactica/checkpoint_reshaping_and_interoperability.py (169:278) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (213:323) duplicated block id: 520 size: 53 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (699:789) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (862:954) duplicated block id: 521 size: 53 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (698:780) - megatron_patch/model/llama2/transformer.py (846:928) duplicated block id: 522 size: 53 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/moe/experts.py (111:163) - megatron_patch/model/qwen2/moe/experts.py (115:167) duplicated block id: 523 size: 53 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (698:780) - megatron_patch/model/qwen1_5_megablocks/transformer.py (737:818) duplicated block id: 524 size: 53 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/moe/experts.py (111:163) - megatron_patch/model/qwen1_5/moe/experts.py (91:143) duplicated block id: 525 size: 53 cleaned lines of code in 2 files: - megatron_patch/model/mistral/transformer.py (312:391) - megatron_patch/model/qwen1_5_megablocks/transformer.py (203:282) duplicated block id: 526 size: 53 cleaned lines of code in 2 files: - megatron_patch/model/llava/transformer.py (312:391) - megatron_patch/model/qwen1_5_megablocks/transformer.py (203:282) duplicated block id: 527 size: 53 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/moe/experts.py (112:164) - megatron_patch/model/mixtral_bak/moe/experts.py (89:141) duplicated block id: 528 size: 53 cleaned lines of code in 2 files: - megatron_patch/model/llama3/language_model.py (499:567) - megatron_patch/model/qwen_vl/language_model.py (551:619) duplicated block id: 529 size: 53 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/language_model.py (132:216) - megatron_patch/model/falcon/language_model.py (131:215) duplicated block id: 530 size: 53 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (698:780) - megatron_patch/model/mistral/transformer.py (843:925) duplicated block id: 531 size: 53 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/moe/experts.py (111:163) - megatron_patch/model/mixtral_bak/moe/experts.py (89:141) duplicated block id: 532 size: 53 cleaned lines of code in 2 files: - megatron_patch/model/llama2/language_model.py (522:590) - megatron_patch/model/llama3/language_model.py (499:567) duplicated block id: 533 size: 53 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (698:780) - megatron_patch/model/qwen_vl/transformer.py (845:927) duplicated block id: 534 size: 53 cleaned lines of code in 2 files: - megatron_patch/model/llama3/language_model.py (499:567) - megatron_patch/model/llava/language_model.py (582:650) duplicated block id: 535 size: 53 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/language_model.py (132:216) - megatron_patch/model/llama/language_model.py (131:215) duplicated block id: 536 size: 53 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/moe/experts.py (112:164) - megatron_patch/model/qwen2/moe/experts.py (115:167) duplicated block id: 537 size: 53 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (698:780) - megatron_patch/model/llava/transformer.py (845:927) duplicated block id: 538 size: 53 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (726:816) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (862:954) duplicated block id: 539 size: 53 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/language_model.py (132:216) - megatron_patch/model/falcon40b/language_model.py (131:215) duplicated block id: 540 size: 53 cleaned lines of code in 2 files: - megatron_patch/model/qwen1_5_megablocks/transformer.py (203:282) - megatron_patch/model/qwen_vl/transformer.py (312:391) duplicated block id: 541 size: 53 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/moe/experts.py (112:164) - megatron_patch/model/qwen1_5/moe/experts.py (91:143) duplicated block id: 542 size: 53 cleaned lines of code in 2 files: - megatron_patch/model/llama3/language_model.py (499:567) - megatron_patch/model/qwen1_5_megablocks/language_model.py (521:589) duplicated block id: 543 size: 53 cleaned lines of code in 2 files: - megatron_patch/model/llama3/language_model.py (499:567) - megatron_patch/model/mistral/language_model.py (534:602) duplicated block id: 544 size: 52 cleaned lines of code in 2 files: - megatron_patch/model/llama/positional_embeddings.py (19:88) - megatron_patch/model/llava/rotary_pos_embedding.py (22:91) duplicated block id: 545 size: 52 cleaned lines of code in 2 files: - megatron_patch/model/falcon/transformer.py (925:1005) - megatron_patch/model/llama/transformer.py (696:764) duplicated block id: 546 size: 52 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer_legacy.py (1534:1612) - megatron_patch/model/llava/transformer.py (1573:1651) duplicated block id: 547 size: 52 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (184:289) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (218:323) duplicated block id: 548 size: 52 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer_legacy.py (1534:1612) - megatron_patch/model/qwen1_5_megablocks/transformer.py (1437:1515) duplicated block id: 549 size: 52 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/reward_model_to_megatron.py (151:258) - toolkits/model_checkpoints_convertor/glm130b/checkpoint_reshaping_and_interoperability.py (122:229) duplicated block id: 550 size: 52 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/layer_specs.py (60:142) - megatron_patch/model/mixtral_bak/layer_specs.py (38:121) duplicated block id: 551 size: 52 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/reward_model_to_megatron.py (151:258) - toolkits/model_checkpoints_convertor/glm/checkpoint_reshaping_and_interoperability.py (122:229) duplicated block id: 552 size: 52 cleaned lines of code in 2 files: - megatron_patch/model/llama2/transformer.py (1574:1652) - megatron_patch/model/llama3/transformer_legacy.py (1534:1612) duplicated block id: 553 size: 52 cleaned lines of code in 2 files: - megatron_patch/model/llama3/model.py (107:195) - megatron_patch/model/qwen2_vl/gpt_model.py (98:186) duplicated block id: 554 size: 52 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer_legacy.py (1534:1612) - megatron_patch/model/qwen_vl/transformer.py (1573:1651) duplicated block id: 555 size: 52 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (190:293) - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (252:355) duplicated block id: 556 size: 52 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/checkpoint_reshaping_and_interoperability.py (151:258) - toolkits/model_checkpoints_convertor/glm130b/checkpoint_reshaping_and_interoperability.py (122:229) duplicated block id: 557 size: 52 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/checkpoint_reshaping_and_interoperability.py (151:258) - toolkits/model_checkpoints_convertor/glm/checkpoint_reshaping_and_interoperability.py (122:229) duplicated block id: 558 size: 52 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer_legacy.py (1534:1612) - megatron_patch/model/mistral/transformer.py (1571:1649) duplicated block id: 559 size: 52 cleaned lines of code in 2 files: - megatron_patch/model/qwen2/model.py (108:196) - megatron_patch/model/qwen2_vl/gpt_model.py (98:186) duplicated block id: 560 size: 52 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (728:799) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (660:731) duplicated block id: 561 size: 52 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (252:355) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (169:272) duplicated block id: 562 size: 52 cleaned lines of code in 2 files: - megatron_patch/model/qwen1_5/model.py (107:195) - megatron_patch/model/qwen2_vl/gpt_model.py (98:186) duplicated block id: 563 size: 51 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (59:112) - toolkits/model_checkpoints_convertor/chatglm/checkpoint_reshaping_and_interoperability.py (49:101) duplicated block id: 564 size: 51 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/transformer.py (638:712) - megatron_patch/model/glm130b/transformer.py (833:907) duplicated block id: 565 size: 51 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/chatglm/checkpoint_reshaping_and_interoperability.py (49:101) - toolkits/model_checkpoints_convertor/falcon/checkpoint_reshaping_and_interoperability.py (60:113) duplicated block id: 566 size: 51 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/chatglm/checkpoint_reshaping_and_interoperability.py (49:101) - toolkits/model_checkpoints_convertor/galactica/checkpoint_reshaping_and_interoperability.py (59:112) duplicated block id: 567 size: 51 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/chatglm/checkpoint_reshaping_and_interoperability.py (49:101) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (88:141) duplicated block id: 568 size: 51 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/chatglm/checkpoint_reshaping_and_interoperability.py (49:101) - toolkits/model_checkpoints_convertor/falcon40b/checkpoint_reshaping_and_interoperability.py (60:113) duplicated block id: 569 size: 51 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/chatglm/checkpoint_reshaping_and_interoperability.py (49:101) - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (83:136) duplicated block id: 570 size: 51 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (59:112) - toolkits/model_checkpoints_convertor/chatglm/checkpoint_reshaping_and_interoperability.py (49:101) duplicated block id: 571 size: 51 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/chatglm/checkpoint_reshaping_and_interoperability.py (49:101) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (87:140) duplicated block id: 572 size: 51 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/deepspeed_to_megatron.py (101:163) - toolkits/model_checkpoints_convertor/bloom/deepspeed_to_megatron_ori.py (100:163) duplicated block id: 573 size: 51 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore_llama3_1.py (500:555) - toolkits/model_checkpoints_convertor/llava/hf2mcore_llava.py (272:327) duplicated block id: 574 size: 51 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/chatglm/checkpoint_reshaping_and_interoperability.py (49:101) - toolkits/model_checkpoints_convertor/yi/checkpoint_reshaping_and_interoperability.py (60:113) duplicated block id: 575 size: 50 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/transformer/attention.py (577:637) - megatron_patch/model/qwen2/transformer/attention.py (436:496) duplicated block id: 576 size: 50 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon40b/checkpoint_reshaping_and_interoperability.py (384:448) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (463:528) duplicated block id: 577 size: 50 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon40b/checkpoint_reshaping_and_interoperability.py (384:448) - toolkits/model_checkpoints_convertor/yi/checkpoint_reshaping_and_interoperability.py (408:473) duplicated block id: 578 size: 50 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/transformer.py (729:802) - megatron_patch/model/falcon40b/transformer.py (844:921) duplicated block id: 579 size: 50 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (433:498) - toolkits/model_checkpoints_convertor/falcon/checkpoint_reshaping_and_interoperability.py (395:459) duplicated block id: 580 size: 50 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (405:470) - toolkits/model_checkpoints_convertor/falcon/checkpoint_reshaping_and_interoperability.py (395:459) duplicated block id: 581 size: 50 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/transformer.py (729:802) - megatron_patch/model/falcon/transformer.py (1105:1182) duplicated block id: 582 size: 50 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon/checkpoint_reshaping_and_interoperability.py (395:459) - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (486:551) duplicated block id: 583 size: 50 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon/checkpoint_reshaping_and_interoperability.py (395:459) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (493:558) duplicated block id: 584 size: 50 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon40b/checkpoint_reshaping_and_interoperability.py (384:448) - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (486:551) duplicated block id: 585 size: 50 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon40b/checkpoint_reshaping_and_interoperability.py (384:448) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (493:558) duplicated block id: 586 size: 50 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (433:498) - toolkits/model_checkpoints_convertor/falcon40b/checkpoint_reshaping_and_interoperability.py (384:448) duplicated block id: 587 size: 50 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore_llama3_1.py (587:641) - toolkits/model_checkpoints_convertor/llava/hf2mcore_llava.py (356:410) duplicated block id: 588 size: 50 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon/checkpoint_reshaping_and_interoperability.py (395:459) - toolkits/model_checkpoints_convertor/yi/checkpoint_reshaping_and_interoperability.py (408:473) duplicated block id: 589 size: 50 cleaned lines of code in 2 files: - megatron_patch/model/qwen2/transformer/attention.py (436:496) - megatron_patch/model/qwen2_vl/attention.py (578:638) duplicated block id: 590 size: 50 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon/checkpoint_reshaping_and_interoperability.py (395:459) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (463:528) duplicated block id: 591 size: 50 cleaned lines of code in 2 files: - toolkits/pretrain_data_preprocessing/preprocess_data_megatron.py (311:376) - toolkits/sft_data_preprocessing/build_idxmap_sft_dataset.py (299:364) duplicated block id: 592 size: 50 cleaned lines of code in 2 files: - megatron_patch/model/qwen2/transformer/attention.py (436:496) - megatron_patch/model/qwen2_vl/attention_vision.py (577:637) duplicated block id: 593 size: 50 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (405:470) - toolkits/model_checkpoints_convertor/falcon40b/checkpoint_reshaping_and_interoperability.py (384:448) duplicated block id: 594 size: 49 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (1367:1423) - megatron_patch/model/llava/transformer.py (1501:1557) duplicated block id: 595 size: 49 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (1367:1423) - megatron_patch/model/mistral/transformer.py (1499:1555) duplicated block id: 596 size: 49 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/transformer.py (661:740) - megatron_patch/model/galactica/transformer.py (553:624) duplicated block id: 597 size: 49 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/transformer.py (661:740) - megatron_patch/model/llama/transformer.py (663:729) duplicated block id: 598 size: 49 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/transformer.py (663:734) - megatron_patch/model/llama/transformer.py (602:673) duplicated block id: 599 size: 49 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (1367:1423) - megatron_patch/model/qwen/transformer.py (1444:1500) duplicated block id: 600 size: 49 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/transformer/attention.py (504:571) - megatron_patch/model/qwen2/transformer/attention.py (365:431) duplicated block id: 601 size: 49 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (589:656) - toolkits/model_checkpoints_convertor/falcon/checkpoint_reshaping_and_interoperability.py (559:627) duplicated block id: 602 size: 49 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (1367:1423) - megatron_patch/model/llama2/transformer.py (1502:1558) duplicated block id: 603 size: 49 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/chatglm/checkpoint_reshaping_and_interoperability.py (150:246) - toolkits/model_checkpoints_convertor/galactica/checkpoint_reshaping_and_interoperability.py (167:263) duplicated block id: 604 size: 49 cleaned lines of code in 2 files: - megatron_patch/model/galactica/transformer.py (128:201) - megatron_patch/model/llama/transformer.py (166:239) duplicated block id: 605 size: 49 cleaned lines of code in 2 files: - megatron_patch/model/qwen2/transformer/attention.py (365:431) - megatron_patch/model/qwen2_vl/attention_vision.py (504:571) duplicated block id: 606 size: 49 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (589:656) - toolkits/model_checkpoints_convertor/falcon40b/checkpoint_reshaping_and_interoperability.py (563:630) duplicated block id: 607 size: 49 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer_legacy.py (1746:1812) - megatron_patch/model/qwen1_5_megablocks/transformer.py (1650:1716) duplicated block id: 608 size: 49 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/glm/checkpoint_reshaping_and_interoperability.py (125:229) - toolkits/model_checkpoints_convertor/starcoder/checkpoint_reshaping_and_interoperability.py (176:280) duplicated block id: 609 size: 49 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/glm130b/checkpoint_reshaping_and_interoperability.py (125:229) - toolkits/model_checkpoints_convertor/starcoder/checkpoint_reshaping_and_interoperability.py (176:280) duplicated block id: 610 size: 49 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (1367:1423) - megatron_patch/model/baichuan2/transformer.py (1502:1558) duplicated block id: 611 size: 49 cleaned lines of code in 2 files: - megatron_patch/model/galactica/transformer.py (553:624) - megatron_patch/model/llama/transformer.py (663:729) duplicated block id: 612 size: 49 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (181:276) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (211:308) duplicated block id: 613 size: 49 cleaned lines of code in 2 files: - megatron_patch/model/qwen2/transformer/attention.py (365:431) - megatron_patch/model/qwen2_vl/attention.py (505:572) duplicated block id: 614 size: 49 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (1367:1423) - megatron_patch/model/qwen_vl/transformer.py (1501:1557) duplicated block id: 615 size: 48 cleaned lines of code in 2 files: - megatron_patch/model/qwen1_5/moe/token_dispatcher.py (238:304) - megatron_patch/model/qwen2/moe/token_dispatcher.py (244:310) duplicated block id: 616 size: 47 cleaned lines of code in 2 files: - megatron_patch/model/llava/gpt_model.py (26:89) - megatron_patch/model/mistral/gpt_model.py (26:89) duplicated block id: 617 size: 47 cleaned lines of code in 2 files: - megatron_patch/model/falcon/transformer.py (187:258) - megatron_patch/model/galactica/transformer.py (130:201) duplicated block id: 618 size: 47 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (183:276) - toolkits/model_checkpoints_convertor/yi/checkpoint_reshaping_and_interoperability.py (168:261) duplicated block id: 619 size: 47 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/reward_model_to_megatron.py (504:575) - toolkits/model_checkpoints_convertor/starcoder/checkpoint_reshaping_and_interoperability.py (520:591) duplicated block id: 620 size: 47 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/chatglm/checkpoint_reshaping_and_interoperability.py (152:246) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (238:331) duplicated block id: 621 size: 47 cleaned lines of code in 2 files: - megatron_patch/model/llama2/transformer.py (905:977) - megatron_patch/model/llama3/transformer_legacy.py (856:927) duplicated block id: 622 size: 47 cleaned lines of code in 2 files: - megatron_patch/model/llava/gpt_model.py (26:89) - megatron_patch/model/qwen1_5_megablocks/gpt_model.py (26:89) duplicated block id: 623 size: 47 cleaned lines of code in 2 files: - megatron_patch/model/llama2/gpt_model.py (26:89) - megatron_patch/model/llava/gpt_model.py (26:89) duplicated block id: 624 size: 47 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (183:276) - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (219:312) duplicated block id: 625 size: 47 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer_legacy.py (856:927) - megatron_patch/model/qwen/transformer.py (850:922) duplicated block id: 626 size: 47 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer_legacy.py (856:927) - megatron_patch/model/mistral/transformer.py (902:974) duplicated block id: 627 size: 47 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (1367:1421) - megatron_patch/model/llama3/transformer_legacy.py (1460:1514) duplicated block id: 628 size: 47 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore.py (329:379) - toolkits/model_checkpoints_convertor/mistral/hf2mcore.py (271:321) duplicated block id: 629 size: 47 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (868:940) - megatron_patch/model/llama3/transformer_legacy.py (856:927) duplicated block id: 630 size: 47 cleaned lines of code in 2 files: - megatron_patch/model/qwen/gpt_model.py (26:89) - megatron_patch/model/qwen_vl/gpt_model.py (26:89) duplicated block id: 631 size: 47 cleaned lines of code in 2 files: - megatron_patch/model/llama2/gpt_model.py (26:89) - megatron_patch/model/qwen_vl/gpt_model.py (26:89) duplicated block id: 632 size: 47 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (183:276) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (238:331) duplicated block id: 633 size: 47 cleaned lines of code in 2 files: - megatron_patch/model/llama3/gpt_model.py (28:91) - megatron_patch/model/qwen_vl/gpt_model.py (26:89) duplicated block id: 634 size: 47 cleaned lines of code in 2 files: - megatron_patch/model/mistral/gpt_model.py (26:89) - megatron_patch/model/qwen_vl/gpt_model.py (26:89) duplicated block id: 635 size: 47 cleaned lines of code in 2 files: - megatron_patch/model/llama3/gpt_model.py (28:91) - megatron_patch/model/llava/gpt_model.py (26:89) duplicated block id: 636 size: 47 cleaned lines of code in 2 files: - megatron_patch/model/qwen1_5_megablocks/gpt_model.py (26:89) - megatron_patch/model/qwen_vl/gpt_model.py (26:89) duplicated block id: 637 size: 47 cleaned lines of code in 2 files: - megatron_patch/model/llava/gpt_model.py (26:89) - megatron_patch/model/qwen/gpt_model.py (26:89) duplicated block id: 638 size: 47 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/chatglm/checkpoint_reshaping_and_interoperability.py (152:246) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (213:308) duplicated block id: 639 size: 47 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/checkpoint_reshaping_and_interoperability.py (505:576) - toolkits/model_checkpoints_convertor/starcoder/checkpoint_reshaping_and_interoperability.py (520:591) duplicated block id: 640 size: 47 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/deepspeed_to_megatron.py (12:65) - toolkits/model_checkpoints_convertor/bloom/deepspeed_to_megatron_ori.py (11:64) duplicated block id: 641 size: 47 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (183:276) - toolkits/model_checkpoints_convertor/chatglm/checkpoint_reshaping_and_interoperability.py (152:246) duplicated block id: 642 size: 47 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer_legacy.py (856:927) - megatron_patch/model/qwen_vl/transformer.py (904:976) duplicated block id: 643 size: 47 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/chatglm/checkpoint_reshaping_and_interoperability.py (152:246) - toolkits/model_checkpoints_convertor/yi/checkpoint_reshaping_and_interoperability.py (168:261) duplicated block id: 644 size: 47 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/chatglm/checkpoint_reshaping_and_interoperability.py (152:246) - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (219:312) duplicated block id: 645 size: 47 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer_legacy.py (856:927) - megatron_patch/model/llava/transformer.py (904:976) duplicated block id: 646 size: 47 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (183:276) - toolkits/model_checkpoints_convertor/galactica/checkpoint_reshaping_and_interoperability.py (169:263) duplicated block id: 647 size: 46 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/gpt_model.py (53:115) - megatron_patch/model/qwen_vl/gpt_model.py (27:89) duplicated block id: 648 size: 46 cleaned lines of code in 2 files: - megatron_patch/model/mixtral_bak/model.py (123:216) - megatron_patch/model/qwen1_5/model.py (147:226) duplicated block id: 649 size: 46 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/gpt_model.py (53:115) - megatron_patch/model/llava/gpt_model.py (27:89) duplicated block id: 650 size: 46 cleaned lines of code in 2 files: - megatron_patch/model/mixtral_bak/model.py (123:216) - megatron_patch/model/qwen2/model.py (148:227) duplicated block id: 651 size: 46 cleaned lines of code in 2 files: - megatron_patch/model/llama3/model.py (147:226) - megatron_patch/model/mixtral_bak/model.py (123:216) duplicated block id: 652 size: 46 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon/configuration_RW.py (83:134) - toolkits/model_checkpoints_convertor/falcon40b/configuration_RW.py (24:75) duplicated block id: 653 size: 46 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (1439:1507) - megatron_patch/model/llama3/transformer_legacy.py (1534:1602) duplicated block id: 654 size: 46 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon/checkpoint_reshaping_and_interoperability.py (188:281) - toolkits/model_checkpoints_convertor/falcon40b/checkpoint_reshaping_and_interoperability.py (187:279) duplicated block id: 655 size: 45 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (708:766) - toolkits/model_checkpoints_convertor/yi/checkpoint_reshaping_and_interoperability.py (606:666) duplicated block id: 656 size: 45 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (184:274) - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (186:276) duplicated block id: 657 size: 45 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/model.py (225:302) - megatron_patch/model/mixtral_bak/model.py (210:287) duplicated block id: 658 size: 45 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (184:274) - toolkits/model_checkpoints_convertor/chatglm/checkpoint_reshaping_and_interoperability.py (156:246) duplicated block id: 659 size: 45 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (1656:1714) - megatron_patch/model/qwen1_5_megablocks/transformer.py (1650:1708) duplicated block id: 660 size: 44 cleaned lines of code in 2 files: - megatron_patch/model/qwen1_5_megablocks/transformer.py (465:521) - megatron_patch/model/qwen_vl/transformer.py (575:631) duplicated block id: 661 size: 44 cleaned lines of code in 2 files: - megatron_patch/model/falcon/transformer.py (1034:1090) - megatron_patch/model/glm130b/transformer.py (849:906) duplicated block id: 662 size: 44 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/transformer.py (654:711) - megatron_patch/model/falcon/transformer.py (1034:1090) duplicated block id: 663 size: 44 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/transformer.py (654:711) - megatron_patch/model/galactica/transformer.py (661:718) duplicated block id: 664 size: 44 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_mha.py (6:69) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_moe.py (7:71) duplicated block id: 665 size: 44 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer/attention.py (99:149) - megatron_patch/model/mixtral/transformer/attention.py (127:177) duplicated block id: 666 size: 44 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/transformer/attention.py (127:177) - megatron_patch/model/qwen2/transformer/attention.py (95:145) duplicated block id: 667 size: 44 cleaned lines of code in 2 files: - megatron_patch/model/llava/transformer.py (575:631) - megatron_patch/model/qwen1_5_megablocks/transformer.py (465:521) duplicated block id: 668 size: 44 cleaned lines of code in 2 files: - megatron_patch/model/galactica/transformer.py (661:718) - megatron_patch/model/glm130b/transformer.py (849:906) duplicated block id: 669 size: 44 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/transformer/attention.py (127:177) - megatron_patch/model/mixtral_bak/transformer/attention.py (73:123) duplicated block id: 670 size: 44 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/transformer/attention.py (127:177) - megatron_patch/model/qwen1_5/transformer/attention.py (99:149) duplicated block id: 671 size: 44 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/transformer.py (654:711) - megatron_patch/model/falcon40b/transformer.py (772:829) duplicated block id: 672 size: 44 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/chatglm/checkpoint_reshaping_and_interoperability.py (538:594) - toolkits/model_checkpoints_convertor/galactica/checkpoint_reshaping_and_interoperability.py (608:664) duplicated block id: 673 size: 44 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/transformer.py (849:906) - megatron_patch/model/llama/transformer.py (827:884) duplicated block id: 674 size: 44 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/transformer.py (772:829) - megatron_patch/model/glm130b/transformer.py (849:906) duplicated block id: 675 size: 44 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/transformer.py (654:711) - megatron_patch/model/llama/transformer.py (827:884) duplicated block id: 676 size: 43 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore.py (544:586) - toolkits/model_checkpoints_convertor/mistral/hf2mcore.py (451:493) duplicated block id: 677 size: 43 cleaned lines of code in 2 files: - megatron_patch/model/mixtral_bak/transformer/attention.py (288:350) - megatron_patch/model/qwen2/transformer/attention.py (311:373) duplicated block id: 678 size: 43 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/language_model.py (211:273) - megatron_patch/model/mistral/language_model.py (161:223) duplicated block id: 679 size: 43 cleaned lines of code in 2 files: - megatron_patch/model/mixtral_bak/transformer/attention.py (288:350) - megatron_patch/model/qwen1_5/transformer/attention.py (315:377) duplicated block id: 680 size: 43 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/language_model.py (211:273) - megatron_patch/model/llava/language_model.py (163:225) duplicated block id: 681 size: 43 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/language_model.py (211:273) - megatron_patch/model/llama2/language_model.py (160:222) duplicated block id: 682 size: 43 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/language_model.py (211:273) - megatron_patch/model/qwen1_5_megablocks/language_model.py (159:221) duplicated block id: 683 size: 43 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/transformer/mlp.py (96:148) - megatron_patch/model/qwen2/transformer/mlp.py (109:161) duplicated block id: 684 size: 43 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer/attention.py (315:377) - megatron_patch/model/mixtral_bak/transformer/attention.py (288:350) duplicated block id: 685 size: 43 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/language_model.py (211:273) - megatron_patch/model/qwen/language_model.py (160:222) duplicated block id: 686 size: 43 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/language_model.py (211:273) - megatron_patch/model/qwen_vl/language_model.py (161:223) duplicated block id: 687 size: 43 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/language_model.py (211:273) - megatron_patch/model/baichuan2/language_model.py (170:232) duplicated block id: 688 size: 42 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (714:780) - megatron_patch/model/baichuan2/transformer.py (825:891) duplicated block id: 689 size: 42 cleaned lines of code in 2 files: - megatron_patch/model/falcon/transformer.py (356:421) - megatron_patch/model/starcoder/transformer.py (352:418) duplicated block id: 690 size: 42 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/configuration_baichuan.py (6:48) - toolkits/model_checkpoints_convertor/baichuan2/configuration_baichuan.py (6:48) duplicated block id: 691 size: 42 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/glm/checkpoint_reshaping_and_interoperability.py (319:370) - toolkits/model_checkpoints_convertor/glm130b/checkpoint_reshaping_and_interoperability.py (314:365) duplicated block id: 692 size: 42 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/language_model.py (543:599) - megatron_patch/model/galactica/language_model.py (593:649) duplicated block id: 693 size: 42 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore_llama3_1.py (39:89) - toolkits/model_checkpoints_convertor/llava/hf2mcore_llava.py (62:112) duplicated block id: 694 size: 41 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer_legacy.py (710:779) - megatron_patch/model/qwen/transformer.py (705:774) duplicated block id: 695 size: 41 cleaned lines of code in 2 files: - megatron_patch/model/bloom/language_model.py (538:593) - megatron_patch/model/llama/language_model.py (641:696) duplicated block id: 696 size: 41 cleaned lines of code in 2 files: - megatron_patch/model/galactica/language_model.py (659:714) - megatron_patch/model/glm130b/language_model.py (568:623) duplicated block id: 697 size: 41 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon/checkpoint_reshaping_and_interoperability.py (798:855) - toolkits/model_checkpoints_convertor/falcon40b/checkpoint_reshaping_and_interoperability.py (774:832) duplicated block id: 698 size: 41 cleaned lines of code in 2 files: - megatron_patch/model/falcon/language_model.py (493:543) - megatron_patch/model/galactica/language_model.py (519:569) duplicated block id: 699 size: 41 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (634:689) - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (606:661) duplicated block id: 700 size: 41 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/language_model.py (568:623) - megatron_patch/model/llama/language_model.py (641:696) duplicated block id: 701 size: 41 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (723:792) - megatron_patch/model/llama3/transformer_legacy.py (710:779) duplicated block id: 702 size: 41 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/language_model.py (604:659) - megatron_patch/model/galactica/language_model.py (659:714) duplicated block id: 703 size: 41 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/language_model.py (604:659) - megatron_patch/model/llama/language_model.py (641:696) duplicated block id: 704 size: 41 cleaned lines of code in 2 files: - megatron_patch/model/bloom/language_model.py (538:593) - megatron_patch/model/galactica/language_model.py (659:714) duplicated block id: 705 size: 41 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (839:898) - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (1045:1105) duplicated block id: 706 size: 41 cleaned lines of code in 2 files: - megatron_patch/model/bloom/gpt_model.py (28:84) - megatron_patch/model/glm130b/gpt_model.py (28:84) duplicated block id: 707 size: 41 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (1376:1423) - megatron_patch/model/qwen1_5_megablocks/transformer.py (1374:1421) duplicated block id: 708 size: 41 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/language_model.py (493:543) - megatron_patch/model/galactica/language_model.py (519:569) duplicated block id: 709 size: 41 cleaned lines of code in 2 files: - megatron_patch/model/galactica/language_model.py (519:569) - megatron_patch/model/llama/language_model.py (493:543) duplicated block id: 710 size: 41 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (899:954) - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (873:928) duplicated block id: 711 size: 41 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (839:898) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (1053:1113) duplicated block id: 712 size: 41 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (612:681) - megatron_patch/model/llama3/transformer_legacy.py (710:779) duplicated block id: 713 size: 40 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (669:719) - megatron_patch/model/mistral/transformer.py (677:727) duplicated block id: 714 size: 40 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_mha.py (6:64) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_mha_to_moe.py (6:63) duplicated block id: 715 size: 40 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (560:609) - megatron_patch/model/mistral/transformer.py (677:727) duplicated block id: 716 size: 40 cleaned lines of code in 2 files: - megatron_patch/model/llama2/transformer.py (678:728) - megatron_patch/model/llama3/transformer_legacy.py (656:706) duplicated block id: 717 size: 40 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer_legacy.py (656:706) - megatron_patch/model/qwen1_5_megablocks/transformer.py (569:619) duplicated block id: 718 size: 40 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/hf2te.py (334:382) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (513:562) duplicated block id: 719 size: 40 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer_legacy.py (656:706) - megatron_patch/model/llava/transformer.py (677:727) duplicated block id: 720 size: 40 cleaned lines of code in 2 files: - megatron_patch/model/mistral/transformer.py (677:727) - megatron_patch/model/qwen/transformer.py (651:701) duplicated block id: 721 size: 40 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer_legacy.py (656:706) - megatron_patch/model/qwen_vl/transformer.py (677:727) duplicated block id: 722 size: 40 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/hf2te.py (306:354) - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (536:585) duplicated block id: 723 size: 40 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/moe/experts.py (686:738) - megatron_patch/model/qwen2/moe/experts.py (307:358) duplicated block id: 724 size: 40 cleaned lines of code in 2 files: - megatron_patch/model/llama2/transformer.py (678:728) - megatron_patch/model/qwen/transformer.py (651:701) duplicated block id: 725 size: 40 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (483:532) - toolkits/model_checkpoints_convertor/baichuan/hf2te.py (334:382) duplicated block id: 726 size: 40 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (455:504) - toolkits/model_checkpoints_convertor/baichuan2/hf2te.py (306:354) duplicated block id: 727 size: 40 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/deepseek/hf2mcore_deepseek_v3_moe.py (86:135) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_moe.py (94:144) duplicated block id: 728 size: 40 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/hf2te.py (334:382) - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (455:504) duplicated block id: 729 size: 40 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (669:719) - megatron_patch/model/llama3/transformer_legacy.py (656:706) duplicated block id: 730 size: 40 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/hf2te.py (306:354) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (513:562) duplicated block id: 731 size: 40 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_mha_to_moe.py (6:63) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_moe.py (7:66) duplicated block id: 732 size: 40 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (560:609) - megatron_patch/model/llama2/transformer.py (678:728) duplicated block id: 733 size: 40 cleaned lines of code in 2 files: - megatron_patch/model/qwen/transformer.py (651:701) - megatron_patch/model/qwen1_5_megablocks/transformer.py (569:619) duplicated block id: 734 size: 40 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (560:609) - megatron_patch/model/qwen1_5_megablocks/transformer.py (569:619) duplicated block id: 735 size: 40 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (669:719) - megatron_patch/model/llama2/transformer.py (678:728) duplicated block id: 736 size: 40 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/moe/experts.py (687:739) - megatron_patch/model/qwen2/moe/experts.py (307:358) duplicated block id: 737 size: 40 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/hf2te.py (306:354) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (543:592) duplicated block id: 738 size: 40 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (669:719) - megatron_patch/model/qwen1_5_megablocks/transformer.py (569:619) duplicated block id: 739 size: 40 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer_legacy.py (656:706) - megatron_patch/model/mistral/transformer.py (677:727) duplicated block id: 740 size: 40 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (483:532) - toolkits/model_checkpoints_convertor/baichuan2/hf2te.py (306:354) duplicated block id: 741 size: 40 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/hf2te.py (334:382) - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (536:585) duplicated block id: 742 size: 40 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/hf2te.py (334:382) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (543:592) duplicated block id: 743 size: 39 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer_legacy.py (526:571) - megatron_patch/model/qwen/transformer.py (521:566) duplicated block id: 744 size: 39 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer/attention.py (450:504) - megatron_patch/model/qwen2/transformer/attention.py (452:506) duplicated block id: 745 size: 39 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer_legacy.py (526:571) - megatron_patch/model/mistral/transformer.py (528:573) duplicated block id: 746 size: 39 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer_legacy.py (526:571) - megatron_patch/model/llava/transformer.py (528:573) duplicated block id: 747 size: 39 cleaned lines of code in 2 files: - megatron_patch/model/llama2/transformer.py (527:572) - megatron_patch/model/llama3/transformer_legacy.py (526:571) duplicated block id: 748 size: 39 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer_legacy.py (1469:1514) - megatron_patch/model/qwen1_5_megablocks/transformer.py (1374:1419) duplicated block id: 749 size: 39 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/galactica/checkpoint_reshaping_and_interoperability.py (504:554) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (556:606) duplicated block id: 750 size: 39 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (496:546) - toolkits/model_checkpoints_convertor/galactica/checkpoint_reshaping_and_interoperability.py (504:554) duplicated block id: 751 size: 39 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (918:973) - megatron_patch/model/starcoder/transformer.py (1081:1136) duplicated block id: 752 size: 39 cleaned lines of code in 2 files: - megatron_patch/model/qwen1_5/transformer/attention.py (450:504) - megatron_patch/model/qwen2/transformer/attention.py (452:506) duplicated block id: 753 size: 39 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (430:475) - megatron_patch/model/llama3/transformer_legacy.py (526:571) duplicated block id: 754 size: 39 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/galactica/checkpoint_reshaping_and_interoperability.py (504:554) - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (549:599) duplicated block id: 755 size: 39 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (468:518) - toolkits/model_checkpoints_convertor/galactica/checkpoint_reshaping_and_interoperability.py (504:554) duplicated block id: 756 size: 39 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer_legacy.py (526:571) - megatron_patch/model/qwen_vl/transformer.py (528:573) duplicated block id: 757 size: 39 cleaned lines of code in 2 files: - megatron_patch/model/qwen2_5_vl/model.py (49:95) - megatron_patch/model/qwen2_vl/model.py (49:95) duplicated block id: 758 size: 39 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (539:584) - megatron_patch/model/llama3/transformer_legacy.py (526:571) duplicated block id: 759 size: 38 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/language_model.py (261:311) - megatron_patch/model/qwen1_5_megablocks/language_model.py (246:296) duplicated block id: 760 size: 38 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/language_model.py (550:601) - megatron_patch/model/starcoder/language_model.py (495:542) duplicated block id: 761 size: 38 cleaned lines of code in 2 files: - megatron_patch/model/llava/rotary_pos_embedding.py (38:91) - megatron_patch/model/qwen1_5_megablocks/rotary_pos_embedding.py (40:93) duplicated block id: 762 size: 38 cleaned lines of code in 2 files: - megatron_patch/model/mixtral_bak/transformer/attention.py (425:479) - megatron_patch/model/qwen2_vl/attention.py (690:744) duplicated block id: 763 size: 38 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/transformer/attention.py (689:743) - megatron_patch/model/qwen2/transformer/attention.py (546:600) duplicated block id: 764 size: 38 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer/attention.py (541:595) - megatron_patch/model/qwen2_vl/attention_vision.py (689:743) duplicated block id: 765 size: 38 cleaned lines of code in 2 files: - megatron_patch/model/llama3/language_model.py (182:236) - megatron_patch/model/qwen_vl/language_model.py (193:247) duplicated block id: 766 size: 38 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (634:684) - toolkits/model_checkpoints_convertor/falcon40b/checkpoint_reshaping_and_interoperability.py (580:630) duplicated block id: 767 size: 38 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/language_model.py (261:311) - megatron_patch/model/qwen_vl/language_model.py (248:298) duplicated block id: 768 size: 38 cleaned lines of code in 2 files: - megatron_patch/model/qwen2/transformer/attention.py (546:600) - megatron_patch/model/qwen2_vl/attention_vision.py (689:743) duplicated block id: 769 size: 38 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/language_model.py (547:599) - megatron_patch/model/falcon/language_model.py (568:620) duplicated block id: 770 size: 38 cleaned lines of code in 2 files: - megatron_patch/model/llama/positional_embeddings.py (35:88) - megatron_patch/model/llama2/rotary_pos_embedding.py (40:93) duplicated block id: 771 size: 38 cleaned lines of code in 2 files: - megatron_patch/model/llama3/language_model.py (182:236) - megatron_patch/model/mistral/language_model.py (193:247) duplicated block id: 772 size: 38 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer/attention.py (541:595) - megatron_patch/model/qwen2_vl/attention.py (690:744) duplicated block id: 773 size: 38 cleaned lines of code in 2 files: - megatron_patch/model/llama2/language_model.py (192:246) - megatron_patch/model/llama3/language_model.py (182:236) duplicated block id: 774 size: 38 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/language_model.py (257:307) - megatron_patch/model/glm130b/language_model.py (261:311) duplicated block id: 775 size: 38 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/language_model.py (261:311) - megatron_patch/model/llama3/language_model.py (242:292) duplicated block id: 776 size: 38 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/language_model.py (587:637) - megatron_patch/model/starcoder/language_model.py (495:542) duplicated block id: 777 size: 38 cleaned lines of code in 2 files: - megatron_patch/model/falcon/transformer.py (130:185) - megatron_patch/model/galactica/transformer.py (73:128) duplicated block id: 778 size: 38 cleaned lines of code in 2 files: - megatron_patch/model/llama3/language_model.py (182:236) - megatron_patch/model/llava/language_model.py (195:249) duplicated block id: 779 size: 38 cleaned lines of code in 2 files: - megatron_patch/model/bloom/language_model.py (521:571) - megatron_patch/model/starcoder/language_model.py (495:542) duplicated block id: 780 size: 38 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/transformer/attention.py (689:743) - megatron_patch/model/mixtral_bak/transformer/attention.py (425:479) duplicated block id: 781 size: 38 cleaned lines of code in 2 files: - megatron_patch/tokenizer/tokenization_baichuan.py (72:123) - megatron_patch/tokenizer/tokenization_yi.py (85:136) duplicated block id: 782 size: 38 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/language_model.py (261:311) - megatron_patch/model/mistral/language_model.py (248:298) duplicated block id: 783 size: 38 cleaned lines of code in 2 files: - megatron_patch/model/qwen1_5/transformer/attention.py (541:595) - megatron_patch/model/qwen2_vl/attention.py (690:744) duplicated block id: 784 size: 38 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/language_model.py (202:256) - megatron_patch/model/llama3/language_model.py (182:236) duplicated block id: 785 size: 38 cleaned lines of code in 2 files: - megatron_patch/model/llama2/rotary_pos_embedding.py (40:93) - megatron_patch/model/llava/rotary_pos_embedding.py (38:91) duplicated block id: 786 size: 38 cleaned lines of code in 2 files: - megatron_patch/model/llama3/language_model.py (182:236) - megatron_patch/model/qwen1_5_megablocks/language_model.py (191:245) duplicated block id: 787 size: 38 cleaned lines of code in 2 files: - megatron_patch/model/mixtral_bak/transformer/attention.py (425:479) - megatron_patch/model/qwen2_vl/attention_vision.py (689:743) duplicated block id: 788 size: 38 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer/attention.py (541:595) - megatron_patch/model/mixtral/transformer/attention.py (689:743) duplicated block id: 789 size: 38 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/language_model.py (261:311) - megatron_patch/model/llama2/language_model.py (247:297) duplicated block id: 790 size: 38 cleaned lines of code in 2 files: - megatron_patch/model/qwen2/transformer/attention.py (546:600) - megatron_patch/model/qwen2_vl/attention.py (690:744) duplicated block id: 791 size: 38 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/transformer/attention.py (689:743) - megatron_patch/model/qwen1_5/transformer/attention.py (541:595) duplicated block id: 792 size: 38 cleaned lines of code in 2 files: - megatron_patch/model/qwen1_5/transformer/attention.py (541:595) - megatron_patch/model/qwen2_vl/attention_vision.py (689:743) duplicated block id: 793 size: 38 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/language_model.py (547:599) - megatron_patch/model/falcon40b/language_model.py (568:620) duplicated block id: 794 size: 38 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/language_model.py (303:353) - megatron_patch/model/glm130b/language_model.py (261:311) duplicated block id: 795 size: 38 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (634:684) - toolkits/model_checkpoints_convertor/falcon/checkpoint_reshaping_and_interoperability.py (577:627) duplicated block id: 796 size: 38 cleaned lines of code in 2 files: - megatron_patch/model/llama/positional_embeddings.py (35:88) - megatron_patch/model/qwen1_5_megablocks/rotary_pos_embedding.py (40:93) duplicated block id: 797 size: 38 cleaned lines of code in 2 files: - megatron_patch/model/llama3/language_model.py (182:236) - megatron_patch/model/qwen/language_model.py (192:246) duplicated block id: 798 size: 38 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/language_model.py (261:311) - megatron_patch/model/llava/language_model.py (250:300) duplicated block id: 799 size: 38 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/language_model.py (547:599) - megatron_patch/model/llama/language_model.py (579:631) duplicated block id: 800 size: 38 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/language_model.py (261:311) - megatron_patch/model/qwen/language_model.py (247:297) duplicated block id: 801 size: 37 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore.py (315:355) - toolkits/model_checkpoints_convertor/qwen/hf2megablocks_qwen1.5.py (308:348) duplicated block id: 802 size: 37 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (301:371) - toolkits/model_checkpoints_convertor/yi/checkpoint_reshaping_and_interoperability.py (288:358) duplicated block id: 803 size: 37 cleaned lines of code in 2 files: - megatron_patch/model/qwen1_5/moe/token_dispatcher.py (338:385) - megatron_patch/model/qwen2/moe/token_dispatcher.py (365:412) duplicated block id: 804 size: 37 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (301:371) - toolkits/model_checkpoints_convertor/galactica/checkpoint_reshaping_and_interoperability.py (290:360) duplicated block id: 805 size: 37 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (586:630) - megatron_patch/model/llama3/transformer_legacy.py (573:617) duplicated block id: 806 size: 37 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer_legacy.py (856:912) - megatron_patch/model/qwen1_5_megablocks/transformer.py (796:852) duplicated block id: 807 size: 37 cleaned lines of code in 2 files: - megatron_patch/model/bloom/language_model.py (305:359) - megatron_patch/model/glm130b/language_model.py (303:356) duplicated block id: 808 size: 36 cleaned lines of code in 2 files: - megatron_patch/model/llama3/layer_specs.py (25:74) - megatron_patch/model/qwen1_5/layer_specs.py (24:74) duplicated block id: 809 size: 36 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (292:367) - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (279:354) duplicated block id: 810 size: 36 cleaned lines of code in 2 files: - megatron_patch/model/llama3_1/layer_specs.py (25:74) - megatron_patch/model/qwen1_5/layer_specs.py (24:74) duplicated block id: 811 size: 36 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/model.py (100:152) - megatron_patch/model/mixtral_bak/model.py (96:148) duplicated block id: 812 size: 36 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/transformer.py (399:450) - megatron_patch/model/galactica/transformer.py (415:465) duplicated block id: 813 size: 36 cleaned lines of code in 2 files: - megatron_patch/model/llama3/language_model.py (565:614) - megatron_patch/model/starcoder/language_model.py (514:563) duplicated block id: 814 size: 36 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/transformer_block.py (548:603) - megatron_patch/model/qwen2/transformer_block.py (440:482) duplicated block id: 815 size: 36 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/language_model.py (303:356) - megatron_patch/model/glm130b/language_model.py (303:355) duplicated block id: 816 size: 36 cleaned lines of code in 2 files: - megatron_patch/model/falcon/transformer.py (782:824) - megatron_patch/model/falcon40b/transformer.py (569:607) duplicated block id: 817 size: 36 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (301:370) - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (339:408) duplicated block id: 818 size: 35 cleaned lines of code in 2 files: - megatron_patch/model/qwen1_5/layer_specs.py (37:74) - megatron_patch/model/qwen2/layer_specs.py (47:84) duplicated block id: 819 size: 35 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (725:770) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (740:785) duplicated block id: 820 size: 35 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/transformer/attention.py (593:637) - megatron_patch/model/qwen1_5/transformer/attention.py (450:494) duplicated block id: 821 size: 35 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer_legacy.py (43:116) - megatron_patch/model/qwen/transformer.py (34:106) duplicated block id: 822 size: 35 cleaned lines of code in 2 files: - megatron_patch/model/mistral/transformer.py (575:616) - megatron_patch/model/qwen1_5_megablocks/transformer.py (465:506) duplicated block id: 823 size: 35 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/gpt_model.py (79:126) - megatron_patch/model/glm130b/gpt_model.py (77:124) duplicated block id: 824 size: 35 cleaned lines of code in 2 files: - megatron_patch/model/bloom/language_model.py (169:217) - megatron_patch/model/llama/language_model.py (167:215) duplicated block id: 825 size: 35 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer_legacy.py (573:614) - megatron_patch/model/llava/transformer.py (575:616) duplicated block id: 826 size: 35 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (586:627) - megatron_patch/model/qwen1_5_megablocks/transformer.py (465:506) duplicated block id: 827 size: 35 cleaned lines of code in 2 files: - megatron_patch/model/mistral/transformer.py (39:112) - megatron_patch/model/qwen/transformer.py (34:106) duplicated block id: 828 size: 35 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer/attention.py (450:494) - megatron_patch/model/mixtral/transformer/attention.py (593:637) duplicated block id: 829 size: 35 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (1080:1128) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (975:1023) duplicated block id: 830 size: 35 cleaned lines of code in 2 files: - megatron_patch/model/llama2/transformer.py (574:615) - megatron_patch/model/llama3/transformer_legacy.py (573:614) duplicated block id: 831 size: 35 cleaned lines of code in 2 files: - megatron_patch/model/llama3_1/layer_specs.py (38:74) - megatron_patch/model/qwen2/layer_specs.py (47:84) duplicated block id: 832 size: 35 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/mlp.py (96:140) - megatron_patch/model/qwen2/transformer/mlp.py (117:161) duplicated block id: 833 size: 35 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon40b/checkpoint_reshaping_and_interoperability.py (459:507) - toolkits/model_checkpoints_convertor/galactica/checkpoint_reshaping_and_interoperability.py (517:565) duplicated block id: 834 size: 35 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/glm/checkpoint_reshaping_and_interoperability.py (416:461) - toolkits/model_checkpoints_convertor/glm130b/checkpoint_reshaping_and_interoperability.py (397:442) duplicated block id: 835 size: 35 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (34:107) - megatron_patch/model/qwen/transformer.py (34:106) duplicated block id: 836 size: 35 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (467:511) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (454:498) duplicated block id: 837 size: 35 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (881:928) - toolkits/model_checkpoints_convertor/falcon/checkpoint_reshaping_and_interoperability.py (808:855) duplicated block id: 838 size: 35 cleaned lines of code in 2 files: - megatron_patch/model/llava/transformer.py (39:112) - megatron_patch/model/qwen/transformer.py (34:106) duplicated block id: 839 size: 35 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore_llama3_1.py (202:239) - toolkits/model_checkpoints_convertor/llava/hf2mcore_llava.py (222:258) duplicated block id: 840 size: 35 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon40b/checkpoint_reshaping_and_interoperability.py (459:507) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (539:586) duplicated block id: 841 size: 35 cleaned lines of code in 2 files: - megatron_patch/model/qwen/transformer.py (34:106) - megatron_patch/model/qwen_vl/transformer.py (39:112) duplicated block id: 842 size: 35 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer_legacy.py (573:614) - megatron_patch/model/mistral/transformer.py (575:616) duplicated block id: 843 size: 35 cleaned lines of code in 2 files: - megatron_patch/model/llama2/transformer.py (39:112) - megatron_patch/model/qwen/transformer.py (34:106) duplicated block id: 844 size: 35 cleaned lines of code in 2 files: - megatron_patch/model/qwen1_5/transformer/attention.py (450:494) - megatron_patch/model/qwen2_vl/attention_vision.py (593:637) duplicated block id: 845 size: 35 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (907:954) - toolkits/model_checkpoints_convertor/falcon40b/checkpoint_reshaping_and_interoperability.py (784:832) duplicated block id: 846 size: 35 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer/attention.py (450:494) - megatron_patch/model/qwen2_vl/attention.py (594:638) duplicated block id: 847 size: 35 cleaned lines of code in 2 files: - megatron_patch/model/bloom/language_model.py (169:217) - megatron_patch/model/falcon/language_model.py (167:215) duplicated block id: 848 size: 35 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (881:928) - toolkits/model_checkpoints_convertor/falcon40b/checkpoint_reshaping_and_interoperability.py (784:832) duplicated block id: 849 size: 35 cleaned lines of code in 2 files: - megatron_patch/model/qwen/transformer.py (34:106) - megatron_patch/model/qwen1_5_megablocks/transformer.py (37:110) duplicated block id: 850 size: 35 cleaned lines of code in 2 files: - megatron_patch/model/qwen1_5/transformer/attention.py (450:494) - megatron_patch/model/qwen2_vl/attention.py (594:638) duplicated block id: 851 size: 35 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (477:518) - megatron_patch/model/qwen1_5_megablocks/transformer.py (465:506) duplicated block id: 852 size: 35 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer/attention.py (450:494) - megatron_patch/model/qwen2_vl/attention_vision.py (593:637) duplicated block id: 853 size: 35 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer_legacy.py (573:614) - megatron_patch/model/qwen_vl/transformer.py (575:616) duplicated block id: 854 size: 35 cleaned lines of code in 2 files: - megatron_patch/model/bloom/language_model.py (169:217) - megatron_patch/model/falcon40b/language_model.py (167:215) duplicated block id: 855 size: 35 cleaned lines of code in 2 files: - megatron_patch/model/llama3/layer_specs.py (37:74) - megatron_patch/model/qwen2/layer_specs.py (47:84) duplicated block id: 856 size: 35 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (907:954) - toolkits/model_checkpoints_convertor/falcon/checkpoint_reshaping_and_interoperability.py (808:855) duplicated block id: 857 size: 34 cleaned lines of code in 2 files: - megatron_patch/model/bloom/language_model.py (538:583) - megatron_patch/model/falcon/language_model.py (630:675) duplicated block id: 858 size: 34 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/moe/experts.py (787:827) - megatron_patch/model/qwen2/moe/experts.py (382:422) duplicated block id: 859 size: 34 cleaned lines of code in 2 files: - megatron_patch/model/llama2/transformer.py (774:829) - megatron_patch/model/mistral/transformer.py (772:826) duplicated block id: 860 size: 34 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (153:197) - megatron_patch/model/galactica/transformer.py (124:168) duplicated block id: 861 size: 34 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/language_model.py (604:649) - megatron_patch/model/falcon/language_model.py (630:675) duplicated block id: 862 size: 34 cleaned lines of code in 2 files: - megatron_patch/model/llama2/transformer.py (1825:1874) - megatron_patch/model/qwen1_5_megablocks/transformer.py (1686:1735) duplicated block id: 863 size: 34 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/moe/experts.py (568:611) - megatron_patch/model/qwen2/moe/experts.py (202:245) duplicated block id: 864 size: 34 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (153:197) - megatron_patch/model/falcon40b/transformer.py (174:218) duplicated block id: 865 size: 34 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (607:661) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (611:665) duplicated block id: 866 size: 34 cleaned lines of code in 2 files: - megatron_patch/model/llava/transformer.py (773:828) - megatron_patch/model/mistral/transformer.py (772:826) duplicated block id: 867 size: 34 cleaned lines of code in 2 files: - megatron_patch/model/mistral/transformer.py (772:826) - megatron_patch/model/qwen_vl/transformer.py (773:828) duplicated block id: 868 size: 34 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (639:678) - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (782:821) duplicated block id: 869 size: 34 cleaned lines of code in 2 files: - megatron_patch/model/falcon/language_model.py (630:675) - megatron_patch/model/glm130b/language_model.py (568:613) duplicated block id: 870 size: 34 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/language_model.py (604:649) - megatron_patch/model/falcon40b/language_model.py (630:675) duplicated block id: 871 size: 34 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/language_model.py (630:675) - megatron_patch/model/glm130b/language_model.py (568:613) duplicated block id: 872 size: 34 cleaned lines of code in 2 files: - megatron_patch/model/mistral/transformer.py (772:826) - megatron_patch/model/qwen1_5_megablocks/transformer.py (665:720) duplicated block id: 873 size: 34 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/moe/experts.py (788:828) - megatron_patch/model/qwen2/moe/experts.py (382:422) duplicated block id: 874 size: 34 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/transformer.py (105:150) - megatron_patch/model/llama/transformer.py (163:208) duplicated block id: 875 size: 34 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon/checkpoint_reshaping_and_interoperability.py (470:515) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (539:584) duplicated block id: 876 size: 34 cleaned lines of code in 2 files: - megatron_patch/model/bloom/language_model.py (538:583) - megatron_patch/model/falcon40b/language_model.py (630:675) duplicated block id: 877 size: 34 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/moe/experts.py (567:610) - megatron_patch/model/qwen2/moe/experts.py (202:245) duplicated block id: 878 size: 34 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon/checkpoint_reshaping_and_interoperability.py (470:515) - toolkits/model_checkpoints_convertor/galactica/checkpoint_reshaping_and_interoperability.py (517:562) duplicated block id: 879 size: 33 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/transformer.py (108:150) - megatron_patch/model/galactica/transformer.py (128:170) duplicated block id: 880 size: 33 cleaned lines of code in 2 files: - megatron_patch/model/falcon/transformer.py (1233:1299) - megatron_patch/model/llama/transformer.py (1027:1093) duplicated block id: 881 size: 33 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (582:624) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (656:698) duplicated block id: 882 size: 33 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (288:354) - toolkits/model_checkpoints_convertor/yi/checkpoint_reshaping_and_interoperability.py (288:354) duplicated block id: 883 size: 33 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/transformer.py (108:150) - megatron_patch/model/falcon40b/transformer.py (178:220) duplicated block id: 884 size: 33 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/chatglm/checkpoint_reshaping_and_interoperability.py (100:138) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (145:182) duplicated block id: 885 size: 33 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore.py (510:542) - toolkits/model_checkpoints_convertor/mistral/hf2mcore.py (417:449) duplicated block id: 886 size: 33 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/checkpoint_reshaping_and_interoperability.py (438:487) - toolkits/model_checkpoints_convertor/starcoder/checkpoint_reshaping_and_interoperability.py (451:501) duplicated block id: 887 size: 33 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (288:354) - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (339:405) duplicated block id: 888 size: 33 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (660:702) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (743:785) duplicated block id: 889 size: 33 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/chatglm/checkpoint_reshaping_and_interoperability.py (100:138) - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (140:177) duplicated block id: 890 size: 33 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/transformer.py (598:642) - megatron_patch/model/llama/transformer.py (586:632) duplicated block id: 891 size: 33 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2.5_vl.py (488:526) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_vl.py (490:528) duplicated block id: 892 size: 33 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (740:781) - toolkits/model_checkpoints_convertor/yi/checkpoint_reshaping_and_interoperability.py (625:666) duplicated block id: 893 size: 33 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/reward_model_to_megatron.py (437:486) - toolkits/model_checkpoints_convertor/starcoder/checkpoint_reshaping_and_interoperability.py (451:501) duplicated block id: 894 size: 33 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (116:153) - toolkits/model_checkpoints_convertor/chatglm/checkpoint_reshaping_and_interoperability.py (100:138) duplicated block id: 895 size: 33 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/transformer.py (972:1038) - megatron_patch/model/llama/transformer.py (1027:1093) duplicated block id: 896 size: 33 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon40b/checkpoint_reshaping_and_interoperability.py (381:426) - toolkits/model_checkpoints_convertor/galactica/checkpoint_reshaping_and_interoperability.py (429:475) duplicated block id: 897 size: 33 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/chatglm/checkpoint_reshaping_and_interoperability.py (100:138) - toolkits/model_checkpoints_convertor/galactica/checkpoint_reshaping_and_interoperability.py (116:155) duplicated block id: 898 size: 33 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (288:354) - toolkits/model_checkpoints_convertor/galactica/checkpoint_reshaping_and_interoperability.py (290:356) duplicated block id: 899 size: 33 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/chatglm/checkpoint_reshaping_and_interoperability.py (100:138) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (144:181) duplicated block id: 900 size: 32 cleaned lines of code in 2 files: - megatron_patch/model/llama2/language_model.py (482:519) - megatron_patch/model/llava/language_model.py (542:579) duplicated block id: 901 size: 32 cleaned lines of code in 2 files: - megatron_patch/model/qwen/transformer.py (568:606) - megatron_patch/model/qwen_vl/transformer.py (575:613) duplicated block id: 902 size: 32 cleaned lines of code in 2 files: - megatron_patch/model/mistral/transformer.py (575:613) - megatron_patch/model/qwen/transformer.py (568:606) duplicated block id: 903 size: 32 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (116:151) - toolkits/model_checkpoints_convertor/chatglm/checkpoint_reshaping_and_interoperability.py (100:134) duplicated block id: 904 size: 32 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/galactica/checkpoint_reshaping_and_interoperability.py (432:475) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (493:535) duplicated block id: 905 size: 32 cleaned lines of code in 2 files: - megatron_patch/model/llama3_1/model.py (80:119) - megatron_patch/model/qwen2/model.py (66:105) duplicated block id: 906 size: 32 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer/attention.py (412:446) - megatron_patch/model/qwen2/transformer/attention.py (416:450) duplicated block id: 907 size: 32 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore_llama3_1.py (150:182) - toolkits/model_checkpoints_convertor/llava/hf2mcore_llava.py (172:204) duplicated block id: 908 size: 32 cleaned lines of code in 2 files: - megatron_patch/model/qwen1_5/moe/experts.py (219:255) - megatron_patch/model/qwen2/moe/experts.py (391:427) duplicated block id: 909 size: 32 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_moe.py (359:390) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_dense_and_moe_gqa.py (611:642) duplicated block id: 910 size: 32 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/transformer.py (1010:1058) - megatron_patch/model/llama/transformer.py (920:968) duplicated block id: 911 size: 32 cleaned lines of code in 2 files: - megatron_patch/model/llava/transformer.py (575:613) - megatron_patch/model/qwen/transformer.py (568:606) duplicated block id: 912 size: 32 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/chatglm/checkpoint_reshaping_and_interoperability.py (100:134) - toolkits/model_checkpoints_convertor/falcon/checkpoint_reshaping_and_interoperability.py (117:152) duplicated block id: 913 size: 32 cleaned lines of code in 2 files: - megatron_patch/model/bloom/language_model.py (376:416) - megatron_patch/model/falcon40b/language_model.py (366:406) duplicated block id: 914 size: 32 cleaned lines of code in 2 files: - megatron_patch/model/bloom/language_model.py (376:416) - megatron_patch/model/galactica/language_model.py (397:437) duplicated block id: 915 size: 32 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/language_model.py (412:451) - megatron_patch/model/glm130b/language_model.py (427:466) duplicated block id: 916 size: 32 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/transformer.py (865:913) - megatron_patch/model/glm130b/transformer.py (1010:1058) duplicated block id: 917 size: 32 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (433:475) - toolkits/model_checkpoints_convertor/galactica/checkpoint_reshaping_and_interoperability.py (432:475) duplicated block id: 918 size: 32 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer_legacy.py (573:611) - megatron_patch/model/qwen/transformer.py (568:606) duplicated block id: 919 size: 32 cleaned lines of code in 2 files: - megatron_patch/model/bloom/language_model.py (408:447) - megatron_patch/model/chatglm/language_model.py (412:451) duplicated block id: 920 size: 32 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/transformer.py (750:798) - megatron_patch/model/glm130b/transformer.py (1010:1058) duplicated block id: 921 size: 32 cleaned lines of code in 2 files: - megatron_patch/model/qwen2_moe/layer_specs.py (68:108) - megatron_patch/model/qwen3_moe/moe_module_specs.py (33:73) duplicated block id: 922 size: 32 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon/checkpoint_reshaping_and_interoperability.py (395:437) - toolkits/model_checkpoints_convertor/galactica/checkpoint_reshaping_and_interoperability.py (432:475) duplicated block id: 923 size: 32 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/galactica/checkpoint_reshaping_and_interoperability.py (432:475) - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (486:528) duplicated block id: 924 size: 32 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (586:624) - megatron_patch/model/qwen/transformer.py (568:606) duplicated block id: 925 size: 32 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (288:351) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (358:421) duplicated block id: 926 size: 32 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (301:364) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (358:421) duplicated block id: 927 size: 32 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (586:625) - toolkits/model_checkpoints_convertor/yi/checkpoint_reshaping_and_interoperability.py (628:667) duplicated block id: 928 size: 32 cleaned lines of code in 2 files: - megatron_patch/model/galactica/transformer.py (754:802) - megatron_patch/model/glm130b/transformer.py (1010:1058) duplicated block id: 929 size: 32 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/galactica/checkpoint_reshaping_and_interoperability.py (432:475) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (463:505) duplicated block id: 930 size: 32 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/chatglm/checkpoint_reshaping_and_interoperability.py (100:134) - toolkits/model_checkpoints_convertor/falcon40b/checkpoint_reshaping_and_interoperability.py (117:152) duplicated block id: 931 size: 32 cleaned lines of code in 2 files: - megatron_patch/model/llava/language_model.py (542:579) - megatron_patch/model/qwen1_5_megablocks/language_model.py (481:518) duplicated block id: 932 size: 32 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_mha_to_moe.py (210:241) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_dense_and_moe_gqa.py (611:642) duplicated block id: 933 size: 32 cleaned lines of code in 2 files: - megatron_patch/model/bloom/language_model.py (376:416) - megatron_patch/model/llama/language_model.py (366:406) duplicated block id: 934 size: 32 cleaned lines of code in 2 files: - megatron_patch/model/bloom/language_model.py (376:416) - megatron_patch/model/falcon/language_model.py (366:406) duplicated block id: 935 size: 32 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (477:515) - megatron_patch/model/qwen/transformer.py (568:606) duplicated block id: 936 size: 32 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (157:197) - megatron_patch/model/llama/transformer.py (166:206) duplicated block id: 937 size: 32 cleaned lines of code in 2 files: - megatron_patch/model/llama2/transformer.py (574:612) - megatron_patch/model/qwen/transformer.py (568:606) duplicated block id: 938 size: 32 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/deepseek/hf2mcore_deepseek_v2_moe.py (22:70) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_dense_and_moe_gqa.py (9:58) duplicated block id: 939 size: 32 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/galactica/checkpoint_reshaping_and_interoperability.py (432:475) - toolkits/model_checkpoints_convertor/yi/checkpoint_reshaping_and_interoperability.py (408:450) duplicated block id: 940 size: 32 cleaned lines of code in 2 files: - megatron_patch/model/falcon/transformer.py (1126:1174) - megatron_patch/model/glm130b/transformer.py (1010:1058) duplicated block id: 941 size: 32 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/chatglm/checkpoint_reshaping_and_interoperability.py (100:134) - toolkits/model_checkpoints_convertor/yi/checkpoint_reshaping_and_interoperability.py (117:152) duplicated block id: 942 size: 32 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/layer_specs.py (121:155) - megatron_patch/model/mixtral/layer_specs.py (155:189) duplicated block id: 943 size: 32 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (405:447) - toolkits/model_checkpoints_convertor/galactica/checkpoint_reshaping_and_interoperability.py (432:475) duplicated block id: 944 size: 32 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore.py (131:170) - toolkits/model_checkpoints_convertor/mistral/hf2mcore.py (99:138) duplicated block id: 945 size: 32 cleaned lines of code in 2 files: - megatron_patch/model/qwen/transformer.py (568:606) - megatron_patch/model/qwen1_5_megablocks/transformer.py (465:503) duplicated block id: 946 size: 32 cleaned lines of code in 2 files: - megatron_patch/model/qwen1_5/transformer/attention.py (412:446) - megatron_patch/model/qwen2/transformer/attention.py (416:450) duplicated block id: 947 size: 32 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (157:197) - megatron_patch/model/chatglm/transformer.py (108:148) duplicated block id: 948 size: 31 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (295:337) - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (358:400) duplicated block id: 949 size: 31 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (1822:1868) - megatron_patch/model/llava/transformer.py (1824:1870) duplicated block id: 950 size: 31 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/transformer.py (110:150) - megatron_patch/model/falcon/transformer.py (187:227) duplicated block id: 951 size: 31 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (1822:1868) - megatron_patch/model/qwen_vl/transformer.py (1824:1870) duplicated block id: 952 size: 31 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (586:624) - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (728:766) duplicated block id: 953 size: 31 cleaned lines of code in 2 files: - megatron_patch/model/falcon/transformer.py (429:497) - megatron_patch/model/starcoder/transformer.py (425:493) duplicated block id: 954 size: 31 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (660:698) - toolkits/model_checkpoints_convertor/yi/checkpoint_reshaping_and_interoperability.py (628:666) duplicated block id: 955 size: 31 cleaned lines of code in 2 files: - megatron_patch/model/qwen2_5_vl/visionmodel.py (18:53) - megatron_patch/model/qwen2_vl/visionmodel.py (18:53) duplicated block id: 956 size: 31 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (586:624) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (743:781) duplicated block id: 957 size: 30 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/hf2te.py (347:382) - toolkits/model_checkpoints_convertor/galactica/checkpoint_reshaping_and_interoperability.py (504:540) duplicated block id: 958 size: 30 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (646:695) - megatron_patch/model/llava/transformer.py (779:828) duplicated block id: 959 size: 30 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (1655:1689) - megatron_patch/model/mistral/transformer.py (1652:1686) duplicated block id: 960 size: 30 cleaned lines of code in 2 files: - megatron_patch/model/llama2/language_model.py (436:471) - megatron_patch/model/qwen_vl/language_model.py (436:471) duplicated block id: 961 size: 30 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (172:235) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (268:331) duplicated block id: 962 size: 30 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (211:274) - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (193:256) duplicated block id: 963 size: 30 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (290:351) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (338:399) duplicated block id: 964 size: 30 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (646:695) - megatron_patch/model/llama2/transformer.py (780:829) duplicated block id: 965 size: 30 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (255:318) - toolkits/model_checkpoints_convertor/yi/checkpoint_reshaping_and_interoperability.py (198:261) duplicated block id: 966 size: 30 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (213:276) - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (255:318) duplicated block id: 967 size: 30 cleaned lines of code in 2 files: - megatron_patch/model/mistral/transformer.py (778:826) - megatron_patch/model/qwen/transformer.py (739:788) duplicated block id: 968 size: 30 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/transformer.py (478:518) - megatron_patch/model/falcon40b/transformer.py (613:653) duplicated block id: 969 size: 30 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (303:364) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (338:399) duplicated block id: 970 size: 30 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore.py (91:125) - toolkits/model_checkpoints_convertor/mistral/hf2mcore.py (65:99) duplicated block id: 971 size: 30 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/checkpoint_reshaping_and_interoperability.py (689:728) - toolkits/model_checkpoints_convertor/starcoder/checkpoint_reshaping_and_interoperability.py (689:728) duplicated block id: 972 size: 30 cleaned lines of code in 2 files: - megatron_patch/model/qwen1_5_megablocks/language_model.py (435:470) - megatron_patch/model/qwen_vl/language_model.py (436:471) duplicated block id: 973 size: 30 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/reward_model_to_megatron.py (688:727) - toolkits/model_checkpoints_convertor/starcoder/checkpoint_reshaping_and_interoperability.py (689:728) duplicated block id: 974 size: 30 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (1620:1676) - megatron_patch/model/llama2/transformer.py (1751:1807) duplicated block id: 975 size: 30 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (757:806) - megatron_patch/model/qwen_vl/transformer.py (779:828) duplicated block id: 976 size: 30 cleaned lines of code in 2 files: - megatron_patch/model/mistral/language_model.py (438:473) - megatron_patch/model/qwen_vl/language_model.py (436:471) duplicated block id: 977 size: 30 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (211:274) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (172:235) duplicated block id: 978 size: 30 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (249:312) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (172:235) duplicated block id: 979 size: 30 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/chatglm/checkpoint_reshaping_and_interoperability.py (183:246) - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (193:256) duplicated block id: 980 size: 30 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (646:695) - megatron_patch/model/qwen_vl/transformer.py (779:828) duplicated block id: 981 size: 30 cleaned lines of code in 2 files: - megatron_patch/model/qwen2_moe/layer_specs.py (335:377) - megatron_patch/model/qwen3_moe/gpt_layer_specs.py (361:403) duplicated block id: 982 size: 30 cleaned lines of code in 2 files: - megatron_patch/model/llama3/language_model.py (427:462) - megatron_patch/model/qwen_vl/language_model.py (436:471) duplicated block id: 983 size: 30 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (1655:1689) - megatron_patch/model/qwen1_5_megablocks/transformer.py (1518:1551) duplicated block id: 984 size: 30 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (211:274) - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (255:318) duplicated block id: 985 size: 30 cleaned lines of code in 2 files: - megatron_patch/model/llava/transformer.py (1750:1806) - megatron_patch/model/qwen/transformer.py (1692:1748) duplicated block id: 986 size: 30 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (1750:1806) - megatron_patch/model/llava/transformer.py (1750:1806) duplicated block id: 987 size: 30 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (1525:1559) - megatron_patch/model/llava/transformer.py (1654:1688) duplicated block id: 988 size: 30 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer_legacy.py (1615:1649) - megatron_patch/model/llava/transformer.py (1654:1688) duplicated block id: 989 size: 30 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (193:256) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (268:331) duplicated block id: 990 size: 30 cleaned lines of code in 2 files: - megatron_patch/model/llava/transformer.py (779:828) - megatron_patch/model/qwen/transformer.py (739:788) duplicated block id: 991 size: 30 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/transformer.py (504:551) - megatron_patch/model/galactica/transformer.py (536:583) duplicated block id: 992 size: 30 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/chatglm/checkpoint_reshaping_and_interoperability.py (183:246) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (172:235) duplicated block id: 993 size: 30 cleaned lines of code in 2 files: - megatron_patch/model/falcon/transformer.py (878:921) - megatron_patch/model/llama/transformer.py (650:693) duplicated block id: 994 size: 30 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/gpt_model.py (61:98) - megatron_patch/model/llava/gpt_model.py (50:89) duplicated block id: 995 size: 30 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (193:256) - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (249:312) duplicated block id: 996 size: 30 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (757:806) - megatron_patch/model/llama2/transformer.py (780:829) duplicated block id: 997 size: 30 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon/checkpoint_reshaping_and_interoperability.py (559:596) - toolkits/model_checkpoints_convertor/yi/checkpoint_reshaping_and_interoperability.py (631:667) duplicated block id: 998 size: 30 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (255:318) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (245:308) duplicated block id: 999 size: 30 cleaned lines of code in 2 files: - megatron_patch/model/llama3/language_model.py (427:462) - megatron_patch/model/llava/language_model.py (444:479) duplicated block id: 1000 size: 30 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer_legacy.py (1710:1766) - megatron_patch/model/qwen_vl/transformer.py (1750:1806) duplicated block id: 1001 size: 30 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (249:312) - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (255:318) duplicated block id: 1002 size: 30 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (1620:1676) - megatron_patch/model/qwen_vl/transformer.py (1750:1806) duplicated block id: 1003 size: 30 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (1620:1676) - megatron_patch/model/mistral/transformer.py (1748:1804) duplicated block id: 1004 size: 30 cleaned lines of code in 2 files: - megatron_patch/model/mixtral_bak/layer_specs.py (106:152) - megatron_patch/model/qwen1_5/layer_specs.py (85:119) duplicated block id: 1005 size: 30 cleaned lines of code in 2 files: - megatron_patch/model/llama2/transformer.py (1655:1689) - megatron_patch/model/llama3/transformer_legacy.py (1615:1649) duplicated block id: 1006 size: 30 cleaned lines of code in 2 files: - megatron_patch/model/llama2/transformer.py (1751:1807) - megatron_patch/model/qwen/transformer.py (1692:1748) duplicated block id: 1007 size: 30 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer_legacy.py (1615:1649) - megatron_patch/model/qwen_vl/transformer.py (1654:1688) duplicated block id: 1008 size: 30 cleaned lines of code in 2 files: - megatron_patch/model/llava/transformer.py (1654:1688) - megatron_patch/model/qwen/transformer.py (1597:1631) duplicated block id: 1009 size: 30 cleaned lines of code in 2 files: - megatron_patch/model/mistral/transformer.py (1748:1804) - megatron_patch/model/qwen/transformer.py (1692:1748) duplicated block id: 1010 size: 30 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/language_model.py (443:478) - megatron_patch/model/llava/language_model.py (444:479) duplicated block id: 1011 size: 30 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/galactica/checkpoint_reshaping_and_interoperability.py (292:353) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (338:399) duplicated block id: 1012 size: 30 cleaned lines of code in 2 files: - megatron_patch/model/qwen/transformer.py (739:788) - megatron_patch/model/qwen1_5_megablocks/transformer.py (671:720) duplicated block id: 1013 size: 30 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer_legacy.py (1710:1766) - megatron_patch/model/llava/transformer.py (1750:1806) duplicated block id: 1014 size: 30 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (646:695) - megatron_patch/model/mistral/transformer.py (778:826) duplicated block id: 1015 size: 30 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (646:695) - megatron_patch/model/qwen1_5_megablocks/transformer.py (671:720) duplicated block id: 1016 size: 30 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/language_model.py (443:478) - megatron_patch/model/qwen_vl/language_model.py (436:471) duplicated block id: 1017 size: 30 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (1525:1559) - megatron_patch/model/qwen_vl/transformer.py (1654:1688) duplicated block id: 1018 size: 30 cleaned lines of code in 2 files: - megatron_patch/model/qwen/transformer.py (739:788) - megatron_patch/model/qwen_vl/transformer.py (779:828) duplicated block id: 1019 size: 30 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (757:806) - megatron_patch/model/qwen1_5_megablocks/transformer.py (671:720) duplicated block id: 1020 size: 30 cleaned lines of code in 2 files: - megatron_patch/model/qwen/transformer.py (1597:1631) - megatron_patch/model/qwen1_5_megablocks/transformer.py (1518:1551) duplicated block id: 1021 size: 30 cleaned lines of code in 2 files: - megatron_patch/model/qwen/transformer.py (1692:1748) - megatron_patch/model/qwen_vl/transformer.py (1750:1806) duplicated block id: 1022 size: 30 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (1750:1806) - megatron_patch/model/mistral/transformer.py (1748:1804) duplicated block id: 1023 size: 30 cleaned lines of code in 2 files: - megatron_patch/model/qwen/language_model.py (433:468) - megatron_patch/model/qwen_vl/language_model.py (436:471) duplicated block id: 1024 size: 30 cleaned lines of code in 2 files: - megatron_patch/model/llava/language_model.py (444:479) - megatron_patch/model/qwen1_5_megablocks/language_model.py (435:470) duplicated block id: 1025 size: 30 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (172:235) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (245:308) duplicated block id: 1026 size: 30 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (213:276) - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (193:256) duplicated block id: 1027 size: 30 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (159:197) - megatron_patch/model/falcon/transformer.py (187:225) duplicated block id: 1028 size: 30 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (1655:1689) - megatron_patch/model/llama2/transformer.py (1655:1689) duplicated block id: 1029 size: 30 cleaned lines of code in 2 files: - megatron_patch/model/llama2/language_model.py (436:471) - megatron_patch/model/llava/language_model.py (444:479) duplicated block id: 1030 size: 30 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer_legacy.py (1710:1766) - megatron_patch/model/mistral/transformer.py (1748:1804) duplicated block id: 1031 size: 30 cleaned lines of code in 2 files: - megatron_patch/model/llama2/transformer.py (1751:1807) - megatron_patch/model/llama3/transformer_legacy.py (1710:1766) duplicated block id: 1032 size: 30 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer_legacy.py (1615:1649) - megatron_patch/model/qwen1_5_megablocks/transformer.py (1518:1551) duplicated block id: 1033 size: 30 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (213:276) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (172:235) duplicated block id: 1034 size: 30 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (1750:1806) - megatron_patch/model/qwen_vl/transformer.py (1750:1806) duplicated block id: 1035 size: 30 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/transformer_layer.py (314:372) - megatron_patch/model/qwen2/transformer_layer.py (193:251) duplicated block id: 1036 size: 30 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (757:806) - megatron_patch/model/llava/transformer.py (779:828) duplicated block id: 1037 size: 30 cleaned lines of code in 2 files: - megatron_patch/model/qwen/transformer.py (1597:1631) - megatron_patch/model/qwen_vl/transformer.py (1654:1688) duplicated block id: 1038 size: 30 cleaned lines of code in 2 files: - megatron_patch/model/llava/language_model.py (444:479) - megatron_patch/model/qwen/language_model.py (433:468) duplicated block id: 1039 size: 30 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (1525:1559) - megatron_patch/model/llama2/transformer.py (1655:1689) duplicated block id: 1040 size: 30 cleaned lines of code in 2 files: - megatron_patch/model/mistral/transformer.py (1652:1686) - megatron_patch/model/qwen/transformer.py (1597:1631) duplicated block id: 1041 size: 30 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (193:256) - toolkits/model_checkpoints_convertor/yi/checkpoint_reshaping_and_interoperability.py (198:261) duplicated block id: 1042 size: 30 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (1655:1689) - megatron_patch/model/llava/transformer.py (1654:1688) duplicated block id: 1043 size: 30 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (1525:1559) - megatron_patch/model/qwen1_5_megablocks/transformer.py (1518:1551) duplicated block id: 1044 size: 30 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (341:402) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (338:399) duplicated block id: 1045 size: 30 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (1620:1676) - megatron_patch/model/llava/transformer.py (1750:1806) duplicated block id: 1046 size: 30 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (172:235) - toolkits/model_checkpoints_convertor/yi/checkpoint_reshaping_and_interoperability.py (198:261) duplicated block id: 1047 size: 30 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/galactica/checkpoint_reshaping_and_interoperability.py (200:263) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (172:235) duplicated block id: 1048 size: 30 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/hf2te.py (319:354) - toolkits/model_checkpoints_convertor/galactica/checkpoint_reshaping_and_interoperability.py (504:540) duplicated block id: 1049 size: 30 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/chatglm/checkpoint_reshaping_and_interoperability.py (183:246) - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (255:318) duplicated block id: 1050 size: 30 cleaned lines of code in 2 files: - megatron_patch/model/llama2/transformer.py (780:829) - megatron_patch/model/qwen/transformer.py (739:788) duplicated block id: 1051 size: 30 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (1655:1689) - megatron_patch/model/qwen_vl/transformer.py (1654:1688) duplicated block id: 1052 size: 30 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon40b/checkpoint_reshaping_and_interoperability.py (563:599) - toolkits/model_checkpoints_convertor/yi/checkpoint_reshaping_and_interoperability.py (631:667) duplicated block id: 1053 size: 30 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (338:399) - toolkits/model_checkpoints_convertor/yi/checkpoint_reshaping_and_interoperability.py (290:351) duplicated block id: 1054 size: 30 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer_legacy.py (1615:1649) - megatron_patch/model/mistral/transformer.py (1652:1686) duplicated block id: 1055 size: 30 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (1750:1806) - megatron_patch/model/llama2/transformer.py (1751:1807) duplicated block id: 1056 size: 30 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/galactica/checkpoint_reshaping_and_interoperability.py (200:263) - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (193:256) duplicated block id: 1057 size: 30 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (338:399) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (360:421) duplicated block id: 1058 size: 30 cleaned lines of code in 2 files: - megatron_patch/model/llama2/transformer.py (1655:1689) - megatron_patch/model/qwen/transformer.py (1597:1631) duplicated block id: 1059 size: 30 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (193:256) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (245:308) duplicated block id: 1060 size: 30 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/galactica/checkpoint_reshaping_and_interoperability.py (200:263) - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (255:318) duplicated block id: 1061 size: 30 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (757:806) - megatron_patch/model/mistral/transformer.py (778:826) duplicated block id: 1062 size: 30 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/gpt_model.py (61:98) - megatron_patch/model/qwen_vl/gpt_model.py (50:89) duplicated block id: 1063 size: 30 cleaned lines of code in 2 files: - megatron_patch/model/llava/language_model.py (444:479) - megatron_patch/model/mistral/language_model.py (438:473) duplicated block id: 1064 size: 30 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (255:318) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (268:331) duplicated block id: 1065 size: 30 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (1525:1559) - megatron_patch/model/mistral/transformer.py (1652:1686) duplicated block id: 1066 size: 29 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon40b/checkpoint_reshaping_and_interoperability.py (459:496) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (569:606) duplicated block id: 1067 size: 29 cleaned lines of code in 2 files: - megatron_patch/model/llava/gpt_model.py (97:133) - megatron_patch/model/qwen/gpt_model.py (96:132) duplicated block id: 1068 size: 29 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/glm/checkpoint_reshaping_and_interoperability.py (469:527) - toolkits/model_checkpoints_convertor/glm130b/checkpoint_reshaping_and_interoperability.py (445:484) duplicated block id: 1069 size: 29 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon/checkpoint_reshaping_and_interoperability.py (798:840) - toolkits/model_checkpoints_convertor/starcoder/checkpoint_reshaping_and_interoperability.py (510:552) duplicated block id: 1070 size: 29 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon/checkpoint_reshaping_and_interoperability.py (559:595) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (663:698) duplicated block id: 1071 size: 29 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (320:371) - toolkits/model_checkpoints_convertor/falcon/checkpoint_reshaping_and_interoperability.py (282:333) duplicated block id: 1072 size: 29 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (223:261) - megatron_patch/model/baichuan2/transformer.py (317:355) duplicated block id: 1073 size: 29 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon/checkpoint_reshaping_and_interoperability.py (282:333) - toolkits/model_checkpoints_convertor/galactica/checkpoint_reshaping_and_interoperability.py (309:360) duplicated block id: 1074 size: 29 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (509:546) - toolkits/model_checkpoints_convertor/falcon40b/checkpoint_reshaping_and_interoperability.py (459:496) duplicated block id: 1075 size: 29 cleaned lines of code in 2 files: - megatron_patch/model/llama3/gpt_model.py (98:134) - megatron_patch/model/llava/gpt_model.py (97:133) duplicated block id: 1076 size: 29 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (317:355) - megatron_patch/model/llava/transformer.py (320:358) duplicated block id: 1077 size: 29 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon40b/checkpoint_reshaping_and_interoperability.py (563:598) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (663:698) duplicated block id: 1078 size: 29 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/transformer/attention.py (409:453) - megatron_patch/model/qwen2_vl/attention_vision.py (407:451) duplicated block id: 1079 size: 29 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon/checkpoint_reshaping_and_interoperability.py (282:333) - toolkits/model_checkpoints_convertor/yi/checkpoint_reshaping_and_interoperability.py (307:358) duplicated block id: 1080 size: 29 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon/checkpoint_reshaping_and_interoperability.py (470:507) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (569:606) duplicated block id: 1081 size: 29 cleaned lines of code in 2 files: - megatron_patch/model/llava/gpt_model.py (97:133) - megatron_patch/model/mistral/gpt_model.py (96:132) duplicated block id: 1082 size: 29 cleaned lines of code in 2 files: - megatron_patch/model/llava/gpt_model.py (97:133) - megatron_patch/model/qwen1_5_megablocks/gpt_model.py (96:132) duplicated block id: 1083 size: 29 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon40b/checkpoint_reshaping_and_interoperability.py (459:496) - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (562:599) duplicated block id: 1084 size: 29 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer/attention.py (314:361) - megatron_patch/model/mixtral/transformer/attention.py (447:494) duplicated block id: 1085 size: 29 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (244:304) - toolkits/model_checkpoints_convertor/baichuan/hf2te.py (132:192) duplicated block id: 1086 size: 29 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (317:355) - megatron_patch/model/qwen/transformer.py (314:352) duplicated block id: 1087 size: 29 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/checkpoint_reshaping_and_interoperability.py (337:368) - toolkits/model_checkpoints_convertor/bloom/reward_model_to_megatron.py (337:368) duplicated block id: 1088 size: 29 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/gpt_model.py (122:158) - megatron_patch/model/qwen_vl/gpt_model.py (97:133) duplicated block id: 1089 size: 29 cleaned lines of code in 2 files: - megatron_patch/model/qwen1_5_megablocks/gpt_model.py (96:132) - megatron_patch/model/qwen_vl/gpt_model.py (97:133) duplicated block id: 1090 size: 29 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (317:355) - megatron_patch/model/mistral/transformer.py (320:358) duplicated block id: 1091 size: 29 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (481:518) - toolkits/model_checkpoints_convertor/falcon40b/checkpoint_reshaping_and_interoperability.py (459:496) duplicated block id: 1092 size: 29 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (828:867) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (930:969) duplicated block id: 1093 size: 29 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/transformer/attention.py (409:453) - megatron_patch/model/qwen2_vl/attention.py (408:452) duplicated block id: 1094 size: 29 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/transformer/attention.py (447:494) - megatron_patch/model/qwen2/transformer/attention.py (310:357) duplicated block id: 1095 size: 29 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/gpt_model.py (122:158) - megatron_patch/model/llava/gpt_model.py (97:133) duplicated block id: 1096 size: 29 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon40b/checkpoint_reshaping_and_interoperability.py (563:598) - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (731:766) duplicated block id: 1097 size: 29 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon40b/checkpoint_reshaping_and_interoperability.py (774:816) - toolkits/model_checkpoints_convertor/starcoder/checkpoint_reshaping_and_interoperability.py (510:552) duplicated block id: 1098 size: 29 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/model.py (118:160) - megatron_patch/model/qwen2/model.py (105:146) duplicated block id: 1099 size: 29 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (975:1015) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (1088:1128) duplicated block id: 1100 size: 29 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/hf2te.py (493:531) - toolkits/model_checkpoints_convertor/baichuan2/hf2te.py (465:503) duplicated block id: 1101 size: 29 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (481:518) - toolkits/model_checkpoints_convertor/falcon/checkpoint_reshaping_and_interoperability.py (470:507) duplicated block id: 1102 size: 29 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/hf2te.py (306:341) - toolkits/model_checkpoints_convertor/yi/checkpoint_reshaping_and_interoperability.py (458:494) duplicated block id: 1103 size: 29 cleaned lines of code in 2 files: - megatron_patch/model/llama2/gpt_model.py (96:132) - megatron_patch/model/llava/gpt_model.py (97:133) duplicated block id: 1104 size: 29 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/transformer.py (396:430) - megatron_patch/model/llama/transformer.py (457:491) duplicated block id: 1105 size: 29 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/deepseek/hf2mcore_deepseek_v2_moe.py (373:402) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_dense_and_moe_gqa.py (525:555) duplicated block id: 1106 size: 29 cleaned lines of code in 2 files: - megatron_patch/model/llama2/gpt_model.py (96:132) - megatron_patch/model/qwen_vl/gpt_model.py (97:133) duplicated block id: 1107 size: 29 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/hf2te.py (334:369) - toolkits/model_checkpoints_convertor/yi/checkpoint_reshaping_and_interoperability.py (458:494) duplicated block id: 1108 size: 29 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (317:355) - megatron_patch/model/qwen1_5_megablocks/transformer.py (211:249) duplicated block id: 1109 size: 29 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon/checkpoint_reshaping_and_interoperability.py (559:595) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (746:781) duplicated block id: 1110 size: 29 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon/checkpoint_reshaping_and_interoperability.py (559:595) - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (731:766) duplicated block id: 1111 size: 29 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (509:546) - toolkits/model_checkpoints_convertor/falcon/checkpoint_reshaping_and_interoperability.py (470:507) duplicated block id: 1112 size: 29 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/hf2te.py (51:82) - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (82:114) duplicated block id: 1113 size: 29 cleaned lines of code in 2 files: - megatron_patch/model/qwen/gpt_model.py (96:132) - megatron_patch/model/qwen_vl/gpt_model.py (97:133) duplicated block id: 1114 size: 29 cleaned lines of code in 2 files: - megatron_patch/model/llama3/gpt_model.py (98:134) - megatron_patch/model/qwen_vl/gpt_model.py (97:133) duplicated block id: 1115 size: 29 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore.py (388:421) - toolkits/model_checkpoints_convertor/mistral/hf2mcore.py (322:355) duplicated block id: 1116 size: 29 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (317:355) - megatron_patch/model/qwen_vl/transformer.py (320:358) duplicated block id: 1117 size: 29 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon/checkpoint_reshaping_and_interoperability.py (470:507) - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (562:599) duplicated block id: 1118 size: 29 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon40b/checkpoint_reshaping_and_interoperability.py (563:598) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (746:781) duplicated block id: 1119 size: 29 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/model.py (118:160) - megatron_patch/model/qwen1_5/model.py (104:145) duplicated block id: 1120 size: 29 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/gpt_model.py (105:142) - megatron_patch/model/llava/gpt_model.py (97:133) duplicated block id: 1121 size: 29 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (317:355) - megatron_patch/model/llama2/transformer.py (320:358) duplicated block id: 1122 size: 29 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/gpt_model.py (105:142) - megatron_patch/model/qwen_vl/gpt_model.py (97:133) duplicated block id: 1123 size: 29 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/model.py (118:160) - megatron_patch/model/llama3/model.py (104:145) duplicated block id: 1124 size: 29 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/transformer/attention.py (447:494) - megatron_patch/model/qwen1_5/transformer/attention.py (314:361) duplicated block id: 1125 size: 29 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (317:355) - megatron_patch/model/llama3/transformer_legacy.py (319:357) duplicated block id: 1126 size: 29 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/hf2te.py (50:81) - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (82:114) duplicated block id: 1127 size: 29 cleaned lines of code in 2 files: - megatron_patch/model/mistral/gpt_model.py (96:132) - megatron_patch/model/qwen_vl/gpt_model.py (97:133) duplicated block id: 1128 size: 28 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/model.py (121:160) - megatron_patch/model/qwen2_vl/gpt_model.py (98:136) duplicated block id: 1129 size: 28 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (520:553) - megatron_patch/model/llava/transformer.py (637:670) duplicated block id: 1130 size: 28 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore.py (10:55) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_mha.py (6:48) duplicated block id: 1131 size: 28 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon40b/checkpoint_reshaping_and_interoperability.py (281:331) - toolkits/model_checkpoints_convertor/yi/checkpoint_reshaping_and_interoperability.py (308:358) duplicated block id: 1132 size: 28 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer_legacy.py (616:649) - megatron_patch/model/qwen1_5_megablocks/transformer.py (529:562) duplicated block id: 1133 size: 28 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon40b/checkpoint_reshaping_and_interoperability.py (281:331) - toolkits/model_checkpoints_convertor/galactica/checkpoint_reshaping_and_interoperability.py (310:360) duplicated block id: 1134 size: 28 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer_legacy.py (223:261) - megatron_patch/model/llava/transformer.py (216:254) duplicated block id: 1135 size: 28 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore.py (10:55) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_moe.py (7:50) duplicated block id: 1136 size: 28 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer_legacy.py (118:149) - megatron_patch/model/qwen/transformer.py (108:139) duplicated block id: 1137 size: 28 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (89:118) - toolkits/model_checkpoints_convertor/baichuan2/hf2te.py (68:98) duplicated block id: 1138 size: 28 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon40b/checkpoint_reshaping_and_interoperability.py (518:556) - toolkits/model_checkpoints_convertor/galactica/checkpoint_reshaping_and_interoperability.py (574:612) duplicated block id: 1139 size: 28 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (832:865) - megatron_patch/model/mistral/transformer.py (962:995) duplicated block id: 1140 size: 28 cleaned lines of code in 2 files: - megatron_patch/model/llama2/transformer.py (855:893) - megatron_patch/model/llama3/transformer_legacy.py (806:844) duplicated block id: 1141 size: 28 cleaned lines of code in 2 files: - megatron_patch/model/qwen/transformer.py (108:139) - megatron_patch/model/qwen1_5_megablocks/transformer.py (112:143) duplicated block id: 1142 size: 28 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (707:745) - megatron_patch/model/llama3/transformer_legacy.py (806:844) duplicated block id: 1143 size: 28 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/moe/experts.py (796:827) - megatron_patch/model/qwen1_5/moe/experts.py (219:250) duplicated block id: 1144 size: 28 cleaned lines of code in 2 files: - megatron_patch/model/mixtral_bak/model.py (31:80) - megatron_patch/model/qwen1_5/model.py (34:83) duplicated block id: 1145 size: 28 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (520:553) - megatron_patch/model/mistral/transformer.py (637:670) duplicated block id: 1146 size: 28 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (857:895) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (930:968) duplicated block id: 1147 size: 28 cleaned lines of code in 2 files: - megatron_patch/model/falcon/transformer.py (925:970) - megatron_patch/model/galactica/transformer.py (586:624) duplicated block id: 1148 size: 28 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/hf2te.py (69:99) - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (113:142) duplicated block id: 1149 size: 28 cleaned lines of code in 2 files: - megatron_patch/model/llava/transformer.py (114:145) - megatron_patch/model/qwen/transformer.py (108:139) duplicated block id: 1150 size: 28 cleaned lines of code in 2 files: - megatron_patch/model/llama2/transformer.py (114:145) - megatron_patch/model/qwen/transformer.py (108:139) duplicated block id: 1151 size: 28 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon/checkpoint_reshaping_and_interoperability.py (282:332) - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (358:408) duplicated block id: 1152 size: 28 cleaned lines of code in 2 files: - megatron_patch/model/llama2/transformer.py (638:671) - megatron_patch/model/qwen/transformer.py (611:644) duplicated block id: 1153 size: 28 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/hf2te.py (68:98) - toolkits/model_checkpoints_convertor/falcon/checkpoint_reshaping_and_interoperability.py (90:119) duplicated block id: 1154 size: 28 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_moe.py (7:50) - toolkits/model_checkpoints_convertor/qwen/hf2megablocks_qwen1.5.py (8:52) duplicated block id: 1155 size: 28 cleaned lines of code in 2 files: - megatron_patch/model/llama2/transformer.py (216:254) - megatron_patch/model/llama3/transformer_legacy.py (223:261) duplicated block id: 1156 size: 28 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer_legacy.py (806:844) - megatron_patch/model/mistral/transformer.py (852:890) duplicated block id: 1157 size: 28 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/hf2te.py (69:99) - toolkits/model_checkpoints_convertor/galactica/checkpoint_reshaping_and_interoperability.py (89:118) duplicated block id: 1158 size: 28 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (832:865) - megatron_patch/model/llava/transformer.py (964:997) duplicated block id: 1159 size: 28 cleaned lines of code in 2 files: - megatron_patch/model/mistral/transformer.py (637:670) - megatron_patch/model/qwen/transformer.py (611:644) duplicated block id: 1160 size: 28 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/checkpoint_reshaping_and_interoperability.py (799:831) - toolkits/model_checkpoints_convertor/starcoder/checkpoint_reshaping_and_interoperability.py (808:840) duplicated block id: 1161 size: 28 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/reward_model_to_megatron.py (798:830) - toolkits/model_checkpoints_convertor/starcoder/checkpoint_reshaping_and_interoperability.py (808:840) duplicated block id: 1162 size: 28 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (520:553) - megatron_patch/model/qwen1_5_megablocks/transformer.py (529:562) duplicated block id: 1163 size: 28 cleaned lines of code in 2 files: - megatron_patch/model/qwen1_5/model.py (202:251) - megatron_patch/model/qwen2_vl/gpt_model.py (190:239) duplicated block id: 1164 size: 28 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/hf2te.py (69:99) - toolkits/model_checkpoints_convertor/falcon40b/checkpoint_reshaping_and_interoperability.py (90:119) duplicated block id: 1165 size: 28 cleaned lines of code in 2 files: - megatron_patch/model/llava/transformer.py (637:670) - megatron_patch/model/qwen/transformer.py (611:644) duplicated block id: 1166 size: 28 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (211:249) - megatron_patch/model/llama3/transformer_legacy.py (223:261) duplicated block id: 1167 size: 28 cleaned lines of code in 2 files: - megatron_patch/model/qwen1_5_megablocks/transformer.py (1686:1729) - megatron_patch/model/qwen_vl/transformer.py (1824:1867) duplicated block id: 1168 size: 28 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (520:553) - megatron_patch/model/qwen_vl/transformer.py (637:670) duplicated block id: 1169 size: 28 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore_llama3_1.py (93:133) - toolkits/model_checkpoints_convertor/llava/hf2mcore_llava.py (117:157) duplicated block id: 1170 size: 28 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon/checkpoint_reshaping_and_interoperability.py (283:333) - toolkits/model_checkpoints_convertor/falcon40b/checkpoint_reshaping_and_interoperability.py (281:331) duplicated block id: 1171 size: 28 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (109:140) - megatron_patch/model/qwen/transformer.py (108:139) duplicated block id: 1172 size: 28 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (321:371) - toolkits/model_checkpoints_convertor/falcon40b/checkpoint_reshaping_and_interoperability.py (281:331) duplicated block id: 1173 size: 28 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer_legacy.py (806:844) - megatron_patch/model/qwen/transformer.py (800:838) duplicated block id: 1174 size: 28 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore.py (57:88) - toolkits/model_checkpoints_convertor/mistral/hf2mcore.py (32:63) duplicated block id: 1175 size: 28 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/checkpoint_reshaping_and_interoperability.py (799:831) - toolkits/model_checkpoints_convertor/bloom/reward_model_to_megatron.py (798:830) duplicated block id: 1176 size: 28 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer_legacy.py (616:649) - megatron_patch/model/llava/transformer.py (637:670) duplicated block id: 1177 size: 28 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_mha_to_moe.py (6:47) - toolkits/model_checkpoints_convertor/qwen/hf2megablocks_qwen1.5.py (8:52) duplicated block id: 1178 size: 28 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (89:118) - toolkits/model_checkpoints_convertor/baichuan2/hf2te.py (68:98) duplicated block id: 1179 size: 28 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/hf2te.py (69:99) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (117:146) duplicated block id: 1180 size: 28 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/hf2te.py (68:98) - toolkits/model_checkpoints_convertor/galactica/checkpoint_reshaping_and_interoperability.py (89:118) duplicated block id: 1181 size: 28 cleaned lines of code in 2 files: - megatron_patch/model/llama3/model.py (202:251) - megatron_patch/model/qwen2_vl/gpt_model.py (190:239) duplicated block id: 1182 size: 28 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer_legacy.py (806:844) - megatron_patch/model/qwen1_5_megablocks/transformer.py (746:784) duplicated block id: 1183 size: 28 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/chatglm/checkpoint_reshaping_and_interoperability.py (561:594) - toolkits/model_checkpoints_convertor/yi/checkpoint_reshaping_and_interoperability.py (648:679) duplicated block id: 1184 size: 28 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/hf2te.py (69:99) - toolkits/model_checkpoints_convertor/yi/checkpoint_reshaping_and_interoperability.py (90:119) duplicated block id: 1185 size: 28 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/hf2te.py (68:98) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (118:147) duplicated block id: 1186 size: 28 cleaned lines of code in 2 files: - megatron_patch/model/qwen/transformer.py (611:644) - megatron_patch/model/qwen1_5_megablocks/transformer.py (529:562) duplicated block id: 1187 size: 28 cleaned lines of code in 2 files: - megatron_patch/model/qwen/transformer.py (611:644) - megatron_patch/model/qwen_vl/transformer.py (637:670) duplicated block id: 1188 size: 28 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/moe/experts.py (797:828) - megatron_patch/model/qwen1_5/moe/experts.py (219:250) duplicated block id: 1189 size: 28 cleaned lines of code in 2 files: - megatron_patch/model/llava/transformer.py (1824:1867) - megatron_patch/model/qwen1_5_megablocks/transformer.py (1686:1729) duplicated block id: 1190 size: 28 cleaned lines of code in 2 files: - megatron_patch/model/falcon/transformer.py (925:970) - megatron_patch/model/falcon40b/transformer.py (693:740) duplicated block id: 1191 size: 28 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/transformer/attention.py (448:494) - megatron_patch/model/mixtral_bak/transformer/attention.py (288:334) duplicated block id: 1192 size: 28 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/hf2te.py (69:99) - toolkits/model_checkpoints_convertor/falcon/checkpoint_reshaping_and_interoperability.py (90:119) duplicated block id: 1193 size: 28 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/hf2te.py (69:99) - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (89:118) duplicated block id: 1194 size: 28 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/mistral/hf2mcore.py (32:63) - toolkits/model_checkpoints_convertor/qwen/hf2megablocks_qwen1.5.py (54:87) duplicated block id: 1195 size: 28 cleaned lines of code in 2 files: - megatron_patch/model/qwen/transformer.py (108:139) - megatron_patch/model/qwen_vl/transformer.py (114:145) duplicated block id: 1196 size: 28 cleaned lines of code in 2 files: - megatron_patch/model/qwen1_5/moe/moe_layer.py (43:80) - megatron_patch/model/qwen2/moe/moe_layer.py (58:96) duplicated block id: 1197 size: 28 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/hf2te.py (69:99) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (118:147) duplicated block id: 1198 size: 28 cleaned lines of code in 2 files: - megatron_patch/model/llama2/transformer.py (638:671) - megatron_patch/model/llama3/transformer_legacy.py (616:649) duplicated block id: 1199 size: 28 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore.py (10:55) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_mha_to_moe.py (6:47) duplicated block id: 1200 size: 28 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer_legacy.py (806:844) - megatron_patch/model/llava/transformer.py (854:892) duplicated block id: 1201 size: 28 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (832:865) - megatron_patch/model/qwen_vl/transformer.py (964:997) duplicated block id: 1202 size: 28 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer_legacy.py (616:649) - megatron_patch/model/mistral/transformer.py (637:670) duplicated block id: 1203 size: 28 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/hf2te.py (68:98) - toolkits/model_checkpoints_convertor/falcon40b/checkpoint_reshaping_and_interoperability.py (90:119) duplicated block id: 1204 size: 28 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (89:118) - toolkits/model_checkpoints_convertor/baichuan/hf2te.py (69:99) duplicated block id: 1205 size: 28 cleaned lines of code in 2 files: - megatron_patch/model/bloom/language_model.py (337:378) - megatron_patch/model/starcoder/language_model.py (320:362) duplicated block id: 1206 size: 28 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer_legacy.py (223:261) - megatron_patch/model/mistral/transformer.py (216:254) duplicated block id: 1207 size: 28 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer_legacy.py (806:844) - megatron_patch/model/qwen_vl/transformer.py (854:892) duplicated block id: 1208 size: 28 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/hf2te.py (68:98) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (117:146) duplicated block id: 1209 size: 28 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_mha.py (6:48) - toolkits/model_checkpoints_convertor/qwen/hf2megablocks_qwen1.5.py (8:52) duplicated block id: 1210 size: 28 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (832:865) - megatron_patch/model/llama2/transformer.py (965:998) duplicated block id: 1211 size: 28 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/galactica/checkpoint_reshaping_and_interoperability.py (631:664) - toolkits/model_checkpoints_convertor/yi/checkpoint_reshaping_and_interoperability.py (648:679) duplicated block id: 1212 size: 28 cleaned lines of code in 2 files: - megatron_patch/model/falcon/transformer.py (647:680) - megatron_patch/model/falcon40b/transformer.py (437:470) duplicated block id: 1213 size: 28 cleaned lines of code in 2 files: - megatron_patch/model/llama3/model.py (34:83) - megatron_patch/model/mixtral_bak/model.py (31:80) duplicated block id: 1214 size: 28 cleaned lines of code in 2 files: - megatron_patch/model/mistral/transformer.py (114:145) - megatron_patch/model/qwen/transformer.py (108:139) duplicated block id: 1215 size: 28 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/hf2te.py (68:98) - toolkits/model_checkpoints_convertor/yi/checkpoint_reshaping_and_interoperability.py (90:119) duplicated block id: 1216 size: 28 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer_legacy.py (223:261) - megatron_patch/model/qwen/transformer.py (210:248) duplicated block id: 1217 size: 28 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer_legacy.py (616:649) - megatron_patch/model/qwen_vl/transformer.py (637:670) duplicated block id: 1218 size: 28 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer_legacy.py (223:261) - megatron_patch/model/qwen_vl/transformer.py (216:254) duplicated block id: 1219 size: 28 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (520:553) - megatron_patch/model/llama2/transformer.py (638:671) duplicated block id: 1220 size: 28 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/hf2te.py (68:98) - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (113:142) duplicated block id: 1221 size: 28 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (1822:1865) - megatron_patch/model/llama2/transformer.py (1825:1868) duplicated block id: 1222 size: 28 cleaned lines of code in 2 files: - megatron_patch/model/qwen2/model.py (203:252) - megatron_patch/model/qwen2_vl/gpt_model.py (190:239) duplicated block id: 1223 size: 28 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (832:865) - megatron_patch/model/qwen/transformer.py (910:943) duplicated block id: 1224 size: 28 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (589:634) - megatron_patch/model/chatglm/transformer.py (486:531) duplicated block id: 1225 size: 27 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore.py (289:315) - toolkits/model_checkpoints_convertor/mistral/hf2mcore.py (229:255) duplicated block id: 1226 size: 27 cleaned lines of code in 2 files: - megatron_patch/model/llava/language_model.py (654:687) - megatron_patch/model/starcoder/language_model.py (520:553) duplicated block id: 1227 size: 27 cleaned lines of code in 2 files: - megatron_patch/model/llama3/language_model.py (571:604) - megatron_patch/model/llava/language_model.py (654:687) duplicated block id: 1228 size: 27 cleaned lines of code in 2 files: - megatron_patch/model/bloom/gpt_model.py (28:65) - megatron_patch/model/chatglm/gpt_model.py (28:65) duplicated block id: 1229 size: 27 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/transformer_block.py (448:479) - megatron_patch/model/qwen2_5_vl/transformer_block.py (507:538) duplicated block id: 1230 size: 27 cleaned lines of code in 2 files: - megatron_patch/model/llama3/language_model.py (571:604) - megatron_patch/model/qwen1_5_megablocks/language_model.py (593:626) duplicated block id: 1231 size: 27 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer_legacy.py (153:187) - megatron_patch/model/qwen_vl/transformer.py (149:183) duplicated block id: 1232 size: 27 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/gpt_model.py (28:65) - megatron_patch/model/glm130b/gpt_model.py (28:65) duplicated block id: 1233 size: 27 cleaned lines of code in 2 files: - megatron_patch/model/bloom/gpt_model.py (89:125) - megatron_patch/model/chatglm/gpt_model.py (90:126) duplicated block id: 1234 size: 27 cleaned lines of code in 2 files: - megatron_patch/model/llama2/language_model.py (594:627) - megatron_patch/model/starcoder/language_model.py (520:553) duplicated block id: 1235 size: 27 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer_legacy.py (153:187) - megatron_patch/model/llava/transformer.py (149:183) duplicated block id: 1236 size: 27 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (152:182) - toolkits/model_checkpoints_convertor/starcoder/checkpoint_reshaping_and_interoperability.py (102:135) duplicated block id: 1237 size: 27 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/language_model.py (587:620) - megatron_patch/model/starcoder/language_model.py (520:553) duplicated block id: 1238 size: 27 cleaned lines of code in 2 files: - megatron_patch/model/llama3/language_model.py (571:604) - megatron_patch/model/mistral/language_model.py (606:639) duplicated block id: 1239 size: 27 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/mistral/hf2mcore.py (219:245) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_dense_and_moe_gqa.py (125:151) duplicated block id: 1240 size: 27 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/language_model.py (587:620) - megatron_patch/model/llama3/language_model.py (571:604) duplicated block id: 1241 size: 27 cleaned lines of code in 2 files: - megatron_patch/model/llava/transformer.py (1824:1865) - megatron_patch/model/qwen/transformer.py (1764:1805) duplicated block id: 1242 size: 27 cleaned lines of code in 2 files: - megatron_patch/model/qwen_vl/language_model.py (623:656) - megatron_patch/model/starcoder/language_model.py (520:553) duplicated block id: 1243 size: 27 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer_legacy.py (153:187) - megatron_patch/model/mistral/transformer.py (149:183) duplicated block id: 1244 size: 27 cleaned lines of code in 2 files: - megatron_patch/model/llama2/transformer.py (149:183) - megatron_patch/model/llama3/transformer_legacy.py (153:187) duplicated block id: 1245 size: 27 cleaned lines of code in 2 files: - megatron_patch/model/llama2/language_model.py (594:627) - megatron_patch/model/llama3/language_model.py (571:604) duplicated block id: 1246 size: 27 cleaned lines of code in 2 files: - megatron_patch/model/bloom/gpt_model.py (89:125) - megatron_patch/model/glm130b/gpt_model.py (88:124) duplicated block id: 1247 size: 27 cleaned lines of code in 2 files: - megatron_patch/model/llama2/transformer.py (1825:1866) - megatron_patch/model/qwen/transformer.py (1764:1805) duplicated block id: 1248 size: 27 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/language_model.py (671:704) - megatron_patch/model/llama3/language_model.py (571:604) duplicated block id: 1249 size: 27 cleaned lines of code in 2 files: - megatron_patch/model/qwen1_5_megablocks/language_model.py (593:626) - megatron_patch/model/starcoder/language_model.py (520:553) duplicated block id: 1250 size: 27 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/language_model.py (671:704) - megatron_patch/model/starcoder/language_model.py (520:553) duplicated block id: 1251 size: 27 cleaned lines of code in 2 files: - megatron_patch/model/qwen/language_model.py (576:609) - megatron_patch/model/starcoder/language_model.py (520:553) duplicated block id: 1252 size: 27 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/chatglm/checkpoint_reshaping_and_interoperability.py (447:485) - toolkits/model_checkpoints_convertor/falcon/checkpoint_reshaping_and_interoperability.py (478:515) duplicated block id: 1253 size: 27 cleaned lines of code in 2 files: - megatron_patch/model/llama3/language_model.py (571:604) - megatron_patch/model/qwen/language_model.py (576:609) duplicated block id: 1254 size: 27 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (574:610) - toolkits/model_checkpoints_convertor/yi/checkpoint_reshaping_and_interoperability.py (500:536) duplicated block id: 1255 size: 27 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer_legacy.py (153:187) - megatron_patch/model/qwen/transformer.py (143:177) duplicated block id: 1256 size: 27 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (1822:1863) - megatron_patch/model/mistral/transformer.py (1822:1863) duplicated block id: 1257 size: 27 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/chatglm/checkpoint_reshaping_and_interoperability.py (447:485) - toolkits/model_checkpoints_convertor/galactica/checkpoint_reshaping_and_interoperability.py (525:562) duplicated block id: 1258 size: 27 cleaned lines of code in 2 files: - megatron_patch/model/mistral/transformer.py (1822:1863) - megatron_patch/model/qwen/transformer.py (1764:1805) duplicated block id: 1259 size: 27 cleaned lines of code in 2 files: - megatron_patch/model/falcon/transformer.py (772:808) - megatron_patch/model/llama/transformer.py (545:580) duplicated block id: 1260 size: 27 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (581:617) - toolkits/model_checkpoints_convertor/yi/checkpoint_reshaping_and_interoperability.py (500:536) duplicated block id: 1261 size: 27 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/transformer_block.py (370:401) - megatron_patch/model/qwen2_5_vl/transformer_block.py (418:449) duplicated block id: 1262 size: 27 cleaned lines of code in 2 files: - megatron_patch/model/mistral/transformer.py (1822:1863) - megatron_patch/model/qwen1_5_megablocks/transformer.py (1686:1727) duplicated block id: 1263 size: 27 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/chatglm/checkpoint_reshaping_and_interoperability.py (447:485) - toolkits/model_checkpoints_convertor/falcon40b/checkpoint_reshaping_and_interoperability.py (467:504) duplicated block id: 1264 size: 27 cleaned lines of code in 2 files: - megatron_patch/model/mistral/language_model.py (606:639) - megatron_patch/model/starcoder/language_model.py (520:553) duplicated block id: 1265 size: 27 cleaned lines of code in 2 files: - megatron_patch/model/qwen/transformer.py (1764:1805) - megatron_patch/model/qwen_vl/transformer.py (1824:1865) duplicated block id: 1266 size: 27 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon40b/checkpoint_reshaping_and_interoperability.py (281:330) - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (359:408) duplicated block id: 1267 size: 27 cleaned lines of code in 2 files: - megatron_patch/model/llama3/language_model.py (571:604) - megatron_patch/model/qwen_vl/language_model.py (623:656) duplicated block id: 1268 size: 27 cleaned lines of code in 2 files: - megatron_patch/model/galactica/transformer.py (415:446) - megatron_patch/model/llama/transformer.py (460:491) duplicated block id: 1269 size: 27 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (144:178) - megatron_patch/model/llama3/transformer_legacy.py (153:187) duplicated block id: 1270 size: 27 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/chatglm/checkpoint_reshaping_and_interoperability.py (447:485) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (547:584) duplicated block id: 1271 size: 27 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_moe.py (212:239) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_moe.py (302:329) duplicated block id: 1272 size: 26 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (148:177) - toolkits/model_checkpoints_convertor/starcoder/checkpoint_reshaping_and_interoperability.py (102:134) duplicated block id: 1273 size: 26 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/checkpoint_reshaping_and_interoperability.py (102:134) - toolkits/model_checkpoints_convertor/chatglm/checkpoint_reshaping_and_interoperability.py (107:138) duplicated block id: 1274 size: 26 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/transformer.py (185:220) - megatron_patch/model/glm130b/transformer.py (198:233) duplicated block id: 1275 size: 26 cleaned lines of code in 2 files: - megatron_patch/model/llava/language_model.py (292:335) - megatron_patch/model/starcoder/language_model.py (290:333) duplicated block id: 1276 size: 26 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (124:153) - toolkits/model_checkpoints_convertor/bloom/checkpoint_reshaping_and_interoperability.py (102:134) duplicated block id: 1277 size: 26 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/language_model.py (345:388) - megatron_patch/model/starcoder/language_model.py (290:333) duplicated block id: 1278 size: 26 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (1318:1354) - megatron_patch/model/llava/transformer.py (1450:1486) duplicated block id: 1279 size: 26 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/checkpoint_reshaping_and_interoperability.py (102:134) - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (148:177) duplicated block id: 1280 size: 26 cleaned lines of code in 2 files: - megatron_patch/model/mistral/language_model.py (290:333) - megatron_patch/model/starcoder/language_model.py (290:333) duplicated block id: 1281 size: 26 cleaned lines of code in 2 files: - megatron_patch/model/llama2/language_model.py (289:332) - megatron_patch/model/starcoder/language_model.py (290:333) duplicated block id: 1282 size: 26 cleaned lines of code in 2 files: - megatron_patch/model/llama3/language_model.py (284:327) - megatron_patch/model/starcoder/language_model.py (290:333) duplicated block id: 1283 size: 26 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon40b/checkpoint_reshaping_and_interoperability.py (678:712) - toolkits/model_checkpoints_convertor/starcoder/checkpoint_reshaping_and_interoperability.py (397:431) duplicated block id: 1284 size: 26 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/reward_model_to_megatron.py (102:134) - toolkits/model_checkpoints_convertor/chatglm/checkpoint_reshaping_and_interoperability.py (107:138) duplicated block id: 1285 size: 26 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/reward_model_to_megatron.py (102:134) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (153:182) duplicated block id: 1286 size: 26 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/hf2te.py (187:227) - toolkits/model_checkpoints_convertor/baichuan2/hf2te.py (171:211) duplicated block id: 1287 size: 26 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/checkpoint_reshaping_and_interoperability.py (102:134) - toolkits/model_checkpoints_convertor/galactica/checkpoint_reshaping_and_interoperability.py (124:155) duplicated block id: 1288 size: 26 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/reward_model_to_megatron.py (102:134) - toolkits/model_checkpoints_convertor/galactica/checkpoint_reshaping_and_interoperability.py (124:155) duplicated block id: 1289 size: 26 cleaned lines of code in 2 files: - megatron_patch/model/falcon/language_model.py (129:174) - megatron_patch/model/glm130b/language_model.py (131:176) duplicated block id: 1290 size: 26 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (124:153) - toolkits/model_checkpoints_convertor/starcoder/checkpoint_reshaping_and_interoperability.py (102:134) duplicated block id: 1291 size: 26 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon/checkpoint_reshaping_and_interoperability.py (355:391) - toolkits/model_checkpoints_convertor/falcon40b/checkpoint_reshaping_and_interoperability.py (345:381) duplicated block id: 1292 size: 26 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/chatglm/checkpoint_reshaping_and_interoperability.py (107:138) - toolkits/model_checkpoints_convertor/starcoder/checkpoint_reshaping_and_interoperability.py (102:134) duplicated block id: 1293 size: 26 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/transformer.py (198:233) - megatron_patch/model/llama/transformer.py (173:208) duplicated block id: 1294 size: 26 cleaned lines of code in 2 files: - megatron_patch/model/qwen/language_model.py (289:332) - megatron_patch/model/starcoder/language_model.py (290:333) duplicated block id: 1295 size: 26 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/reward_model_to_megatron.py (102:134) - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (148:177) duplicated block id: 1296 size: 26 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/language_model.py (299:342) - megatron_patch/model/starcoder/language_model.py (290:333) duplicated block id: 1297 size: 26 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (1318:1354) - megatron_patch/model/llama2/transformer.py (1451:1487) duplicated block id: 1298 size: 26 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (1318:1354) - megatron_patch/model/llama3/transformer_legacy.py (1409:1445) duplicated block id: 1299 size: 26 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/checkpoint_reshaping_and_interoperability.py (102:134) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (153:182) duplicated block id: 1300 size: 26 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/language_model.py (129:174) - megatron_patch/model/glm130b/language_model.py (131:176) duplicated block id: 1301 size: 26 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer/mlp.py (79:110) - megatron_patch/model/mixtral_bak/transformer/mlp.py (57:88) duplicated block id: 1302 size: 26 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (100:131) - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (90:121) duplicated block id: 1303 size: 26 cleaned lines of code in 2 files: - megatron_patch/model/galactica/transformer.py (135:170) - megatron_patch/model/glm130b/transformer.py (198:233) duplicated block id: 1304 size: 26 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/transformer.py (115:150) - megatron_patch/model/glm130b/transformer.py (198:233) duplicated block id: 1305 size: 26 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (358:393) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (274:308) duplicated block id: 1306 size: 26 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (124:153) - toolkits/model_checkpoints_convertor/bloom/reward_model_to_megatron.py (102:134) duplicated block id: 1307 size: 26 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (1318:1354) - megatron_patch/model/baichuan2/transformer.py (1451:1487) duplicated block id: 1308 size: 26 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (1318:1354) - megatron_patch/model/qwen1_5_megablocks/transformer.py (1336:1372) duplicated block id: 1309 size: 26 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/reward_model_to_megatron.py (102:134) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (152:181) duplicated block id: 1310 size: 26 cleaned lines of code in 2 files: - megatron_patch/model/qwen1_5_megablocks/language_model.py (288:331) - megatron_patch/model/starcoder/language_model.py (290:333) duplicated block id: 1311 size: 26 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/language_model.py (131:176) - megatron_patch/model/llama/language_model.py (129:174) duplicated block id: 1312 size: 26 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (1318:1354) - megatron_patch/model/qwen/transformer.py (1393:1429) duplicated block id: 1313 size: 26 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (153:182) - toolkits/model_checkpoints_convertor/starcoder/checkpoint_reshaping_and_interoperability.py (102:134) duplicated block id: 1314 size: 26 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore.py (282:307) - toolkits/model_checkpoints_convertor/qwen/hf2megablocks_qwen1.5.py (276:301) duplicated block id: 1315 size: 26 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (1318:1354) - megatron_patch/model/mistral/transformer.py (1448:1484) duplicated block id: 1316 size: 26 cleaned lines of code in 2 files: - megatron_patch/model/falcon/transformer.py (192:227) - megatron_patch/model/glm130b/transformer.py (198:233) duplicated block id: 1317 size: 26 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (178:215) - megatron_patch/model/starcoder/transformer.py (157:194) duplicated block id: 1318 size: 26 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/transformer/mlp.py (81:112) - megatron_patch/model/mixtral_bak/transformer/mlp.py (57:88) duplicated block id: 1319 size: 26 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon/checkpoint_reshaping_and_interoperability.py (693:727) - toolkits/model_checkpoints_convertor/starcoder/checkpoint_reshaping_and_interoperability.py (397:431) duplicated block id: 1320 size: 26 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/model.py (167:221) - megatron_patch/model/mixtral/model.py (127:191) duplicated block id: 1321 size: 26 cleaned lines of code in 2 files: - megatron_patch/model/qwen_vl/language_model.py (290:333) - megatron_patch/model/starcoder/language_model.py (290:333) duplicated block id: 1322 size: 26 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/galactica/checkpoint_reshaping_and_interoperability.py (124:155) - toolkits/model_checkpoints_convertor/starcoder/checkpoint_reshaping_and_interoperability.py (102:134) duplicated block id: 1323 size: 26 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/transformer_block.py (276:304) - megatron_patch/model/qwen2/transformer_block.py (228:255) duplicated block id: 1324 size: 26 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (1318:1354) - megatron_patch/model/qwen_vl/transformer.py (1450:1486) duplicated block id: 1325 size: 26 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/checkpoint_reshaping_and_interoperability.py (102:134) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (152:181) duplicated block id: 1326 size: 26 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/transformer.py (556:599) - megatron_patch/model/llama/transformer.py (507:550) duplicated block id: 1327 size: 26 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/model.py (229:259) - megatron_patch/model/mixtral/model.py (193:223) duplicated block id: 1328 size: 25 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore.py (57:84) - toolkits/model_checkpoints_convertor/starcoder/checkpoint_reshaping_and_interoperability.py (28:52) duplicated block id: 1329 size: 25 cleaned lines of code in 2 files: - megatron_patch/model/qwen1_5/moe/token_dispatcher.py (147:176) - megatron_patch/model/qwen2/moe/token_dispatcher.py (150:179) duplicated block id: 1330 size: 25 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/glm/checkpoint_reshaping_and_interoperability.py (25:49) - toolkits/model_checkpoints_convertor/qwen/hf2megablocks_qwen1.5.py (54:83) duplicated block id: 1331 size: 25 cleaned lines of code in 2 files: - megatron_patch/model/falcon/transformer.py (855:891) - megatron_patch/model/qwen_vl/transformer.py (875:911) duplicated block id: 1332 size: 25 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (493:526) - toolkits/model_checkpoints_convertor/yi/checkpoint_reshaping_and_interoperability.py (500:533) duplicated block id: 1333 size: 25 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/mistral/hf2mcore.py (32:59) - toolkits/model_checkpoints_convertor/starcoder/checkpoint_reshaping_and_interoperability.py (28:52) duplicated block id: 1334 size: 25 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/language_model.py (194:227) - megatron_patch/model/glm130b/language_model.py (197:230) duplicated block id: 1335 size: 25 cleaned lines of code in 2 files: - megatron_patch/model/bloom/language_model.py (418:447) - megatron_patch/model/falcon40b/language_model.py (416:445) duplicated block id: 1336 size: 25 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (728:764) - megatron_patch/model/falcon/transformer.py (855:891) duplicated block id: 1337 size: 25 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (307:354) - toolkits/model_checkpoints_convertor/falcon/checkpoint_reshaping_and_interoperability.py (282:329) duplicated block id: 1338 size: 25 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (1025:1063) - megatron_patch/model/galactica/transformer.py (754:792) duplicated block id: 1339 size: 25 cleaned lines of code in 2 files: - megatron_patch/model/falcon/transformer.py (855:891) - megatron_patch/model/qwen1_5_megablocks/transformer.py (767:803) duplicated block id: 1340 size: 25 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/language_model.py (235:268) - megatron_patch/model/glm130b/language_model.py (197:230) duplicated block id: 1341 size: 25 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/language_model.py (406:435) - megatron_patch/model/llama/language_model.py (377:406) duplicated block id: 1342 size: 25 cleaned lines of code in 2 files: - megatron_patch/model/bloom/language_model.py (418:447) - megatron_patch/model/galactica/language_model.py (447:476) duplicated block id: 1343 size: 25 cleaned lines of code in 2 files: - megatron_patch/model/falcon/transformer.py (855:891) - megatron_patch/model/qwen/transformer.py (821:857) duplicated block id: 1344 size: 25 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (1025:1063) - megatron_patch/model/falcon/transformer.py (1126:1164) duplicated block id: 1345 size: 25 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/glm130b/checkpoint_reshaping_and_interoperability.py (25:49) - toolkits/model_checkpoints_convertor/qwen/hf2megablocks_qwen1.5.py (54:83) duplicated block id: 1346 size: 25 cleaned lines of code in 2 files: - megatron_patch/model/falcon/transformer.py (855:891) - megatron_patch/model/llama2/transformer.py (876:912) duplicated block id: 1347 size: 25 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/glm130b/checkpoint_reshaping_and_interoperability.py (25:49) - toolkits/model_checkpoints_convertor/mistral/hf2mcore.py (32:59) duplicated block id: 1348 size: 25 cleaned lines of code in 2 files: - megatron_patch/model/falcon/language_model.py (416:445) - megatron_patch/model/glm130b/language_model.py (437:466) duplicated block id: 1349 size: 25 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (521:554) - toolkits/model_checkpoints_convertor/yi/checkpoint_reshaping_and_interoperability.py (500:533) duplicated block id: 1350 size: 25 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/language_model.py (422:451) - megatron_patch/model/falcon40b/language_model.py (416:445) duplicated block id: 1351 size: 25 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/transformer_config.py (356:391) - megatron_patch/model/mixtral_bak/transformer_config.py (142:176) duplicated block id: 1352 size: 25 cleaned lines of code in 2 files: - megatron_patch/model/bloom/language_model.py (418:447) - megatron_patch/model/llama/language_model.py (416:445) duplicated block id: 1353 size: 25 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/glm130b/checkpoint_reshaping_and_interoperability.py (25:49) - toolkits/model_checkpoints_convertor/llama/hf2mcore.py (57:84) duplicated block id: 1354 size: 25 cleaned lines of code in 2 files: - megatron_patch/model/falcon/language_model.py (377:406) - megatron_patch/model/glm130b/language_model.py (406:435) duplicated block id: 1355 size: 25 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (1025:1063) - megatron_patch/model/glm130b/transformer.py (1010:1048) duplicated block id: 1356 size: 25 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/checkpoint_reshaping_and_interoperability.py (28:52) - toolkits/model_checkpoints_convertor/mistral/hf2mcore.py (32:59) duplicated block id: 1357 size: 25 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (839:875) - megatron_patch/model/falcon/transformer.py (855:891) duplicated block id: 1358 size: 25 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/language_model.py (437:466) - megatron_patch/model/llama/language_model.py (416:445) duplicated block id: 1359 size: 25 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/language_model.py (197:230) - megatron_patch/model/qwen1_5_megablocks/language_model.py (183:216) duplicated block id: 1360 size: 25 cleaned lines of code in 2 files: - megatron_patch/model/mixtral_bak/transformer/attention.py (355:394) - megatron_patch/model/qwen2/transformer/attention.py (467:506) duplicated block id: 1361 size: 25 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/language_model.py (197:230) - megatron_patch/model/mistral/language_model.py (185:218) duplicated block id: 1362 size: 25 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/checkpoint_reshaping_and_interoperability.py (102:130) - toolkits/model_checkpoints_convertor/falcon40b/checkpoint_reshaping_and_interoperability.py (125:152) duplicated block id: 1363 size: 25 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/language_model.py (197:230) - megatron_patch/model/qwen/language_model.py (184:217) duplicated block id: 1364 size: 25 cleaned lines of code in 2 files: - megatron_patch/model/bloom/language_model.py (418:447) - megatron_patch/model/falcon/language_model.py (416:445) duplicated block id: 1365 size: 25 cleaned lines of code in 2 files: - megatron_patch/model/qwen2/moe/experts.py (277:303) - megatron_patch/model/qwen2/transformer/mlp.py (124:150) duplicated block id: 1366 size: 25 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/language_model.py (422:451) - megatron_patch/model/galactica/language_model.py (447:476) duplicated block id: 1367 size: 25 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/reward_model_to_megatron.py (102:130) - toolkits/model_checkpoints_convertor/falcon40b/checkpoint_reshaping_and_interoperability.py (125:152) duplicated block id: 1368 size: 25 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon40b/checkpoint_reshaping_and_interoperability.py (125:152) - toolkits/model_checkpoints_convertor/starcoder/checkpoint_reshaping_and_interoperability.py (102:130) duplicated block id: 1369 size: 25 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (530:563) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (533:566) duplicated block id: 1370 size: 25 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/moe/moe_layer.py (134:164) - megatron_patch/model/mixtral/moe/moe_layer.py (127:157) duplicated block id: 1371 size: 25 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (832:861) - megatron_patch/model/baichuan2/transformer.py (928:957) duplicated block id: 1372 size: 25 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (124:151) - toolkits/model_checkpoints_convertor/starcoder/checkpoint_reshaping_and_interoperability.py (102:130) duplicated block id: 1373 size: 25 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (164:197) - megatron_patch/model/glm130b/transformer.py (198:231) duplicated block id: 1374 size: 25 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore_llama3_1.py (143:169) - toolkits/model_checkpoints_convertor/llama/hf2mcore_llama3_1.py (196:220) duplicated block id: 1375 size: 25 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2megablocks_qwen1.5.py (54:83) - toolkits/model_checkpoints_convertor/starcoder/checkpoint_reshaping_and_interoperability.py (28:52) duplicated block id: 1376 size: 25 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/reward_model_to_megatron.py (28:52) - toolkits/model_checkpoints_convertor/mistral/hf2mcore.py (32:59) duplicated block id: 1377 size: 25 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (1025:1063) - megatron_patch/model/falcon40b/transformer.py (865:903) duplicated block id: 1378 size: 25 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (124:151) - toolkits/model_checkpoints_convertor/bloom/reward_model_to_megatron.py (102:130) duplicated block id: 1379 size: 25 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/checkpoint_reshaping_and_interoperability.py (28:52) - toolkits/model_checkpoints_convertor/llama/hf2mcore.py (57:84) duplicated block id: 1380 size: 25 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/checkpoint_reshaping_and_interoperability.py (102:130) - toolkits/model_checkpoints_convertor/falcon/checkpoint_reshaping_and_interoperability.py (125:152) duplicated block id: 1381 size: 25 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/language_model.py (416:445) - megatron_patch/model/glm130b/language_model.py (437:466) duplicated block id: 1382 size: 25 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/reward_model_to_megatron.py (102:130) - toolkits/model_checkpoints_convertor/yi/checkpoint_reshaping_and_interoperability.py (125:152) duplicated block id: 1383 size: 25 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (1025:1063) - megatron_patch/model/llama/transformer.py (920:958) duplicated block id: 1384 size: 25 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon/checkpoint_reshaping_and_interoperability.py (125:152) - toolkits/model_checkpoints_convertor/starcoder/checkpoint_reshaping_and_interoperability.py (102:130) duplicated block id: 1385 size: 25 cleaned lines of code in 2 files: - megatron_patch/model/qwen1_5/moe/router.py (42:105) - megatron_patch/model/qwen2/moe/router.py (117:180) duplicated block id: 1386 size: 25 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (1025:1063) - megatron_patch/model/chatglm/transformer.py (750:788) duplicated block id: 1387 size: 25 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/reward_model_to_megatron.py (102:130) - toolkits/model_checkpoints_convertor/falcon/checkpoint_reshaping_and_interoperability.py (125:152) duplicated block id: 1388 size: 25 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (124:151) - toolkits/model_checkpoints_convertor/bloom/checkpoint_reshaping_and_interoperability.py (102:130) duplicated block id: 1389 size: 25 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/starcoder/checkpoint_reshaping_and_interoperability.py (102:130) - toolkits/model_checkpoints_convertor/yi/checkpoint_reshaping_and_interoperability.py (125:152) duplicated block id: 1390 size: 25 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/language_model.py (377:406) - megatron_patch/model/glm130b/language_model.py (406:435) duplicated block id: 1391 size: 25 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer/attention.py (465:504) - megatron_patch/model/mixtral_bak/transformer/attention.py (355:394) duplicated block id: 1392 size: 25 cleaned lines of code in 2 files: - megatron_patch/model/qwen2_moe/layer_specs.py (157:184) - megatron_patch/model/qwen3_moe/gpt_layer_specs.py (143:170) duplicated block id: 1393 size: 25 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/language_model.py (422:451) - megatron_patch/model/llama/language_model.py (416:445) duplicated block id: 1394 size: 25 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/checkpoint_reshaping_and_interoperability.py (28:52) - toolkits/model_checkpoints_convertor/qwen/hf2megablocks_qwen1.5.py (54:83) duplicated block id: 1395 size: 25 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/mlp.py (103:129) - megatron_patch/model/qwen2/moe/experts.py (277:303) duplicated block id: 1396 size: 25 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/reward_model_to_megatron.py (28:52) - toolkits/model_checkpoints_convertor/llama/hf2mcore.py (57:84) duplicated block id: 1397 size: 25 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/reward_model_to_megatron.py (28:52) - toolkits/model_checkpoints_convertor/qwen/hf2megablocks_qwen1.5.py (54:83) duplicated block id: 1398 size: 25 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (390:419) - megatron_patch/model/chatglm/transformer.py (332:361) duplicated block id: 1399 size: 25 cleaned lines of code in 2 files: - megatron_patch/model/galactica/language_model.py (408:437) - megatron_patch/model/glm130b/language_model.py (406:435) duplicated block id: 1400 size: 25 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/language_model.py (197:230) - megatron_patch/model/qwen_vl/language_model.py (185:218) duplicated block id: 1401 size: 25 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/language_model.py (132:175) - megatron_patch/model/glm130b/language_model.py (133:176) duplicated block id: 1402 size: 25 cleaned lines of code in 2 files: - megatron_patch/model/falcon/transformer.py (855:891) - megatron_patch/model/mistral/transformer.py (873:909) duplicated block id: 1403 size: 25 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/deepseek/hf2mcore_deepseek_v2_moe.py (147:175) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_moe.py (290:316) duplicated block id: 1404 size: 25 cleaned lines of code in 2 files: - megatron_patch/model/galactica/language_model.py (447:476) - megatron_patch/model/glm130b/language_model.py (437:466) duplicated block id: 1405 size: 25 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/checkpoint_reshaping_and_interoperability.py (102:130) - toolkits/model_checkpoints_convertor/yi/checkpoint_reshaping_and_interoperability.py (125:152) duplicated block id: 1406 size: 25 cleaned lines of code in 2 files: - megatron_patch/model/mixtral_bak/transformer/attention.py (355:394) - megatron_patch/model/qwen1_5/transformer/attention.py (465:504) duplicated block id: 1407 size: 25 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/transformer/mlp.py (111:137) - megatron_patch/model/qwen2/moe/experts.py (277:303) duplicated block id: 1408 size: 25 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/glm/checkpoint_reshaping_and_interoperability.py (25:49) - toolkits/model_checkpoints_convertor/llama/hf2mcore.py (57:84) duplicated block id: 1409 size: 25 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/glm/checkpoint_reshaping_and_interoperability.py (25:49) - toolkits/model_checkpoints_convertor/mistral/hf2mcore.py (32:59) duplicated block id: 1410 size: 25 cleaned lines of code in 2 files: - megatron_patch/model/falcon/transformer.py (855:891) - megatron_patch/model/llava/transformer.py (875:911) duplicated block id: 1411 size: 25 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/language_model.py (422:451) - megatron_patch/model/falcon/language_model.py (416:445) duplicated block id: 1412 size: 25 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/language_model.py (197:230) - megatron_patch/model/llava/language_model.py (187:220) duplicated block id: 1413 size: 25 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/language_model.py (197:230) - megatron_patch/model/llama2/language_model.py (184:217) duplicated block id: 1414 size: 24 cleaned lines of code in 2 files: - megatron_patch/model/bloom/language_model.py (480:513) - megatron_patch/model/llama/language_model.py (578:611) duplicated block id: 1415 size: 24 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon/checkpoint_reshaping_and_interoperability.py (282:326) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (377:421) duplicated block id: 1416 size: 24 cleaned lines of code in 2 files: - megatron_patch/model/galactica/language_model.py (659:692) - megatron_patch/model/starcoder/language_model.py (509:542) duplicated block id: 1417 size: 24 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/reward_model_to_megatron.py (540:575) - toolkits/model_checkpoints_convertor/glm130b/checkpoint_reshaping_and_interoperability.py (271:306) duplicated block id: 1418 size: 24 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/hf2te.py (132:187) - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (282:337) duplicated block id: 1419 size: 24 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (927:954) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (1003:1031) duplicated block id: 1420 size: 24 cleaned lines of code in 2 files: - megatron_patch/model/mixtral_bak/model.py (123:185) - megatron_patch/model/qwen2_vl/gpt_model.py (138:186) duplicated block id: 1421 size: 24 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/glm/checkpoint_reshaping_and_interoperability.py (271:306) - toolkits/model_checkpoints_convertor/starcoder/checkpoint_reshaping_and_interoperability.py (556:591) duplicated block id: 1422 size: 24 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/hf2te.py (132:187) - toolkits/model_checkpoints_convertor/galactica/checkpoint_reshaping_and_interoperability.py (233:288) duplicated block id: 1423 size: 24 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (669:700) - megatron_patch/model/chatglm/transformer.py (564:592) duplicated block id: 1424 size: 24 cleaned lines of code in 2 files: - megatron_patch/model/bloom/language_model.py (480:513) - megatron_patch/model/falcon/language_model.py (567:600) duplicated block id: 1425 size: 24 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (308:354) - toolkits/model_checkpoints_convertor/falcon40b/checkpoint_reshaping_and_interoperability.py (281:327) duplicated block id: 1426 size: 24 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (901:928) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (1003:1031) duplicated block id: 1427 size: 24 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/language_model.py (630:663) - megatron_patch/model/starcoder/language_model.py (509:542) duplicated block id: 1428 size: 24 cleaned lines of code in 2 files: - megatron_patch/model/qwen_vl/language_model.py (436:465) - megatron_patch/model/starcoder/language_model.py (403:432) duplicated block id: 1429 size: 24 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/mistral/hf2mcore.py (271:297) - toolkits/model_checkpoints_convertor/qwen/hf2megablocks_qwen1.5.py (322:348) duplicated block id: 1430 size: 24 cleaned lines of code in 2 files: - megatron_patch/model/qwen1_5/model.py (65:95) - megatron_patch/model/qwen2/model.py (66:96) duplicated block id: 1431 size: 24 cleaned lines of code in 2 files: - megatron_patch/model/bloom/language_model.py (183:217) - megatron_patch/model/galactica/language_model.py (211:246) duplicated block id: 1432 size: 24 cleaned lines of code in 2 files: - megatron_patch/model/llama3/language_model.py (427:456) - megatron_patch/model/starcoder/language_model.py (403:432) duplicated block id: 1433 size: 24 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/reward_model_to_megatron.py (540:575) - toolkits/model_checkpoints_convertor/glm/checkpoint_reshaping_and_interoperability.py (271:306) duplicated block id: 1434 size: 24 cleaned lines of code in 2 files: - megatron_patch/model/galactica/language_model.py (211:246) - megatron_patch/model/llama/language_model.py (181:215) duplicated block id: 1435 size: 24 cleaned lines of code in 2 files: - megatron_patch/model/bloom/language_model.py (480:513) - megatron_patch/model/falcon40b/language_model.py (567:600) duplicated block id: 1436 size: 24 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/hf2te.py (132:187) - toolkits/model_checkpoints_convertor/yi/checkpoint_reshaping_and_interoperability.py (231:286) duplicated block id: 1437 size: 24 cleaned lines of code in 2 files: - megatron_patch/model/qwen1_5/moe/router.py (116:151) - megatron_patch/model/qwen2/moe/router.py (190:225) duplicated block id: 1438 size: 24 cleaned lines of code in 2 files: - megatron_patch/model/llama/language_model.py (641:674) - megatron_patch/model/starcoder/language_model.py (509:542) duplicated block id: 1439 size: 24 cleaned lines of code in 2 files: - megatron_patch/model/falcon/language_model.py (630:663) - megatron_patch/model/starcoder/language_model.py (509:542) duplicated block id: 1440 size: 24 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore.py (281:304) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_mha.py (112:135) duplicated block id: 1441 size: 24 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/hf2te.py (132:187) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (301:356) duplicated block id: 1442 size: 24 cleaned lines of code in 2 files: - megatron_patch/model/mistral/language_model.py (438:467) - megatron_patch/model/starcoder/language_model.py (403:432) duplicated block id: 1443 size: 24 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon/checkpoint_reshaping_and_interoperability.py (282:326) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (355:399) duplicated block id: 1444 size: 24 cleaned lines of code in 2 files: - megatron_patch/model/qwen/language_model.py (433:462) - megatron_patch/model/starcoder/language_model.py (403:432) duplicated block id: 1445 size: 24 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/model.py (101:129) - megatron_patch/model/mixtral/model.py (83:111) duplicated block id: 1446 size: 24 cleaned lines of code in 2 files: - megatron_patch/model/llama2/language_model.py (436:465) - megatron_patch/model/starcoder/language_model.py (403:432) duplicated block id: 1447 size: 24 cleaned lines of code in 2 files: - megatron_patch/model/llama3/model.py (65:95) - megatron_patch/model/llama3_1/model.py (80:110) duplicated block id: 1448 size: 24 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/language_model.py (181:215) - megatron_patch/model/galactica/language_model.py (211:246) duplicated block id: 1449 size: 24 cleaned lines of code in 2 files: - megatron_patch/model/llama3_1/model.py (80:110) - megatron_patch/model/qwen1_5/model.py (65:95) duplicated block id: 1450 size: 24 cleaned lines of code in 2 files: - megatron_patch/model/llama3/model.py (65:95) - megatron_patch/model/qwen2/model.py (66:96) duplicated block id: 1451 size: 24 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/language_model.py (182:216) - megatron_patch/model/galactica/language_model.py (211:246) duplicated block id: 1452 size: 24 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/language_model.py (443:472) - megatron_patch/model/starcoder/language_model.py (403:432) duplicated block id: 1453 size: 24 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/language_model.py (231:266) - megatron_patch/model/starcoder/language_model.py (219:253) duplicated block id: 1454 size: 24 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/checkpoint_reshaping_and_interoperability.py (541:576) - toolkits/model_checkpoints_convertor/glm/checkpoint_reshaping_and_interoperability.py (271:306) duplicated block id: 1455 size: 24 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon/checkpoint_reshaping_and_interoperability.py (828:855) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (1003:1031) duplicated block id: 1456 size: 24 cleaned lines of code in 2 files: - megatron_patch/model/qwen1_5_megablocks/language_model.py (435:464) - megatron_patch/model/starcoder/language_model.py (403:432) duplicated block id: 1457 size: 24 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/checkpoint_reshaping_and_interoperability.py (541:576) - toolkits/model_checkpoints_convertor/glm130b/checkpoint_reshaping_and_interoperability.py (271:306) duplicated block id: 1458 size: 24 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon40b/checkpoint_reshaping_and_interoperability.py (804:832) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (1003:1031) duplicated block id: 1459 size: 24 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/transformer.py (446:474) - megatron_patch/model/galactica/transformer.py (411:439) duplicated block id: 1460 size: 24 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/glm130b/checkpoint_reshaping_and_interoperability.py (271:306) - toolkits/model_checkpoints_convertor/starcoder/checkpoint_reshaping_and_interoperability.py (556:591) duplicated block id: 1461 size: 24 cleaned lines of code in 2 files: - megatron_patch/model/llava/language_model.py (444:473) - megatron_patch/model/starcoder/language_model.py (403:432) duplicated block id: 1462 size: 24 cleaned lines of code in 2 files: - megatron_patch/model/falcon/language_model.py (181:215) - megatron_patch/model/galactica/language_model.py (211:246) duplicated block id: 1463 size: 24 cleaned lines of code in 2 files: - megatron_patch/model/qwen1_5/moe/experts.py (188:217) - megatron_patch/model/qwen2/moe/experts.py (354:384) duplicated block id: 1464 size: 23 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/mistral/hf2mcore.py (71:95) - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (65:88) duplicated block id: 1465 size: 23 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/checkpoint_reshaping_and_interoperability.py (505:537) - toolkits/model_checkpoints_convertor/falcon40b/checkpoint_reshaping_and_interoperability.py (784:816) duplicated block id: 1466 size: 23 cleaned lines of code in 2 files: - megatron_patch/model/qwen1_5_megablocks/language_model.py (521:552) - megatron_patch/model/starcoder/language_model.py (452:483) duplicated block id: 1467 size: 23 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (38:88) - megatron_patch/model/baichuan2/transformer.py (34:84) duplicated block id: 1468 size: 23 cleaned lines of code in 2 files: - megatron_patch/model/galactica/language_model.py (166:206) - megatron_patch/model/llama/language_model.py (137:177) duplicated block id: 1469 size: 23 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/transformer.py (391:417) - megatron_patch/model/galactica/transformer.py (340:388) duplicated block id: 1470 size: 23 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/language_model.py (599:630) - megatron_patch/model/starcoder/language_model.py (452:483) duplicated block id: 1471 size: 23 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (38:88) - megatron_patch/model/mistral/transformer.py (39:89) duplicated block id: 1472 size: 23 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (38:88) - megatron_patch/model/qwen/transformer.py (34:84) duplicated block id: 1473 size: 23 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (881:913) - toolkits/model_checkpoints_convertor/bloom/reward_model_to_megatron.py (504:536) duplicated block id: 1474 size: 23 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/reward_model_to_megatron.py (504:536) - toolkits/model_checkpoints_convertor/falcon40b/checkpoint_reshaping_and_interoperability.py (784:816) duplicated block id: 1475 size: 23 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/hf2te.py (68:92) - toolkits/model_checkpoints_convertor/chatglm/checkpoint_reshaping_and_interoperability.py (79:101) duplicated block id: 1476 size: 23 cleaned lines of code in 2 files: - megatron_patch/model/falcon/language_model.py (137:177) - megatron_patch/model/galactica/language_model.py (166:206) duplicated block id: 1477 size: 23 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/checkpoint_reshaping_and_interoperability.py (505:537) - toolkits/model_checkpoints_convertor/falcon/checkpoint_reshaping_and_interoperability.py (808:840) duplicated block id: 1478 size: 23 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon40b/checkpoint_reshaping_and_interoperability.py (281:324) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (378:421) duplicated block id: 1479 size: 23 cleaned lines of code in 2 files: - megatron_patch/model/qwen_vl/language_model.py (551:582) - megatron_patch/model/starcoder/language_model.py (452:483) duplicated block id: 1480 size: 23 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (38:88) - megatron_patch/model/llava/transformer.py (39:89) duplicated block id: 1481 size: 23 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore.py (97:121) - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (65:88) duplicated block id: 1482 size: 23 cleaned lines of code in 2 files: - megatron_patch/model/falcon/transformer.py (690:721) - megatron_patch/model/starcoder/transformer.py (656:688) duplicated block id: 1483 size: 23 cleaned lines of code in 2 files: - megatron_patch/model/qwen1_5/moe/router.py (208:252) - megatron_patch/model/qwen2/moe/router.py (294:339) duplicated block id: 1484 size: 23 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/language_model.py (515:546) - megatron_patch/model/starcoder/language_model.py (452:483) duplicated block id: 1485 size: 23 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/language_model.py (137:177) - megatron_patch/model/galactica/language_model.py (166:206) duplicated block id: 1486 size: 23 cleaned lines of code in 2 files: - megatron_patch/model/bloom/language_model.py (481:513) - megatron_patch/model/galactica/language_model.py (597:629) duplicated block id: 1487 size: 23 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/language_model.py (138:178) - megatron_patch/model/galactica/language_model.py (166:206) duplicated block id: 1488 size: 23 cleaned lines of code in 2 files: - megatron_patch/model/qwen/language_model.py (504:535) - megatron_patch/model/starcoder/language_model.py (452:483) duplicated block id: 1489 size: 23 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/transformer_block.py (166:203) - megatron_patch/model/qwen2/transformer_block.py (98:133) duplicated block id: 1490 size: 23 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (65:88) - toolkits/model_checkpoints_convertor/qwen/hf2megablocks_qwen1.5.py (97:121) duplicated block id: 1491 size: 23 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (454:481) - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (524:549) duplicated block id: 1492 size: 23 cleaned lines of code in 2 files: - megatron_patch/model/bloom/language_model.py (481:513) - megatron_patch/model/chatglm/language_model.py (547:579) duplicated block id: 1493 size: 23 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/mistral/hf2mcore.py (223:245) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_dense_and_moe_gqa.py (174:197) duplicated block id: 1494 size: 23 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/transformer.py (478:507) - megatron_patch/model/llama/transformer.py (602:632) duplicated block id: 1495 size: 23 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (635:664) - megatron_patch/model/starcoder/transformer.py (849:878) duplicated block id: 1496 size: 23 cleaned lines of code in 2 files: - megatron_patch/model/galactica/transformer.py (889:931) - megatron_patch/model/llama/transformer.py (1060:1102) duplicated block id: 1497 size: 23 cleaned lines of code in 2 files: - megatron_patch/model/mistral/language_model.py (534:565) - megatron_patch/model/starcoder/language_model.py (452:483) duplicated block id: 1498 size: 23 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (38:88) - megatron_patch/model/qwen_vl/transformer.py (39:89) duplicated block id: 1499 size: 23 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (38:88) - megatron_patch/model/llama2/transformer.py (39:89) duplicated block id: 1500 size: 23 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (38:88) - megatron_patch/model/qwen1_5_megablocks/transformer.py (37:87) duplicated block id: 1501 size: 23 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_dense_and_moe_gqa.py (129:151) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_dense_and_moe_gqa.py (174:197) duplicated block id: 1502 size: 23 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/transformer.py (881:923) - megatron_patch/model/llama/transformer.py (1060:1102) duplicated block id: 1503 size: 23 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/language_model.py (413:446) - megatron_patch/model/starcoder/language_model.py (379:412) duplicated block id: 1504 size: 23 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/reward_model_to_megatron.py (504:536) - toolkits/model_checkpoints_convertor/falcon/checkpoint_reshaping_and_interoperability.py (808:840) duplicated block id: 1505 size: 23 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (907:939) - toolkits/model_checkpoints_convertor/bloom/checkpoint_reshaping_and_interoperability.py (505:537) duplicated block id: 1506 size: 23 cleaned lines of code in 2 files: - megatron_patch/model/llava/language_model.py (582:613) - megatron_patch/model/starcoder/language_model.py (452:483) duplicated block id: 1507 size: 23 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (38:88) - megatron_patch/model/llama3/transformer_legacy.py (43:93) duplicated block id: 1508 size: 23 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (907:939) - toolkits/model_checkpoints_convertor/starcoder/checkpoint_reshaping_and_interoperability.py (520:552) duplicated block id: 1509 size: 23 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llava/hf2mcore_llava.py (167:191) - toolkits/model_checkpoints_convertor/llava/hf2mcore_llava.py (218:240) duplicated block id: 1510 size: 23 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon40b/checkpoint_reshaping_and_interoperability.py (281:324) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (356:399) duplicated block id: 1511 size: 23 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (881:913) - toolkits/model_checkpoints_convertor/bloom/checkpoint_reshaping_and_interoperability.py (505:537) duplicated block id: 1512 size: 23 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/deepseek/hf2mcore_deepseek_v2_moe.py (72:102) - toolkits/model_checkpoints_convertor/deepseek/hf2mcore_deepseek_v3_moe.py (137:166) duplicated block id: 1513 size: 23 cleaned lines of code in 2 files: - megatron_patch/model/llava/language_model.py (542:568) - megatron_patch/model/mistral/language_model.py (484:510) duplicated block id: 1514 size: 23 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/transformer.py (613:642) - megatron_patch/model/glm130b/transformer.py (663:693) duplicated block id: 1515 size: 23 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_mha.py (113:135) - toolkits/model_checkpoints_convertor/qwen/hf2megablocks_qwen1.5.py (276:298) duplicated block id: 1516 size: 23 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/hf2te.py (69:93) - toolkits/model_checkpoints_convertor/chatglm/checkpoint_reshaping_and_interoperability.py (79:101) duplicated block id: 1517 size: 23 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/transformer.py (478:507) - megatron_patch/model/glm130b/transformer.py (663:693) duplicated block id: 1518 size: 23 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (907:939) - toolkits/model_checkpoints_convertor/bloom/reward_model_to_megatron.py (504:536) duplicated block id: 1519 size: 23 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/transformer_block.py (577:603) - megatron_patch/model/qwen2_5_vl/transformer_block.py (652:678) duplicated block id: 1520 size: 23 cleaned lines of code in 2 files: - megatron_patch/model/bloom/language_model.py (394:427) - megatron_patch/model/starcoder/language_model.py (379:412) duplicated block id: 1521 size: 23 cleaned lines of code in 2 files: - megatron_patch/model/llama2/language_model.py (522:553) - megatron_patch/model/starcoder/language_model.py (452:483) duplicated block id: 1522 size: 23 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore.py (456:482) - toolkits/model_checkpoints_convertor/qwen/hf2megablocks_qwen1.5.py (477:503) duplicated block id: 1523 size: 23 cleaned lines of code in 2 files: - megatron_patch/model/qwen2/transformer_block.py (98:133) - megatron_patch/model/qwen2_5_vl/transformer_block.py (205:242) duplicated block id: 1524 size: 23 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/transformer_block.py (526:569) - megatron_patch/model/qwen2_5_vl/transformer_block.py (595:638) duplicated block id: 1525 size: 23 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (881:913) - toolkits/model_checkpoints_convertor/starcoder/checkpoint_reshaping_and_interoperability.py (520:552) duplicated block id: 1526 size: 23 cleaned lines of code in 2 files: - megatron_patch/model/qwen2/transformer_block.py (456:482) - megatron_patch/model/qwen2_5_vl/transformer_block.py (652:678) duplicated block id: 1527 size: 23 cleaned lines of code in 2 files: - megatron_patch/model/llama3/language_model.py (499:530) - megatron_patch/model/starcoder/language_model.py (452:483) duplicated block id: 1528 size: 22 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/hf2te.py (132:177) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (278:323) duplicated block id: 1529 size: 22 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (32:53) - toolkits/model_checkpoints_convertor/chatglm/checkpoint_reshaping_and_interoperability.py (28:49) duplicated block id: 1530 size: 22 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (628:661) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (785:818) duplicated block id: 1531 size: 22 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/transformer.py (569:593) - megatron_patch/model/llama/transformer.py (555:580) duplicated block id: 1532 size: 22 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (299:339) - toolkits/model_checkpoints_convertor/chatglm/checkpoint_reshaping_and_interoperability.py (256:296) duplicated block id: 1533 size: 22 cleaned lines of code in 2 files: - megatron_patch/model/llama2/language_model.py (490:515) - megatron_patch/model/qwen_vl/language_model.py (519:544) duplicated block id: 1534 size: 22 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (672:697) - toolkits/model_checkpoints_convertor/bloom/reward_model_to_megatron.py (318:343) duplicated block id: 1535 size: 22 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/hf2te.py (492:520) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (747:775) duplicated block id: 1536 size: 22 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/language_model.py (243:273) - megatron_patch/model/llama3/language_model.py (182:212) duplicated block id: 1537 size: 22 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (941:969) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (1053:1082) duplicated block id: 1538 size: 22 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon40b/checkpoint_reshaping_and_interoperability.py (215:260) - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (282:327) duplicated block id: 1539 size: 22 cleaned lines of code in 2 files: - megatron_patch/model/qwen1_5_megablocks/language_model.py (489:514) - megatron_patch/model/qwen_vl/language_model.py (519:544) duplicated block id: 1540 size: 22 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_mha.py (292:317) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_moe.py (401:425) duplicated block id: 1541 size: 22 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer/attention.py (99:124) - megatron_patch/model/qwen2_vl/attention.py (114:139) duplicated block id: 1542 size: 22 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_mha.py (140:169) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_moe.py (178:206) duplicated block id: 1543 size: 22 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/language_model.py (494:520) - megatron_patch/model/llava/language_model.py (444:470) duplicated block id: 1544 size: 22 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon40b/checkpoint_reshaping_and_interoperability.py (215:260) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (301:346) duplicated block id: 1545 size: 22 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon/checkpoint_reshaping_and_interoperability.py (217:262) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (301:346) duplicated block id: 1546 size: 22 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (589:621) - megatron_patch/model/falcon40b/transformer.py (621:653) duplicated block id: 1547 size: 22 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/chatglm/checkpoint_reshaping_and_interoperability.py (28:49) - toolkits/model_checkpoints_convertor/falcon40b/checkpoint_reshaping_and_interoperability.py (33:54) duplicated block id: 1548 size: 22 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/chatglm/checkpoint_reshaping_and_interoperability.py (447:476) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (577:606) duplicated block id: 1549 size: 22 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/transformer.py (583:611) - megatron_patch/model/falcon/transformer.py (953:981) duplicated block id: 1550 size: 22 cleaned lines of code in 2 files: - megatron_patch/model/qwen1_5/model.py (68:97) - megatron_patch/model/qwen2_vl/gpt_model.py (58:86) duplicated block id: 1551 size: 22 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon/checkpoint_reshaping_and_interoperability.py (217:262) - toolkits/model_checkpoints_convertor/galactica/checkpoint_reshaping_and_interoperability.py (233:278) duplicated block id: 1552 size: 22 cleaned lines of code in 2 files: - megatron_patch/model/llava/language_model.py (550:575) - megatron_patch/model/qwen_vl/language_model.py (519:544) duplicated block id: 1553 size: 22 cleaned lines of code in 2 files: - megatron_patch/model/qwen2/transformer/attention.py (95:120) - megatron_patch/model/qwen2_vl/attention.py (114:139) duplicated block id: 1554 size: 22 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/hf2te.py (132:177) - toolkits/model_checkpoints_convertor/falcon40b/checkpoint_reshaping_and_interoperability.py (215:260) duplicated block id: 1555 size: 22 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/chatglm/checkpoint_reshaping_and_interoperability.py (28:49) - toolkits/model_checkpoints_convertor/yi/checkpoint_reshaping_and_interoperability.py (33:54) duplicated block id: 1556 size: 22 cleaned lines of code in 2 files: - toolkits/pretrain_data_preprocessing/preprocess_data_megatron.py (226:248) - toolkits/sft_data_preprocessing/build_idxmap_sft_dataset.py (219:241) duplicated block id: 1557 size: 22 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/chatglm/checkpoint_reshaping_and_interoperability.py (28:49) - toolkits/model_checkpoints_convertor/falcon/checkpoint_reshaping_and_interoperability.py (33:54) duplicated block id: 1558 size: 22 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2.5_vl.py (661:682) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_dense_and_moe_gqa.py (668:689) duplicated block id: 1559 size: 22 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (489:518) - toolkits/model_checkpoints_convertor/chatglm/checkpoint_reshaping_and_interoperability.py (447:476) duplicated block id: 1560 size: 22 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/glm/checkpoint_reshaping_and_interoperability.py (552:578) - toolkits/model_checkpoints_convertor/glm130b/checkpoint_reshaping_and_interoperability.py (509:535) duplicated block id: 1561 size: 22 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (970:995) - megatron_patch/model/glm130b/transformer.py (895:920) duplicated block id: 1562 size: 22 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon/checkpoint_reshaping_and_interoperability.py (217:262) - toolkits/model_checkpoints_convertor/yi/checkpoint_reshaping_and_interoperability.py (231:276) duplicated block id: 1563 size: 22 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/hf2te.py (492:520) - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (732:760) duplicated block id: 1564 size: 22 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon40b/checkpoint_reshaping_and_interoperability.py (215:260) - toolkits/model_checkpoints_convertor/galactica/checkpoint_reshaping_and_interoperability.py (233:278) duplicated block id: 1565 size: 22 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon40b/checkpoint_reshaping_and_interoperability.py (215:260) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (278:323) duplicated block id: 1566 size: 22 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (244:289) - toolkits/model_checkpoints_convertor/falcon40b/checkpoint_reshaping_and_interoperability.py (215:260) duplicated block id: 1567 size: 22 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/hf2te.py (492:520) - toolkits/model_checkpoints_convertor/falcon/checkpoint_reshaping_and_interoperability.py (560:589) duplicated block id: 1568 size: 22 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_dense_and_moe_gqa.py (566:587) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_dense_and_moe_gqa.py (667:688) duplicated block id: 1569 size: 22 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer_legacy.py (744:779) - megatron_patch/model/mistral/transformer.py (778:813) duplicated block id: 1570 size: 22 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_dense_and_moe_gqa.py (668:689) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_vl.py (668:689) duplicated block id: 1571 size: 22 cleaned lines of code in 2 files: - megatron_patch/model/qwen2/transformer/attention.py (95:120) - megatron_patch/model/qwen2_vl/attention_vision.py (113:138) duplicated block id: 1572 size: 22 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (656:689) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (785:818) duplicated block id: 1573 size: 22 cleaned lines of code in 2 files: - toolkits/pretrain_data_preprocessing/preprocess_data_megatron.py (274:302) - toolkits/sft_data_preprocessing/build_idxmap_sft_dataset.py (269:297) duplicated block id: 1574 size: 22 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon/checkpoint_reshaping_and_interoperability.py (217:262) - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (282:327) duplicated block id: 1575 size: 22 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon/checkpoint_reshaping_and_interoperability.py (217:262) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (278:323) duplicated block id: 1576 size: 22 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2.5_vl.py (134:155) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2.5_vl.py (182:204) duplicated block id: 1577 size: 22 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon40b/checkpoint_reshaping_and_interoperability.py (215:260) - toolkits/model_checkpoints_convertor/yi/checkpoint_reshaping_and_interoperability.py (231:276) duplicated block id: 1578 size: 22 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/chatglm/checkpoint_reshaping_and_interoperability.py (28:49) - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (56:77) duplicated block id: 1579 size: 22 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_vl.py (134:155) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_vl.py (182:204) duplicated block id: 1580 size: 22 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/chatglm/checkpoint_reshaping_and_interoperability.py (28:49) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (61:82) duplicated block id: 1581 size: 22 cleaned lines of code in 2 files: - megatron_patch/model/llama2/transformer.py (780:815) - megatron_patch/model/llama3/transformer_legacy.py (744:779) duplicated block id: 1582 size: 22 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/hf2te.py (131:176) - toolkits/model_checkpoints_convertor/chatglm/checkpoint_reshaping_and_interoperability.py (216:261) duplicated block id: 1583 size: 22 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/hf2te.py (132:177) - toolkits/model_checkpoints_convertor/falcon/checkpoint_reshaping_and_interoperability.py (217:262) duplicated block id: 1584 size: 22 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/language_model.py (494:520) - megatron_patch/model/qwen_vl/language_model.py (436:462) duplicated block id: 1585 size: 22 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (32:53) - toolkits/model_checkpoints_convertor/chatglm/checkpoint_reshaping_and_interoperability.py (28:49) duplicated block id: 1586 size: 22 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2.5_vl.py (182:204) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_vl.py (134:155) duplicated block id: 1587 size: 22 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer_legacy.py (744:779) - megatron_patch/model/llava/transformer.py (779:814) duplicated block id: 1588 size: 22 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (1045:1074) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (941:969) duplicated block id: 1589 size: 22 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/transformer/attention.py (127:152) - megatron_patch/model/qwen2_vl/attention.py (114:139) duplicated block id: 1590 size: 22 cleaned lines of code in 2 files: - megatron_patch/model/mixtral_bak/transformer/attention.py (73:98) - megatron_patch/model/qwen2_vl/attention.py (114:139) duplicated block id: 1591 size: 22 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/chatglm/checkpoint_reshaping_and_interoperability.py (28:49) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (60:81) duplicated block id: 1592 size: 22 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer_legacy.py (744:779) - megatron_patch/model/qwen1_5_megablocks/transformer.py (671:706) duplicated block id: 1593 size: 22 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/hf2te.py (492:520) - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (590:618) duplicated block id: 1594 size: 22 cleaned lines of code in 2 files: - megatron_patch/model/llama3/model.py (68:97) - megatron_patch/model/qwen2_vl/gpt_model.py (58:86) duplicated block id: 1595 size: 22 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/chatglm/checkpoint_reshaping_and_interoperability.py (447:476) - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (570:599) duplicated block id: 1596 size: 22 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2.5_vl.py (134:155) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_vl.py (182:204) duplicated block id: 1597 size: 22 cleaned lines of code in 2 files: - megatron_patch/model/qwen1_5/transformer/attention.py (99:124) - megatron_patch/model/qwen2_vl/attention.py (114:139) duplicated block id: 1598 size: 22 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (517:546) - toolkits/model_checkpoints_convertor/chatglm/checkpoint_reshaping_and_interoperability.py (447:476) duplicated block id: 1599 size: 22 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (244:289) - toolkits/model_checkpoints_convertor/falcon/checkpoint_reshaping_and_interoperability.py (217:262) duplicated block id: 1600 size: 22 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/language_model.py (494:520) - megatron_patch/model/starcoder/language_model.py (403:429) duplicated block id: 1601 size: 22 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer/attention.py (99:124) - megatron_patch/model/qwen2_vl/attention_vision.py (113:138) duplicated block id: 1602 size: 22 cleaned lines of code in 2 files: - megatron_patch/model/mixtral_bak/transformer/attention.py (73:98) - megatron_patch/model/qwen2_vl/attention_vision.py (113:138) duplicated block id: 1603 size: 22 cleaned lines of code in 2 files: - megatron_patch/model/qwen1_5/transformer/attention.py (99:124) - megatron_patch/model/qwen2_vl/attention_vision.py (113:138) duplicated block id: 1604 size: 22 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (286:326) - toolkits/model_checkpoints_convertor/chatglm/checkpoint_reshaping_and_interoperability.py (256:296) duplicated block id: 1605 size: 22 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/hf2te.py (492:520) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (664:692) duplicated block id: 1606 size: 22 cleaned lines of code in 2 files: - megatron_patch/tokenizer/tokenization_baichuan.py (184:231) - megatron_patch/tokenizer/tokenization_yi.py (208:255) duplicated block id: 1607 size: 22 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/chatglm/checkpoint_reshaping_and_interoperability.py (28:49) - toolkits/model_checkpoints_convertor/galactica/checkpoint_reshaping_and_interoperability.py (32:53) duplicated block id: 1608 size: 22 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/hf2te.py (492:520) - toolkits/model_checkpoints_convertor/yi/checkpoint_reshaping_and_interoperability.py (632:660) duplicated block id: 1609 size: 22 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/transformer.py (583:611) - megatron_patch/model/llama/transformer.py (712:740) duplicated block id: 1610 size: 22 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer_legacy.py (744:779) - megatron_patch/model/qwen_vl/transformer.py (779:814) duplicated block id: 1611 size: 22 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/hf2te.py (492:520) - toolkits/model_checkpoints_convertor/falcon40b/checkpoint_reshaping_and_interoperability.py (564:592) duplicated block id: 1612 size: 21 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/transformer/attention.py (608:637) - megatron_patch/model/mixtral_bak/transformer/attention.py (355:384) duplicated block id: 1613 size: 21 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (591:618) - toolkits/model_checkpoints_convertor/baichuan2/hf2te.py (465:492) duplicated block id: 1614 size: 21 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/hf2te.py (98:120) - toolkits/model_checkpoints_convertor/starcoder/checkpoint_reshaping_and_interoperability.py (109:130) duplicated block id: 1615 size: 21 cleaned lines of code in 2 files: - megatron_patch/model/galactica/language_model.py (597:626) - megatron_patch/model/glm130b/language_model.py (510:539) duplicated block id: 1616 size: 21 cleaned lines of code in 2 files: - megatron_patch/model/qwen2/model.py (69:96) - megatron_patch/model/qwen2_vl/gpt_model.py (58:84) duplicated block id: 1617 size: 21 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (61:83) - toolkits/model_checkpoints_convertor/mistral/hf2mcore.py (67:91) duplicated block id: 1618 size: 21 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/language_model.py (510:539) - megatron_patch/model/llama/language_model.py (579:608) duplicated block id: 1619 size: 21 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (628:660) - toolkits/model_checkpoints_convertor/bloom/checkpoint_reshaping_and_interoperability.py (302:334) duplicated block id: 1620 size: 21 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (115:139) - megatron_patch/model/llava/transformer.py (117:141) duplicated block id: 1621 size: 21 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore.py (93:117) - toolkits/model_checkpoints_convertor/yi/checkpoint_reshaping_and_interoperability.py (62:84) duplicated block id: 1622 size: 21 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (61:83) - toolkits/model_checkpoints_convertor/qwen/hf2megablocks_qwen1.5.py (92:117) duplicated block id: 1623 size: 21 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (115:139) - megatron_patch/model/llama3/transformer_legacy.py (121:145) duplicated block id: 1624 size: 21 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/checkpoint_reshaping_and_interoperability.py (302:334) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (785:817) duplicated block id: 1625 size: 21 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (35:56) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (62:82) duplicated block id: 1626 size: 21 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/transformer/attention.py (59:97) - megatron_patch/model/qwen2/transformer/attention.py (37:67) duplicated block id: 1627 size: 21 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2megablocks_qwen1.5.py (92:117) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (90:112) duplicated block id: 1628 size: 21 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/galactica/checkpoint_reshaping_and_interoperability.py (34:54) - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (35:56) duplicated block id: 1629 size: 21 cleaned lines of code in 2 files: - megatron_patch/model/falcon/transformer.py (891:921) - megatron_patch/model/galactica/transformer.py (553:583) duplicated block id: 1630 size: 21 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (115:139) - megatron_patch/model/qwen_vl/transformer.py (117:141) duplicated block id: 1631 size: 21 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon40b/checkpoint_reshaping_and_interoperability.py (35:55) - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (35:56) duplicated block id: 1632 size: 21 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/hf2te.py (99:121) - toolkits/model_checkpoints_convertor/falcon/checkpoint_reshaping_and_interoperability.py (131:152) duplicated block id: 1633 size: 21 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_dense_and_moe_gqa.py (567:587) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_vl.py (668:688) duplicated block id: 1634 size: 21 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon/checkpoint_reshaping_and_interoperability.py (62:84) - toolkits/model_checkpoints_convertor/llama/hf2mcore.py (93:117) duplicated block id: 1635 size: 21 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (848:884) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (911:947) duplicated block id: 1636 size: 21 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/transformer.py (399:423) - megatron_patch/model/falcon40b/transformer.py (450:474) duplicated block id: 1637 size: 21 cleaned lines of code in 2 files: - megatron_patch/model/mixtral_bak/transformer/attention.py (355:384) - megatron_patch/model/qwen2_vl/attention.py (609:638) duplicated block id: 1638 size: 21 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_dense_and_moe_gqa.py (27:52) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_vl.py (43:68) duplicated block id: 1639 size: 21 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/chatglm/checkpoint_reshaping_and_interoperability.py (51:73) - toolkits/model_checkpoints_convertor/mistral/hf2mcore.py (67:91) duplicated block id: 1640 size: 21 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer/attention.py (41:71) - megatron_patch/model/qwen2_vl/attention_vision.py (45:83) duplicated block id: 1641 size: 21 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/glm/checkpoint_reshaping_and_interoperability.py (230:269) - toolkits/model_checkpoints_convertor/starcoder/checkpoint_reshaping_and_interoperability.py (282:321) duplicated block id: 1642 size: 21 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/transformer.py (450:474) - megatron_patch/model/llama/transformer.py (460:484) duplicated block id: 1643 size: 21 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/hf2te.py (99:121) - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (154:175) duplicated block id: 1644 size: 21 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (85:107) - toolkits/model_checkpoints_convertor/qwen/hf2megablocks_qwen1.5.py (92:117) duplicated block id: 1645 size: 21 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/hf2te.py (98:120) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (158:179) duplicated block id: 1646 size: 21 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (85:107) - toolkits/model_checkpoints_convertor/mistral/hf2mcore.py (67:91) duplicated block id: 1647 size: 21 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/moe/shared_experts.py (158:180) - megatron_patch/model/mixtral/transformer/mlp.py (115:137) duplicated block id: 1648 size: 21 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/hf2te.py (98:120) - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (154:175) duplicated block id: 1649 size: 21 cleaned lines of code in 2 files: - megatron_patch/model/galactica/language_model.py (258:289) - megatron_patch/model/glm130b/language_model.py (243:274) duplicated block id: 1650 size: 21 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/deepseek/hf2mcore_deepseek_v2_moe.py (39:64) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_vl.py (43:68) duplicated block id: 1651 size: 21 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer/attention.py (41:71) - megatron_patch/model/mixtral/transformer/attention.py (59:97) duplicated block id: 1652 size: 21 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/hf2te.py (98:120) - toolkits/model_checkpoints_convertor/chatglm/checkpoint_reshaping_and_interoperability.py (113:134) duplicated block id: 1653 size: 21 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore.py (93:117) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (89:111) duplicated block id: 1654 size: 21 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/hf2te.py (99:121) - toolkits/model_checkpoints_convertor/bloom/checkpoint_reshaping_and_interoperability.py (109:130) duplicated block id: 1655 size: 21 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon/checkpoint_reshaping_and_interoperability.py (62:84) - toolkits/model_checkpoints_convertor/qwen/hf2megablocks_qwen1.5.py (92:117) duplicated block id: 1656 size: 21 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (130:151) - toolkits/model_checkpoints_convertor/baichuan/hf2te.py (99:121) duplicated block id: 1657 size: 21 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/hf2te.py (98:120) - toolkits/model_checkpoints_convertor/bloom/reward_model_to_megatron.py (109:130) duplicated block id: 1658 size: 21 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon/checkpoint_reshaping_and_interoperability.py (35:55) - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (35:56) duplicated block id: 1659 size: 21 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/transformer.py (521:551) - megatron_patch/model/falcon40b/transformer.py (661:691) duplicated block id: 1660 size: 21 cleaned lines of code in 2 files: - megatron_patch/model/qwen2/transformer/attention.py (37:67) - megatron_patch/model/qwen2_vl/attention.py (45:83) duplicated block id: 1661 size: 21 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/hf2te.py (99:121) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (159:180) duplicated block id: 1662 size: 21 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/language_model.py (547:576) - megatron_patch/model/glm130b/language_model.py (510:539) duplicated block id: 1663 size: 21 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (35:56) - toolkits/model_checkpoints_convertor/yi/checkpoint_reshaping_and_interoperability.py (35:55) duplicated block id: 1664 size: 21 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/hf2te.py (98:120) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (159:180) duplicated block id: 1665 size: 21 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/mlp.py (107:129) - megatron_patch/model/deepseek_v2/moe/shared_experts.py (158:180) duplicated block id: 1666 size: 21 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/hf2te.py (465:492) - toolkits/model_checkpoints_convertor/yi/checkpoint_reshaping_and_interoperability.py (633:660) duplicated block id: 1667 size: 21 cleaned lines of code in 2 files: - megatron_patch/model/falcon/language_model.py (568:597) - megatron_patch/model/glm130b/language_model.py (510:539) duplicated block id: 1668 size: 21 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2megablocks_qwen1.5.py (92:117) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (89:111) duplicated block id: 1669 size: 21 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore.py (93:117) - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (85:107) duplicated block id: 1670 size: 21 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/hf2te.py (465:492) - toolkits/model_checkpoints_convertor/falcon40b/checkpoint_reshaping_and_interoperability.py (565:592) duplicated block id: 1671 size: 21 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/hf2te.py (465:492) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (748:775) duplicated block id: 1672 size: 21 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (130:151) - toolkits/model_checkpoints_convertor/baichuan2/hf2te.py (98:120) duplicated block id: 1673 size: 21 cleaned lines of code in 2 files: - megatron_patch/model/qwen1_5/moe/token_dispatcher.py (54:118) - megatron_patch/model/qwen2/moe/token_dispatcher.py (55:119) duplicated block id: 1674 size: 21 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (868:895) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (1053:1081) duplicated block id: 1675 size: 21 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2.5_vl.py (661:681) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_dense_and_moe_gqa.py (567:587) duplicated block id: 1676 size: 21 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/hf2te.py (465:492) - toolkits/model_checkpoints_convertor/falcon/checkpoint_reshaping_and_interoperability.py (561:589) duplicated block id: 1677 size: 21 cleaned lines of code in 2 files: - megatron_patch/model/mixtral_bak/transformer/attention.py (355:384) - megatron_patch/model/qwen2_vl/attention_vision.py (608:637) duplicated block id: 1678 size: 21 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/mistral/hf2mcore.py (67:91) - toolkits/model_checkpoints_convertor/yi/checkpoint_reshaping_and_interoperability.py (62:84) duplicated block id: 1679 size: 21 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (115:139) - megatron_patch/model/mistral/transformer.py (117:141) duplicated block id: 1680 size: 21 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (61:83) - toolkits/model_checkpoints_convertor/llama/hf2mcore.py (93:117) duplicated block id: 1681 size: 21 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (34:54) - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (35:56) duplicated block id: 1682 size: 21 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/hf2te.py (465:492) - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (733:760) duplicated block id: 1683 size: 21 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/galactica/checkpoint_reshaping_and_interoperability.py (61:83) - toolkits/model_checkpoints_convertor/qwen/hf2megablocks_qwen1.5.py (92:117) duplicated block id: 1684 size: 21 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/hf2te.py (99:121) - toolkits/model_checkpoints_convertor/yi/checkpoint_reshaping_and_interoperability.py (131:152) duplicated block id: 1685 size: 21 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2.5_vl.py (43:68) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_dense_and_moe_gqa.py (27:52) duplicated block id: 1686 size: 21 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (1158:1201) - megatron_patch/model/falcon40b/transformer.py (995:1038) duplicated block id: 1687 size: 21 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/language_model.py (243:274) - megatron_patch/model/llama/language_model.py (227:258) duplicated block id: 1688 size: 21 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2megablocks_qwen1.5.py (92:117) - toolkits/model_checkpoints_convertor/yi/checkpoint_reshaping_and_interoperability.py (62:84) duplicated block id: 1689 size: 21 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/moe/shared_experts.py (158:180) - megatron_patch/model/qwen2/transformer/mlp.py (128:150) duplicated block id: 1690 size: 21 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_dense_and_moe_gqa.py (717:737) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_moe.py (599:620) duplicated block id: 1691 size: 21 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (551:583) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (680:712) duplicated block id: 1692 size: 21 cleaned lines of code in 2 files: - megatron_patch/model/falcon/transformer.py (568:590) - megatron_patch/model/falcon40b/transformer.py (391:414) duplicated block id: 1693 size: 21 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/checkpoint_reshaping_and_interoperability.py (260:299) - toolkits/model_checkpoints_convertor/glm130b/checkpoint_reshaping_and_interoperability.py (230:269) duplicated block id: 1694 size: 21 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/hf2te.py (98:120) - toolkits/model_checkpoints_convertor/bloom/checkpoint_reshaping_and_interoperability.py (109:130) duplicated block id: 1695 size: 21 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/transformer/attention.py (59:97) - megatron_patch/model/qwen1_5/transformer/attention.py (41:71) duplicated block id: 1696 size: 21 cleaned lines of code in 2 files: - megatron_patch/model/bloom/language_model.py (245:276) - megatron_patch/model/glm130b/language_model.py (243:274) duplicated block id: 1697 size: 21 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/hf2te.py (99:121) - toolkits/model_checkpoints_convertor/starcoder/checkpoint_reshaping_and_interoperability.py (109:130) duplicated block id: 1698 size: 21 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (61:83) - toolkits/model_checkpoints_convertor/mistral/hf2mcore.py (67:91) duplicated block id: 1699 size: 21 cleaned lines of code in 2 files: - megatron_patch/model/bloom/language_model.py (481:510) - megatron_patch/model/glm130b/language_model.py (510:539) duplicated block id: 1700 size: 21 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon40b/checkpoint_reshaping_and_interoperability.py (62:84) - toolkits/model_checkpoints_convertor/llama/hf2mcore.py (93:117) duplicated block id: 1701 size: 21 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/language_model.py (211:241) - megatron_patch/model/llama3/language_model.py (149:179) duplicated block id: 1702 size: 21 cleaned lines of code in 2 files: - megatron_patch/model/mixtral_bak/moe/router.py (119:150) - megatron_patch/model/qwen2/moe/router.py (194:225) duplicated block id: 1703 size: 21 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (1158:1201) - megatron_patch/model/llama/transformer.py (1050:1093) duplicated block id: 1704 size: 21 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/hf2te.py (99:121) - toolkits/model_checkpoints_convertor/chatglm/checkpoint_reshaping_and_interoperability.py (113:134) duplicated block id: 1705 size: 21 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/hf2te.py (99:121) - toolkits/model_checkpoints_convertor/falcon40b/checkpoint_reshaping_and_interoperability.py (131:152) duplicated block id: 1706 size: 21 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/hf2te.py (98:120) - toolkits/model_checkpoints_convertor/yi/checkpoint_reshaping_and_interoperability.py (131:152) duplicated block id: 1707 size: 21 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (775:811) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (848:884) duplicated block id: 1708 size: 21 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (656:688) - toolkits/model_checkpoints_convertor/bloom/checkpoint_reshaping_and_interoperability.py (302:334) duplicated block id: 1709 size: 21 cleaned lines of code in 2 files: - megatron_patch/model/llama3_1/model.py (83:110) - megatron_patch/model/qwen2_vl/gpt_model.py (58:84) duplicated block id: 1710 size: 21 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (551:583) - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (748:781) duplicated block id: 1711 size: 21 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/language_model.py (568:597) - megatron_patch/model/glm130b/language_model.py (510:539) duplicated block id: 1712 size: 21 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/hf2te.py (98:120) - toolkits/model_checkpoints_convertor/falcon/checkpoint_reshaping_and_interoperability.py (131:152) duplicated block id: 1713 size: 21 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/galactica/checkpoint_reshaping_and_interoperability.py (61:83) - toolkits/model_checkpoints_convertor/mistral/hf2mcore.py (67:91) duplicated block id: 1714 size: 21 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (61:83) - toolkits/model_checkpoints_convertor/llama/hf2mcore.py (93:117) duplicated block id: 1715 size: 21 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (115:139) - megatron_patch/model/baichuan2/transformer.py (112:136) duplicated block id: 1716 size: 21 cleaned lines of code in 2 files: - megatron_patch/model/falcon/language_model.py (227:258) - megatron_patch/model/glm130b/language_model.py (243:274) duplicated block id: 1717 size: 21 cleaned lines of code in 2 files: - megatron_patch/model/falcon/transformer.py (891:921) - megatron_patch/model/falcon40b/transformer.py (661:691) duplicated block id: 1718 size: 21 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer/attention.py (41:71) - megatron_patch/model/qwen2_vl/attention.py (45:83) duplicated block id: 1719 size: 21 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (58:78) - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (35:56) duplicated block id: 1720 size: 21 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/mistral/hf2mcore.py (67:91) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (90:112) duplicated block id: 1721 size: 21 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon40b/checkpoint_reshaping_and_interoperability.py (62:84) - toolkits/model_checkpoints_convertor/mistral/hf2mcore.py (67:91) duplicated block id: 1722 size: 21 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (115:139) - megatron_patch/model/qwen1_5_megablocks/transformer.py (115:139) duplicated block id: 1723 size: 21 cleaned lines of code in 2 files: - megatron_patch/model/mixtral_bak/moe/router.py (119:150) - megatron_patch/model/qwen1_5/moe/router.py (120:151) duplicated block id: 1724 size: 21 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/hf2te.py (99:121) - toolkits/model_checkpoints_convertor/galactica/checkpoint_reshaping_and_interoperability.py (130:151) duplicated block id: 1725 size: 21 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (825:856) - megatron_patch/model/llama3/transformer_legacy.py (813:844) duplicated block id: 1726 size: 21 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/reward_model_to_megatron.py (260:299) - toolkits/model_checkpoints_convertor/glm130b/checkpoint_reshaping_and_interoperability.py (230:269) duplicated block id: 1727 size: 21 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon40b/checkpoint_reshaping_and_interoperability.py (62:84) - toolkits/model_checkpoints_convertor/qwen/hf2megablocks_qwen1.5.py (92:117) duplicated block id: 1728 size: 21 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (130:151) - toolkits/model_checkpoints_convertor/baichuan2/hf2te.py (98:120) duplicated block id: 1729 size: 21 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (115:139) - megatron_patch/model/qwen/transformer.py (111:135) duplicated block id: 1730 size: 21 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/checkpoint_reshaping_and_interoperability.py (260:299) - toolkits/model_checkpoints_convertor/glm/checkpoint_reshaping_and_interoperability.py (230:269) duplicated block id: 1731 size: 21 cleaned lines of code in 2 files: - megatron_patch/model/qwen2/transformer/attention.py (37:67) - megatron_patch/model/qwen2_vl/attention_vision.py (45:83) duplicated block id: 1732 size: 21 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/chatglm/checkpoint_reshaping_and_interoperability.py (51:73) - toolkits/model_checkpoints_convertor/qwen/hf2megablocks_qwen1.5.py (92:117) duplicated block id: 1733 size: 21 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore.py (748:774) - toolkits/model_checkpoints_convertor/qwen/hf2megablocks_qwen1.5.py (628:651) duplicated block id: 1734 size: 21 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/hf2te.py (99:121) - toolkits/model_checkpoints_convertor/bloom/reward_model_to_megatron.py (109:130) duplicated block id: 1735 size: 21 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (1076:1101) - megatron_patch/model/starcoder/transformer.py (1225:1250) duplicated block id: 1736 size: 21 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/hf2te.py (99:121) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (158:179) duplicated block id: 1737 size: 21 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (34:54) - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (35:56) duplicated block id: 1738 size: 21 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (748:784) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (848:884) duplicated block id: 1739 size: 21 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (61:83) - toolkits/model_checkpoints_convertor/qwen/hf2megablocks_qwen1.5.py (92:117) duplicated block id: 1740 size: 21 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (1158:1201) - megatron_patch/model/falcon/transformer.py (1256:1299) duplicated block id: 1741 size: 21 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/galactica/checkpoint_reshaping_and_interoperability.py (61:83) - toolkits/model_checkpoints_convertor/llama/hf2mcore.py (93:117) duplicated block id: 1742 size: 21 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (889:925) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (848:884) duplicated block id: 1743 size: 21 cleaned lines of code in 2 files: - megatron_patch/model/falcon/transformer.py (568:590) - megatron_patch/model/galactica/transformer.py (340:363) duplicated block id: 1744 size: 21 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/chatglm/checkpoint_reshaping_and_interoperability.py (51:73) - toolkits/model_checkpoints_convertor/llama/hf2mcore.py (93:117) duplicated block id: 1745 size: 21 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (868:895) - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (1045:1073) duplicated block id: 1746 size: 21 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/language_model.py (227:258) - megatron_patch/model/glm130b/language_model.py (243:274) duplicated block id: 1747 size: 21 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/hf2te.py (465:492) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (665:692) duplicated block id: 1748 size: 21 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/hf2te.py (98:120) - toolkits/model_checkpoints_convertor/galactica/checkpoint_reshaping_and_interoperability.py (130:151) duplicated block id: 1749 size: 21 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (667:689) - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (782:804) duplicated block id: 1750 size: 21 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (782:804) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (796:818) duplicated block id: 1751 size: 21 cleaned lines of code in 2 files: - megatron_patch/model/galactica/language_model.py (166:203) - megatron_patch/model/glm130b/language_model.py (139:176) duplicated block id: 1752 size: 21 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon/checkpoint_reshaping_and_interoperability.py (62:84) - toolkits/model_checkpoints_convertor/mistral/hf2mcore.py (67:91) duplicated block id: 1753 size: 21 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore.py (93:117) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (90:112) duplicated block id: 1754 size: 21 cleaned lines of code in 2 files: - megatron_patch/model/qwen1_5/transformer/attention.py (41:71) - megatron_patch/model/qwen2_vl/attention.py (45:83) duplicated block id: 1755 size: 21 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (35:56) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (63:83) duplicated block id: 1756 size: 21 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/glm130b/checkpoint_reshaping_and_interoperability.py (230:269) - toolkits/model_checkpoints_convertor/starcoder/checkpoint_reshaping_and_interoperability.py (282:321) duplicated block id: 1757 size: 21 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (968:996) - megatron_patch/model/starcoder/transformer.py (1029:1057) duplicated block id: 1758 size: 21 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/hf2te.py (98:120) - toolkits/model_checkpoints_convertor/falcon40b/checkpoint_reshaping_and_interoperability.py (131:152) duplicated block id: 1759 size: 21 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/transformer.py (521:551) - megatron_patch/model/llama/transformer.py (663:693) duplicated block id: 1760 size: 21 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/hf2te.py (99:121) - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (130:151) duplicated block id: 1761 size: 21 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (115:139) - megatron_patch/model/llama2/transformer.py (117:141) duplicated block id: 1762 size: 21 cleaned lines of code in 2 files: - megatron_patch/model/falcon/transformer.py (744:777) - megatron_patch/model/starcoder/transformer.py (716:751) duplicated block id: 1763 size: 21 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/transformer.py (521:551) - megatron_patch/model/falcon/transformer.py (891:921) duplicated block id: 1764 size: 21 cleaned lines of code in 2 files: - megatron_patch/model/qwen1_5/transformer/attention.py (41:71) - megatron_patch/model/qwen2_vl/attention_vision.py (45:83) duplicated block id: 1765 size: 21 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/deepseek/hf2mcore_deepseek_v2_moe.py (66:92) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_moe.py (134:160) duplicated block id: 1766 size: 21 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/moe/shared_experts.py (158:180) - megatron_patch/model/qwen2/moe/experts.py (281:303) duplicated block id: 1767 size: 21 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/mistral/hf2mcore.py (67:91) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (89:111) duplicated block id: 1768 size: 21 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/deepseek/hf2mcore_deepseek_v2_moe.py (39:64) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2.5_vl.py (43:68) duplicated block id: 1769 size: 21 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/reward_model_to_megatron.py (260:299) - toolkits/model_checkpoints_convertor/glm/checkpoint_reshaping_and_interoperability.py (230:269) duplicated block id: 1770 size: 20 cleaned lines of code in 2 files: - megatron_patch/model/mixtral_bak/transformer/attention.py (46:72) - megatron_patch/model/qwen2_vl/attention.py (85:111) duplicated block id: 1771 size: 20 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/model.py (65:89) - megatron_patch/model/mixtral_bak/model.py (64:88) duplicated block id: 1772 size: 20 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (390:412) - megatron_patch/model/falcon40b/transformer.py (391:413) duplicated block id: 1773 size: 20 cleaned lines of code in 2 files: - megatron_patch/model/llava/gpt_model.py (26:56) - megatron_patch/model/starcoder/gpt_model.py (28:58) duplicated block id: 1774 size: 20 cleaned lines of code in 2 files: - megatron_patch/model/qwen_vl/gpt_model.py (26:56) - megatron_patch/model/starcoder/gpt_model.py (28:58) duplicated block id: 1775 size: 20 cleaned lines of code in 2 files: - megatron_patch/model/mistral/gpt_model.py (26:56) - megatron_patch/model/starcoder/gpt_model.py (28:58) duplicated block id: 1776 size: 20 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (802:827) - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (1099:1124) duplicated block id: 1777 size: 20 cleaned lines of code in 2 files: - megatron_patch/model/llava_mcore/llava_spec.py (35:55) - megatron_patch/model/qwen1_5/layer_specs.py (38:57) duplicated block id: 1778 size: 20 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer/attention.py (72:98) - megatron_patch/model/qwen2_vl/attention.py (85:111) duplicated block id: 1779 size: 20 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (82:104) - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (63:84) duplicated block id: 1780 size: 20 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/chatglm/checkpoint_reshaping_and_interoperability.py (258:296) - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (339:377) duplicated block id: 1781 size: 20 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/language_model.py (635:663) - megatron_patch/model/llama3/language_model.py (565:593) duplicated block id: 1782 size: 20 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2.5_vl.py (605:624) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_dense_and_moe_gqa.py (568:587) duplicated block id: 1783 size: 20 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2.5_vl.py (605:624) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_dense_and_moe_gqa.py (669:688) duplicated block id: 1784 size: 20 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/chatglm/checkpoint_reshaping_and_interoperability.py (258:296) - toolkits/model_checkpoints_convertor/galactica/checkpoint_reshaping_and_interoperability.py (290:328) duplicated block id: 1785 size: 20 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/moe/experts.py (80:109) - megatron_patch/model/qwen2/moe/experts.py (82:111) duplicated block id: 1786 size: 20 cleaned lines of code in 2 files: - megatron_patch/model/llama/language_model.py (646:674) - megatron_patch/model/llama3/language_model.py (565:593) duplicated block id: 1787 size: 20 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/transformer/attention.py (99:125) - megatron_patch/model/qwen2/transformer/attention.py (68:94) duplicated block id: 1788 size: 20 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_vl.py (612:631) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_vl.py (669:688) duplicated block id: 1789 size: 20 cleaned lines of code in 2 files: - megatron_patch/model/qwen1_5/transformer/attention.py (72:98) - megatron_patch/model/qwen2_vl/attention.py (85:111) duplicated block id: 1790 size: 20 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (802:827) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (994:1019) duplicated block id: 1791 size: 20 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/chatglm/checkpoint_reshaping_and_interoperability.py (258:296) - toolkits/model_checkpoints_convertor/yi/checkpoint_reshaping_and_interoperability.py (288:326) duplicated block id: 1792 size: 20 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/transformer/attention.py (99:125) - megatron_patch/model/mixtral_bak/transformer/attention.py (46:72) duplicated block id: 1793 size: 20 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/hf2te.py (332:354) - toolkits/model_checkpoints_convertor/falcon/checkpoint_reshaping_and_interoperability.py (470:493) duplicated block id: 1794 size: 20 cleaned lines of code in 2 files: - megatron_patch/model/qwen1_5/layer_specs.py (89:112) - megatron_patch/model/qwen2/layer_specs.py (99:122) duplicated block id: 1795 size: 20 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2.5_vl.py (360:380) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_vl.py (359:379) duplicated block id: 1796 size: 20 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_dense_and_moe_gqa.py (669:688) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_vl.py (612:631) duplicated block id: 1797 size: 20 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/transformer.py (332:354) - megatron_patch/model/galactica/transformer.py (340:362) duplicated block id: 1798 size: 20 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (1615:1661) - megatron_patch/model/glm130b/transformer.py (1190:1235) duplicated block id: 1799 size: 20 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/galactica/checkpoint_reshaping_and_interoperability.py (529:554) - toolkits/model_checkpoints_convertor/yi/checkpoint_reshaping_and_interoperability.py (500:525) duplicated block id: 1800 size: 20 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/checkpoint_reshaping_and_interoperability.py (313:334) - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (782:803) duplicated block id: 1801 size: 20 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/moe/experts.py (81:110) - megatron_patch/model/qwen2/moe/experts.py (82:111) duplicated block id: 1802 size: 20 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/deepspeed_to_megatron.py (69:95) - toolkits/model_checkpoints_convertor/bloom/deepspeed_to_megatron_ori.py (66:92) duplicated block id: 1803 size: 20 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/hf2te.py (332:354) - toolkits/model_checkpoints_convertor/falcon40b/checkpoint_reshaping_and_interoperability.py (459:482) duplicated block id: 1804 size: 20 cleaned lines of code in 2 files: - megatron_patch/model/falcon/language_model.py (635:663) - megatron_patch/model/llama3/language_model.py (565:593) duplicated block id: 1805 size: 20 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_dense_and_moe_gqa.py (568:587) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_vl.py (612:631) duplicated block id: 1806 size: 20 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/transformer/attention.py (99:125) - megatron_patch/model/qwen1_5/transformer/attention.py (72:98) duplicated block id: 1807 size: 20 cleaned lines of code in 2 files: - megatron_patch/model/bloom/language_model.py (287:313) - megatron_patch/model/starcoder/language_model.py (271:298) duplicated block id: 1808 size: 20 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/moe/experts.py (638:658) - megatron_patch/model/qwen2/moe/experts.py (273:293) duplicated block id: 1809 size: 20 cleaned lines of code in 2 files: - megatron_patch/model/llama3_1/layer_specs.py (39:58) - megatron_patch/model/llava_mcore/llava_spec.py (35:55) duplicated block id: 1810 size: 20 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon/checkpoint_reshaping_and_interoperability.py (599:630) - toolkits/model_checkpoints_convertor/starcoder/checkpoint_reshaping_and_interoperability.py (324:355) duplicated block id: 1811 size: 20 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/reward_model_to_megatron.py (798:820) - toolkits/model_checkpoints_convertor/glm130b/checkpoint_reshaping_and_interoperability.py (507:529) duplicated block id: 1812 size: 20 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (390:412) - megatron_patch/model/llama/transformer.py (378:400) duplicated block id: 1813 size: 20 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/transformer.py (332:354) - megatron_patch/model/llama/transformer.py (378:400) duplicated block id: 1814 size: 20 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (551:576) - toolkits/model_checkpoints_convertor/yi/checkpoint_reshaping_and_interoperability.py (500:525) duplicated block id: 1815 size: 20 cleaned lines of code in 2 files: - megatron_patch/model/llava_mcore/layer_specs.py (81:100) - megatron_patch/model/qwen2_vl/layer_specs.py (77:96) duplicated block id: 1816 size: 20 cleaned lines of code in 2 files: - megatron_patch/model/bloom/language_model.py (543:571) - megatron_patch/model/llama3/language_model.py (565:593) duplicated block id: 1817 size: 20 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer/attention.py (72:98) - megatron_patch/model/mixtral/transformer/attention.py (99:125) duplicated block id: 1818 size: 20 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon40b/checkpoint_reshaping_and_interoperability.py (471:496) - toolkits/model_checkpoints_convertor/yi/checkpoint_reshaping_and_interoperability.py (500:525) duplicated block id: 1819 size: 20 cleaned lines of code in 2 files: - megatron_patch/model/falcon/transformer.py (568:589) - megatron_patch/model/llama/transformer.py (378:400) duplicated block id: 1820 size: 20 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/chatglm/checkpoint_reshaping_and_interoperability.py (30:49) - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (35:54) duplicated block id: 1821 size: 20 cleaned lines of code in 2 files: - megatron_patch/model/mixtral_bak/layer_specs.py (110:145) - megatron_patch/model/qwen2/layer_specs.py (99:122) duplicated block id: 1822 size: 20 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/hf2te.py (51:73) - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (63:84) duplicated block id: 1823 size: 20 cleaned lines of code in 2 files: - megatron_patch/model/falcon/language_model.py (269:295) - megatron_patch/model/starcoder/language_model.py (271:298) duplicated block id: 1824 size: 20 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/hf2te.py (50:72) - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (63:84) duplicated block id: 1825 size: 20 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/language_model.py (609:637) - megatron_patch/model/llama3/language_model.py (565:593) duplicated block id: 1826 size: 20 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/chatglm/checkpoint_reshaping_and_interoperability.py (258:296) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (358:396) duplicated block id: 1827 size: 20 cleaned lines of code in 2 files: - megatron_patch/model/qwen2/transformer/attention.py (68:94) - megatron_patch/model/qwen2_vl/attention.py (85:111) duplicated block id: 1828 size: 20 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/transformer.py (332:354) - megatron_patch/model/falcon40b/transformer.py (391:413) duplicated block id: 1829 size: 20 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer/attention.py (72:98) - megatron_patch/model/qwen2_vl/attention_vision.py (85:111) duplicated block id: 1830 size: 20 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/transformer.py (1190:1235) - megatron_patch/model/llama3/transformer_legacy.py (1705:1751) duplicated block id: 1831 size: 20 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore.py (234:257) - toolkits/model_checkpoints_convertor/mistral/hf2mcore.py (172:195) duplicated block id: 1832 size: 20 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (983:1008) - megatron_patch/model/starcoder/transformer.py (1144:1169) duplicated block id: 1833 size: 20 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/transformer.py (332:354) - megatron_patch/model/falcon/transformer.py (568:589) duplicated block id: 1834 size: 20 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (390:412) - megatron_patch/model/galactica/transformer.py (340:362) duplicated block id: 1835 size: 20 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/moe/moe_layer.py (131:153) - megatron_patch/model/qwen3_moe/moe/moe_layer.py (63:85) duplicated block id: 1836 size: 20 cleaned lines of code in 2 files: - megatron_patch/model/qwen2/transformer/attention.py (68:94) - megatron_patch/model/qwen2_vl/attention_vision.py (85:111) duplicated block id: 1837 size: 20 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/hf2te.py (360:382) - toolkits/model_checkpoints_convertor/falcon40b/checkpoint_reshaping_and_interoperability.py (459:482) duplicated block id: 1838 size: 20 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2.5_vl.py (605:624) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_vl.py (669:688) duplicated block id: 1839 size: 20 cleaned lines of code in 2 files: - megatron_patch/model/llama3/layer_specs.py (38:57) - megatron_patch/model/llava_mcore/llava_spec.py (35:55) duplicated block id: 1840 size: 20 cleaned lines of code in 2 files: - megatron_patch/model/qwen1_5_megablocks/gpt_model.py (26:56) - megatron_patch/model/starcoder/gpt_model.py (28:58) duplicated block id: 1841 size: 20 cleaned lines of code in 2 files: - megatron_patch/model/mixtral_bak/transformer/attention.py (46:72) - megatron_patch/model/qwen2_vl/attention_vision.py (85:111) duplicated block id: 1842 size: 20 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/hf2te.py (360:382) - toolkits/model_checkpoints_convertor/falcon/checkpoint_reshaping_and_interoperability.py (470:493) duplicated block id: 1843 size: 20 cleaned lines of code in 2 files: - megatron_patch/model/qwen1_5/transformer/attention.py (72:98) - megatron_patch/model/qwen2_vl/attention_vision.py (85:111) duplicated block id: 1844 size: 20 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/chatglm/checkpoint_reshaping_and_interoperability.py (451:476) - toolkits/model_checkpoints_convertor/yi/checkpoint_reshaping_and_interoperability.py (500:525) duplicated block id: 1845 size: 20 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/glm130b/checkpoint_reshaping_and_interoperability.py (507:529) - toolkits/model_checkpoints_convertor/starcoder/checkpoint_reshaping_and_interoperability.py (808:830) duplicated block id: 1846 size: 20 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon40b/checkpoint_reshaping_and_interoperability.py (602:633) - toolkits/model_checkpoints_convertor/starcoder/checkpoint_reshaping_and_interoperability.py (324:355) duplicated block id: 1847 size: 20 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon/checkpoint_reshaping_and_interoperability.py (482:507) - toolkits/model_checkpoints_convertor/yi/checkpoint_reshaping_and_interoperability.py (500:525) duplicated block id: 1848 size: 20 cleaned lines of code in 2 files: - megatron_patch/model/falcon/transformer.py (265:310) - megatron_patch/model/glm130b/transformer.py (275:319) duplicated block id: 1849 size: 20 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (156:180) - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (180:204) duplicated block id: 1850 size: 20 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (49:70) - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (29:49) duplicated block id: 1851 size: 20 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/checkpoint_reshaping_and_interoperability.py (799:821) - toolkits/model_checkpoints_convertor/glm130b/checkpoint_reshaping_and_interoperability.py (507:529) duplicated block id: 1852 size: 20 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (390:412) - megatron_patch/model/falcon/transformer.py (568:589) duplicated block id: 1853 size: 20 cleaned lines of code in 2 files: - megatron_patch/model/llama2/gpt_model.py (26:56) - megatron_patch/model/starcoder/gpt_model.py (28:58) duplicated block id: 1854 size: 20 cleaned lines of code in 2 files: - megatron_patch/model/galactica/language_model.py (300:326) - megatron_patch/model/starcoder/language_model.py (271:298) duplicated block id: 1855 size: 20 cleaned lines of code in 2 files: - megatron_patch/model/falcon/transformer.py (656:680) - megatron_patch/model/galactica/transformer.py (411:435) duplicated block id: 1856 size: 20 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/language_model.py (285:311) - megatron_patch/model/starcoder/language_model.py (271:298) duplicated block id: 1857 size: 20 cleaned lines of code in 2 files: - megatron_patch/model/galactica/language_model.py (664:692) - megatron_patch/model/llama3/language_model.py (565:593) duplicated block id: 1858 size: 20 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2.5_vl.py (605:624) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2.5_vl.py (662:681) duplicated block id: 1859 size: 20 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/language_model.py (573:601) - megatron_patch/model/llama3/language_model.py (565:593) duplicated block id: 1860 size: 20 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/language_model.py (269:295) - megatron_patch/model/starcoder/language_model.py (271:298) duplicated block id: 1861 size: 20 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (1745:1791) - megatron_patch/model/glm130b/transformer.py (1190:1235) duplicated block id: 1862 size: 20 cleaned lines of code in 2 files: - megatron_patch/model/llava_mcore/llava_spec.py (35:55) - megatron_patch/model/qwen2/layer_specs.py (48:67) duplicated block id: 1863 size: 20 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/deepseek/hf2mcore_deepseek_v3_moe.py (579:598) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_moe.py (579:598) duplicated block id: 1864 size: 20 cleaned lines of code in 2 files: - megatron_patch/model/llama3/gpt_model.py (28:58) - megatron_patch/model/starcoder/gpt_model.py (28:58) duplicated block id: 1865 size: 20 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (551:573) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (763:785) duplicated block id: 1866 size: 20 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/moe/experts.py (637:657) - megatron_patch/model/qwen2/moe/experts.py (273:293) duplicated block id: 1867 size: 20 cleaned lines of code in 2 files: - megatron_patch/model/qwen/gpt_model.py (26:56) - megatron_patch/model/starcoder/gpt_model.py (28:58) duplicated block id: 1868 size: 20 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/transformer.py (391:413) - megatron_patch/model/llama/transformer.py (378:400) duplicated block id: 1869 size: 20 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/transformer.py (1190:1235) - megatron_patch/model/qwen/transformer.py (1687:1733) duplicated block id: 1870 size: 20 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2.5_vl.py (662:681) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_vl.py (612:631) duplicated block id: 1871 size: 20 cleaned lines of code in 2 files: - megatron_patch/model/llama/language_model.py (269:295) - megatron_patch/model/starcoder/language_model.py (271:298) duplicated block id: 1872 size: 19 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llava/hf2mcore_llava.py (41:62) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_mha_to_moe.py (28:49) duplicated block id: 1873 size: 19 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (85:104) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (93:111) duplicated block id: 1874 size: 19 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (65:84) - toolkits/model_checkpoints_convertor/yi/checkpoint_reshaping_and_interoperability.py (66:84) duplicated block id: 1875 size: 19 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/chatglm/checkpoint_reshaping_and_interoperability.py (561:580) - toolkits/model_checkpoints_convertor/falcon40b/checkpoint_reshaping_and_interoperability.py (580:599) duplicated block id: 1876 size: 19 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/mistral/hf2mcore.py (229:247) - toolkits/model_checkpoints_convertor/qwen/hf2megablocks_qwen1.5.py (283:301) duplicated block id: 1877 size: 19 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/moe/router.py (106:136) - megatron_patch/model/qwen2/moe/router.py (189:218) duplicated block id: 1878 size: 19 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer_legacy.py (710:743) - megatron_patch/model/qwen1_5_megablocks/transformer.py (623:656) duplicated block id: 1879 size: 19 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (606:625) - toolkits/model_checkpoints_convertor/chatglm/checkpoint_reshaping_and_interoperability.py (561:580) duplicated block id: 1880 size: 19 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (65:83) - toolkits/model_checkpoints_convertor/baichuan/hf2te.py (54:73) duplicated block id: 1881 size: 19 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (35:59) - toolkits/model_checkpoints_convertor/utils/__init__.py (80:101) duplicated block id: 1882 size: 19 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (634:653) - toolkits/model_checkpoints_convertor/galactica/checkpoint_reshaping_and_interoperability.py (631:650) duplicated block id: 1883 size: 19 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_mha.py (181:203) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_mha_to_moe.py (76:98) duplicated block id: 1884 size: 19 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon/configuration_RW.py (25:44) - toolkits/model_checkpoints_convertor/falcon/configuration_RW.py (84:103) duplicated block id: 1885 size: 19 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/galactica/checkpoint_reshaping_and_interoperability.py (65:83) - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (65:84) duplicated block id: 1886 size: 19 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/hf2te.py (53:72) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (94:112) duplicated block id: 1887 size: 19 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/transformer/mlp.py (124:148) - megatron_patch/model/qwen1_5/transformer/mlp.py (134:158) duplicated block id: 1888 size: 19 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/hf2te.py (53:72) - toolkits/model_checkpoints_convertor/chatglm/checkpoint_reshaping_and_interoperability.py (55:73) duplicated block id: 1889 size: 19 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/language_model.py (124:164) - megatron_patch/model/starcoder/language_model.py (77:117) duplicated block id: 1890 size: 19 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/hf2te.py (54:73) - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (89:107) duplicated block id: 1891 size: 19 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/hf2te.py (54:73) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (94:112) duplicated block id: 1892 size: 19 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/hf2te.py (53:72) - toolkits/model_checkpoints_convertor/galactica/checkpoint_reshaping_and_interoperability.py (65:83) duplicated block id: 1893 size: 19 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/layer_specs.py (22:53) - megatron_patch/model/mixtral/layer_specs.py (29:60) duplicated block id: 1894 size: 19 cleaned lines of code in 2 files: - megatron_patch/model/mistral/transformer.py (730:763) - megatron_patch/model/qwen_vl/transformer.py (731:764) duplicated block id: 1895 size: 19 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/mlp.py (116:140) - megatron_patch/model/qwen1_5/transformer/mlp.py (134:158) duplicated block id: 1896 size: 19 cleaned lines of code in 2 files: - megatron_patch/model/llama3/language_model.py (68:108) - megatron_patch/model/starcoder/language_model.py (77:117) duplicated block id: 1897 size: 19 cleaned lines of code in 2 files: - megatron_patch/model/mistral/language_model.py (80:120) - megatron_patch/model/starcoder/language_model.py (77:117) duplicated block id: 1898 size: 19 cleaned lines of code in 2 files: - megatron_patch/model/qwen_vl/language_model.py (80:120) - megatron_patch/model/starcoder/language_model.py (77:117) duplicated block id: 1899 size: 19 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (65:83) - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (85:104) duplicated block id: 1900 size: 19 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/gpt_model.py (53:82) - megatron_patch/model/starcoder/gpt_model.py (29:58) duplicated block id: 1901 size: 19 cleaned lines of code in 2 files: - toolkits/pretrain_data_preprocessing/preprocess_data_megatron.py (417:437) - toolkits/sft_data_preprocessing/build_idxmap_sft_dataset.py (389:409) duplicated block id: 1902 size: 19 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_mha.py (119:137) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2.5_vl.py (188:206) duplicated block id: 1903 size: 19 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/hf2te.py (53:72) - toolkits/model_checkpoints_convertor/falcon40b/checkpoint_reshaping_and_interoperability.py (66:84) duplicated block id: 1904 size: 19 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore.py (36:57) - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (35:56) duplicated block id: 1905 size: 19 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (634:653) - toolkits/model_checkpoints_convertor/yi/checkpoint_reshaping_and_interoperability.py (648:667) duplicated block id: 1906 size: 19 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/hf2te.py (54:73) - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (65:83) duplicated block id: 1907 size: 19 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (65:83) - toolkits/model_checkpoints_convertor/baichuan2/hf2te.py (53:72) duplicated block id: 1908 size: 19 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer_legacy.py (710:743) - megatron_patch/model/mistral/transformer.py (730:763) duplicated block id: 1909 size: 19 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore_llama3_1.py (150:169) - toolkits/model_checkpoints_convertor/llava/hf2mcore_llava.py (222:240) duplicated block id: 1910 size: 19 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer/mlp.py (118:142) - megatron_patch/model/mixtral/transformer/mlp.py (124:148) duplicated block id: 1911 size: 19 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/mlp.py (13:60) - megatron_patch/model/mixtral/transformer/mlp.py (25:71) duplicated block id: 1912 size: 19 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/hf2te.py (54:73) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (93:111) duplicated block id: 1913 size: 19 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/mlp.py (116:140) - megatron_patch/model/llama3/transformer/mlp.py (118:142) duplicated block id: 1914 size: 19 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/hf2te.py (53:72) - toolkits/model_checkpoints_convertor/mistral/hf2mcore.py (71:91) duplicated block id: 1915 size: 19 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/language_model.py (205:230) - megatron_patch/model/llama3/language_model.py (182:207) duplicated block id: 1916 size: 19 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/galactica/checkpoint_reshaping_and_interoperability.py (65:83) - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (85:104) duplicated block id: 1917 size: 19 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (656:684) - toolkits/model_checkpoints_convertor/starcoder/checkpoint_reshaping_and_interoperability.py (324:352) duplicated block id: 1918 size: 19 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore.py (36:57) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (40:61) duplicated block id: 1919 size: 19 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon/checkpoint_reshaping_and_interoperability.py (577:596) - toolkits/model_checkpoints_convertor/galactica/checkpoint_reshaping_and_interoperability.py (631:650) duplicated block id: 1920 size: 19 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/hf2te.py (54:73) - toolkits/model_checkpoints_convertor/yi/checkpoint_reshaping_and_interoperability.py (66:84) duplicated block id: 1921 size: 19 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/transformer/mlp.py (96:118) - megatron_patch/model/qwen1_5/transformer/mlp.py (110:132) duplicated block id: 1922 size: 19 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/checkpoint_reshaping_and_interoperability.py (302:330) - toolkits/model_checkpoints_convertor/falcon/checkpoint_reshaping_and_interoperability.py (599:627) duplicated block id: 1923 size: 19 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/hf2te.py (54:73) - toolkits/model_checkpoints_convertor/mistral/hf2mcore.py (71:91) duplicated block id: 1924 size: 19 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/mlp.py (13:60) - megatron_patch/model/llama3/transformer/mlp.py (24:69) duplicated block id: 1925 size: 19 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/hf2te.py (54:73) - toolkits/model_checkpoints_convertor/llama/hf2mcore.py (97:117) duplicated block id: 1926 size: 19 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/galactica/checkpoint_reshaping_and_interoperability.py (504:527) - toolkits/model_checkpoints_convertor/yi/checkpoint_reshaping_and_interoperability.py (471:494) duplicated block id: 1927 size: 19 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/hf2te.py (53:72) - toolkits/model_checkpoints_convertor/falcon/checkpoint_reshaping_and_interoperability.py (66:84) duplicated block id: 1928 size: 19 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_mha.py (119:137) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_vl.py (188:206) duplicated block id: 1929 size: 19 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_mha_to_moe.py (77:100) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_moe.py (237:259) duplicated block id: 1930 size: 19 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer/mlp.py (94:116) - megatron_patch/model/qwen2/transformer/mlp.py (109:131) duplicated block id: 1931 size: 19 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/hf2te.py (53:72) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (93:111) duplicated block id: 1932 size: 19 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (65:84) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (94:112) duplicated block id: 1933 size: 19 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (154:176) - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (162:184) duplicated block id: 1934 size: 19 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/chatglm/checkpoint_reshaping_and_interoperability.py (561:580) - toolkits/model_checkpoints_convertor/falcon/checkpoint_reshaping_and_interoperability.py (577:596) duplicated block id: 1935 size: 19 cleaned lines of code in 2 files: - megatron_patch/model/llama2/transformer.py (1825:1855) - megatron_patch/model/llama3/transformer_legacy.py (1782:1812) duplicated block id: 1936 size: 19 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore_llama3_1.py (202:220) - toolkits/model_checkpoints_convertor/llava/hf2mcore_llava.py (172:191) duplicated block id: 1937 size: 19 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/chatglm/checkpoint_reshaping_and_interoperability.py (55:73) - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (65:84) duplicated block id: 1938 size: 19 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/checkpoint_reshaping_and_interoperability.py (302:330) - toolkits/model_checkpoints_convertor/falcon40b/checkpoint_reshaping_and_interoperability.py (602:630) duplicated block id: 1939 size: 19 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/hf2te.py (54:73) - toolkits/model_checkpoints_convertor/qwen/hf2megablocks_qwen1.5.py (97:117) duplicated block id: 1940 size: 19 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (85:104) - toolkits/model_checkpoints_convertor/qwen/hf2megablocks_qwen1.5.py (97:117) duplicated block id: 1941 size: 19 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer_legacy.py (710:743) - megatron_patch/model/llava/transformer.py (731:764) duplicated block id: 1942 size: 19 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (89:107) - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (65:84) duplicated block id: 1943 size: 19 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/hf2te.py (54:73) - toolkits/model_checkpoints_convertor/galactica/checkpoint_reshaping_and_interoperability.py (65:83) duplicated block id: 1944 size: 19 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (723:756) - megatron_patch/model/mistral/transformer.py (730:763) duplicated block id: 1945 size: 19 cleaned lines of code in 2 files: - megatron_patch/model/llava/language_model.py (82:122) - megatron_patch/model/starcoder/language_model.py (77:117) duplicated block id: 1946 size: 19 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (65:84) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (93:111) duplicated block id: 1947 size: 19 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon40b/checkpoint_reshaping_and_interoperability.py (66:84) - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (65:84) duplicated block id: 1948 size: 19 cleaned lines of code in 2 files: - megatron_patch/model/qwen1_5_megablocks/language_model.py (78:118) - megatron_patch/model/starcoder/language_model.py (77:117) duplicated block id: 1949 size: 19 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (612:645) - megatron_patch/model/mistral/transformer.py (730:763) duplicated block id: 1950 size: 19 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (65:83) - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (65:84) duplicated block id: 1951 size: 19 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (612:645) - megatron_patch/model/qwen1_5_megablocks/transformer.py (623:656) duplicated block id: 1952 size: 19 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/hf2te.py (54:73) - toolkits/model_checkpoints_convertor/falcon40b/checkpoint_reshaping_and_interoperability.py (66:84) duplicated block id: 1953 size: 19 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/hf2te.py (54:73) - toolkits/model_checkpoints_convertor/falcon/checkpoint_reshaping_and_interoperability.py (66:84) duplicated block id: 1954 size: 19 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (65:83) - toolkits/model_checkpoints_convertor/baichuan2/hf2te.py (53:72) duplicated block id: 1955 size: 19 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/transformer.py (534:558) - megatron_patch/model/llama/transformer.py (486:509) duplicated block id: 1956 size: 19 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llava/hf2mcore_llava.py (41:62) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_mha.py (29:50) duplicated block id: 1957 size: 19 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (85:104) - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (89:107) duplicated block id: 1958 size: 19 cleaned lines of code in 2 files: - megatron_patch/model/llama2/transformer.py (732:765) - megatron_patch/model/llama3/transformer_legacy.py (710:743) duplicated block id: 1959 size: 19 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer_legacy.py (1782:1812) - megatron_patch/model/qwen_vl/transformer.py (1824:1854) duplicated block id: 1960 size: 19 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon/configuration_RW.py (25:44) - toolkits/model_checkpoints_convertor/falcon40b/configuration_RW.py (25:44) duplicated block id: 1961 size: 19 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/chatglm/checkpoint_reshaping_and_interoperability.py (55:73) - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (85:104) duplicated block id: 1962 size: 19 cleaned lines of code in 2 files: - megatron_patch/model/llama2/language_model.py (79:119) - megatron_patch/model/starcoder/language_model.py (77:117) duplicated block id: 1963 size: 19 cleaned lines of code in 2 files: - megatron_patch/model/qwen/language_model.py (79:119) - megatron_patch/model/starcoder/language_model.py (77:117) duplicated block id: 1964 size: 19 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (785:813) - toolkits/model_checkpoints_convertor/starcoder/checkpoint_reshaping_and_interoperability.py (324:352) duplicated block id: 1965 size: 19 cleaned lines of code in 2 files: - megatron_patch/model/bloom/language_model.py (449:469) - megatron_patch/model/chatglm/language_model.py (503:523) duplicated block id: 1966 size: 19 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (634:653) - toolkits/model_checkpoints_convertor/chatglm/checkpoint_reshaping_and_interoperability.py (561:580) duplicated block id: 1967 size: 19 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2megablocks_qwen1.5.py (33:54) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (40:61) duplicated block id: 1968 size: 19 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/hf2te.py (53:72) - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (89:107) duplicated block id: 1969 size: 19 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/hf2te.py (53:72) - toolkits/model_checkpoints_convertor/yi/checkpoint_reshaping_and_interoperability.py (66:84) duplicated block id: 1970 size: 19 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (723:756) - megatron_patch/model/qwen1_5_megablocks/transformer.py (623:656) duplicated block id: 1971 size: 19 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/transformer/attention.py (354:392) - megatron_patch/model/qwen2_vl/attention.py (346:384) duplicated block id: 1972 size: 19 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2megablocks_qwen1.5.py (33:54) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (39:60) duplicated block id: 1973 size: 19 cleaned lines of code in 2 files: - megatron_patch/model/llama2/transformer.py (732:765) - megatron_patch/model/qwen/transformer.py (705:738) duplicated block id: 1974 size: 19 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (606:625) - toolkits/model_checkpoints_convertor/galactica/checkpoint_reshaping_and_interoperability.py (631:650) duplicated block id: 1975 size: 19 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (85:104) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (94:112) duplicated block id: 1976 size: 19 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (628:656) - toolkits/model_checkpoints_convertor/starcoder/checkpoint_reshaping_and_interoperability.py (324:352) duplicated block id: 1977 size: 19 cleaned lines of code in 2 files: - megatron_patch/model/llava_mcore/llava_spec.py (69:87) - megatron_patch/model/llava_mcore/vision/vit_layer_specs.py (65:83) duplicated block id: 1978 size: 19 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/hf2te.py (53:72) - toolkits/model_checkpoints_convertor/llama/hf2mcore.py (97:117) duplicated block id: 1979 size: 19 cleaned lines of code in 2 files: - megatron_patch/model/mistral/transformer.py (730:763) - megatron_patch/model/qwen/transformer.py (705:738) duplicated block id: 1980 size: 19 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (65:83) - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (85:104) duplicated block id: 1981 size: 19 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon/checkpoint_reshaping_and_interoperability.py (599:627) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (785:813) duplicated block id: 1982 size: 19 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/language_model.py (89:129) - megatron_patch/model/starcoder/language_model.py (77:117) duplicated block id: 1983 size: 19 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llava/hf2mcore_llava.py (41:62) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_moe.py (31:52) duplicated block id: 1984 size: 19 cleaned lines of code in 2 files: - megatron_patch/model/llava/transformer.py (731:764) - megatron_patch/model/mistral/transformer.py (730:763) duplicated block id: 1985 size: 19 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (35:56) - toolkits/model_checkpoints_convertor/qwen/hf2megablocks_qwen1.5.py (33:54) duplicated block id: 1986 size: 19 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon/checkpoint_reshaping_and_interoperability.py (66:84) - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (65:84) duplicated block id: 1987 size: 19 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (65:83) - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (65:84) duplicated block id: 1988 size: 19 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore.py (97:117) - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (85:104) duplicated block id: 1989 size: 19 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (377:405) - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (361:390) duplicated block id: 1990 size: 19 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer_legacy.py (710:743) - megatron_patch/model/qwen_vl/transformer.py (731:764) duplicated block id: 1991 size: 19 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon40b/checkpoint_reshaping_and_interoperability.py (580:599) - toolkits/model_checkpoints_convertor/galactica/checkpoint_reshaping_and_interoperability.py (631:650) duplicated block id: 1992 size: 19 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/transformer/attention.py (354:392) - megatron_patch/model/qwen2_vl/attention_vision.py (345:383) duplicated block id: 1993 size: 19 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon40b/checkpoint_reshaping_and_interoperability.py (602:630) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (785:813) duplicated block id: 1994 size: 19 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/hf2te.py (54:73) - toolkits/model_checkpoints_convertor/chatglm/checkpoint_reshaping_and_interoperability.py (55:73) duplicated block id: 1995 size: 19 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore.py (36:57) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (39:60) duplicated block id: 1996 size: 19 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (723:756) - megatron_patch/model/llama2/transformer.py (732:765) duplicated block id: 1997 size: 19 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer_legacy.py (1782:1812) - megatron_patch/model/mistral/transformer.py (1822:1852) duplicated block id: 1998 size: 19 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (85:104) - toolkits/model_checkpoints_convertor/mistral/hf2mcore.py (71:91) duplicated block id: 1999 size: 19 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/hf2te.py (53:72) - toolkits/model_checkpoints_convertor/qwen/hf2megablocks_qwen1.5.py (97:117) duplicated block id: 2000 size: 19 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer_legacy.py (1782:1812) - megatron_patch/model/llava/transformer.py (1824:1854) duplicated block id: 2001 size: 19 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (612:645) - megatron_patch/model/llama2/transformer.py (732:765) duplicated block id: 2002 size: 19 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon40b/checkpoint_reshaping_and_interoperability.py (66:84) - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (85:104) duplicated block id: 2003 size: 19 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon/checkpoint_reshaping_and_interoperability.py (66:84) - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (85:104) duplicated block id: 2004 size: 19 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (85:104) - toolkits/model_checkpoints_convertor/yi/checkpoint_reshaping_and_interoperability.py (66:84) duplicated block id: 2005 size: 19 cleaned lines of code in 2 files: - megatron_patch/model/qwen/transformer.py (705:738) - megatron_patch/model/qwen1_5_megablocks/transformer.py (623:656) duplicated block id: 2006 size: 18 cleaned lines of code in 2 files: - megatron_patch/model/llama3_1/model.py (188:216) - megatron_patch/model/qwen2_vl/gpt_model.py (121:149) duplicated block id: 2007 size: 18 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/hf2te.py (33:50) - toolkits/model_checkpoints_convertor/falcon/checkpoint_reshaping_and_interoperability.py (43:60) duplicated block id: 2008 size: 18 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/transformer.py (1195:1235) - megatron_patch/model/mistral/transformer.py (1748:1789) duplicated block id: 2009 size: 18 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_mha.py (182:203) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_moe.py (237:258) duplicated block id: 2010 size: 18 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/glm/checkpoint_reshaping_and_interoperability.py (552:572) - toolkits/model_checkpoints_convertor/starcoder/checkpoint_reshaping_and_interoperability.py (810:830) duplicated block id: 2011 size: 18 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/reward_model_to_megatron.py (442:472) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (856:887) duplicated block id: 2012 size: 18 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (28:47) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_moe.py (31:50) duplicated block id: 2013 size: 18 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore.py (36:55) - toolkits/model_checkpoints_convertor/llava/hf2mcore_llava.py (41:60) duplicated block id: 2014 size: 18 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/language_model.py (261:283) - megatron_patch/model/qwen/language_model.py (262:285) duplicated block id: 2015 size: 18 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (640:662) - megatron_patch/model/mistral/transformer.py (648:670) duplicated block id: 2016 size: 18 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (634:652) - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (748:766) duplicated block id: 2017 size: 18 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (640:662) - megatron_patch/model/llama3/transformer_legacy.py (627:649) duplicated block id: 2018 size: 18 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/language_model.py (318:341) - megatron_patch/model/falcon40b/language_model.py (261:283) duplicated block id: 2019 size: 18 cleaned lines of code in 2 files: - megatron_patch/model/llama3/language_model.py (189:214) - megatron_patch/model/starcoder/language_model.py (198:223) duplicated block id: 2020 size: 18 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (28:47) - toolkits/model_checkpoints_convertor/qwen/hf2megablocks_qwen1.5.py (33:52) duplicated block id: 2021 size: 18 cleaned lines of code in 2 files: - megatron_patch/model/llama/language_model.py (652:674) - megatron_patch/model/llama2/language_model.py (594:616) duplicated block id: 2022 size: 18 cleaned lines of code in 2 files: - toolkits/pretrain_data_preprocessing/preprocess_data_megatron.py (250:271) - toolkits/sft_data_preprocessing/build_idxmap_sft_dataset.py (243:264) duplicated block id: 2023 size: 18 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (35:54) - toolkits/model_checkpoints_convertor/utils/__init__.py (82:101) duplicated block id: 2024 size: 18 cleaned lines of code in 2 files: - megatron_patch/model/llama/language_model.py (652:674) - megatron_patch/model/mistral/language_model.py (606:628) duplicated block id: 2025 size: 18 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (28:47) - toolkits/model_checkpoints_convertor/llava/hf2mcore_llava.py (41:60) duplicated block id: 2026 size: 18 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (1562:1583) - megatron_patch/model/mistral/transformer.py (1689:1710) duplicated block id: 2027 size: 18 cleaned lines of code in 2 files: - megatron_patch/model/qwen2/transformer/mlp.py (200:217) - megatron_patch/model/qwen2/transformer/mlp.py (251:268) duplicated block id: 2028 size: 18 cleaned lines of code in 2 files: - megatron_patch/model/qwen1_5/transformer/mlp.py (197:214) - megatron_patch/model/qwen2/transformer/mlp.py (251:268) duplicated block id: 2029 size: 18 cleaned lines of code in 2 files: - megatron_patch/model/llama/language_model.py (652:674) - megatron_patch/model/llava/language_model.py (654:676) duplicated block id: 2030 size: 18 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/reward_model_to_megatron.py (800:820) - toolkits/model_checkpoints_convertor/glm/checkpoint_reshaping_and_interoperability.py (552:572) duplicated block id: 2031 size: 18 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon40b/checkpoint_reshaping_and_interoperability.py (580:598) - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (551:569) duplicated block id: 2032 size: 18 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer_legacy.py (1652:1673) - megatron_patch/model/mistral/transformer.py (1689:1710) duplicated block id: 2033 size: 18 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (1692:1713) - megatron_patch/model/qwen_vl/transformer.py (1691:1712) duplicated block id: 2034 size: 18 cleaned lines of code in 2 files: - megatron_patch/model/bloom/language_model.py (549:571) - megatron_patch/model/qwen/language_model.py (576:598) duplicated block id: 2035 size: 18 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (28:47) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_mha_to_moe.py (28:47) duplicated block id: 2036 size: 18 cleaned lines of code in 2 files: - megatron_patch/model/llama/language_model.py (261:283) - megatron_patch/model/qwen/language_model.py (262:285) duplicated block id: 2037 size: 18 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (28:47) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_mha.py (29:48) duplicated block id: 2038 size: 18 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/language_model.py (277:299) - megatron_patch/model/llava/language_model.py (265:288) duplicated block id: 2039 size: 18 cleaned lines of code in 2 files: - megatron_patch/model/qwen_vl/language_model.py (200:225) - megatron_patch/model/starcoder/language_model.py (198:223) duplicated block id: 2040 size: 18 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/language_model.py (277:299) - megatron_patch/model/qwen_vl/language_model.py (263:286) duplicated block id: 2041 size: 18 cleaned lines of code in 2 files: - megatron_patch/model/llava/transformer.py (1603:1630) - megatron_patch/model/starcoder/transformer.py (1200:1227) duplicated block id: 2042 size: 18 cleaned lines of code in 2 files: - megatron_patch/model/falcon/language_model.py (641:663) - megatron_patch/model/qwen1_5_megablocks/language_model.py (593:615) duplicated block id: 2043 size: 18 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore.py (36:55) - toolkits/model_checkpoints_convertor/utils/__init__.py (82:101) duplicated block id: 2044 size: 18 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/deepseek/hf2mcore_deepseek_v2_moe.py (384:401) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_moe.py (318:336) duplicated block id: 2045 size: 18 cleaned lines of code in 2 files: - megatron_patch/model/falcon/language_model.py (384:406) - megatron_patch/model/starcoder/language_model.py (379:401) duplicated block id: 2046 size: 18 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/deepseek/hf2mcore_deepseek_v2_moe.py (384:401) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_mha_to_moe.py (169:187) duplicated block id: 2047 size: 18 cleaned lines of code in 2 files: - megatron_patch/model/galactica/language_model.py (292:314) - megatron_patch/model/qwen1_5_megablocks/language_model.py (261:284) duplicated block id: 2048 size: 18 cleaned lines of code in 2 files: - megatron_patch/model/bloom/language_model.py (131:165) - megatron_patch/model/glm130b/language_model.py (131:165) duplicated block id: 2049 size: 18 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (28:47) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (40:59) duplicated block id: 2050 size: 18 cleaned lines of code in 2 files: - megatron_patch/model/bloom/language_model.py (279:301) - megatron_patch/model/qwen1_5_megablocks/language_model.py (261:284) duplicated block id: 2051 size: 18 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/language_model.py (261:283) - megatron_patch/model/llama3/language_model.py (257:280) duplicated block id: 2052 size: 18 cleaned lines of code in 2 files: - megatron_patch/model/bloom/language_model.py (131:165) - megatron_patch/model/llama/language_model.py (129:163) duplicated block id: 2053 size: 18 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/language_model.py (272:295) - megatron_patch/model/llama/language_model.py (261:283) duplicated block id: 2054 size: 18 cleaned lines of code in 2 files: - megatron_patch/model/qwen/transformer.py (1546:1573) - megatron_patch/model/starcoder/transformer.py (1200:1227) duplicated block id: 2055 size: 18 cleaned lines of code in 2 files: - megatron_patch/model/llama/language_model.py (261:283) - megatron_patch/model/llama2/language_model.py (262:285) duplicated block id: 2056 size: 18 cleaned lines of code in 2 files: - megatron_patch/model/galactica/language_model.py (670:692) - megatron_patch/model/llava/language_model.py (654:676) duplicated block id: 2057 size: 18 cleaned lines of code in 2 files: - megatron_patch/model/qwen_vl/transformer.py (1603:1630) - megatron_patch/model/starcoder/transformer.py (1200:1227) duplicated block id: 2058 size: 18 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/language_model.py (579:601) - megatron_patch/model/qwen1_5_megablocks/language_model.py (593:615) duplicated block id: 2059 size: 18 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/chatglm/checkpoint_reshaping_and_interoperability.py (561:579) - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (748:766) duplicated block id: 2060 size: 18 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (640:662) - megatron_patch/model/qwen/transformer.py (622:644) duplicated block id: 2061 size: 18 cleaned lines of code in 2 files: - megatron_patch/model/falcon/language_model.py (641:663) - megatron_patch/model/mistral/language_model.py (606:628) duplicated block id: 2062 size: 18 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/language_model.py (261:283) - megatron_patch/model/glm130b/language_model.py (276:299) duplicated block id: 2063 size: 18 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/language_model.py (318:341) - megatron_patch/model/chatglm/language_model.py (277:299) duplicated block id: 2064 size: 18 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (640:662) - megatron_patch/model/qwen1_5_megablocks/transformer.py (540:562) duplicated block id: 2065 size: 18 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (28:47) - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (35:54) duplicated block id: 2066 size: 18 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/gpt_model.py (26:50) - megatron_patch/model/falcon40b/gpt_model.py (28:51) duplicated block id: 2067 size: 18 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (551:569) - toolkits/model_checkpoints_convertor/yi/checkpoint_reshaping_and_interoperability.py (648:666) duplicated block id: 2068 size: 18 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (667:684) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (714:731) duplicated block id: 2069 size: 18 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (1562:1583) - megatron_patch/model/qwen1_5_megablocks/transformer.py (1554:1575) duplicated block id: 2070 size: 18 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_mha.py (29:48) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (40:59) duplicated block id: 2071 size: 18 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_mha.py (29:48) - toolkits/model_checkpoints_convertor/utils/__init__.py (82:101) duplicated block id: 2072 size: 18 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/language_model.py (579:601) - megatron_patch/model/qwen/language_model.py (576:598) duplicated block id: 2073 size: 18 cleaned lines of code in 2 files: - megatron_patch/model/llama/language_model.py (652:674) - megatron_patch/model/qwen/language_model.py (576:598) duplicated block id: 2074 size: 18 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/language_model.py (272:295) - megatron_patch/model/falcon/language_model.py (261:283) duplicated block id: 2075 size: 18 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/hf2te.py (33:50) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (71:88) duplicated block id: 2076 size: 18 cleaned lines of code in 2 files: - megatron_patch/model/bloom/language_model.py (232:260) - megatron_patch/model/starcoder/language_model.py (219:246) duplicated block id: 2077 size: 18 cleaned lines of code in 2 files: - megatron_patch/model/falcon/language_model.py (261:283) - megatron_patch/model/llama2/language_model.py (262:285) duplicated block id: 2078 size: 18 cleaned lines of code in 2 files: - megatron_patch/model/falcon/language_model.py (641:663) - megatron_patch/model/llama2/language_model.py (594:616) duplicated block id: 2079 size: 18 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore.py (288:305) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2.5_vl.py (139:156) duplicated block id: 2080 size: 18 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/language_model.py (671:693) - megatron_patch/model/chatglm/language_model.py (615:637) duplicated block id: 2081 size: 18 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/language_model.py (318:341) - megatron_patch/model/galactica/language_model.py (292:314) duplicated block id: 2082 size: 18 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (714:731) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (796:813) duplicated block id: 2083 size: 18 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (856:887) - toolkits/model_checkpoints_convertor/starcoder/checkpoint_reshaping_and_interoperability.py (456:486) duplicated block id: 2084 size: 18 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/checkpoint_reshaping_and_interoperability.py (801:821) - toolkits/model_checkpoints_convertor/glm/checkpoint_reshaping_and_interoperability.py (552:572) duplicated block id: 2085 size: 18 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (640:662) - megatron_patch/model/llava/transformer.py (648:670) duplicated block id: 2086 size: 18 cleaned lines of code in 2 files: - megatron_patch/model/bloom/language_model.py (279:301) - megatron_patch/model/qwen_vl/language_model.py (263:286) duplicated block id: 2087 size: 18 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/galactica/checkpoint_reshaping_and_interoperability.py (631:649) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (680:698) duplicated block id: 2088 size: 18 cleaned lines of code in 2 files: - megatron_patch/model/falcon/language_model.py (261:283) - megatron_patch/model/llava/language_model.py (265:288) duplicated block id: 2089 size: 18 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (1692:1713) - megatron_patch/model/mistral/transformer.py (1689:1710) duplicated block id: 2090 size: 18 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llava/hf2mcore_llava.py (41:60) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (39:58) duplicated block id: 2091 size: 18 cleaned lines of code in 2 files: - megatron_patch/model/qwen2/transformer_block.py (358:379) - megatron_patch/model/qwen2_5_vl/transformer_block.py (509:530) duplicated block id: 2092 size: 18 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/language_model.py (277:299) - megatron_patch/model/mistral/language_model.py (263:286) duplicated block id: 2093 size: 18 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer/mlp.py (181:198) - megatron_patch/model/qwen2/transformer/mlp.py (251:268) duplicated block id: 2094 size: 18 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/language_model.py (641:663) - megatron_patch/model/mistral/language_model.py (606:628) duplicated block id: 2095 size: 18 cleaned lines of code in 2 files: - megatron_patch/model/qwen/transformer.py (1634:1655) - megatron_patch/model/qwen_vl/transformer.py (1691:1712) duplicated block id: 2096 size: 18 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/transformer.py (577:597) - megatron_patch/model/qwen1_5_megablocks/transformer.py (665:685) duplicated block id: 2097 size: 18 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/transformer.py (577:597) - megatron_patch/model/llava/transformer.py (773:793) duplicated block id: 2098 size: 18 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon40b/checkpoint_reshaping_and_interoperability.py (613:630) - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (782:799) duplicated block id: 2099 size: 18 cleaned lines of code in 2 files: - megatron_patch/model/bloom/language_model.py (549:571) - megatron_patch/model/llama2/language_model.py (594:616) duplicated block id: 2100 size: 18 cleaned lines of code in 2 files: - megatron_patch/model/galactica/language_model.py (292:314) - megatron_patch/model/llama2/language_model.py (262:285) duplicated block id: 2101 size: 18 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_mha.py (29:48) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (39:58) duplicated block id: 2102 size: 18 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/hf2te.py (34:51) - toolkits/model_checkpoints_convertor/galactica/checkpoint_reshaping_and_interoperability.py (42:59) duplicated block id: 2103 size: 18 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2.5_vl.py (139:156) - toolkits/model_checkpoints_convertor/qwen/hf2megablocks_qwen1.5.py (282:299) duplicated block id: 2104 size: 18 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/language_model.py (318:341) - megatron_patch/model/llama/language_model.py (261:283) duplicated block id: 2105 size: 18 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/checkpoint_reshaping_and_interoperability.py (443:473) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (856:887) duplicated block id: 2106 size: 18 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/mlp.py (170:187) - megatron_patch/model/qwen2/transformer/mlp.py (200:217) duplicated block id: 2107 size: 18 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_mha_to_moe.py (169:187) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_dense_and_moe_gqa.py (536:554) duplicated block id: 2108 size: 18 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (606:624) - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (551:569) duplicated block id: 2109 size: 18 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (640:662) - megatron_patch/model/llama2/transformer.py (649:671) duplicated block id: 2110 size: 18 cleaned lines of code in 2 files: - megatron_patch/model/falcon/transformer.py (791:812) - megatron_patch/model/llava/transformer.py (773:793) duplicated block id: 2111 size: 18 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/hf2te.py (33:50) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (70:87) duplicated block id: 2112 size: 18 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/language_model.py (261:283) - megatron_patch/model/qwen_vl/language_model.py (263:286) duplicated block id: 2113 size: 18 cleaned lines of code in 2 files: - megatron_patch/model/falcon/transformer.py (791:812) - megatron_patch/model/qwen1_5_megablocks/transformer.py (665:685) duplicated block id: 2114 size: 18 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (252:301) - megatron_patch/model/galactica/transformer.py (198:246) duplicated block id: 2115 size: 18 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (634:652) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (763:781) duplicated block id: 2116 size: 18 cleaned lines of code in 2 files: - megatron_patch/model/mixtral_bak/model.py (192:216) - megatron_patch/model/qwen2_vl/gpt_model.py (190:214) duplicated block id: 2117 size: 18 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/chatglm/checkpoint_reshaping_and_interoperability.py (561:579) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (763:781) duplicated block id: 2118 size: 18 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/language_model.py (579:601) - megatron_patch/model/qwen_vl/language_model.py (623:645) duplicated block id: 2119 size: 18 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (28:47) - toolkits/model_checkpoints_convertor/utils/__init__.py (82:101) duplicated block id: 2120 size: 18 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/mlp.py (170:187) - megatron_patch/model/qwen1_5/transformer/mlp.py (197:214) duplicated block id: 2121 size: 18 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (927:947) - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (1108:1128) duplicated block id: 2122 size: 18 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2.5_vl.py (251:269) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_vl.py (253:271) duplicated block id: 2123 size: 18 cleaned lines of code in 2 files: - megatron_patch/model/galactica/language_model.py (670:692) - megatron_patch/model/qwen/language_model.py (576:598) duplicated block id: 2124 size: 18 cleaned lines of code in 2 files: - megatron_patch/model/mistral/language_model.py (200:225) - megatron_patch/model/starcoder/language_model.py (198:223) duplicated block id: 2125 size: 18 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer_legacy.py (1652:1673) - megatron_patch/model/qwen1_5_megablocks/transformer.py (1554:1575) duplicated block id: 2126 size: 18 cleaned lines of code in 2 files: - megatron_patch/model/llama2/language_model.py (199:224) - megatron_patch/model/starcoder/language_model.py (198:223) duplicated block id: 2127 size: 18 cleaned lines of code in 2 files: - megatron_patch/model/galactica/language_model.py (670:692) - megatron_patch/model/qwen_vl/language_model.py (623:645) duplicated block id: 2128 size: 18 cleaned lines of code in 2 files: - megatron_patch/model/llama3/model.py (130:158) - megatron_patch/model/llama3_1/model.py (188:216) duplicated block id: 2129 size: 18 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/hf2te.py (34:51) - toolkits/model_checkpoints_convertor/yi/checkpoint_reshaping_and_interoperability.py (43:60) duplicated block id: 2130 size: 18 cleaned lines of code in 2 files: - megatron_patch/model/qwen1_5_megablocks/transformer.py (1467:1494) - megatron_patch/model/starcoder/transformer.py (1200:1227) duplicated block id: 2131 size: 18 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_moe.py (250:268) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_moe.py (340:358) duplicated block id: 2132 size: 18 cleaned lines of code in 2 files: - megatron_patch/model/galactica/language_model.py (292:314) - megatron_patch/model/qwen_vl/language_model.py (263:286) duplicated block id: 2133 size: 18 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/gpt_model.py (26:50) - megatron_patch/model/galactica/gpt_model.py (28:51) duplicated block id: 2134 size: 18 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (28:47) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_moe.py (31:50) duplicated block id: 2135 size: 18 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/galactica/checkpoint_reshaping_and_interoperability.py (631:649) - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (748:766) duplicated block id: 2136 size: 18 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer/mlp.py (114:136) - megatron_patch/model/mixtral_bak/transformer/mlp.py (89:111) duplicated block id: 2137 size: 18 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/language_model.py (641:663) - megatron_patch/model/qwen/language_model.py (576:598) duplicated block id: 2138 size: 18 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/transformer_block.py (475:494) - megatron_patch/model/qwen2/transformer_block.py (381:400) duplicated block id: 2139 size: 18 cleaned lines of code in 2 files: - megatron_patch/model/llama/language_model.py (261:283) - megatron_patch/model/qwen_vl/language_model.py (263:286) duplicated block id: 2140 size: 18 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (1692:1713) - megatron_patch/model/llava/transformer.py (1691:1712) duplicated block id: 2141 size: 18 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/transformer_block.py (450:471) - megatron_patch/model/qwen2/transformer_block.py (358:379) duplicated block id: 2142 size: 18 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/language_model.py (671:693) - megatron_patch/model/glm130b/language_model.py (579:601) duplicated block id: 2143 size: 18 cleaned lines of code in 2 files: - megatron_patch/model/galactica/language_model.py (670:692) - megatron_patch/model/qwen1_5_megablocks/language_model.py (593:615) duplicated block id: 2144 size: 18 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/language_model.py (318:341) - megatron_patch/model/bloom/language_model.py (279:301) duplicated block id: 2145 size: 18 cleaned lines of code in 2 files: - megatron_patch/model/falcon/language_model.py (261:283) - megatron_patch/model/llama3/language_model.py (257:280) duplicated block id: 2146 size: 18 cleaned lines of code in 2 files: - megatron_patch/model/galactica/language_model.py (292:314) - megatron_patch/model/glm130b/language_model.py (276:299) duplicated block id: 2147 size: 18 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (1099:1120) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (824:845) duplicated block id: 2148 size: 18 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (324:349) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (458:484) duplicated block id: 2149 size: 18 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon/checkpoint_reshaping_and_interoperability.py (610:627) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (714:731) duplicated block id: 2150 size: 18 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_mha_to_moe.py (28:47) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (39:58) duplicated block id: 2151 size: 18 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/transformer.py (577:597) - megatron_patch/model/llama2/transformer.py (774:794) duplicated block id: 2152 size: 18 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (703:726) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (698:721) duplicated block id: 2153 size: 18 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (42:59) - toolkits/model_checkpoints_convertor/baichuan2/hf2te.py (33:50) duplicated block id: 2154 size: 18 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (802:823) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (824:845) duplicated block id: 2155 size: 18 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/hf2te.py (34:51) - toolkits/model_checkpoints_convertor/falcon40b/checkpoint_reshaping_and_interoperability.py (43:60) duplicated block id: 2156 size: 18 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (640:662) - megatron_patch/model/qwen_vl/transformer.py (648:670) duplicated block id: 2157 size: 18 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/language_model.py (615:637) - megatron_patch/model/llama2/language_model.py (594:616) duplicated block id: 2158 size: 18 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (42:59) - toolkits/model_checkpoints_convertor/baichuan/hf2te.py (34:51) duplicated block id: 2159 size: 18 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon/checkpoint_reshaping_and_interoperability.py (828:848) - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (1108:1128) duplicated block id: 2160 size: 18 cleaned lines of code in 2 files: - megatron_patch/model/falcon/language_model.py (261:283) - megatron_patch/model/qwen/language_model.py (262:285) duplicated block id: 2161 size: 18 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/moe/experts.py (180:207) - megatron_patch/model/qwen2/moe/experts.py (170:197) duplicated block id: 2162 size: 18 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (28:47) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (40:59) duplicated block id: 2163 size: 18 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/moe/moe_layer.py (127:146) - megatron_patch/model/qwen3_moe/moe/moe_layer.py (66:85) duplicated block id: 2164 size: 18 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (28:47) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (28:47) duplicated block id: 2165 size: 18 cleaned lines of code in 2 files: - megatron_patch/model/llava/language_model.py (202:227) - megatron_patch/model/starcoder/language_model.py (198:223) duplicated block id: 2166 size: 18 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon40b/checkpoint_reshaping_and_interoperability.py (613:630) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (714:731) duplicated block id: 2167 size: 18 cleaned lines of code in 2 files: - megatron_patch/model/llama2/transformer.py (1692:1713) - megatron_patch/model/llama3/transformer_legacy.py (1652:1673) duplicated block id: 2168 size: 18 cleaned lines of code in 2 files: - megatron_patch/model/bloom/language_model.py (279:301) - megatron_patch/model/llava/language_model.py (265:288) duplicated block id: 2169 size: 18 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/transformer.py (577:597) - megatron_patch/model/qwen_vl/transformer.py (773:793) duplicated block id: 2170 size: 18 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (1692:1713) - megatron_patch/model/llama2/transformer.py (1692:1713) duplicated block id: 2171 size: 18 cleaned lines of code in 2 files: - megatron_patch/model/qwen/transformer.py (1634:1655) - megatron_patch/model/qwen1_5_megablocks/transformer.py (1554:1575) duplicated block id: 2172 size: 18 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/language_model.py (615:637) - megatron_patch/model/qwen_vl/language_model.py (623:645) duplicated block id: 2173 size: 18 cleaned lines of code in 2 files: - megatron_patch/model/bloom/language_model.py (549:571) - megatron_patch/model/qwen_vl/language_model.py (623:645) duplicated block id: 2174 size: 18 cleaned lines of code in 2 files: - megatron_patch/model/llama/language_model.py (261:283) - megatron_patch/model/llava/language_model.py (265:288) duplicated block id: 2175 size: 18 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_moe.py (31:50) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (39:58) duplicated block id: 2176 size: 18 cleaned lines of code in 2 files: - megatron_patch/model/falcon/language_model.py (261:283) - megatron_patch/model/mistral/language_model.py (263:286) duplicated block id: 2177 size: 18 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (28:47) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_mha_to_moe.py (28:47) duplicated block id: 2178 size: 18 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (824:845) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (994:1015) duplicated block id: 2179 size: 18 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llava/hf2mcore_llava.py (41:60) - toolkits/model_checkpoints_convertor/qwen/hf2megablocks_qwen1.5.py (33:52) duplicated block id: 2180 size: 18 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (391:429) - megatron_patch/model/starcoder/transformer.py (283:322) duplicated block id: 2181 size: 18 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/language_model.py (671:693) - megatron_patch/model/galactica/language_model.py (670:692) duplicated block id: 2182 size: 18 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/hf2te.py (34:51) - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (66:83) duplicated block id: 2183 size: 18 cleaned lines of code in 2 files: - megatron_patch/model/falcon/language_model.py (641:663) - megatron_patch/model/qwen_vl/language_model.py (623:645) duplicated block id: 2184 size: 18 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore.py (491:508) - toolkits/model_checkpoints_convertor/mistral/hf2mcore.py (398:415) duplicated block id: 2185 size: 18 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/gpt_model.py (26:50) - megatron_patch/model/falcon/gpt_model.py (28:51) duplicated block id: 2186 size: 18 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (639:656) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (714:731) duplicated block id: 2187 size: 18 cleaned lines of code in 2 files: - megatron_patch/model/falcon/transformer.py (791:812) - megatron_patch/model/mistral/transformer.py (772:792) duplicated block id: 2188 size: 18 cleaned lines of code in 2 files: - megatron_patch/model/llama/language_model.py (261:283) - megatron_patch/model/mistral/language_model.py (263:286) duplicated block id: 2189 size: 18 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer/mlp.py (181:198) - megatron_patch/model/mixtral/transformer/mlp.py (178:195) duplicated block id: 2190 size: 18 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/language_model.py (261:283) - megatron_patch/model/qwen1_5_megablocks/language_model.py (261:284) duplicated block id: 2191 size: 18 cleaned lines of code in 2 files: - megatron_patch/model/bloom/language_model.py (279:301) - megatron_patch/model/mistral/language_model.py (263:286) duplicated block id: 2192 size: 18 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/language_model.py (587:609) - megatron_patch/model/falcon40b/language_model.py (641:663) duplicated block id: 2193 size: 18 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/hf2te.py (33:50) - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (66:83) duplicated block id: 2194 size: 18 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/language_model.py (587:609) - megatron_patch/model/chatglm/language_model.py (615:637) duplicated block id: 2195 size: 18 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/hf2te.py (34:51) - toolkits/model_checkpoints_convertor/falcon/checkpoint_reshaping_and_interoperability.py (43:60) duplicated block id: 2196 size: 18 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (802:823) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (1107:1128) duplicated block id: 2197 size: 18 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/glm/checkpoint_reshaping_and_interoperability.py (380:404) - toolkits/model_checkpoints_convertor/glm130b/checkpoint_reshaping_and_interoperability.py (370:393) duplicated block id: 2198 size: 18 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/language_model.py (384:406) - megatron_patch/model/starcoder/language_model.py (379:401) duplicated block id: 2199 size: 18 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (782:799) - toolkits/model_checkpoints_convertor/starcoder/checkpoint_reshaping_and_interoperability.py (335:352) duplicated block id: 2200 size: 18 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_moe.py (31:50) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (40:59) duplicated block id: 2201 size: 18 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (35:54) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_moe.py (31:50) duplicated block id: 2202 size: 18 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/language_model.py (277:299) - megatron_patch/model/llama2/language_model.py (262:285) duplicated block id: 2203 size: 18 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/language_model.py (261:283) - megatron_patch/model/llava/language_model.py (265:288) duplicated block id: 2204 size: 18 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore.py (596:615) - toolkits/model_checkpoints_convertor/qwen/hf2megablocks_qwen1.5.py (606:625) duplicated block id: 2205 size: 18 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/language_model.py (209:234) - megatron_patch/model/starcoder/language_model.py (198:223) duplicated block id: 2206 size: 18 cleaned lines of code in 2 files: - megatron_patch/model/bloom/language_model.py (279:301) - megatron_patch/model/llama2/language_model.py (262:285) duplicated block id: 2207 size: 18 cleaned lines of code in 2 files: - megatron_patch/model/llama3_1/model.py (188:216) - megatron_patch/model/qwen1_5/model.py (130:158) duplicated block id: 2208 size: 18 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/hf2te.py (34:51) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (71:88) duplicated block id: 2209 size: 18 cleaned lines of code in 2 files: - megatron_patch/model/falcon/transformer.py (791:812) - megatron_patch/model/llama2/transformer.py (774:794) duplicated block id: 2210 size: 18 cleaned lines of code in 2 files: - megatron_patch/model/llama2/transformer.py (1604:1631) - megatron_patch/model/starcoder/transformer.py (1200:1227) duplicated block id: 2211 size: 18 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/language_model.py (671:693) - megatron_patch/model/llama/language_model.py (652:674) duplicated block id: 2212 size: 18 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (28:47) - toolkits/model_checkpoints_convertor/qwen/hf2megablocks_qwen1.5.py (33:52) duplicated block id: 2213 size: 18 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/gpt_model.py (49:71) - megatron_patch/model/starcoder/gpt_model.py (52:75) duplicated block id: 2214 size: 18 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (28:47) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (39:58) duplicated block id: 2215 size: 18 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer_legacy.py (1652:1673) - megatron_patch/model/llava/transformer.py (1691:1712) duplicated block id: 2216 size: 18 cleaned lines of code in 2 files: - megatron_patch/model/bloom/language_model.py (131:165) - megatron_patch/model/falcon40b/language_model.py (129:163) duplicated block id: 2217 size: 18 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/language_model.py (272:295) - megatron_patch/model/falcon40b/language_model.py (261:283) duplicated block id: 2218 size: 18 cleaned lines of code in 2 files: - megatron_patch/model/bloom/language_model.py (549:571) - megatron_patch/model/mistral/language_model.py (606:628) duplicated block id: 2219 size: 18 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/language_model.py (272:295) - megatron_patch/model/chatglm/language_model.py (277:299) duplicated block id: 2220 size: 18 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/moe/router.py (107:136) - megatron_patch/model/qwen1_5/moe/router.py (116:144) duplicated block id: 2221 size: 18 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (1692:1713) - megatron_patch/model/qwen1_5_megablocks/transformer.py (1554:1575) duplicated block id: 2222 size: 18 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/galactica/checkpoint_reshaping_and_interoperability.py (631:649) - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (551:569) duplicated block id: 2223 size: 18 cleaned lines of code in 2 files: - megatron_patch/model/llama2/transformer.py (1692:1713) - megatron_patch/model/qwen/transformer.py (1634:1655) duplicated block id: 2224 size: 18 cleaned lines of code in 2 files: - megatron_patch/model/llama/language_model.py (261:283) - megatron_patch/model/llama3/language_model.py (257:280) duplicated block id: 2225 size: 18 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/language_model.py (671:693) - megatron_patch/model/bloom/language_model.py (549:571) duplicated block id: 2226 size: 18 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_mha_to_moe.py (28:47) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (40:59) duplicated block id: 2227 size: 18 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/checkpoint_reshaping_and_interoperability.py (313:330) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (714:731) duplicated block id: 2228 size: 18 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/language_model.py (276:299) - megatron_patch/model/llama/language_model.py (261:283) duplicated block id: 2229 size: 18 cleaned lines of code in 2 files: - megatron_patch/model/galactica/language_model.py (292:314) - megatron_patch/model/llama3/language_model.py (257:280) duplicated block id: 2230 size: 18 cleaned lines of code in 2 files: - megatron_patch/model/llava/transformer.py (1691:1712) - megatron_patch/model/qwen/transformer.py (1634:1655) duplicated block id: 2231 size: 18 cleaned lines of code in 2 files: - megatron_patch/model/llama/language_model.py (652:674) - megatron_patch/model/qwen_vl/language_model.py (623:645) duplicated block id: 2232 size: 18 cleaned lines of code in 2 files: - megatron_patch/model/llama/language_model.py (261:283) - megatron_patch/model/qwen1_5_megablocks/language_model.py (261:284) duplicated block id: 2233 size: 18 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/language_model.py (579:601) - megatron_patch/model/llama2/language_model.py (594:616) duplicated block id: 2234 size: 18 cleaned lines of code in 2 files: - megatron_patch/model/falcon/language_model.py (261:283) - megatron_patch/model/qwen1_5_megablocks/language_model.py (261:284) duplicated block id: 2235 size: 18 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/language_model.py (615:637) - megatron_patch/model/qwen1_5_megablocks/language_model.py (593:615) duplicated block id: 2236 size: 18 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/galactica/checkpoint_reshaping_and_interoperability.py (631:649) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (763:781) duplicated block id: 2237 size: 18 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/mlp.py (170:187) - megatron_patch/model/llama3/transformer/mlp.py (181:198) duplicated block id: 2238 size: 18 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/language_model.py (272:295) - megatron_patch/model/galactica/language_model.py (292:314) duplicated block id: 2239 size: 18 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/language_model.py (587:609) - megatron_patch/model/galactica/language_model.py (670:692) duplicated block id: 2240 size: 18 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (42:59) - toolkits/model_checkpoints_convertor/baichuan2/hf2te.py (33:50) duplicated block id: 2241 size: 18 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore.py (36:55) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (28:47) duplicated block id: 2242 size: 18 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llava/hf2mcore_llava.py (41:60) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (40:59) duplicated block id: 2243 size: 18 cleaned lines of code in 2 files: - megatron_patch/model/llama/language_model.py (384:406) - megatron_patch/model/starcoder/language_model.py (379:401) duplicated block id: 2244 size: 18 cleaned lines of code in 2 files: - megatron_patch/model/mistral/transformer.py (1601:1628) - megatron_patch/model/starcoder/transformer.py (1200:1227) duplicated block id: 2245 size: 18 cleaned lines of code in 2 files: - megatron_patch/model/qwen/language_model.py (199:224) - megatron_patch/model/starcoder/language_model.py (198:223) duplicated block id: 2246 size: 18 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (714:731) - toolkits/model_checkpoints_convertor/starcoder/checkpoint_reshaping_and_interoperability.py (335:352) duplicated block id: 2247 size: 18 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/moe/experts.py (179:206) - megatron_patch/model/qwen1_5/moe/experts.py (146:173) duplicated block id: 2248 size: 18 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (1562:1583) - megatron_patch/model/llama2/transformer.py (1692:1713) duplicated block id: 2249 size: 18 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/language_model.py (671:693) - megatron_patch/model/falcon40b/language_model.py (641:663) duplicated block id: 2250 size: 18 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2megablocks_qwen1.5.py (33:52) - toolkits/model_checkpoints_convertor/utils/__init__.py (82:101) duplicated block id: 2251 size: 18 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/language_model.py (277:299) - megatron_patch/model/qwen1_5_megablocks/language_model.py (261:284) duplicated block id: 2252 size: 18 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_moe.py (318:336) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_dense_and_moe_gqa.py (536:554) duplicated block id: 2253 size: 18 cleaned lines of code in 2 files: - megatron_patch/model/qwen1_5_megablocks/language_model.py (198:223) - megatron_patch/model/starcoder/language_model.py (198:223) duplicated block id: 2254 size: 18 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/language_model.py (277:299) - megatron_patch/model/llama3/language_model.py (257:280) duplicated block id: 2255 size: 18 cleaned lines of code in 2 files: - megatron_patch/model/bloom/gpt_model.py (49:71) - megatron_patch/model/starcoder/gpt_model.py (52:75) duplicated block id: 2256 size: 18 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/language_model.py (641:663) - megatron_patch/model/llava/language_model.py (654:676) duplicated block id: 2257 size: 18 cleaned lines of code in 2 files: - megatron_patch/model/bloom/language_model.py (279:301) - megatron_patch/model/glm130b/language_model.py (276:299) duplicated block id: 2258 size: 18 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon/checkpoint_reshaping_and_interoperability.py (577:595) - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (551:569) duplicated block id: 2259 size: 18 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (824:845) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (1107:1128) duplicated block id: 2260 size: 18 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/hf2te.py (33:50) - toolkits/model_checkpoints_convertor/galactica/checkpoint_reshaping_and_interoperability.py (42:59) duplicated block id: 2261 size: 18 cleaned lines of code in 2 files: - megatron_patch/model/mixtral_bak/transformer/mlp.py (89:111) - megatron_patch/model/qwen1_5/transformer/mlp.py (130:152) duplicated block id: 2262 size: 18 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/moe/experts.py (179:206) - megatron_patch/model/qwen2/moe/experts.py (170:197) duplicated block id: 2263 size: 18 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon/checkpoint_reshaping_and_interoperability.py (610:627) - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (782:799) duplicated block id: 2264 size: 18 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer_legacy.py (1652:1673) - megatron_patch/model/qwen_vl/transformer.py (1691:1712) duplicated block id: 2265 size: 18 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/chatglm/checkpoint_reshaping_and_interoperability.py (561:579) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (680:698) duplicated block id: 2266 size: 18 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/language_model.py (277:299) - megatron_patch/model/glm130b/language_model.py (276:299) duplicated block id: 2267 size: 18 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (634:652) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (680:698) duplicated block id: 2268 size: 18 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/language_model.py (277:299) - megatron_patch/model/qwen/language_model.py (262:285) duplicated block id: 2269 size: 18 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon40b/checkpoint_reshaping_and_interoperability.py (804:824) - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (1108:1128) duplicated block id: 2270 size: 18 cleaned lines of code in 2 files: - megatron_patch/model/bloom/language_model.py (131:165) - megatron_patch/model/falcon/language_model.py (129:163) duplicated block id: 2271 size: 18 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (1562:1583) - megatron_patch/model/qwen_vl/transformer.py (1691:1712) duplicated block id: 2272 size: 18 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (28:47) - toolkits/model_checkpoints_convertor/utils/__init__.py (82:101) duplicated block id: 2273 size: 18 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/moe/experts.py (180:207) - megatron_patch/model/qwen1_5/moe/experts.py (146:173) duplicated block id: 2274 size: 18 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/language_model.py (579:601) - megatron_patch/model/mistral/language_model.py (606:628) duplicated block id: 2275 size: 18 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/language_model.py (587:609) - megatron_patch/model/llama/language_model.py (652:674) duplicated block id: 2276 size: 18 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (901:921) - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (1108:1128) duplicated block id: 2277 size: 18 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/language_model.py (641:663) - megatron_patch/model/qwen1_5_megablocks/language_model.py (593:615) duplicated block id: 2278 size: 18 cleaned lines of code in 2 files: - megatron_patch/model/bloom/language_model.py (549:571) - megatron_patch/model/qwen1_5_megablocks/language_model.py (593:615) duplicated block id: 2279 size: 18 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/chatglm/checkpoint_reshaping_and_interoperability.py (260:296) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (338:374) duplicated block id: 2280 size: 18 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (35:54) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_mha_to_moe.py (28:47) duplicated block id: 2281 size: 18 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_moe.py (31:50) - toolkits/model_checkpoints_convertor/utils/__init__.py (82:101) duplicated block id: 2282 size: 18 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/language_model.py (318:341) - megatron_patch/model/falcon/language_model.py (261:283) duplicated block id: 2283 size: 18 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (531:553) - megatron_patch/model/baichuan2/transformer.py (640:662) duplicated block id: 2284 size: 18 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/language_model.py (671:693) - megatron_patch/model/falcon/language_model.py (641:663) duplicated block id: 2285 size: 18 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/language_model.py (615:637) - megatron_patch/model/mistral/language_model.py (606:628) duplicated block id: 2286 size: 18 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (1469:1496) - megatron_patch/model/starcoder/transformer.py (1200:1227) duplicated block id: 2287 size: 18 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llava/hf2mcore_llava.py (41:60) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (28:47) duplicated block id: 2288 size: 18 cleaned lines of code in 2 files: - megatron_patch/model/mistral/transformer.py (1689:1710) - megatron_patch/model/qwen/transformer.py (1634:1655) duplicated block id: 2289 size: 18 cleaned lines of code in 2 files: - megatron_patch/model/falcon/language_model.py (641:663) - megatron_patch/model/llava/language_model.py (654:676) duplicated block id: 2290 size: 18 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (337:362) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (458:484) duplicated block id: 2291 size: 18 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/transformer.py (1195:1235) - megatron_patch/model/llama2/transformer.py (1751:1792) duplicated block id: 2292 size: 18 cleaned lines of code in 2 files: - megatron_patch/model/falcon/language_model.py (261:283) - megatron_patch/model/qwen_vl/language_model.py (263:286) duplicated block id: 2293 size: 18 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer_legacy.py (1564:1591) - megatron_patch/model/starcoder/transformer.py (1200:1227) duplicated block id: 2294 size: 18 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (28:47) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (39:58) duplicated block id: 2295 size: 18 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (1604:1631) - megatron_patch/model/starcoder/transformer.py (1200:1227) duplicated block id: 2296 size: 18 cleaned lines of code in 2 files: - megatron_patch/model/llama3_1/model.py (188:216) - megatron_patch/model/qwen2/model.py (131:159) duplicated block id: 2297 size: 18 cleaned lines of code in 2 files: - megatron_patch/model/galactica/language_model.py (292:314) - megatron_patch/model/llava/language_model.py (265:288) duplicated block id: 2298 size: 18 cleaned lines of code in 2 files: - megatron_patch/model/bloom/language_model.py (549:571) - megatron_patch/model/llava/language_model.py (654:676) duplicated block id: 2299 size: 18 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/hf2te.py (33:50) - toolkits/model_checkpoints_convertor/falcon40b/checkpoint_reshaping_and_interoperability.py (43:60) duplicated block id: 2300 size: 18 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/transformer/mlp.py (178:195) - megatron_patch/model/qwen2/transformer/mlp.py (200:217) duplicated block id: 2301 size: 18 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (1562:1583) - megatron_patch/model/llava/transformer.py (1691:1712) duplicated block id: 2302 size: 18 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (39:58) - toolkits/model_checkpoints_convertor/utils/__init__.py (82:101) duplicated block id: 2303 size: 18 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/transformer.py (577:597) - megatron_patch/model/mistral/transformer.py (772:792) duplicated block id: 2304 size: 18 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (634:652) - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (551:569) duplicated block id: 2305 size: 18 cleaned lines of code in 2 files: - megatron_patch/model/galactica/language_model.py (292:314) - megatron_patch/model/mistral/language_model.py (263:286) duplicated block id: 2306 size: 18 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/language_model.py (587:609) - megatron_patch/model/bloom/language_model.py (549:571) duplicated block id: 2307 size: 18 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/hf2te.py (34:51) - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (42:59) duplicated block id: 2308 size: 18 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/language_model.py (261:283) - megatron_patch/model/mistral/language_model.py (263:286) duplicated block id: 2309 size: 18 cleaned lines of code in 2 files: - megatron_patch/model/falcon/language_model.py (641:663) - megatron_patch/model/qwen/language_model.py (576:598) duplicated block id: 2310 size: 18 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore.py (700:719) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_moe.py (510:529) duplicated block id: 2311 size: 18 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/transformer.py (1195:1235) - megatron_patch/model/llava/transformer.py (1750:1791) duplicated block id: 2312 size: 18 cleaned lines of code in 2 files: - megatron_patch/model/falcon/transformer.py (791:812) - megatron_patch/model/qwen_vl/transformer.py (773:793) duplicated block id: 2313 size: 18 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llava/hf2mcore_llava.py (41:60) - toolkits/model_checkpoints_convertor/utils/__init__.py (82:101) duplicated block id: 2314 size: 18 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (28:47) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_mha.py (29:48) duplicated block id: 2315 size: 18 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/hf2te.py (34:51) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (70:87) duplicated block id: 2316 size: 18 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_vl.py (139:156) - toolkits/model_checkpoints_convertor/qwen/hf2megablocks_qwen1.5.py (282:299) duplicated block id: 2317 size: 18 cleaned lines of code in 2 files: - megatron_patch/model/falcon/language_model.py (261:283) - megatron_patch/model/glm130b/language_model.py (276:299) duplicated block id: 2318 size: 18 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/language_model.py (579:601) - megatron_patch/model/llava/language_model.py (654:676) duplicated block id: 2319 size: 18 cleaned lines of code in 2 files: - megatron_patch/model/bloom/language_model.py (279:301) - megatron_patch/model/llama3/language_model.py (257:280) duplicated block id: 2320 size: 18 cleaned lines of code in 2 files: - megatron_patch/model/galactica/language_model.py (670:692) - megatron_patch/model/mistral/language_model.py (606:628) duplicated block id: 2321 size: 18 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/language_model.py (587:609) - megatron_patch/model/glm130b/language_model.py (579:601) duplicated block id: 2322 size: 18 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore.py (36:55) - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (28:47) duplicated block id: 2323 size: 18 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon/checkpoint_reshaping_and_interoperability.py (530:552) - toolkits/model_checkpoints_convertor/galactica/checkpoint_reshaping_and_interoperability.py (588:612) duplicated block id: 2324 size: 18 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/hf2te.py (33:50) - toolkits/model_checkpoints_convertor/yi/checkpoint_reshaping_and_interoperability.py (43:60) duplicated block id: 2325 size: 18 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/transformer/mlp.py (178:195) - megatron_patch/model/qwen1_5/transformer/mlp.py (197:214) duplicated block id: 2326 size: 18 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/language_model.py (261:283) - megatron_patch/model/llama2/language_model.py (262:285) duplicated block id: 2327 size: 18 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/language_model.py (641:663) - megatron_patch/model/llama2/language_model.py (594:616) duplicated block id: 2328 size: 18 cleaned lines of code in 2 files: - megatron_patch/model/galactica/language_model.py (292:314) - megatron_patch/model/qwen/language_model.py (262:285) duplicated block id: 2329 size: 18 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/transformer.py (1195:1235) - megatron_patch/model/qwen_vl/transformer.py (1750:1791) duplicated block id: 2330 size: 18 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (35:54) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_mha.py (29:48) duplicated block id: 2331 size: 18 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (35:54) - toolkits/model_checkpoints_convertor/llava/hf2mcore_llava.py (41:60) duplicated block id: 2332 size: 18 cleaned lines of code in 2 files: - megatron_patch/model/galactica/language_model.py (415:437) - megatron_patch/model/starcoder/language_model.py (379:401) duplicated block id: 2333 size: 18 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/language_model.py (641:663) - megatron_patch/model/qwen_vl/language_model.py (623:645) duplicated block id: 2334 size: 18 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/language_model.py (587:609) - megatron_patch/model/falcon/language_model.py (641:663) duplicated block id: 2335 size: 18 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore.py (288:305) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_vl.py (139:156) duplicated block id: 2336 size: 18 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (35:54) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (28:47) duplicated block id: 2337 size: 18 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/language_model.py (615:637) - megatron_patch/model/qwen/language_model.py (576:598) duplicated block id: 2338 size: 18 cleaned lines of code in 2 files: - megatron_patch/model/bloom/language_model.py (279:301) - megatron_patch/model/qwen/language_model.py (262:285) duplicated block id: 2339 size: 18 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/language_model.py (272:295) - megatron_patch/model/bloom/language_model.py (279:301) duplicated block id: 2340 size: 18 cleaned lines of code in 2 files: - megatron_patch/model/llama/language_model.py (652:674) - megatron_patch/model/qwen1_5_megablocks/language_model.py (593:615) duplicated block id: 2341 size: 18 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/language_model.py (615:637) - megatron_patch/model/llava/language_model.py (654:676) duplicated block id: 2342 size: 18 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/chatglm/checkpoint_reshaping_and_interoperability.py (561:579) - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (551:569) duplicated block id: 2343 size: 18 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_mha_to_moe.py (28:47) - toolkits/model_checkpoints_convertor/utils/__init__.py (82:101) duplicated block id: 2344 size: 18 cleaned lines of code in 2 files: - megatron_patch/model/galactica/language_model.py (670:692) - megatron_patch/model/llama2/language_model.py (594:616) duplicated block id: 2345 size: 17 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/language_model.py (230:255) - megatron_patch/model/glm130b/language_model.py (232:258) duplicated block id: 2346 size: 17 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon/checkpoint_reshaping_and_interoperability.py (741:769) - toolkits/model_checkpoints_convertor/starcoder/checkpoint_reshaping_and_interoperability.py (458:486) duplicated block id: 2347 size: 17 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2.5_vl.py (140:156) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_dense_and_moe_gqa.py (180:197) duplicated block id: 2348 size: 17 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (783:811) - toolkits/model_checkpoints_convertor/bloom/reward_model_to_megatron.py (442:469) duplicated block id: 2349 size: 17 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_dense_and_moe_gqa.py (38:58) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_moe.py (112:132) duplicated block id: 2350 size: 17 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/language_model.py (232:258) - megatron_patch/model/qwen_vl/language_model.py (221:246) duplicated block id: 2351 size: 17 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/language_model.py (250:273) - megatron_patch/model/starcoder/language_model.py (198:221) duplicated block id: 2352 size: 17 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/mistral/hf2mcore.py (229:245) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_vl.py (140:156) duplicated block id: 2353 size: 17 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (783:811) - toolkits/model_checkpoints_convertor/bloom/checkpoint_reshaping_and_interoperability.py (443:470) duplicated block id: 2354 size: 17 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_vl.py (188:204) - toolkits/model_checkpoints_convertor/qwen/hf2megablocks_qwen1.5.py (282:298) duplicated block id: 2355 size: 17 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/language_model.py (233:259) - megatron_patch/model/llama/language_model.py (217:242) duplicated block id: 2356 size: 17 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/mlp.py (103:119) - megatron_patch/model/mixtral/moe/experts.py (642:658) duplicated block id: 2357 size: 17 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon/checkpoint_reshaping_and_interoperability.py (741:769) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (858:887) duplicated block id: 2358 size: 17 cleaned lines of code in 2 files: - megatron_patch/model/llama3/language_model.py (606:625) - megatron_patch/model/qwen1_5_megablocks/language_model.py (631:650) duplicated block id: 2359 size: 17 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_dense_and_moe_gqa.py (135:151) - toolkits/model_checkpoints_convertor/qwen/hf2megablocks_qwen1.5.py (283:299) duplicated block id: 2360 size: 17 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (919:947) - toolkits/model_checkpoints_convertor/starcoder/checkpoint_reshaping_and_interoperability.py (456:483) duplicated block id: 2361 size: 17 cleaned lines of code in 2 files: - megatron_patch/model/mixtral_bak/moe/token_dispatcher.py (284:302) - megatron_patch/model/qwen1_5/moe/token_dispatcher.py (247:265) duplicated block id: 2362 size: 17 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/moe/experts.py (83:109) - megatron_patch/model/qwen1_5/moe/experts.py (62:87) duplicated block id: 2363 size: 17 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/checkpoint_reshaping_and_interoperability.py (445:473) - toolkits/model_checkpoints_convertor/falcon40b/checkpoint_reshaping_and_interoperability.py (726:754) duplicated block id: 2364 size: 17 cleaned lines of code in 2 files: - megatron_patch/model/llama3/language_model.py (606:625) - megatron_patch/model/qwen/language_model.py (614:633) duplicated block id: 2365 size: 17 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/mistral/hf2mcore.py (268:284) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_dense_and_moe_gqa.py (177:193) duplicated block id: 2366 size: 17 cleaned lines of code in 2 files: - megatron_patch/model/bloom/language_model.py (133:165) - megatron_patch/model/chatglm/language_model.py (132:164) duplicated block id: 2367 size: 17 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_dense_and_moe_gqa.py (180:197) - toolkits/model_checkpoints_convertor/qwen/hf2megablocks_qwen1.5.py (283:299) duplicated block id: 2368 size: 17 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon40b/checkpoint_reshaping_and_interoperability.py (726:754) - toolkits/model_checkpoints_convertor/starcoder/checkpoint_reshaping_and_interoperability.py (458:486) duplicated block id: 2369 size: 17 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/deepseek/hf2mcore_deepseek_v3_moe.py (52:74) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_moe.py (58:82) duplicated block id: 2370 size: 17 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/mistral/hf2mcore.py (229:245) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2.5_vl.py (140:156) duplicated block id: 2371 size: 17 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/language_model.py (233:259) - megatron_patch/model/falcon40b/language_model.py (217:242) duplicated block id: 2372 size: 17 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/reward_model_to_megatron.py (444:472) - toolkits/model_checkpoints_convertor/falcon40b/checkpoint_reshaping_and_interoperability.py (726:754) duplicated block id: 2373 size: 17 cleaned lines of code in 2 files: - megatron_patch/model/llama3/language_model.py (606:625) - megatron_patch/model/mistral/language_model.py (644:663) duplicated block id: 2374 size: 17 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/language_model.py (709:728) - megatron_patch/model/llama3/language_model.py (606:625) duplicated block id: 2375 size: 17 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_dense_and_moe_gqa.py (135:151) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_vl.py (140:156) duplicated block id: 2376 size: 17 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/mistral/hf2mcore.py (268:284) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_dense_and_moe_gqa.py (132:148) duplicated block id: 2377 size: 17 cleaned lines of code in 2 files: - megatron_patch/model/mixtral_bak/moe/token_dispatcher.py (284:302) - megatron_patch/model/qwen2/moe/token_dispatcher.py (253:271) duplicated block id: 2378 size: 17 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (756:784) - toolkits/model_checkpoints_convertor/bloom/reward_model_to_megatron.py (442:469) duplicated block id: 2379 size: 17 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/language_model.py (232:258) - megatron_patch/model/qwen/language_model.py (220:245) duplicated block id: 2380 size: 17 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/deepseek/hf2mcore_deepseek_v2_moe.py (50:70) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_moe.py (112:132) duplicated block id: 2381 size: 17 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore.py (288:304) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_vl.py (188:204) duplicated block id: 2382 size: 17 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/reward_model_to_megatron.py (444:472) - toolkits/model_checkpoints_convertor/falcon/checkpoint_reshaping_and_interoperability.py (741:769) duplicated block id: 2383 size: 17 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore_llama3_1.py (567:584) - toolkits/model_checkpoints_convertor/llava/hf2mcore_llava.py (336:353) duplicated block id: 2384 size: 17 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/language_model.py (232:258) - megatron_patch/model/llama3/language_model.py (210:235) duplicated block id: 2385 size: 17 cleaned lines of code in 2 files: - megatron_patch/model/mistral/language_model.py (492:510) - megatron_patch/model/qwen_vl/language_model.py (519:537) duplicated block id: 2386 size: 17 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/language_model.py (217:242) - megatron_patch/model/starcoder/language_model.py (221:246) duplicated block id: 2387 size: 17 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon40b/checkpoint_reshaping_and_interoperability.py (726:754) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (858:887) duplicated block id: 2388 size: 17 cleaned lines of code in 2 files: - megatron_patch/model/qwen2_5_vl/visionmodel.py (69:88) - megatron_patch/model/qwen2_vl/visionmodel.py (69:88) duplicated block id: 2389 size: 17 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/deepseek/hf2mcore_deepseek_v3_moe.py (103:123) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_dense_and_moe_gqa.py (38:58) duplicated block id: 2390 size: 17 cleaned lines of code in 2 files: - megatron_patch/model/llama3/language_model.py (606:625) - megatron_patch/model/llava/language_model.py (692:711) duplicated block id: 2391 size: 17 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/moe/experts.py (642:658) - megatron_patch/model/qwen2/transformer/mlp.py (124:140) duplicated block id: 2392 size: 17 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/moe/experts.py (83:109) - megatron_patch/model/mixtral_bak/moe/experts.py (60:85) duplicated block id: 2393 size: 17 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (255:301) - megatron_patch/model/glm130b/transformer.py (268:312) duplicated block id: 2394 size: 17 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore.py (289:305) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_dense_and_moe_gqa.py (135:151) duplicated block id: 2395 size: 17 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2.5_vl.py (140:156) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_dense_and_moe_gqa.py (135:151) duplicated block id: 2396 size: 17 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (897:925) - toolkits/model_checkpoints_convertor/starcoder/checkpoint_reshaping_and_interoperability.py (456:483) duplicated block id: 2397 size: 17 cleaned lines of code in 2 files: - megatron_patch/model/llama2/language_model.py (632:651) - megatron_patch/model/llama3/language_model.py (606:625) duplicated block id: 2398 size: 17 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (756:784) - toolkits/model_checkpoints_convertor/starcoder/checkpoint_reshaping_and_interoperability.py (456:483) duplicated block id: 2399 size: 17 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/moe/experts.py (84:110) - megatron_patch/model/qwen1_5/moe/experts.py (62:87) duplicated block id: 2400 size: 17 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/language_model.py (625:644) - megatron_patch/model/llama3/language_model.py (606:625) duplicated block id: 2401 size: 17 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/transformer.py (540:561) - megatron_patch/model/starcoder/transformer.py (857:878) duplicated block id: 2402 size: 17 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/mlp.py (103:119) - megatron_patch/model/deepseek_v2/moe/experts.py (641:657) duplicated block id: 2403 size: 17 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (783:811) - toolkits/model_checkpoints_convertor/starcoder/checkpoint_reshaping_and_interoperability.py (456:483) duplicated block id: 2404 size: 17 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/language_model.py (232:258) - megatron_patch/model/mistral/language_model.py (221:246) duplicated block id: 2405 size: 17 cleaned lines of code in 2 files: - megatron_patch/model/llama/language_model.py (217:242) - megatron_patch/model/starcoder/language_model.py (221:246) duplicated block id: 2406 size: 17 cleaned lines of code in 2 files: - megatron_patch/model/falcon/transformer.py (660:680) - megatron_patch/model/llama/transformer.py (460:480) duplicated block id: 2407 size: 17 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/mistral/hf2mcore.py (226:242) - toolkits/model_checkpoints_convertor/mistral/hf2mcore.py (268:284) duplicated block id: 2408 size: 17 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/reward_model_to_megatron.py (442:469) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (919:947) duplicated block id: 2409 size: 17 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (114:133) - megatron_patch/model/galactica/transformer.py (91:110) duplicated block id: 2410 size: 17 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2.5_vl.py (188:204) - toolkits/model_checkpoints_convertor/qwen/hf2megablocks_qwen1.5.py (282:298) duplicated block id: 2411 size: 17 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/moe/experts.py (641:657) - megatron_patch/model/mixtral/transformer/mlp.py (111:127) duplicated block id: 2412 size: 17 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/language_model.py (232:258) - megatron_patch/model/llava/language_model.py (223:248) duplicated block id: 2413 size: 17 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (880:901) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (796:817) duplicated block id: 2414 size: 17 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/deepseek/hf2mcore_deepseek_v2_moe.py (50:70) - toolkits/model_checkpoints_convertor/deepseek/hf2mcore_deepseek_v3_moe.py (103:123) duplicated block id: 2415 size: 17 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore.py (289:305) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_dense_and_moe_gqa.py (180:197) duplicated block id: 2416 size: 17 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_mha.py (119:135) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2.5_vl.py (139:155) duplicated block id: 2417 size: 17 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_dense_and_moe_gqa.py (180:197) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_vl.py (140:156) duplicated block id: 2418 size: 17 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/language_model.py (233:259) - megatron_patch/model/falcon/language_model.py (217:242) duplicated block id: 2419 size: 17 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/checkpoint_reshaping_and_interoperability.py (445:473) - toolkits/model_checkpoints_convertor/falcon/checkpoint_reshaping_and_interoperability.py (741:769) duplicated block id: 2420 size: 17 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/hf2te.py (238:262) - toolkits/model_checkpoints_convertor/baichuan2/hf2te.py (222:247) duplicated block id: 2421 size: 17 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (114:133) - megatron_patch/model/falcon40b/transformer.py (141:160) duplicated block id: 2422 size: 17 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/language_model.py (232:258) - megatron_patch/model/llama2/language_model.py (220:245) duplicated block id: 2423 size: 17 cleaned lines of code in 2 files: - megatron_patch/model/falcon/language_model.py (217:242) - megatron_patch/model/starcoder/language_model.py (221:246) duplicated block id: 2424 size: 17 cleaned lines of code in 2 files: - megatron_patch/model/galactica/transformer.py (201:246) - megatron_patch/model/glm130b/transformer.py (268:312) duplicated block id: 2425 size: 17 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/moe/experts.py (84:110) - megatron_patch/model/mixtral_bak/moe/experts.py (60:85) duplicated block id: 2426 size: 17 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/transformer.py (399:419) - megatron_patch/model/falcon/transformer.py (660:680) duplicated block id: 2427 size: 17 cleaned lines of code in 2 files: - megatron_patch/model/qwen3_moe/moe/router.py (51:70) - megatron_patch/model/qwen3_moe/moe/router.py (94:112) duplicated block id: 2428 size: 17 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/checkpoint_reshaping_and_interoperability.py (443:470) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (919:947) duplicated block id: 2429 size: 17 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/checkpoint_reshaping_and_interoperability.py (443:470) - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (897:925) duplicated block id: 2430 size: 17 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (767:788) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (796:817) duplicated block id: 2431 size: 17 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/reward_model_to_megatron.py (442:469) - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (897:925) duplicated block id: 2432 size: 17 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore.py (288:304) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2.5_vl.py (188:204) duplicated block id: 2433 size: 17 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_mha.py (119:135) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_vl.py (139:155) duplicated block id: 2434 size: 17 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/moe/experts.py (642:658) - megatron_patch/model/mixtral/transformer/mlp.py (111:127) duplicated block id: 2435 size: 17 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (756:784) - toolkits/model_checkpoints_convertor/bloom/checkpoint_reshaping_and_interoperability.py (443:470) duplicated block id: 2436 size: 17 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/language_model.py (232:258) - megatron_patch/model/qwen1_5_megablocks/language_model.py (219:244) duplicated block id: 2437 size: 17 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/moe/experts.py (641:657) - megatron_patch/model/qwen2/transformer/mlp.py (124:140) duplicated block id: 2438 size: 17 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (114:133) - megatron_patch/model/falcon/transformer.py (148:167) duplicated block id: 2439 size: 17 cleaned lines of code in 2 files: - megatron_patch/model/llama3/language_model.py (606:625) - megatron_patch/model/qwen_vl/language_model.py (661:680) duplicated block id: 2440 size: 17 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (643:664) - megatron_patch/model/chatglm/transformer.py (540:561) duplicated block id: 2441 size: 16 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (340:370) - megatron_patch/model/starcoder/transformer.py (463:493) duplicated block id: 2442 size: 16 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (449:479) - megatron_patch/model/falcon/transformer.py (467:497) duplicated block id: 2443 size: 16 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (561:583) - toolkits/model_checkpoints_convertor/yi/checkpoint_reshaping_and_interoperability.py (581:604) duplicated block id: 2444 size: 16 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (741:757) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (772:789) duplicated block id: 2445 size: 16 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/mistral/hf2mcore.py (226:241) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_dense_and_moe_gqa.py (238:253) duplicated block id: 2446 size: 16 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_dense_and_moe_gqa.py (132:147) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_dense_and_moe_gqa.py (238:253) duplicated block id: 2447 size: 16 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/hf2te.py (379:402) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (409:432) duplicated block id: 2448 size: 16 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (744:760) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (772:789) duplicated block id: 2449 size: 16 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (262:301) - megatron_patch/model/falcon40b/transformer.py (258:296) duplicated block id: 2450 size: 16 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (678:700) - toolkits/model_checkpoints_convertor/yi/checkpoint_reshaping_and_interoperability.py (581:604) duplicated block id: 2451 size: 16 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (182:208) - toolkits/model_checkpoints_convertor/falcon40b/checkpoint_reshaping_and_interoperability.py (187:213) duplicated block id: 2452 size: 16 cleaned lines of code in 2 files: - megatron_patch/model/llama2/transformer.py (1787:1807) - megatron_patch/model/qwen1_5_megablocks/transformer.py (1650:1670) duplicated block id: 2453 size: 16 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/language_model.py (600:620) - megatron_patch/model/glm130b/language_model.py (543:563) duplicated block id: 2454 size: 16 cleaned lines of code in 2 files: - megatron_patch/model/falcon/transformer.py (467:497) - megatron_patch/model/llama2/transformer.py (437:467) duplicated block id: 2455 size: 16 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/transformer.py (128:148) - megatron_patch/model/llama2/transformer.py (119:135) duplicated block id: 2456 size: 16 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (364:386) - megatron_patch/model/glm130b/transformer.py (398:420) duplicated block id: 2457 size: 16 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore_llama3_1.py (184:199) - toolkits/model_checkpoints_convertor/llava/hf2mcore_llava.py (206:221) duplicated block id: 2458 size: 16 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (626:649) - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (740:763) duplicated block id: 2459 size: 16 cleaned lines of code in 2 files: - megatron_patch/model/llama2/transformer.py (265:282) - megatron_patch/model/llama3/transformer_legacy.py (269:286) duplicated block id: 2460 size: 16 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (449:479) - megatron_patch/model/starcoder/transformer.py (463:493) duplicated block id: 2461 size: 16 cleaned lines of code in 2 files: - megatron_patch/model/galactica/transformer.py (310:332) - megatron_patch/model/glm130b/transformer.py (398:420) duplicated block id: 2462 size: 16 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (899:924) - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (1080:1105) duplicated block id: 2463 size: 16 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/mistral/hf2mcore.py (268:283) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_dense_and_moe_gqa.py (238:253) duplicated block id: 2464 size: 16 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/transformer.py (128:148) - megatron_patch/model/llama3/transformer_legacy.py (123:139) duplicated block id: 2465 size: 16 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (899:924) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (975:1000) duplicated block id: 2466 size: 16 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/language_model.py (566:588) - megatron_patch/model/starcoder/language_model.py (450:472) duplicated block id: 2467 size: 16 cleaned lines of code in 2 files: - megatron_patch/model/falcon/transformer.py (467:497) - megatron_patch/model/qwen1_5_megablocks/transformer.py (328:358) duplicated block id: 2468 size: 16 cleaned lines of code in 2 files: - megatron_patch/model/llava_mcore/llava_model.py (226:252) - megatron_patch/model/qwen2_vl/model.py (140:166) duplicated block id: 2469 size: 16 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (561:583) - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (678:700) duplicated block id: 2470 size: 16 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer/mlp.py (72:89) - megatron_patch/model/qwen2/transformer/mlp.py (78:96) duplicated block id: 2471 size: 16 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/language_model.py (680:696) - megatron_patch/model/llama/language_model.py (691:707) duplicated block id: 2472 size: 16 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer_legacy.py (269:286) - megatron_patch/model/llava/transformer.py (265:282) duplicated block id: 2473 size: 16 cleaned lines of code in 2 files: - megatron_patch/model/falcon/transformer.py (532:556) - megatron_patch/model/starcoder/transformer.py (510:533) duplicated block id: 2474 size: 16 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (260:277) - megatron_patch/model/llama3/transformer_legacy.py (269:286) duplicated block id: 2475 size: 16 cleaned lines of code in 2 files: - megatron_patch/model/mistral/transformer.py (1784:1804) - megatron_patch/model/qwen1_5_megablocks/transformer.py (1650:1670) duplicated block id: 2476 size: 16 cleaned lines of code in 2 files: - megatron_patch/model/bloom/language_model.py (78:97) - megatron_patch/model/chatglm/language_model.py (77:96) duplicated block id: 2477 size: 16 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/transformer.py (128:148) - megatron_patch/model/qwen1_5_megablocks/transformer.py (117:133) duplicated block id: 2478 size: 16 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (117:133) - megatron_patch/model/glm130b/transformer.py (128:148) duplicated block id: 2479 size: 16 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (425:447) - toolkits/model_checkpoints_convertor/baichuan2/hf2te.py (279:301) duplicated block id: 2480 size: 16 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/hf2te.py (351:374) - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (422:445) duplicated block id: 2481 size: 16 cleaned lines of code in 2 files: - megatron_patch/model/qwen1_5/moe/moe_layer.py (85:100) - megatron_patch/model/qwen2/moe/moe_layer.py (109:125) duplicated block id: 2482 size: 16 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon40b/checkpoint_reshaping_and_interoperability.py (726:751) - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (899:925) duplicated block id: 2483 size: 16 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/language_model.py (77:96) - megatron_patch/model/galactica/language_model.py (76:95) duplicated block id: 2484 size: 16 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (785:811) - toolkits/model_checkpoints_convertor/falcon/checkpoint_reshaping_and_interoperability.py (741:766) duplicated block id: 2485 size: 16 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/moe/router.py (232:270) - megatron_patch/model/qwen1_5/moe/router.py (214:250) duplicated block id: 2486 size: 16 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (453:475) - toolkits/model_checkpoints_convertor/baichuan2/hf2te.py (279:301) duplicated block id: 2487 size: 16 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/hf2te.py (279:301) - toolkits/model_checkpoints_convertor/falcon40b/checkpoint_reshaping_and_interoperability.py (404:426) duplicated block id: 2488 size: 16 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/deepseek/hf2mcore_deepseek_v3_moe.py (137:156) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_moe.py (140:160) duplicated block id: 2489 size: 16 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (785:811) - toolkits/model_checkpoints_convertor/falcon40b/checkpoint_reshaping_and_interoperability.py (726:751) duplicated block id: 2490 size: 16 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/transformer.py (128:148) - megatron_patch/model/llava/transformer.py (119:135) duplicated block id: 2491 size: 16 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/hf2te.py (379:402) - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (422:445) duplicated block id: 2492 size: 16 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2.5_vl.py (189:204) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_dense_and_moe_gqa.py (180:195) duplicated block id: 2493 size: 16 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_mha.py (120:135) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_dense_and_moe_gqa.py (180:195) duplicated block id: 2494 size: 16 cleaned lines of code in 2 files: - megatron_patch/model/llama2/transformer.py (437:467) - megatron_patch/model/starcoder/transformer.py (463:493) duplicated block id: 2495 size: 16 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/mistral/hf2mcore.py (229:244) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2.5_vl.py (189:204) duplicated block id: 2496 size: 16 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/transformer.py (33:69) - megatron_patch/model/galactica/transformer.py (35:70) duplicated block id: 2497 size: 16 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_dense_and_moe_gqa.py (177:192) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_dense_and_moe_gqa.py (238:253) duplicated block id: 2498 size: 16 cleaned lines of code in 2 files: - megatron_patch/model/galactica/language_model.py (629:649) - megatron_patch/model/glm130b/language_model.py (543:563) duplicated block id: 2499 size: 16 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon/checkpoint_reshaping_and_interoperability.py (741:766) - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (899:925) duplicated block id: 2500 size: 16 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/transformer.py (258:296) - megatron_patch/model/glm130b/transformer.py (275:312) duplicated block id: 2501 size: 16 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2.5_vl.py (189:204) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_dense_and_moe_gqa.py (135:150) duplicated block id: 2502 size: 16 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/reward_model_to_megatron.py (744:766) - toolkits/model_checkpoints_convertor/glm130b/checkpoint_reshaping_and_interoperability.py (451:474) duplicated block id: 2503 size: 16 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/hf2te.py (282:305) - toolkits/model_checkpoints_convertor/baichuan2/hf2te.py (254:277) duplicated block id: 2504 size: 16 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer/mlp.py (72:89) - megatron_patch/model/qwen1_5/transformer/mlp.py (79:97) duplicated block id: 2505 size: 16 cleaned lines of code in 2 files: - megatron_patch/model/qwen/transformer.py (141:162) - megatron_patch/model/qwen1_5_megablocks/transformer.py (145:166) duplicated block id: 2506 size: 16 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/transformer.py (564:581) - megatron_patch/model/starcoder/transformer.py (882:899) duplicated block id: 2507 size: 16 cleaned lines of code in 2 files: - megatron_patch/model/falcon/transformer.py (467:497) - megatron_patch/model/mistral/transformer.py (438:468) duplicated block id: 2508 size: 16 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (561:583) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (673:695) duplicated block id: 2509 size: 16 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (293:313) - megatron_patch/model/llama3/transformer_legacy.py (297:317) duplicated block id: 2510 size: 16 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/hf2te.py (351:374) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (409:432) duplicated block id: 2511 size: 16 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/hf2te.py (279:301) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (513:535) duplicated block id: 2512 size: 16 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (873:898) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (975:1000) duplicated block id: 2513 size: 16 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/transformer/mlp.py (74:91) - megatron_patch/model/qwen2/transformer/mlp.py (78:96) duplicated block id: 2514 size: 16 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/transformer.py (305:327) - megatron_patch/model/glm130b/transformer.py (398:420) duplicated block id: 2515 size: 16 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/language_model.py (513:535) - megatron_patch/model/glm130b/language_model.py (508:530) duplicated block id: 2516 size: 16 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (790:809) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (957:978) duplicated block id: 2517 size: 16 cleaned lines of code in 2 files: - megatron_patch/model/qwen/transformer.py (431:461) - megatron_patch/model/starcoder/transformer.py (463:493) duplicated block id: 2518 size: 16 cleaned lines of code in 2 files: - megatron_patch/model/qwen1_5_megablocks/transformer.py (328:358) - megatron_patch/model/starcoder/transformer.py (463:493) duplicated block id: 2519 size: 16 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (340:370) - megatron_patch/model/falcon/transformer.py (467:497) duplicated block id: 2520 size: 16 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/language_model.py (76:95) - megatron_patch/model/starcoder/language_model.py (64:83) duplicated block id: 2521 size: 16 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/language_model.py (77:96) - megatron_patch/model/starcoder/language_model.py (64:83) duplicated block id: 2522 size: 16 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/language_model.py (508:530) - megatron_patch/model/qwen/language_model.py (502:524) duplicated block id: 2523 size: 16 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (441:464) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (457:480) duplicated block id: 2524 size: 16 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/hf2te.py (307:329) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (483:505) duplicated block id: 2525 size: 16 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/hf2te.py (307:329) - toolkits/model_checkpoints_convertor/falcon/checkpoint_reshaping_and_interoperability.py (415:437) duplicated block id: 2526 size: 16 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/hf2te.py (279:301) - toolkits/model_checkpoints_convertor/yi/checkpoint_reshaping_and_interoperability.py (428:450) duplicated block id: 2527 size: 16 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer_legacy.py (269:286) - megatron_patch/model/mistral/transformer.py (265:282) duplicated block id: 2528 size: 16 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (857:873) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (772:789) duplicated block id: 2529 size: 16 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore.py (328:343) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_moe.py (156:171) duplicated block id: 2530 size: 16 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer_legacy.py (269:286) - megatron_patch/model/qwen/transformer.py (259:276) duplicated block id: 2531 size: 16 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/moe/router.py (232:270) - megatron_patch/model/qwen2/moe/router.py (300:337) duplicated block id: 2532 size: 16 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/mistral/hf2mcore.py (229:244) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_mha.py (120:135) duplicated block id: 2533 size: 16 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/hf2te.py (279:301) - toolkits/model_checkpoints_convertor/galactica/checkpoint_reshaping_and_interoperability.py (452:475) duplicated block id: 2534 size: 16 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/hf2te.py (307:329) - toolkits/model_checkpoints_convertor/yi/checkpoint_reshaping_and_interoperability.py (428:450) duplicated block id: 2535 size: 16 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (589:611) - toolkits/model_checkpoints_convertor/yi/checkpoint_reshaping_and_interoperability.py (581:604) duplicated block id: 2536 size: 16 cleaned lines of code in 2 files: - megatron_patch/model/falcon/transformer.py (467:497) - megatron_patch/model/qwen/transformer.py (431:461) duplicated block id: 2537 size: 16 cleaned lines of code in 2 files: - megatron_patch/model/llama/language_model.py (76:95) - megatron_patch/model/starcoder/language_model.py (64:83) duplicated block id: 2538 size: 16 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/transformer.py (128:148) - megatron_patch/model/qwen_vl/transformer.py (119:135) duplicated block id: 2539 size: 16 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/transformer.py (360:382) - megatron_patch/model/glm130b/transformer.py (398:420) duplicated block id: 2540 size: 16 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon/checkpoint_reshaping_and_interoperability.py (741:766) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (921:947) duplicated block id: 2541 size: 16 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer_legacy.py (269:286) - megatron_patch/model/qwen_vl/transformer.py (265:282) duplicated block id: 2542 size: 16 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/transformer.py (398:420) - megatron_patch/model/llama/transformer.py (348:370) duplicated block id: 2543 size: 16 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/mistral/hf2mcore.py (229:244) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_vl.py (189:204) duplicated block id: 2544 size: 16 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/hf2te.py (279:301) - toolkits/model_checkpoints_convertor/falcon/checkpoint_reshaping_and_interoperability.py (415:437) duplicated block id: 2545 size: 16 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/language_model.py (508:530) - megatron_patch/model/llama3/language_model.py (497:519) duplicated block id: 2546 size: 16 cleaned lines of code in 2 files: - megatron_patch/model/falcon/language_model.py (600:620) - megatron_patch/model/glm130b/language_model.py (543:563) duplicated block id: 2547 size: 16 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/language_model.py (680:696) - megatron_patch/model/galactica/language_model.py (709:725) duplicated block id: 2548 size: 16 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/language_model.py (77:96) - megatron_patch/model/falcon40b/language_model.py (76:95) duplicated block id: 2549 size: 16 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/language_model.py (77:96) - megatron_patch/model/starcoder/language_model.py (64:83) duplicated block id: 2550 size: 16 cleaned lines of code in 2 files: - megatron_patch/model/llava_mcore/llava_model.py (226:252) - megatron_patch/model/qwen2_5_vl/model.py (140:166) duplicated block id: 2551 size: 16 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/transformer/mlp.py (74:91) - megatron_patch/model/qwen1_5/transformer/mlp.py (79:97) duplicated block id: 2552 size: 16 cleaned lines of code in 2 files: - megatron_patch/model/falcon/language_model.py (566:588) - megatron_patch/model/starcoder/language_model.py (450:472) duplicated block id: 2553 size: 16 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/language_model.py (77:96) - megatron_patch/model/llama/language_model.py (76:95) duplicated block id: 2554 size: 16 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (817:838) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (957:978) duplicated block id: 2555 size: 16 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (453:475) - toolkits/model_checkpoints_convertor/baichuan/hf2te.py (307:329) duplicated block id: 2556 size: 16 cleaned lines of code in 2 files: - megatron_patch/model/falcon/transformer.py (37:73) - megatron_patch/model/galactica/transformer.py (35:70) duplicated block id: 2557 size: 16 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/deepseek/hf2mcore_deepseek_v3_moe.py (663:680) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_moe.py (653:672) duplicated block id: 2558 size: 16 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (262:301) - megatron_patch/model/falcon/transformer.py (265:303) duplicated block id: 2559 size: 16 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (589:611) - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (678:700) duplicated block id: 2560 size: 16 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (899:924) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (1088:1113) duplicated block id: 2561 size: 16 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/hf2te.py (307:329) - toolkits/model_checkpoints_convertor/falcon40b/checkpoint_reshaping_and_interoperability.py (404:426) duplicated block id: 2562 size: 16 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (262:301) - megatron_patch/model/llama/transformer.py (246:284) duplicated block id: 2563 size: 16 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/transformer.py (128:148) - megatron_patch/model/mistral/transformer.py (119:135) duplicated block id: 2564 size: 16 cleaned lines of code in 2 files: - megatron_patch/model/falcon/transformer.py (467:497) - megatron_patch/model/llama3/transformer_legacy.py (436:466) duplicated block id: 2565 size: 16 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (740:763) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (630:653) duplicated block id: 2566 size: 16 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (492:518) - megatron_patch/model/llama/transformer.py (491:517) duplicated block id: 2567 size: 16 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (644:661) - toolkits/model_checkpoints_convertor/bloom/reward_model_to_megatron.py (318:335) duplicated block id: 2568 size: 16 cleaned lines of code in 2 files: - megatron_patch/model/galactica/transformer.py (35:70) - megatron_patch/model/llama/transformer.py (38:73) duplicated block id: 2569 size: 16 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/hf2te.py (307:329) - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (425:447) duplicated block id: 2570 size: 16 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (758:784) - toolkits/model_checkpoints_convertor/falcon40b/checkpoint_reshaping_and_interoperability.py (726:751) duplicated block id: 2571 size: 16 cleaned lines of code in 2 files: - megatron_patch/model/mistral/transformer.py (438:468) - megatron_patch/model/starcoder/transformer.py (463:493) duplicated block id: 2572 size: 16 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/reward_model_to_megatron.py (616:638) - toolkits/model_checkpoints_convertor/starcoder/checkpoint_reshaping_and_interoperability.py (630:652) duplicated block id: 2573 size: 16 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/hf2te.py (307:329) - toolkits/model_checkpoints_convertor/galactica/checkpoint_reshaping_and_interoperability.py (452:475) duplicated block id: 2574 size: 16 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_moe.py (156:171) - toolkits/model_checkpoints_convertor/qwen/hf2megablocks_qwen1.5.py (321:336) duplicated block id: 2575 size: 16 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer_legacy.py (436:466) - megatron_patch/model/starcoder/transformer.py (463:493) duplicated block id: 2576 size: 16 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (38:73) - megatron_patch/model/falcon/transformer.py (37:73) duplicated block id: 2577 size: 16 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (114:130) - megatron_patch/model/glm130b/transformer.py (128:148) duplicated block id: 2578 size: 16 cleaned lines of code in 2 files: - megatron_patch/model/falcon/language_model.py (680:696) - megatron_patch/model/llama/language_model.py (691:707) duplicated block id: 2579 size: 16 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/reward_model_to_megatron.py (744:766) - toolkits/model_checkpoints_convertor/glm/checkpoint_reshaping_and_interoperability.py (485:517) duplicated block id: 2580 size: 16 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_mha.py (120:135) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_dense_and_moe_gqa.py (135:150) duplicated block id: 2581 size: 16 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/checkpoint_reshaping_and_interoperability.py (617:639) - toolkits/model_checkpoints_convertor/starcoder/checkpoint_reshaping_and_interoperability.py (630:652) duplicated block id: 2582 size: 16 cleaned lines of code in 2 files: - megatron_patch/model/qwen_vl/transformer.py (438:468) - megatron_patch/model/starcoder/transformer.py (463:493) duplicated block id: 2583 size: 16 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/transformer.py (275:312) - megatron_patch/model/llama/transformer.py (246:284) duplicated block id: 2584 size: 16 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (933:954) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (957:978) duplicated block id: 2585 size: 16 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (673:695) - toolkits/model_checkpoints_convertor/yi/checkpoint_reshaping_and_interoperability.py (581:604) duplicated block id: 2586 size: 16 cleaned lines of code in 2 files: - megatron_patch/model/falcon/transformer.py (467:497) - megatron_patch/model/qwen_vl/transformer.py (438:468) duplicated block id: 2587 size: 16 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/hf2te.py (307:329) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (513:535) duplicated block id: 2588 size: 16 cleaned lines of code in 2 files: - megatron_patch/model/llava/transformer.py (1786:1806) - megatron_patch/model/qwen1_5_megablocks/transformer.py (1650:1670) duplicated block id: 2589 size: 16 cleaned lines of code in 2 files: - megatron_patch/model/llama/language_model.py (577:599) - megatron_patch/model/starcoder/language_model.py (450:472) duplicated block id: 2590 size: 16 cleaned lines of code in 2 files: - megatron_patch/model/falcon/transformer.py (467:497) - megatron_patch/model/llava/transformer.py (438:468) duplicated block id: 2591 size: 16 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/language_model.py (77:96) - megatron_patch/model/glm130b/language_model.py (77:96) duplicated block id: 2592 size: 16 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon40b/checkpoint_reshaping_and_interoperability.py (726:751) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (921:947) duplicated block id: 2593 size: 16 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (678:700) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (673:695) duplicated block id: 2594 size: 16 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/language_model.py (543:563) - megatron_patch/model/llama/language_model.py (611:631) duplicated block id: 2595 size: 16 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/reward_model_to_megatron.py (318:335) - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (787:804) duplicated block id: 2596 size: 16 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/checkpoint_reshaping_and_interoperability.py (745:767) - toolkits/model_checkpoints_convertor/glm/checkpoint_reshaping_and_interoperability.py (485:517) duplicated block id: 2597 size: 16 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/hf2te.py (279:301) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (483:505) duplicated block id: 2598 size: 16 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/reward_model_to_megatron.py (318:335) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (801:818) duplicated block id: 2599 size: 16 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (758:784) - toolkits/model_checkpoints_convertor/falcon/checkpoint_reshaping_and_interoperability.py (741:766) duplicated block id: 2600 size: 16 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/transformer.py (128:148) - megatron_patch/model/qwen/transformer.py (113:129) duplicated block id: 2601 size: 16 cleaned lines of code in 2 files: - megatron_patch/model/galactica/language_model.py (76:95) - megatron_patch/model/starcoder/language_model.py (64:83) duplicated block id: 2602 size: 16 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (38:73) - megatron_patch/model/falcon40b/transformer.py (33:69) duplicated block id: 2603 size: 16 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_dense_and_moe_gqa.py (180:195) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_vl.py (189:204) duplicated block id: 2604 size: 16 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/language_model.py (77:96) - megatron_patch/model/falcon/language_model.py (76:95) duplicated block id: 2605 size: 16 cleaned lines of code in 2 files: - megatron_patch/model/llava/transformer.py (438:468) - megatron_patch/model/starcoder/transformer.py (463:493) duplicated block id: 2606 size: 16 cleaned lines of code in 2 files: - megatron_patch/model/falcon/transformer.py (265:303) - megatron_patch/model/galactica/transformer.py (208:246) duplicated block id: 2607 size: 16 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (38:73) - megatron_patch/model/galactica/transformer.py (35:70) duplicated block id: 2608 size: 16 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (669:686) - megatron_patch/model/starcoder/transformer.py (882:899) duplicated block id: 2609 size: 16 cleaned lines of code in 2 files: - megatron_patch/model/falcon/language_model.py (680:696) - megatron_patch/model/galactica/language_model.py (709:725) duplicated block id: 2610 size: 16 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/transformer.py (33:69) - megatron_patch/model/llama/transformer.py (38:73) duplicated block id: 2611 size: 16 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/language_model.py (597:619) - megatron_patch/model/glm130b/language_model.py (508:530) duplicated block id: 2612 size: 16 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (589:611) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (673:695) duplicated block id: 2613 size: 16 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (38:73) - megatron_patch/model/llama/transformer.py (38:73) duplicated block id: 2614 size: 16 cleaned lines of code in 2 files: - megatron_patch/model/bloom/language_model.py (78:97) - megatron_patch/model/starcoder/language_model.py (64:83) duplicated block id: 2615 size: 16 cleaned lines of code in 2 files: - megatron_patch/model/qwen1_5_megablocks/transformer.py (1650:1670) - megatron_patch/model/qwen_vl/transformer.py (1786:1806) duplicated block id: 2616 size: 16 cleaned lines of code in 2 files: - megatron_patch/model/falcon/language_model.py (76:95) - megatron_patch/model/starcoder/language_model.py (64:83) duplicated block id: 2617 size: 16 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_dense_and_moe_gqa.py (135:150) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_vl.py (189:204) duplicated block id: 2618 size: 16 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/hf2te.py (279:301) - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (506:528) duplicated block id: 2619 size: 16 cleaned lines of code in 2 files: - megatron_patch/model/falcon/transformer.py (37:73) - megatron_patch/model/llama/transformer.py (38:73) duplicated block id: 2620 size: 16 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_dense_and_moe_gqa.py (362:378) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_moe.py (390:406) duplicated block id: 2621 size: 16 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (182:208) - toolkits/model_checkpoints_convertor/falcon/checkpoint_reshaping_and_interoperability.py (188:215) duplicated block id: 2622 size: 16 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/hf2te.py (307:329) - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (506:528) duplicated block id: 2623 size: 16 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/checkpoint_reshaping_and_interoperability.py (745:767) - toolkits/model_checkpoints_convertor/glm130b/checkpoint_reshaping_and_interoperability.py (451:474) duplicated block id: 2624 size: 15 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/moe/moe_layer.py (32:53) - megatron_patch/model/qwen2/moe/moe_layer.py (41:62) duplicated block id: 2625 size: 15 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon40b/checkpoint_reshaping_and_interoperability.py (189:213) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (238:262) duplicated block id: 2626 size: 15 cleaned lines of code in 2 files: - megatron_patch/model/llama3/model.py (147:172) - megatron_patch/model/mixtral/model.py (127:152) duplicated block id: 2627 size: 15 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_moe.py (156:170) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2.5_vl.py (188:202) duplicated block id: 2628 size: 15 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon/checkpoint_reshaping_and_interoperability.py (35:49) - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (55:70) duplicated block id: 2629 size: 15 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (607:634) - megatron_patch/model/galactica/transformer.py (536:563) duplicated block id: 2630 size: 15 cleaned lines of code in 2 files: - megatron_patch/model/llama/transformer.py (1057:1092) - megatron_patch/model/starcoder/transformer.py (1306:1341) duplicated block id: 2631 size: 15 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer/attention.py (412:427) - megatron_patch/model/qwen2_vl/attention_vision.py (556:571) duplicated block id: 2632 size: 15 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/hf2te.py (131:161) - toolkits/model_checkpoints_convertor/yi/checkpoint_reshaping_and_interoperability.py (231:261) duplicated block id: 2633 size: 15 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (637:656) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (643:662) duplicated block id: 2634 size: 15 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon40b/checkpoint_reshaping_and_interoperability.py (215:245) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (205:235) duplicated block id: 2635 size: 15 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (238:262) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (241:265) duplicated block id: 2636 size: 15 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (321:345) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (241:265) duplicated block id: 2637 size: 15 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_mha.py (119:133) - toolkits/model_checkpoints_convertor/qwen/hf2megablocks_qwen1.5.py (321:335) duplicated block id: 2638 size: 15 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (510:528) - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (583:601) duplicated block id: 2639 size: 15 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/chatglm/checkpoint_reshaping_and_interoperability.py (233:257) - toolkits/model_checkpoints_convertor/glm130b/checkpoint_reshaping_and_interoperability.py (203:227) duplicated block id: 2640 size: 15 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_vl.py (188:202) - toolkits/model_checkpoints_convertor/qwen/hf2megablocks_qwen1.5.py (321:335) duplicated block id: 2641 size: 15 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/language_model.py (611:630) - megatron_patch/model/starcoder/language_model.py (551:569) duplicated block id: 2642 size: 15 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/hf2te.py (131:161) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (278:308) duplicated block id: 2643 size: 15 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon/checkpoint_reshaping_and_interoperability.py (191:215) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (218:242) duplicated block id: 2644 size: 15 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer_legacy.py (1519:1533) - megatron_patch/model/llava/transformer.py (1556:1570) duplicated block id: 2645 size: 15 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (246:276) - toolkits/model_checkpoints_convertor/qwen/hf2megablocks_qwen1.5.py (389:419) duplicated block id: 2646 size: 15 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (184:208) - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (321:345) duplicated block id: 2647 size: 15 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (186:210) - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (259:283) duplicated block id: 2648 size: 15 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/mistral/hf2mcore.py (271:285) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_moe.py (157:171) duplicated block id: 2649 size: 15 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_moe.py (156:170) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_vl.py (139:153) duplicated block id: 2650 size: 15 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon/checkpoint_reshaping_and_interoperability.py (191:215) - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (321:345) duplicated block id: 2651 size: 15 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/language_model.py (281:302) - megatron_patch/model/llama3/language_model.py (215:236) duplicated block id: 2652 size: 15 cleaned lines of code in 2 files: - megatron_patch/model/falcon/transformer.py (878:901) - megatron_patch/model/glm130b/transformer.py (711:734) duplicated block id: 2653 size: 15 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llava/hf2mcore_llava.py (525:544) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_dense_and_moe_gqa.py (421:441) duplicated block id: 2654 size: 15 cleaned lines of code in 2 files: - megatron_patch/model/qwen2/transformer/attention.py (506:532) - megatron_patch/model/qwen2_vl/attention_vision.py (649:675) duplicated block id: 2655 size: 15 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/transformer/attention.py (556:571) - megatron_patch/model/qwen1_5/transformer/attention.py (412:427) duplicated block id: 2656 size: 15 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon40b/checkpoint_reshaping_and_interoperability.py (518:537) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (595:614) duplicated block id: 2657 size: 15 cleaned lines of code in 2 files: - megatron_patch/model/mixtral_bak/transformer/attention.py (99:115) - megatron_patch/model/qwen2_vl/attention_vision.py (140:156) duplicated block id: 2658 size: 15 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/hf2te.py (132:162) - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (226:256) duplicated block id: 2659 size: 15 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/model.py (145:160) - megatron_patch/model/llama3_1/model.py (188:203) duplicated block id: 2660 size: 15 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer_legacy.py (1519:1533) - megatron_patch/model/qwen1_5_megablocks/transformer.py (1420:1434) duplicated block id: 2661 size: 15 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon40b/checkpoint_reshaping_and_interoperability.py (189:213) - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (321:345) duplicated block id: 2662 size: 15 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/chatglm/checkpoint_reshaping_and_interoperability.py (216:246) - toolkits/model_checkpoints_convertor/falcon/checkpoint_reshaping_and_interoperability.py (217:247) duplicated block id: 2663 size: 15 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/hf2te.py (131:161) - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (282:312) duplicated block id: 2664 size: 15 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (1692:1714) - megatron_patch/model/qwen_vl/transformer.py (1824:1846) duplicated block id: 2665 size: 15 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_mha.py (119:133) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_moe.py (156:170) duplicated block id: 2666 size: 15 cleaned lines of code in 2 files: - megatron_patch/model/llama3_1/model.py (221:246) - megatron_patch/model/qwen1_5/model.py (161:186) duplicated block id: 2667 size: 15 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (222:246) - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (321:345) duplicated block id: 2668 size: 15 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (184:208) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (238:262) duplicated block id: 2669 size: 15 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (645:660) - megatron_patch/model/starcoder/transformer.py (758:773) duplicated block id: 2670 size: 15 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (188:205) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (189:206) duplicated block id: 2671 size: 15 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/model.py (127:152) - megatron_patch/model/qwen2/model.py (148:173) duplicated block id: 2672 size: 15 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/language_model.py (647:665) - megatron_patch/model/starcoder/language_model.py (551:569) duplicated block id: 2673 size: 15 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/model.py (167:192) - megatron_patch/model/qwen1_5/model.py (147:172) duplicated block id: 2674 size: 15 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore.py (288:302) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_moe.py (156:170) duplicated block id: 2675 size: 15 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/chatglm/checkpoint_reshaping_and_interoperability.py (216:246) - toolkits/model_checkpoints_convertor/qwen/hf2megablocks_qwen1.5.py (389:419) duplicated block id: 2676 size: 15 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/transformer/mlp.py (25:65) - megatron_patch/model/qwen1_5/transformer/mlp.py (24:63) duplicated block id: 2677 size: 15 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (288:318) - toolkits/model_checkpoints_convertor/qwen/hf2megablocks_qwen1.5.py (389:419) duplicated block id: 2678 size: 15 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/hf2te.py (131:161) - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (288:318) duplicated block id: 2679 size: 15 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/reward_model_to_megatron.py (618:638) - toolkits/model_checkpoints_convertor/glm130b/checkpoint_reshaping_and_interoperability.py (348:368) duplicated block id: 2680 size: 15 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (259:283) - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (222:246) duplicated block id: 2681 size: 15 cleaned lines of code in 2 files: - megatron_patch/model/qwen2/transformer/attention.py (506:532) - megatron_patch/model/qwen2_vl/attention.py (650:676) duplicated block id: 2682 size: 15 cleaned lines of code in 2 files: - megatron_patch/model/llava/transformer.py (1750:1786) - megatron_patch/model/starcoder/transformer.py (1305:1339) duplicated block id: 2683 size: 15 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (154:172) - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (186:204) duplicated block id: 2684 size: 15 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer/attention.py (412:427) - megatron_patch/model/mixtral/transformer/attention.py (556:571) duplicated block id: 2685 size: 15 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon40b/checkpoint_reshaping_and_interoperability.py (35:49) - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (55:70) duplicated block id: 2686 size: 15 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon40b/checkpoint_reshaping_and_interoperability.py (189:213) - toolkits/model_checkpoints_convertor/galactica/checkpoint_reshaping_and_interoperability.py (173:197) duplicated block id: 2687 size: 15 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/galactica/checkpoint_reshaping_and_interoperability.py (173:197) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (238:262) duplicated block id: 2688 size: 15 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (1692:1714) - megatron_patch/model/llama2/transformer.py (1825:1847) duplicated block id: 2689 size: 15 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer_legacy.py (1519:1533) - megatron_patch/model/qwen_vl/transformer.py (1556:1570) duplicated block id: 2690 size: 15 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/hf2te.py (131:161) - toolkits/model_checkpoints_convertor/falcon40b/checkpoint_reshaping_and_interoperability.py (215:245) duplicated block id: 2691 size: 15 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/hf2te.py (131:161) - toolkits/model_checkpoints_convertor/qwen/hf2megablocks_qwen1.5.py (389:419) duplicated block id: 2692 size: 15 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore.py (288:302) - toolkits/model_checkpoints_convertor/llama/hf2mcore.py (328:342) duplicated block id: 2693 size: 15 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/moe/moe_layer.py (31:52) - megatron_patch/model/qwen2/moe/moe_layer.py (41:62) duplicated block id: 2694 size: 15 cleaned lines of code in 2 files: - megatron_patch/model/bloom/language_model.py (581:599) - megatron_patch/model/starcoder/language_model.py (551:569) duplicated block id: 2695 size: 15 cleaned lines of code in 2 files: - megatron_patch/model/qwen2/transformer_block.py (239:255) - megatron_patch/model/qwen2_5_vl/transformer_block.py (334:351) duplicated block id: 2696 size: 15 cleaned lines of code in 2 files: - megatron_patch/model/qwen/transformer.py (738:753) - megatron_patch/model/starcoder/transformer.py (758:773) duplicated block id: 2697 size: 15 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/moe/router.py (233:270) - megatron_patch/model/mixtral_bak/moe/router.py (202:237) duplicated block id: 2698 size: 15 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/moe/router.py (111:136) - megatron_patch/model/mixtral_bak/moe/router.py (119:143) duplicated block id: 2699 size: 15 cleaned lines of code in 2 files: - megatron_patch/model/llama3_1/model.py (221:246) - megatron_patch/model/qwen2/model.py (162:187) duplicated block id: 2700 size: 15 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2.5_vl.py (139:153) - toolkits/model_checkpoints_convertor/qwen/hf2megablocks_qwen1.5.py (321:335) duplicated block id: 2701 size: 15 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/hf2te.py (131:161) - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (226:256) duplicated block id: 2702 size: 15 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/chatglm/checkpoint_reshaping_and_interoperability.py (233:257) - toolkits/model_checkpoints_convertor/glm/checkpoint_reshaping_and_interoperability.py (203:227) duplicated block id: 2703 size: 15 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore.py (328:342) - toolkits/model_checkpoints_convertor/qwen/hf2megablocks_qwen1.5.py (282:296) duplicated block id: 2704 size: 15 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (756:771) - megatron_patch/model/starcoder/transformer.py (758:773) duplicated block id: 2705 size: 15 cleaned lines of code in 2 files: - megatron_patch/model/galactica/transformer.py (60:84) - megatron_patch/model/glm130b/transformer.py (90:114) duplicated block id: 2706 size: 15 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/transformer/attention.py (153:169) - megatron_patch/model/qwen2_vl/attention_vision.py (140:156) duplicated block id: 2707 size: 15 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (259:283) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (241:265) duplicated block id: 2708 size: 15 cleaned lines of code in 2 files: - megatron_patch/model/qwen2/transformer/attention.py (121:137) - megatron_patch/model/qwen2_vl/attention.py (141:157) duplicated block id: 2709 size: 15 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon/checkpoint_reshaping_and_interoperability.py (217:247) - toolkits/model_checkpoints_convertor/qwen/hf2megablocks_qwen1.5.py (389:419) duplicated block id: 2710 size: 15 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (55:70) - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (58:72) duplicated block id: 2711 size: 15 cleaned lines of code in 2 files: - megatron_patch/model/qwen_vl/transformer.py (1750:1786) - megatron_patch/model/starcoder/transformer.py (1305:1339) duplicated block id: 2712 size: 15 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/model.py (127:152) - megatron_patch/model/qwen1_5/model.py (147:172) duplicated block id: 2713 size: 15 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/hf2te.py (132:162) - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (246:276) duplicated block id: 2714 size: 15 cleaned lines of code in 2 files: - megatron_patch/model/falcon/transformer.py (1263:1298) - megatron_patch/model/starcoder/transformer.py (1306:1341) duplicated block id: 2715 size: 15 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer_legacy.py (1519:1533) - megatron_patch/model/mistral/transformer.py (1554:1568) duplicated block id: 2716 size: 15 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_moe.py (156:170) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_vl.py (188:202) duplicated block id: 2717 size: 15 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (186:210) - toolkits/model_checkpoints_convertor/falcon/checkpoint_reshaping_and_interoperability.py (191:215) duplicated block id: 2718 size: 15 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/transformer.py (583:600) - megatron_patch/model/galactica/transformer.py (607:624) duplicated block id: 2719 size: 15 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2megablocks_qwen1.5.py (389:419) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (278:308) duplicated block id: 2720 size: 15 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon/checkpoint_reshaping_and_interoperability.py (217:247) - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (288:318) duplicated block id: 2721 size: 15 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer/attention.py (412:427) - megatron_patch/model/qwen2_vl/attention.py (557:572) duplicated block id: 2722 size: 15 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/model.py (127:152) - megatron_patch/model/qwen2_vl/gpt_model.py (138:163) duplicated block id: 2723 size: 15 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/chatglm/checkpoint_reshaping_and_interoperability.py (30:44) - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (55:70) duplicated block id: 2724 size: 15 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (321:345) - toolkits/model_checkpoints_convertor/yi/checkpoint_reshaping_and_interoperability.py (171:195) duplicated block id: 2725 size: 15 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/chatglm/checkpoint_reshaping_and_interoperability.py (233:257) - toolkits/model_checkpoints_convertor/starcoder/checkpoint_reshaping_and_interoperability.py (254:278) duplicated block id: 2726 size: 15 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon40b/checkpoint_reshaping_and_interoperability.py (189:213) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (218:242) duplicated block id: 2727 size: 15 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore.py (288:302) - toolkits/model_checkpoints_convertor/qwen/hf2megablocks_qwen1.5.py (321:335) duplicated block id: 2728 size: 15 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon/checkpoint_reshaping_and_interoperability.py (191:215) - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (259:283) duplicated block id: 2729 size: 15 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (1692:1714) - megatron_patch/model/mistral/transformer.py (1822:1844) duplicated block id: 2730 size: 15 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/hf2te.py (132:162) - toolkits/model_checkpoints_convertor/chatglm/checkpoint_reshaping_and_interoperability.py (216:246) duplicated block id: 2731 size: 15 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/glm130b/checkpoint_reshaping_and_interoperability.py (348:368) - toolkits/model_checkpoints_convertor/starcoder/checkpoint_reshaping_and_interoperability.py (632:652) duplicated block id: 2732 size: 15 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2megablocks_qwen1.5.py (282:296) - toolkits/model_checkpoints_convertor/qwen/hf2megablocks_qwen1.5.py (321:335) duplicated block id: 2733 size: 15 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/model.py (167:192) - megatron_patch/model/qwen2/model.py (148:173) duplicated block id: 2734 size: 15 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore.py (328:342) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_vl.py (188:202) duplicated block id: 2735 size: 15 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/hf2te.py (148:172) - toolkits/model_checkpoints_convertor/starcoder/checkpoint_reshaping_and_interoperability.py (254:278) duplicated block id: 2736 size: 15 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/galactica/checkpoint_reshaping_and_interoperability.py (233:263) - toolkits/model_checkpoints_convertor/qwen/hf2megablocks_qwen1.5.py (389:419) duplicated block id: 2737 size: 15 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/mlp.py (13:54) - megatron_patch/model/qwen2/transformer/mlp.py (23:62) duplicated block id: 2738 size: 15 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (222:246) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (238:262) duplicated block id: 2739 size: 15 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon/configuration_RW.py (49:65) - toolkits/model_checkpoints_convertor/falcon40b/configuration_RW.py (47:63) duplicated block id: 2740 size: 15 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/chatglm/checkpoint_reshaping_and_interoperability.py (156:180) - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (321:345) duplicated block id: 2741 size: 15 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/hf2te.py (379:400) - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (491:512) duplicated block id: 2742 size: 15 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (246:276) - toolkits/model_checkpoints_convertor/falcon40b/checkpoint_reshaping_and_interoperability.py (215:245) duplicated block id: 2743 size: 15 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (238:262) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (218:242) duplicated block id: 2744 size: 15 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_mha.py (268:282) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_dense_and_moe_gqa.py (574:588) duplicated block id: 2745 size: 15 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (282:312) - toolkits/model_checkpoints_convertor/qwen/hf2megablocks_qwen1.5.py (389:419) duplicated block id: 2746 size: 15 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/hf2te.py (131:161) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (301:331) duplicated block id: 2747 size: 15 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (205:235) - toolkits/model_checkpoints_convertor/qwen/hf2megablocks_qwen1.5.py (389:419) duplicated block id: 2748 size: 15 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon40b/checkpoint_reshaping_and_interoperability.py (189:213) - toolkits/model_checkpoints_convertor/yi/checkpoint_reshaping_and_interoperability.py (171:195) duplicated block id: 2749 size: 15 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (55:70) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (62:76) duplicated block id: 2750 size: 15 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_moe.py (156:170) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2.5_vl.py (139:153) duplicated block id: 2751 size: 15 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/reward_model_to_megatron.py (370:392) - toolkits/model_checkpoints_convertor/starcoder/checkpoint_reshaping_and_interoperability.py (393:415) duplicated block id: 2752 size: 15 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/galactica/checkpoint_reshaping_and_interoperability.py (173:197) - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (321:345) duplicated block id: 2753 size: 15 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/checkpoint_reshaping_and_interoperability.py (371:393) - toolkits/model_checkpoints_convertor/starcoder/checkpoint_reshaping_and_interoperability.py (393:415) duplicated block id: 2754 size: 15 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer_legacy.py (1710:1746) - megatron_patch/model/starcoder/transformer.py (1305:1339) duplicated block id: 2755 size: 15 cleaned lines of code in 2 files: - megatron_patch/model/qwen1_5/transformer/attention.py (412:427) - megatron_patch/model/qwen2_vl/attention_vision.py (556:571) duplicated block id: 2756 size: 15 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/hf2te.py (132:162) - toolkits/model_checkpoints_convertor/qwen/hf2megablocks_qwen1.5.py (389:419) duplicated block id: 2757 size: 15 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon/configuration_RW.py (49:65) - toolkits/model_checkpoints_convertor/falcon/configuration_RW.py (106:122) duplicated block id: 2758 size: 15 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/model.py (167:192) - megatron_patch/model/mixtral_bak/model.py (123:148) duplicated block id: 2759 size: 15 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon/checkpoint_reshaping_and_interoperability.py (191:215) - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (222:246) duplicated block id: 2760 size: 15 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon40b/checkpoint_reshaping_and_interoperability.py (215:245) - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (288:318) duplicated block id: 2761 size: 15 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/galactica/checkpoint_reshaping_and_interoperability.py (173:197) - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (259:283) duplicated block id: 2762 size: 15 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/hf2te.py (148:172) - toolkits/model_checkpoints_convertor/bloom/reward_model_to_megatron.py (232:256) duplicated block id: 2763 size: 15 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (1557:1571) - megatron_patch/model/llama3/transformer_legacy.py (1519:1533) duplicated block id: 2764 size: 15 cleaned lines of code in 2 files: - megatron_patch/model/llama3/model.py (34:66) - megatron_patch/model/mixtral/model.py (31:63) duplicated block id: 2765 size: 15 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon/checkpoint_reshaping_and_interoperability.py (191:215) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (241:265) duplicated block id: 2766 size: 15 cleaned lines of code in 2 files: - megatron_patch/model/llama3_1/model.py (221:246) - megatron_patch/model/mixtral_bak/model.py (137:176) duplicated block id: 2767 size: 15 cleaned lines of code in 2 files: - megatron_patch/model/qwen1_5/transformer/attention.py (412:427) - megatron_patch/model/qwen2_vl/attention.py (557:572) duplicated block id: 2768 size: 15 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/transformer/mlp.py (25:65) - megatron_patch/model/qwen2/transformer/mlp.py (23:62) duplicated block id: 2769 size: 15 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon/checkpoint_reshaping_and_interoperability.py (217:247) - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (226:256) duplicated block id: 2770 size: 15 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/checkpoint_reshaping_and_interoperability.py (619:639) - toolkits/model_checkpoints_convertor/glm130b/checkpoint_reshaping_and_interoperability.py (348:368) duplicated block id: 2771 size: 15 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (246:276) - toolkits/model_checkpoints_convertor/baichuan2/hf2te.py (131:161) duplicated block id: 2772 size: 15 cleaned lines of code in 2 files: - megatron_patch/model/llama3_1/model.py (221:246) - megatron_patch/model/qwen2_vl/gpt_model.py (152:177) duplicated block id: 2773 size: 15 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/mlp.py (13:54) - megatron_patch/model/qwen1_5/transformer/mlp.py (24:63) duplicated block id: 2774 size: 15 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_moe.py (338:352) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_dense_and_moe_gqa.py (590:604) duplicated block id: 2775 size: 15 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2.5_vl.py (188:202) - toolkits/model_checkpoints_convertor/qwen/hf2megablocks_qwen1.5.py (321:335) duplicated block id: 2776 size: 15 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (244:274) - toolkits/model_checkpoints_convertor/baichuan2/hf2te.py (131:161) duplicated block id: 2777 size: 15 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/model.py (31:63) - megatron_patch/model/qwen1_5/model.py (34:66) duplicated block id: 2778 size: 15 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (226:256) - toolkits/model_checkpoints_convertor/qwen/hf2megablocks_qwen1.5.py (389:419) duplicated block id: 2779 size: 15 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/model.py (167:192) - megatron_patch/model/qwen2_vl/gpt_model.py (138:163) duplicated block id: 2780 size: 15 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/model.py (31:63) - megatron_patch/model/mixtral_bak/model.py (31:63) duplicated block id: 2781 size: 15 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/chatglm/checkpoint_reshaping_and_interoperability.py (156:180) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (238:262) duplicated block id: 2782 size: 15 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore_llama3_1.py (719:738) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_dense_and_moe_gqa.py (421:441) duplicated block id: 2783 size: 15 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore.py (328:342) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2.5_vl.py (139:153) duplicated block id: 2784 size: 15 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon/checkpoint_reshaping_and_interoperability.py (217:247) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (205:235) duplicated block id: 2785 size: 15 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (321:345) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (218:242) duplicated block id: 2786 size: 15 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon/checkpoint_reshaping_and_interoperability.py (191:215) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (238:262) duplicated block id: 2787 size: 15 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (589:610) - megatron_patch/model/glm130b/transformer.py (671:693) duplicated block id: 2788 size: 15 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/hf2te.py (148:172) - toolkits/model_checkpoints_convertor/bloom/checkpoint_reshaping_and_interoperability.py (232:256) duplicated block id: 2789 size: 15 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (560:575) - megatron_patch/model/glm130b/transformer.py (612:627) duplicated block id: 2790 size: 15 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon40b/checkpoint_reshaping_and_interoperability.py (189:213) - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (259:283) duplicated block id: 2791 size: 15 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (246:276) - toolkits/model_checkpoints_convertor/falcon/checkpoint_reshaping_and_interoperability.py (217:247) duplicated block id: 2792 size: 15 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/transformer/attention.py (649:675) - megatron_patch/model/qwen2/transformer/attention.py (506:532) duplicated block id: 2793 size: 15 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (1620:1656) - megatron_patch/model/starcoder/transformer.py (1305:1339) duplicated block id: 2794 size: 15 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/chatglm/checkpoint_reshaping_and_interoperability.py (216:246) - toolkits/model_checkpoints_convertor/falcon40b/checkpoint_reshaping_and_interoperability.py (215:245) duplicated block id: 2795 size: 15 cleaned lines of code in 2 files: - megatron_patch/model/llama2/transformer.py (1751:1787) - megatron_patch/model/starcoder/transformer.py (1305:1339) duplicated block id: 2796 size: 15 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon/checkpoint_reshaping_and_interoperability.py (191:215) - toolkits/model_checkpoints_convertor/yi/checkpoint_reshaping_and_interoperability.py (171:195) duplicated block id: 2797 size: 15 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (186:210) - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (321:345) duplicated block id: 2798 size: 15 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (244:274) - toolkits/model_checkpoints_convertor/qwen/hf2megablocks_qwen1.5.py (389:419) duplicated block id: 2799 size: 15 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (186:210) - toolkits/model_checkpoints_convertor/falcon40b/checkpoint_reshaping_and_interoperability.py (189:213) duplicated block id: 2800 size: 15 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (55:70) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (63:77) duplicated block id: 2801 size: 15 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/hf2te.py (131:161) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (205:235) duplicated block id: 2802 size: 15 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_moe.py (156:170) - toolkits/model_checkpoints_convertor/qwen/hf2megablocks_qwen1.5.py (282:296) duplicated block id: 2803 size: 15 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (55:70) - toolkits/model_checkpoints_convertor/yi/checkpoint_reshaping_and_interoperability.py (35:49) duplicated block id: 2804 size: 15 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (259:283) - toolkits/model_checkpoints_convertor/yi/checkpoint_reshaping_and_interoperability.py (171:195) duplicated block id: 2805 size: 15 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/transformer/attention.py (153:169) - megatron_patch/model/qwen2_vl/attention.py (141:157) duplicated block id: 2806 size: 15 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer/attention.py (125:141) - megatron_patch/model/qwen2_vl/attention.py (141:157) duplicated block id: 2807 size: 15 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/hf2te.py (132:162) - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (288:318) duplicated block id: 2808 size: 15 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (1750:1786) - megatron_patch/model/starcoder/transformer.py (1305:1339) duplicated block id: 2809 size: 15 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/model.py (167:192) - megatron_patch/model/llama3/model.py (147:172) duplicated block id: 2810 size: 15 cleaned lines of code in 2 files: - toolkits/pretrain_data_preprocessing/preprocess_data_megatron.py (165:181) - toolkits/sft_data_preprocessing/build_idxmap_sft_dataset.py (161:176) duplicated block id: 2811 size: 15 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (1692:1714) - megatron_patch/model/llava/transformer.py (1824:1846) duplicated block id: 2812 size: 15 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/checkpoint_reshaping_and_interoperability.py (318:334) - toolkits/model_checkpoints_convertor/bloom/reward_model_to_megatron.py (318:334) duplicated block id: 2813 size: 15 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2megablocks_qwen1.5.py (389:419) - toolkits/model_checkpoints_convertor/yi/checkpoint_reshaping_and_interoperability.py (231:261) duplicated block id: 2814 size: 15 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore.py (328:342) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_mha.py (119:133) duplicated block id: 2815 size: 15 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/galactica/checkpoint_reshaping_and_interoperability.py (34:48) - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (55:70) duplicated block id: 2816 size: 15 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (820:839) - megatron_patch/model/starcoder/transformer.py (966:985) duplicated block id: 2817 size: 15 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/chatglm/checkpoint_reshaping_and_interoperability.py (156:180) - toolkits/model_checkpoints_convertor/falcon/checkpoint_reshaping_and_interoperability.py (191:215) duplicated block id: 2818 size: 15 cleaned lines of code in 2 files: - megatron_patch/model/mixtral_bak/moe/experts.py (171:208) - megatron_patch/model/qwen1_5/moe/experts.py (179:200) duplicated block id: 2819 size: 15 cleaned lines of code in 2 files: - megatron_patch/model/qwen1_5/transformer/attention.py (125:141) - megatron_patch/model/qwen2_vl/attention_vision.py (140:156) duplicated block id: 2820 size: 15 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer_legacy.py (1519:1533) - megatron_patch/model/qwen/transformer.py (1499:1513) duplicated block id: 2821 size: 15 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_mha_to_moe.py (189:203) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_dense_and_moe_gqa.py (590:604) duplicated block id: 2822 size: 15 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (184:208) - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (259:283) duplicated block id: 2823 size: 15 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer/mlp.py (24:63) - megatron_patch/model/qwen2/transformer/mlp.py (23:62) duplicated block id: 2824 size: 15 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/transformer.py (583:600) - megatron_patch/model/falcon40b/transformer.py (721:740) duplicated block id: 2825 size: 15 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/hf2te.py (351:372) - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (491:512) duplicated block id: 2826 size: 15 cleaned lines of code in 2 files: - megatron_patch/model/mixtral_bak/transformer/attention.py (99:115) - megatron_patch/model/qwen2_vl/attention.py (141:157) duplicated block id: 2827 size: 15 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore.py (328:342) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_vl.py (139:153) duplicated block id: 2828 size: 15 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/hf2te.py (148:172) - toolkits/model_checkpoints_convertor/glm/checkpoint_reshaping_and_interoperability.py (203:227) duplicated block id: 2829 size: 15 cleaned lines of code in 2 files: - megatron_patch/model/llama3_1/model.py (80:98) - megatron_patch/model/mixtral_bak/model.py (62:80) duplicated block id: 2830 size: 15 cleaned lines of code in 2 files: - megatron_patch/model/qwen1_5/moe/token_dispatcher.py (435:472) - megatron_patch/model/qwen2/moe/token_dispatcher.py (472:508) duplicated block id: 2831 size: 15 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (259:283) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (218:242) duplicated block id: 2832 size: 15 cleaned lines of code in 2 files: - megatron_patch/model/mixtral_bak/model.py (62:80) - megatron_patch/model/qwen2/model.py (66:84) duplicated block id: 2833 size: 15 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore.py (328:342) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2.5_vl.py (188:202) duplicated block id: 2834 size: 15 cleaned lines of code in 2 files: - megatron_patch/model/bloom/language_model.py (480:501) - megatron_patch/model/starcoder/language_model.py (451:472) duplicated block id: 2835 size: 15 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (34:48) - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (55:70) duplicated block id: 2836 size: 15 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (337:358) - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (442:464) duplicated block id: 2837 size: 15 cleaned lines of code in 2 files: - megatron_patch/model/mistral/transformer.py (1748:1784) - megatron_patch/model/starcoder/transformer.py (1305:1339) duplicated block id: 2838 size: 15 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/checkpoint_reshaping_and_interoperability.py (232:256) - toolkits/model_checkpoints_convertor/chatglm/checkpoint_reshaping_and_interoperability.py (233:257) duplicated block id: 2839 size: 15 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer/attention.py (125:141) - megatron_patch/model/qwen2_vl/attention_vision.py (140:156) duplicated block id: 2840 size: 15 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/deepseek/hf2mcore_deepseek_v2_moe.py (426:442) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_dense_and_moe_gqa.py (711:725) duplicated block id: 2841 size: 15 cleaned lines of code in 2 files: - megatron_patch/model/qwen2/transformer/attention.py (121:137) - megatron_patch/model/qwen2_vl/attention_vision.py (140:156) duplicated block id: 2842 size: 15 cleaned lines of code in 2 files: - megatron_patch/model/mixtral_bak/moe/router.py (202:237) - megatron_patch/model/qwen1_5/moe/router.py (215:250) duplicated block id: 2843 size: 15 cleaned lines of code in 2 files: - megatron_patch/model/falcon/transformer.py (829:846) - megatron_patch/model/starcoder/transformer.py (779:796) duplicated block id: 2844 size: 15 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (1165:1200) - megatron_patch/model/starcoder/transformer.py (1306:1341) duplicated block id: 2845 size: 15 cleaned lines of code in 2 files: - megatron_patch/model/llama2/transformer.py (1557:1571) - megatron_patch/model/llama3/transformer_legacy.py (1519:1533) duplicated block id: 2846 size: 15 cleaned lines of code in 2 files: - megatron_patch/model/qwen/transformer.py (1692:1728) - megatron_patch/model/starcoder/transformer.py (1305:1339) duplicated block id: 2847 size: 15 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon/checkpoint_reshaping_and_interoperability.py (191:215) - toolkits/model_checkpoints_convertor/galactica/checkpoint_reshaping_and_interoperability.py (173:197) duplicated block id: 2848 size: 15 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/chatglm/checkpoint_reshaping_and_interoperability.py (156:180) - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (259:283) duplicated block id: 2849 size: 15 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/chatglm/checkpoint_reshaping_and_interoperability.py (156:180) - toolkits/model_checkpoints_convertor/falcon40b/checkpoint_reshaping_and_interoperability.py (189:213) duplicated block id: 2850 size: 15 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon40b/checkpoint_reshaping_and_interoperability.py (189:213) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (241:265) duplicated block id: 2851 size: 15 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/reward_model_to_megatron.py (232:256) - toolkits/model_checkpoints_convertor/chatglm/checkpoint_reshaping_and_interoperability.py (233:257) duplicated block id: 2852 size: 15 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (442:464) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (324:345) duplicated block id: 2853 size: 15 cleaned lines of code in 2 files: - megatron_patch/model/qwen1_5/transformer/attention.py (125:141) - megatron_patch/model/qwen2_vl/attention.py (141:157) duplicated block id: 2854 size: 15 cleaned lines of code in 2 files: - megatron_patch/model/llama3/model.py (161:186) - megatron_patch/model/llama3_1/model.py (221:246) duplicated block id: 2855 size: 15 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/hf2te.py (131:161) - toolkits/model_checkpoints_convertor/falcon/checkpoint_reshaping_and_interoperability.py (217:247) duplicated block id: 2856 size: 15 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/hf2te.py (131:161) - toolkits/model_checkpoints_convertor/galactica/checkpoint_reshaping_and_interoperability.py (233:263) duplicated block id: 2857 size: 15 cleaned lines of code in 2 files: - toolkits/pretrain_data_preprocessing/clean_raw_text.py (44:61) - toolkits/pretrain_data_preprocessing/preprocess_wudao2.py (42:59) duplicated block id: 2858 size: 15 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_vl.py (139:153) - toolkits/model_checkpoints_convertor/qwen/hf2megablocks_qwen1.5.py (321:335) duplicated block id: 2859 size: 15 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon40b/checkpoint_reshaping_and_interoperability.py (189:213) - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (222:246) duplicated block id: 2860 size: 15 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2.5_vl.py (529:544) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_vl.py (535:550) duplicated block id: 2861 size: 15 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer/mlp.py (24:63) - megatron_patch/model/qwen1_5/transformer/mlp.py (24:63) duplicated block id: 2862 size: 15 cleaned lines of code in 2 files: - megatron_patch/model/llama3_1/model.py (46:78) - megatron_patch/model/qwen2/model.py (33:65) duplicated block id: 2863 size: 15 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer_legacy.py (743:758) - megatron_patch/model/starcoder/transformer.py (758:773) duplicated block id: 2864 size: 15 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/hf2te.py (132:162) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (205:235) duplicated block id: 2865 size: 15 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (238:262) - toolkits/model_checkpoints_convertor/yi/checkpoint_reshaping_and_interoperability.py (171:195) duplicated block id: 2866 size: 15 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (186:210) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (238:262) duplicated block id: 2867 size: 15 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon40b/checkpoint_reshaping_and_interoperability.py (215:245) - toolkits/model_checkpoints_convertor/qwen/hf2megablocks_qwen1.5.py (389:419) duplicated block id: 2868 size: 15 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon40b/checkpoint_reshaping_and_interoperability.py (215:245) - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (226:256) duplicated block id: 2869 size: 15 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/transformer.py (1195:1229) - megatron_patch/model/starcoder/transformer.py (1305:1339) duplicated block id: 2870 size: 15 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2megablocks_qwen1.5.py (389:419) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (301:331) duplicated block id: 2871 size: 15 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/hf2te.py (148:172) - toolkits/model_checkpoints_convertor/glm130b/checkpoint_reshaping_and_interoperability.py (203:227) duplicated block id: 2872 size: 15 cleaned lines of code in 2 files: - megatron_patch/model/mixtral_bak/moe/router.py (202:237) - megatron_patch/model/qwen2/moe/router.py (301:337) duplicated block id: 2873 size: 15 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/transformer.py (1002:1037) - megatron_patch/model/starcoder/transformer.py (1306:1341) duplicated block id: 2874 size: 15 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (589:610) - megatron_patch/model/llama/transformer.py (610:632) duplicated block id: 2875 size: 15 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (34:48) - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (55:70) duplicated block id: 2876 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/llama3_1/model.py (258:274) - megatron_patch/model/qwen1_5/model.py (196:212) duplicated block id: 2877 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/model.py (273:305) - megatron_patch/model/qwen1_5/model.py (220:251) duplicated block id: 2878 size: 14 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore.py (289:302) - toolkits/model_checkpoints_convertor/mistral/hf2mcore.py (271:284) duplicated block id: 2879 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (1008:1025) - megatron_patch/model/baichuan2/transformer.py (1140:1157) duplicated block id: 2880 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/galactica/language_model.py (332:346) - megatron_patch/model/llama2/language_model.py (301:315) duplicated block id: 2881 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (1165:1198) - megatron_patch/model/mistral/transformer.py (1749:1784) duplicated block id: 2882 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/qwen2_moe/layer_specs.py (124:152) - megatron_patch/model/qwen3_moe/gpt_layer_specs.py (74:102) duplicated block id: 2883 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/language_model.py (317:331) - megatron_patch/model/qwen1_5_megablocks/language_model.py (300:314) duplicated block id: 2884 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/transformer.py (1002:1035) - megatron_patch/model/qwen/transformer.py (1693:1728) duplicated block id: 2885 size: 14 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/mistral/hf2mcore.py (229:242) - toolkits/model_checkpoints_convertor/qwen/hf2megablocks_qwen1.5.py (322:335) duplicated block id: 2886 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/language_model.py (311:325) - megatron_patch/model/galactica/language_model.py (332:346) duplicated block id: 2887 size: 14 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/mistral/hf2mcore.py (229:242) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_moe.py (157:170) duplicated block id: 2888 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/mixtral_bak/layer_specs.py (92:105) - megatron_patch/model/qwen1_5/layer_specs.py (69:82) duplicated block id: 2889 size: 14 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (386:405) - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (441:461) duplicated block id: 2890 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/galactica/gpt_model.py (83:100) - megatron_patch/model/glm130b/gpt_model.py (77:95) duplicated block id: 2891 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/galactica/transformer.py (393:409) - megatron_patch/model/llama/transformer.py (430:446) duplicated block id: 2892 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/language_model.py (547:567) - megatron_patch/model/mistral/language_model.py (534:554) duplicated block id: 2893 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer_legacy.py (1767:1781) - megatron_patch/model/mistral/transformer.py (1806:1820) duplicated block id: 2894 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/language_model.py (301:315) - megatron_patch/model/qwen_vl/language_model.py (302:316) duplicated block id: 2895 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/language_model.py (532:547) - megatron_patch/model/qwen1_5_megablocks/language_model.py (463:478) duplicated block id: 2896 size: 14 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_moe.py (157:170) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_dense_and_moe_gqa.py (180:193) duplicated block id: 2897 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/mlp.py (66:80) - megatron_patch/model/qwen2/transformer/mlp.py (103:117) duplicated block id: 2898 size: 14 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (788:803) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (888:905) duplicated block id: 2899 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/galactica/language_model.py (332:346) - megatron_patch/model/starcoder/language_model.py (302:316) duplicated block id: 2900 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/language_model.py (568:588) - megatron_patch/model/llama2/language_model.py (522:542) duplicated block id: 2901 size: 14 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon/checkpoint_reshaping_and_interoperability.py (828:844) - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (811:827) duplicated block id: 2902 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/language_model.py (316:330) - megatron_patch/model/llama3/language_model.py (296:310) duplicated block id: 2903 size: 14 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_dense_and_moe_gqa.py (350:365) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_vl.py (220:234) duplicated block id: 2904 size: 14 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_mha.py (268:281) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2.5_vl.py (611:624) duplicated block id: 2905 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/llama/transformer.py (1057:1090) - megatron_patch/model/qwen_vl/transformer.py (1751:1786) duplicated block id: 2906 size: 14 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/checkpoint_reshaping_and_interoperability.py (714:735) - toolkits/model_checkpoints_convertor/glm130b/checkpoint_reshaping_and_interoperability.py (422:442) duplicated block id: 2907 size: 14 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_mha.py (268:281) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2.5_vl.py (668:681) duplicated block id: 2908 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/transformer/attention.py (577:591) - megatron_patch/model/qwen1_5/transformer/attention.py (432:446) duplicated block id: 2909 size: 14 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/chatglm/checkpoint_reshaping_and_interoperability.py (298:323) - toolkits/model_checkpoints_convertor/galactica/checkpoint_reshaping_and_interoperability.py (335:360) duplicated block id: 2910 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/falcon/transformer.py (1263:1296) - megatron_patch/model/qwen_vl/transformer.py (1751:1786) duplicated block id: 2911 size: 14 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (726:745) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (829:845) duplicated block id: 2912 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/transformer.py (227:247) - megatron_patch/model/glm130b/transformer.py (245:264) duplicated block id: 2913 size: 14 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/chatglm/checkpoint_reshaping_and_interoperability.py (298:323) - toolkits/model_checkpoints_convertor/falcon40b/checkpoint_reshaping_and_interoperability.py (306:331) duplicated block id: 2914 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/language_model.py (311:325) - megatron_patch/model/falcon40b/language_model.py (301:315) duplicated block id: 2915 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/language_model.py (532:547) - megatron_patch/model/baichuan2/language_model.py (471:486) duplicated block id: 2916 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/falcon/language_model.py (568:588) - megatron_patch/model/qwen1_5_megablocks/language_model.py (521:541) duplicated block id: 2917 size: 14 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_dense_and_moe_gqa.py (341:360) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_moe.py (368:388) duplicated block id: 2918 size: 14 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (346:371) - toolkits/model_checkpoints_convertor/chatglm/checkpoint_reshaping_and_interoperability.py (298:323) duplicated block id: 2919 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/llama2/transformer.py (1809:1823) - megatron_patch/model/qwen1_5_megablocks/transformer.py (1671:1685) duplicated block id: 2920 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/gpt_model.py (27:45) - megatron_patch/model/llama/gpt_model.py (28:46) duplicated block id: 2921 size: 14 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_dense_and_moe_gqa.py (135:148) - toolkits/model_checkpoints_convertor/qwen/hf2megablocks_qwen1.5.py (322:335) duplicated block id: 2922 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/llama3_1/model.py (258:274) - megatron_patch/model/mixtral_bak/model.py (186:202) duplicated block id: 2923 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/language_model.py (568:588) - megatron_patch/model/mistral/language_model.py (534:554) duplicated block id: 2924 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/language_model.py (316:330) - megatron_patch/model/mistral/language_model.py (302:316) duplicated block id: 2925 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/bloom/language_model.py (319:333) - megatron_patch/model/qwen1_5_megablocks/language_model.py (300:314) duplicated block id: 2926 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/transformer.py (583:597) - megatron_patch/model/starcoder/transformer.py (759:773) duplicated block id: 2927 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (1807:1821) - megatron_patch/model/qwen_vl/transformer.py (1808:1822) duplicated block id: 2928 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/language_model.py (599:619) - megatron_patch/model/chatglm/language_model.py (547:567) duplicated block id: 2929 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/moe/moe_layer.py (64:89) - megatron_patch/model/qwen2/moe/moe_layer.py (70:93) duplicated block id: 2930 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/mistral/transformer.py (778:792) - megatron_patch/model/starcoder/transformer.py (759:773) duplicated block id: 2931 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/language_model.py (311:325) - megatron_patch/model/chatglm/language_model.py (317:331) duplicated block id: 2932 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/llama2/transformer.py (1809:1823) - megatron_patch/model/llama3/transformer_legacy.py (1767:1781) duplicated block id: 2933 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (757:780) - megatron_patch/model/llama3/transformer_legacy.py (856:878) duplicated block id: 2934 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/qwen1_5_megablocks/transformer.py (1671:1685) - megatron_patch/model/qwen_vl/transformer.py (1808:1822) duplicated block id: 2935 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/bloom/language_model.py (319:333) - megatron_patch/model/qwen/language_model.py (301:315) duplicated block id: 2936 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/transformer/mlp.py (124:142) - megatron_patch/model/mixtral_bak/transformer/mlp.py (93:111) duplicated block id: 2937 size: 14 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_dense_and_moe_gqa.py (180:193) - toolkits/model_checkpoints_convertor/qwen/hf2megablocks_qwen1.5.py (322:335) duplicated block id: 2938 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (757:771) - megatron_patch/model/falcon40b/transformer.py (583:597) duplicated block id: 2939 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/galactica/language_model.py (597:617) - megatron_patch/model/qwen_vl/language_model.py (551:571) duplicated block id: 2940 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/llama/language_model.py (579:599) - megatron_patch/model/mistral/language_model.py (534:554) duplicated block id: 2941 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/language_model.py (301:315) - megatron_patch/model/llava/language_model.py (304:318) duplicated block id: 2942 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/language_model.py (510:530) - megatron_patch/model/qwen1_5_megablocks/language_model.py (521:541) duplicated block id: 2943 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/llama/language_model.py (301:315) - megatron_patch/model/llama2/language_model.py (301:315) duplicated block id: 2944 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/mistral/transformer.py (1806:1820) - megatron_patch/model/qwen1_5_megablocks/transformer.py (1671:1685) duplicated block id: 2945 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (1165:1198) - megatron_patch/model/qwen/transformer.py (1693:1728) duplicated block id: 2946 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/llama2/language_model.py (179:197) - megatron_patch/model/starcoder/language_model.py (178:196) duplicated block id: 2947 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (1620:1653) - megatron_patch/model/qwen1_5_megablocks/transformer.py (1613:1646) duplicated block id: 2948 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (1168:1201) - megatron_patch/model/galactica/transformer.py (889:922) duplicated block id: 2949 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (1008:1025) - megatron_patch/model/llava/transformer.py (1137:1154) duplicated block id: 2950 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/transformer.py (1005:1038) - megatron_patch/model/galactica/transformer.py (889:922) duplicated block id: 2951 size: 14 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/mistral/hf2mcore.py (271:284) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_mha.py (120:133) duplicated block id: 2952 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/llama/language_model.py (301:315) - megatron_patch/model/qwen_vl/language_model.py (302:316) duplicated block id: 2953 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/layer_specs.py (62:88) - megatron_patch/model/qwen2/layer_specs.py (49:62) duplicated block id: 2954 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/falcon/gpt_model.py (83:100) - megatron_patch/model/glm130b/gpt_model.py (77:95) duplicated block id: 2955 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/transformer_block.py (23:42) - megatron_patch/model/qwen2_5_vl/transformer_block.py (22:41) duplicated block id: 2956 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/llama/language_model.py (579:599) - megatron_patch/model/qwen/language_model.py (504:524) duplicated block id: 2957 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/language_model.py (189:207) - megatron_patch/model/starcoder/language_model.py (178:196) duplicated block id: 2958 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/language_model.py (317:331) - megatron_patch/model/starcoder/language_model.py (302:316) duplicated block id: 2959 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/qwen/transformer.py (1749:1763) - megatron_patch/model/qwen_vl/transformer.py (1808:1822) duplicated block id: 2960 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/galactica/language_model.py (597:617) - megatron_patch/model/llama2/language_model.py (522:542) duplicated block id: 2961 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/transformer.py (881:914) - megatron_patch/model/falcon/transformer.py (1266:1299) duplicated block id: 2962 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/language_model.py (532:547) - megatron_patch/model/qwen/language_model.py (461:476) duplicated block id: 2963 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/gpt_model.py (83:100) - megatron_patch/model/glm130b/gpt_model.py (77:95) duplicated block id: 2964 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/qwen1_5_megablocks/language_model.py (178:196) - megatron_patch/model/starcoder/language_model.py (178:196) duplicated block id: 2965 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/llama3/layer_specs.py (39:52) - megatron_patch/model/mixtral_bak/layer_specs.py (40:66) duplicated block id: 2966 size: 14 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore.py (329:342) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_dense_and_moe_gqa.py (180:193) duplicated block id: 2967 size: 14 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (703:721) - toolkits/model_checkpoints_convertor/yi/checkpoint_reshaping_and_interoperability.py (606:626) duplicated block id: 2968 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (1008:1025) - megatron_patch/model/mistral/transformer.py (1135:1152) duplicated block id: 2969 size: 14 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/mistral/hf2mcore.py (271:284) - toolkits/model_checkpoints_convertor/qwen/hf2megablocks_qwen1.5.py (283:296) duplicated block id: 2970 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/falcon/language_model.py (192:210) - megatron_patch/model/glm130b/language_model.py (192:210) duplicated block id: 2971 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/llama/language_model.py (301:315) - megatron_patch/model/llava/language_model.py (304:318) duplicated block id: 2972 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/mixtral_bak/layer_specs.py (40:66) - megatron_patch/model/qwen2/layer_specs.py (49:62) duplicated block id: 2973 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/mlp.py (82:96) - megatron_patch/model/qwen1_5/transformer/mlp.py (81:95) duplicated block id: 2974 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/llama/transformer.py (564:580) - megatron_patch/model/qwen_vl/transformer.py (773:789) duplicated block id: 2975 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/language_model.py (357:371) - megatron_patch/model/glm130b/language_model.py (316:330) duplicated block id: 2976 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/falcon/language_model.py (301:315) - megatron_patch/model/llava/language_model.py (304:318) duplicated block id: 2977 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/language_model.py (301:315) - megatron_patch/model/llama3/language_model.py (296:310) duplicated block id: 2978 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (1677:1691) - megatron_patch/model/mistral/transformer.py (1806:1820) duplicated block id: 2979 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/mixtral_bak/transformer/attention.py (24:45) - megatron_patch/model/qwen2_vl/attention_vision.py (58:83) duplicated block id: 2980 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer/attention.py (432:446) - megatron_patch/model/mixtral/transformer/attention.py (577:591) duplicated block id: 2981 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/falcon/language_model.py (301:315) - megatron_patch/model/qwen1_5_megablocks/language_model.py (300:314) duplicated block id: 2982 size: 14 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (929:948) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (888:905) duplicated block id: 2983 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/falcon/language_model.py (568:588) - megatron_patch/model/llava/language_model.py (582:602) duplicated block id: 2984 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/llama3_1/model.py (258:274) - megatron_patch/model/qwen2/model.py (197:213) duplicated block id: 2985 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/qwen/transformer.py (1692:1725) - megatron_patch/model/qwen1_5_megablocks/transformer.py (1613:1646) duplicated block id: 2986 size: 14 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore.py (748:763) - toolkits/model_checkpoints_convertor/mistral/hf2mcore.py (500:515) duplicated block id: 2987 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (1008:1025) - megatron_patch/model/qwen/transformer.py (1083:1100) duplicated block id: 2988 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/falcon/transformer.py (1263:1296) - megatron_patch/model/llama2/transformer.py (1752:1787) duplicated block id: 2989 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (1008:1025) - megatron_patch/model/qwen1_5_megablocks/transformer.py (1022:1039) duplicated block id: 2990 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/falcon/transformer.py (797:812) - megatron_patch/model/qwen/transformer.py (739:753) duplicated block id: 2991 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/language_model.py (192:210) - megatron_patch/model/llama/language_model.py (192:210) duplicated block id: 2992 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (1621:1656) - megatron_patch/model/falcon40b/transformer.py (1002:1035) duplicated block id: 2993 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/galactica/transformer.py (619:636) - megatron_patch/model/llama/transformer.py (795:811) duplicated block id: 2994 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/language_model.py (311:325) - megatron_patch/model/glm130b/language_model.py (316:330) duplicated block id: 2995 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/transformer.py (245:264) - megatron_patch/model/llama/transformer.py (215:235) duplicated block id: 2996 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/bloom/language_model.py (481:501) - megatron_patch/model/mistral/language_model.py (534:554) duplicated block id: 2997 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/mlp.py (82:96) - megatron_patch/model/qwen2/transformer/mlp.py (80:94) duplicated block id: 2998 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/language_model.py (357:371) - megatron_patch/model/bloom/language_model.py (319:333) duplicated block id: 2999 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/bloom/language_model.py (481:501) - megatron_patch/model/llava/language_model.py (582:602) duplicated block id: 3000 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/falcon/language_model.py (568:588) - megatron_patch/model/llama2/language_model.py (522:542) duplicated block id: 3001 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (1165:1198) - megatron_patch/model/glm130b/transformer.py (1196:1229) duplicated block id: 3002 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/falcon/language_model.py (301:315) - megatron_patch/model/starcoder/language_model.py (302:316) duplicated block id: 3003 size: 14 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon40b/checkpoint_reshaping_and_interoperability.py (804:820) - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (811:827) duplicated block id: 3004 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/qwen1_5_megablocks/transformer.py (671:685) - megatron_patch/model/starcoder/transformer.py (759:773) duplicated block id: 3005 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/bloom/language_model.py (319:333) - megatron_patch/model/mistral/language_model.py (302:316) duplicated block id: 3006 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/layer_specs.py (62:88) - megatron_patch/model/qwen1_5/layer_specs.py (39:52) duplicated block id: 3007 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/language_model.py (510:530) - megatron_patch/model/qwen_vl/language_model.py (551:571) duplicated block id: 3008 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/falcon/transformer.py (1263:1296) - megatron_patch/model/mistral/transformer.py (1749:1784) duplicated block id: 3009 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/transformer_layer.py (59:95) - megatron_patch/model/qwen2/transformer_layer.py (42:78) duplicated block id: 3010 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/language_model.py (547:567) - megatron_patch/model/llama2/language_model.py (522:542) duplicated block id: 3011 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/qwen1_5_megablocks/transformer.py (1613:1646) - megatron_patch/model/starcoder/transformer.py (1305:1337) duplicated block id: 3012 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/llava_mcore/llava_spec.py (36:50) - megatron_patch/model/mixtral/layer_specs.py (62:88) duplicated block id: 3013 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/gpt_model.py (79:97) - megatron_patch/model/llama/gpt_model.py (82:100) duplicated block id: 3014 size: 14 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (815:832) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (888:905) duplicated block id: 3015 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/falcon/language_model.py (301:315) - megatron_patch/model/mistral/language_model.py (302:316) duplicated block id: 3016 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/falcon/transformer.py (1263:1296) - megatron_patch/model/glm130b/transformer.py (1196:1229) duplicated block id: 3017 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/mixtral_bak/transformer/mlp.py (93:111) - megatron_patch/model/qwen2/transformer/mlp.py (137:155) duplicated block id: 3018 size: 14 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (370:390) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (457:477) duplicated block id: 3019 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/language_model.py (568:588) - megatron_patch/model/qwen/language_model.py (504:524) duplicated block id: 3020 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/language_model.py (568:588) - megatron_patch/model/llava/language_model.py (582:602) duplicated block id: 3021 size: 14 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/reward_model_to_megatron.py (618:635) - toolkits/model_checkpoints_convertor/glm/checkpoint_reshaping_and_interoperability.py (353:370) duplicated block id: 3022 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/gpt_model.py (79:97) - megatron_patch/model/falcon40b/gpt_model.py (83:100) duplicated block id: 3023 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/language_model.py (281:301) - megatron_patch/model/glm130b/language_model.py (237:258) duplicated block id: 3024 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (646:660) - megatron_patch/model/falcon/transformer.py (797:812) duplicated block id: 3025 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/falcon/language_model.py (301:315) - megatron_patch/model/qwen/language_model.py (301:315) duplicated block id: 3026 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/language_model.py (510:530) - megatron_patch/model/starcoder/language_model.py (452:472) duplicated block id: 3027 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/language_model.py (532:547) - megatron_patch/model/llama2/language_model.py (464:479) duplicated block id: 3028 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/qwen_vl/transformer.py (779:793) - megatron_patch/model/starcoder/transformer.py (759:773) duplicated block id: 3029 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/language_model.py (599:619) - megatron_patch/model/falcon40b/language_model.py (568:588) duplicated block id: 3030 size: 14 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (753:772) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (829:845) duplicated block id: 3031 size: 14 cleaned lines of code in 2 files: - megatron_patch/data/utils.py (201:216) - megatron_patch/data/utils.py (387:402) duplicated block id: 3032 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/language_model.py (212:230) - megatron_patch/model/starcoder/language_model.py (198:216) duplicated block id: 3033 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/language_model.py (532:547) - megatron_patch/model/mistral/language_model.py (466:481) duplicated block id: 3034 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/galactica/language_model.py (597:617) - megatron_patch/model/qwen/language_model.py (504:524) duplicated block id: 3035 size: 14 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/reward_model_to_megatron.py (713:734) - toolkits/model_checkpoints_convertor/glm/checkpoint_reshaping_and_interoperability.py (441:461) duplicated block id: 3036 size: 14 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore.py (329:342) - toolkits/model_checkpoints_convertor/mistral/hf2mcore.py (229:242) duplicated block id: 3037 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/language_model.py (317:331) - megatron_patch/model/llama2/language_model.py (301:315) duplicated block id: 3038 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/language_model.py (317:331) - megatron_patch/model/mistral/language_model.py (302:316) duplicated block id: 3039 size: 14 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/chatglm/checkpoint_reshaping_and_interoperability.py (298:323) - toolkits/model_checkpoints_convertor/yi/checkpoint_reshaping_and_interoperability.py (333:358) duplicated block id: 3040 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/llama2/transformer.py (780:794) - megatron_patch/model/starcoder/transformer.py (759:773) duplicated block id: 3041 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/llama/language_model.py (579:599) - megatron_patch/model/llama2/language_model.py (522:542) duplicated block id: 3042 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/moe/moe_layer.py (65:90) - megatron_patch/model/qwen1_5/moe/moe_layer.py (54:77) duplicated block id: 3043 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/galactica/language_model.py (597:617) - megatron_patch/model/llama3/language_model.py (499:519) duplicated block id: 3044 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/bloom/language_model.py (481:501) - megatron_patch/model/qwen1_5_megablocks/language_model.py (521:541) duplicated block id: 3045 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/llama/language_model.py (560:575) - megatron_patch/model/llama2/language_model.py (499:514) duplicated block id: 3046 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/mlp.py (66:80) - megatron_patch/model/qwen1_5/transformer/mlp.py (104:118) duplicated block id: 3047 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/falcon/language_model.py (568:588) - megatron_patch/model/qwen_vl/language_model.py (551:571) duplicated block id: 3048 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/language_model.py (192:210) - megatron_patch/model/glm130b/language_model.py (192:210) duplicated block id: 3049 size: 14 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/mistral/hf2mcore.py (271:284) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_vl.py (140:153) duplicated block id: 3050 size: 14 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/mistral/hf2mcore.py (271:284) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_vl.py (189:202) duplicated block id: 3051 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/transformer.py (1002:1035) - megatron_patch/model/llava/transformer.py (1751:1786) duplicated block id: 3052 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/galactica/language_model.py (332:346) - megatron_patch/model/qwen1_5_megablocks/language_model.py (300:314) duplicated block id: 3053 size: 14 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/mistral/hf2mcore.py (271:284) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2.5_vl.py (189:202) duplicated block id: 3054 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/transformer.py (1002:1035) - megatron_patch/model/llama2/transformer.py (1752:1787) duplicated block id: 3055 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/falcon/language_model.py (301:315) - megatron_patch/model/llama3/language_model.py (296:310) duplicated block id: 3056 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/llama/transformer.py (1057:1090) - megatron_patch/model/mistral/transformer.py (1749:1784) duplicated block id: 3057 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (1751:1786) - megatron_patch/model/falcon40b/transformer.py (1002:1035) duplicated block id: 3058 size: 14 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/mistral/hf2mcore.py (271:284) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2.5_vl.py (140:153) duplicated block id: 3059 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/language_model.py (357:371) - megatron_patch/model/llama/language_model.py (301:315) duplicated block id: 3060 size: 14 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/glm/checkpoint_reshaping_and_interoperability.py (353:370) - toolkits/model_checkpoints_convertor/starcoder/checkpoint_reshaping_and_interoperability.py (632:649) duplicated block id: 3061 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/llava_mcore/llava_spec.py (36:50) - megatron_patch/model/mixtral_bak/layer_specs.py (40:66) duplicated block id: 3062 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/llava_mcore/layer_specs.py (14:34) - megatron_patch/model/qwen3_moe/gpt_layer_specs.py (37:59) duplicated block id: 3063 size: 14 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_moe.py (157:170) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_dense_and_moe_gqa.py (135:148) duplicated block id: 3064 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (1751:1786) - megatron_patch/model/falcon/transformer.py (1263:1296) duplicated block id: 3065 size: 14 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2.5_vl.py (220:234) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_dense_and_moe_gqa.py (350:365) duplicated block id: 3066 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/language_model.py (316:330) - megatron_patch/model/qwen1_5_megablocks/language_model.py (300:314) duplicated block id: 3067 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/bloom/language_model.py (481:501) - megatron_patch/model/llama2/language_model.py (522:542) duplicated block id: 3068 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/transformer.py (1002:1035) - megatron_patch/model/glm130b/transformer.py (1196:1229) duplicated block id: 3069 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/falcon/transformer.py (797:812) - megatron_patch/model/llama3/transformer_legacy.py (744:758) duplicated block id: 3070 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/gpt_model.py (79:97) - megatron_patch/model/falcon/gpt_model.py (83:100) duplicated block id: 3071 size: 14 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/deepseek/hf2mcore_deepseek_v2_moe.py (161:175) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_moe.py (212:226) duplicated block id: 3072 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/bloom/language_model.py (319:333) - megatron_patch/model/llama2/language_model.py (301:315) duplicated block id: 3073 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (1807:1821) - megatron_patch/model/mistral/transformer.py (1806:1820) duplicated block id: 3074 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/language_model.py (599:619) - megatron_patch/model/galactica/language_model.py (597:617) duplicated block id: 3075 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/galactica/language_model.py (597:617) - megatron_patch/model/qwen1_5_megablocks/language_model.py (521:541) duplicated block id: 3076 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/transformer.py (1196:1229) - megatron_patch/model/llama/transformer.py (1057:1090) duplicated block id: 3077 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/mlp.py (82:96) - megatron_patch/model/mixtral/transformer/mlp.py (75:89) duplicated block id: 3078 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/llama/language_model.py (301:315) - megatron_patch/model/llama3/language_model.py (296:310) duplicated block id: 3079 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/language_model.py (547:567) - megatron_patch/model/llava/language_model.py (582:602) duplicated block id: 3080 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/galactica/language_model.py (597:617) - megatron_patch/model/starcoder/language_model.py (452:472) duplicated block id: 3081 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (1807:1821) - megatron_patch/model/llava/transformer.py (1808:1822) duplicated block id: 3082 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/language_model.py (316:330) - megatron_patch/model/llava/language_model.py (304:318) duplicated block id: 3083 size: 14 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_mha.py (228:243) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_moe.py (293:308) duplicated block id: 3084 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/falcon/transformer.py (797:812) - megatron_patch/model/starcoder/transformer.py (759:773) duplicated block id: 3085 size: 14 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (829:845) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (889:908) duplicated block id: 3086 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/galactica/language_model.py (332:346) - megatron_patch/model/qwen_vl/language_model.py (302:316) duplicated block id: 3087 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/llama/transformer.py (1057:1090) - megatron_patch/model/llama3/transformer_legacy.py (1711:1746) duplicated block id: 3088 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer/attention.py (432:446) - megatron_patch/model/qwen2_vl/attention.py (578:592) duplicated block id: 3089 size: 14 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/checkpoint_reshaping_and_interoperability.py (619:636) - toolkits/model_checkpoints_convertor/glm/checkpoint_reshaping_and_interoperability.py (353:370) duplicated block id: 3090 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/llama/transformer.py (1057:1090) - megatron_patch/model/qwen/transformer.py (1693:1728) duplicated block id: 3091 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/llama/transformer.py (1057:1090) - megatron_patch/model/llama2/transformer.py (1752:1787) duplicated block id: 3092 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (1008:1025) - megatron_patch/model/llama3/transformer_legacy.py (1087:1104) duplicated block id: 3093 size: 14 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/checkpoint_reshaping_and_interoperability.py (714:735) - toolkits/model_checkpoints_convertor/glm/checkpoint_reshaping_and_interoperability.py (441:461) duplicated block id: 3094 size: 14 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_mha.py (268:281) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_dense_and_moe_gqa.py (675:688) duplicated block id: 3095 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/language_model.py (510:530) - megatron_patch/model/llava/language_model.py (582:602) duplicated block id: 3096 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (1621:1656) - megatron_patch/model/llama/transformer.py (1057:1090) duplicated block id: 3097 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/language_model.py (193:211) - megatron_patch/model/glm130b/language_model.py (192:210) duplicated block id: 3098 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/language_model.py (515:535) - megatron_patch/model/falcon40b/language_model.py (568:588) duplicated block id: 3099 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/qwen1_5/transformer/attention.py (432:446) - megatron_patch/model/qwen2_vl/attention.py (578:592) duplicated block id: 3100 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/llama3/layer_specs.py (39:52) - megatron_patch/model/mixtral/layer_specs.py (62:88) duplicated block id: 3101 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/galactica/language_model.py (332:346) - megatron_patch/model/qwen/language_model.py (301:315) duplicated block id: 3102 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/transformer.py (1195:1227) - megatron_patch/model/qwen1_5_megablocks/transformer.py (1613:1646) duplicated block id: 3103 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/transformer.py (40:76) - megatron_patch/model/starcoder/transformer.py (42:80) duplicated block id: 3104 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/language_model.py (547:567) - megatron_patch/model/starcoder/language_model.py (452:472) duplicated block id: 3105 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/language_model.py (515:535) - megatron_patch/model/bloom/language_model.py (481:501) duplicated block id: 3106 size: 14 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore.py (269:282) - toolkits/model_checkpoints_convertor/mistral/hf2mcore.py (207:220) duplicated block id: 3107 size: 14 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (867:886) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (829:845) duplicated block id: 3108 size: 14 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/reward_model_to_megatron.py (713:734) - toolkits/model_checkpoints_convertor/glm130b/checkpoint_reshaping_and_interoperability.py (422:442) duplicated block id: 3109 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/falcon/transformer.py (1266:1299) - megatron_patch/model/galactica/transformer.py (889:922) duplicated block id: 3110 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/language_model.py (599:619) - megatron_patch/model/bloom/language_model.py (481:501) duplicated block id: 3111 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/llama3/model.py (196:212) - megatron_patch/model/llama3_1/model.py (258:274) duplicated block id: 3112 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (440:456) - megatron_patch/model/llama/transformer.py (430:446) duplicated block id: 3113 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/llama/language_model.py (560:575) - megatron_patch/model/llava/language_model.py (559:574) duplicated block id: 3114 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/galactica/language_model.py (597:617) - megatron_patch/model/mistral/language_model.py (534:554) duplicated block id: 3115 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/llava/transformer.py (779:793) - megatron_patch/model/starcoder/transformer.py (759:773) duplicated block id: 3116 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer_legacy.py (1710:1743) - megatron_patch/model/qwen1_5_megablocks/transformer.py (1613:1646) duplicated block id: 3117 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/language_model.py (230:248) - megatron_patch/model/starcoder/language_model.py (178:196) duplicated block id: 3118 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/qwen_vl/language_model.py (180:198) - megatron_patch/model/starcoder/language_model.py (178:196) duplicated block id: 3119 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/language_model.py (547:567) - megatron_patch/model/llama3/language_model.py (499:519) duplicated block id: 3120 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/mlp.py (82:96) - megatron_patch/model/llama3/transformer/mlp.py (73:87) duplicated block id: 3121 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (1807:1821) - megatron_patch/model/llama2/transformer.py (1809:1823) duplicated block id: 3122 size: 14 cleaned lines of code in 2 files: - megatron_patch/tokenizer/tokenization_baichuan.py (144:163) - megatron_patch/tokenizer/tokenization_yi.py (163:182) duplicated block id: 3123 size: 14 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (386:405) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (457:477) duplicated block id: 3124 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/llama/language_model.py (579:599) - megatron_patch/model/llama3/language_model.py (499:519) duplicated block id: 3125 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/mistral/language_model.py (180:198) - megatron_patch/model/starcoder/language_model.py (178:196) duplicated block id: 3126 size: 14 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_mha.py (319:333) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_moe.py (538:552) duplicated block id: 3127 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/llava/transformer.py (1808:1822) - megatron_patch/model/qwen/transformer.py (1749:1763) duplicated block id: 3128 size: 14 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/chatglm/checkpoint_reshaping_and_interoperability.py (515:536) - toolkits/model_checkpoints_convertor/falcon40b/checkpoint_reshaping_and_interoperability.py (529:550) duplicated block id: 3129 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (757:771) - megatron_patch/model/falcon/transformer.py (797:812) duplicated block id: 3130 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/llama/language_model.py (301:315) - megatron_patch/model/mistral/language_model.py (302:316) duplicated block id: 3131 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/falcon/transformer.py (1263:1296) - megatron_patch/model/qwen/transformer.py (1693:1728) duplicated block id: 3132 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/moe/moe_layer.py (65:90) - megatron_patch/model/qwen2/moe/moe_layer.py (70:93) duplicated block id: 3133 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (1165:1198) - megatron_patch/model/llama2/transformer.py (1752:1787) duplicated block id: 3134 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/language_model.py (316:330) - megatron_patch/model/qwen_vl/language_model.py (302:316) duplicated block id: 3135 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (1165:1198) - megatron_patch/model/llava/transformer.py (1751:1786) duplicated block id: 3136 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/bloom/language_model.py (481:501) - megatron_patch/model/qwen/language_model.py (504:524) duplicated block id: 3137 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/language_model.py (510:530) - megatron_patch/model/llama2/language_model.py (522:542) duplicated block id: 3138 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/language_model.py (568:588) - megatron_patch/model/llama3/language_model.py (499:519) duplicated block id: 3139 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/language_model.py (599:619) - megatron_patch/model/llama/language_model.py (579:599) duplicated block id: 3140 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/language_model.py (547:567) - megatron_patch/model/qwen_vl/language_model.py (551:571) duplicated block id: 3141 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/falcon/transformer.py (234:254) - megatron_patch/model/glm130b/transformer.py (245:264) duplicated block id: 3142 size: 14 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/mistral/hf2mcore.py (500:515) - toolkits/model_checkpoints_convertor/qwen/hf2megablocks_qwen1.5.py (628:643) duplicated block id: 3143 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/language_model.py (301:315) - megatron_patch/model/mistral/language_model.py (302:316) duplicated block id: 3144 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/qwen1_5/transformer/attention.py (432:446) - megatron_patch/model/qwen2_vl/attention_vision.py (577:591) duplicated block id: 3145 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/language_model.py (515:535) - megatron_patch/model/falcon/language_model.py (568:588) duplicated block id: 3146 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/model.py (273:305) - megatron_patch/model/qwen2_vl/gpt_model.py (208:239) duplicated block id: 3147 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/language_model.py (316:330) - megatron_patch/model/llama2/language_model.py (301:315) duplicated block id: 3148 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/transformer.py (613:627) - megatron_patch/model/llama/transformer.py (570:584) duplicated block id: 3149 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/galactica/language_model.py (223:241) - megatron_patch/model/glm130b/language_model.py (192:210) duplicated block id: 3150 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/language_model.py (515:535) - megatron_patch/model/chatglm/language_model.py (547:567) duplicated block id: 3151 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/galactica/language_model.py (332:346) - megatron_patch/model/llava/language_model.py (304:318) duplicated block id: 3152 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/language_model.py (316:330) - megatron_patch/model/starcoder/language_model.py (302:316) duplicated block id: 3153 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/falcon/transformer.py (1263:1296) - megatron_patch/model/llava/transformer.py (1751:1786) duplicated block id: 3154 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/transformer.py (1002:1035) - megatron_patch/model/llama3/transformer_legacy.py (1711:1746) duplicated block id: 3155 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (851:865) - megatron_patch/model/qwen1_5_megablocks/transformer.py (868:882) duplicated block id: 3156 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/language_model.py (515:535) - megatron_patch/model/llama/language_model.py (579:599) duplicated block id: 3157 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/llama/transformer.py (1057:1090) - megatron_patch/model/llava/transformer.py (1751:1786) duplicated block id: 3158 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/bloom/language_model.py (319:333) - megatron_patch/model/llama3/language_model.py (296:310) duplicated block id: 3159 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (1165:1198) - megatron_patch/model/llama3/transformer_legacy.py (1711:1746) duplicated block id: 3160 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/language_model.py (532:547) - megatron_patch/model/llama3/language_model.py (455:470) duplicated block id: 3161 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/galactica/language_model.py (597:617) - megatron_patch/model/llava/language_model.py (582:602) duplicated block id: 3162 size: 14 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/chatglm/checkpoint_reshaping_and_interoperability.py (515:536) - toolkits/model_checkpoints_convertor/galactica/checkpoint_reshaping_and_interoperability.py (585:606) duplicated block id: 3163 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/transformer.py (1002:1035) - megatron_patch/model/mistral/transformer.py (1749:1784) duplicated block id: 3164 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (1008:1025) - megatron_patch/model/qwen_vl/transformer.py (1137:1154) duplicated block id: 3165 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/language_model.py (311:325) - megatron_patch/model/falcon/language_model.py (301:315) duplicated block id: 3166 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/galactica/language_model.py (332:346) - megatron_patch/model/llama3/language_model.py (296:310) duplicated block id: 3167 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/gpt_model.py (77:95) - megatron_patch/model/llama/gpt_model.py (82:100) duplicated block id: 3168 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/llama2/transformer.py (1809:1823) - megatron_patch/model/qwen/transformer.py (1749:1763) duplicated block id: 3169 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/transformer.py (1002:1035) - megatron_patch/model/qwen_vl/transformer.py (1751:1786) duplicated block id: 3170 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/llama3_1/layer_specs.py (40:53) - megatron_patch/model/mixtral/layer_specs.py (62:88) duplicated block id: 3171 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/bloom/language_model.py (319:333) - megatron_patch/model/starcoder/language_model.py (302:316) duplicated block id: 3172 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (1168:1201) - megatron_patch/model/chatglm/transformer.py (881:914) duplicated block id: 3173 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/llama/transformer.py (564:580) - megatron_patch/model/llama2/transformer.py (774:790) duplicated block id: 3174 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/mlp.py (116:134) - megatron_patch/model/mixtral_bak/transformer/mlp.py (93:111) duplicated block id: 3175 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/llama/language_model.py (579:599) - megatron_patch/model/qwen_vl/language_model.py (551:571) duplicated block id: 3176 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/language_model.py (515:535) - megatron_patch/model/galactica/language_model.py (597:617) duplicated block id: 3177 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/mixtral_bak/moe/experts.py (20:49) - megatron_patch/model/qwen1_5/moe/experts.py (20:48) duplicated block id: 3178 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/bloom/language_model.py (319:333) - megatron_patch/model/llava/language_model.py (304:318) duplicated block id: 3179 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/llama/language_model.py (579:599) - megatron_patch/model/llava/language_model.py (582:602) duplicated block id: 3180 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/language_model.py (301:315) - megatron_patch/model/llama2/language_model.py (301:315) duplicated block id: 3181 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/gpt_model.py (79:97) - megatron_patch/model/galactica/gpt_model.py (83:100) duplicated block id: 3182 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/llama/transformer.py (564:580) - megatron_patch/model/qwen1_5_megablocks/transformer.py (665:681) duplicated block id: 3183 size: 14 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore.py (329:342) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_dense_and_moe_gqa.py (135:148) duplicated block id: 3184 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (1750:1783) - megatron_patch/model/qwen1_5_megablocks/transformer.py (1613:1646) duplicated block id: 3185 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/transformer.py (583:597) - megatron_patch/model/llama3/transformer_legacy.py (744:758) duplicated block id: 3186 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/language_model.py (311:325) - megatron_patch/model/llama/language_model.py (301:315) duplicated block id: 3187 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/language_model.py (317:331) - megatron_patch/model/llama3/language_model.py (296:310) duplicated block id: 3188 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (1677:1691) - megatron_patch/model/llava/transformer.py (1808:1822) duplicated block id: 3189 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/transformer.py (583:597) - megatron_patch/model/qwen/transformer.py (739:753) duplicated block id: 3190 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/language_model.py (547:567) - megatron_patch/model/qwen/language_model.py (504:524) duplicated block id: 3191 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/language_model.py (301:315) - megatron_patch/model/qwen1_5_megablocks/language_model.py (300:314) duplicated block id: 3192 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (795:814) - megatron_patch/model/starcoder/transformer.py (940:959) duplicated block id: 3193 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (561:575) - megatron_patch/model/llama/transformer.py (570:584) duplicated block id: 3194 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/mixtral_bak/transformer/attention.py (24:45) - megatron_patch/model/qwen2_vl/attention.py (58:83) duplicated block id: 3195 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/falcon/language_model.py (568:588) - megatron_patch/model/qwen/language_model.py (504:524) duplicated block id: 3196 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/model.py (273:305) - megatron_patch/model/qwen2/model.py (221:252) duplicated block id: 3197 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/transformer.py (881:914) - megatron_patch/model/falcon40b/transformer.py (1005:1038) duplicated block id: 3198 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer_legacy.py (1767:1781) - megatron_patch/model/qwen_vl/transformer.py (1808:1822) duplicated block id: 3199 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer/attention.py (432:446) - megatron_patch/model/qwen2_vl/attention_vision.py (577:591) duplicated block id: 3200 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (1621:1656) - megatron_patch/model/bloom/transformer.py (1165:1198) duplicated block id: 3201 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/language_model.py (568:588) - megatron_patch/model/qwen1_5_megablocks/language_model.py (521:541) duplicated block id: 3202 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/model.py (273:305) - megatron_patch/model/llama3/model.py (220:251) duplicated block id: 3203 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (851:865) - megatron_patch/model/llama3/transformer_legacy.py (934:948) duplicated block id: 3204 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/mistral/transformer.py (1806:1820) - megatron_patch/model/qwen/transformer.py (1749:1763) duplicated block id: 3205 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/falcon/language_model.py (568:588) - megatron_patch/model/mistral/language_model.py (534:554) duplicated block id: 3206 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (492:510) - megatron_patch/model/falcon40b/transformer.py (539:558) duplicated block id: 3207 size: 14 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (370:390) - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (441:461) duplicated block id: 3208 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/llama/transformer.py (564:580) - megatron_patch/model/llava/transformer.py (773:789) duplicated block id: 3209 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (1677:1691) - megatron_patch/model/llama2/transformer.py (1809:1823) duplicated block id: 3210 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/mixtral_bak/layer_specs.py (40:66) - megatron_patch/model/qwen1_5/layer_specs.py (39:52) duplicated block id: 3211 size: 14 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (927:943) - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (811:827) duplicated block id: 3212 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/bloom/language_model.py (319:333) - megatron_patch/model/qwen_vl/language_model.py (302:316) duplicated block id: 3213 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (1165:1198) - megatron_patch/model/qwen_vl/transformer.py (1751:1786) duplicated block id: 3214 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/language_model.py (317:331) - megatron_patch/model/qwen/language_model.py (301:315) duplicated block id: 3215 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer_legacy.py (153:172) - megatron_patch/model/qwen1_5_megablocks/transformer.py (147:166) duplicated block id: 3216 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/layer_specs.py (114:127) - megatron_patch/model/qwen1_5/layer_specs.py (69:82) duplicated block id: 3217 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/moe/moe_layer.py (64:89) - megatron_patch/model/qwen1_5/moe/moe_layer.py (54:77) duplicated block id: 3218 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/llama/language_model.py (301:315) - megatron_patch/model/qwen1_5_megablocks/language_model.py (300:314) duplicated block id: 3219 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/bloom/language_model.py (481:501) - megatron_patch/model/llama3/language_model.py (499:519) duplicated block id: 3220 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/llama/language_model.py (560:575) - megatron_patch/model/qwen_vl/language_model.py (528:543) duplicated block id: 3221 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/llava/transformer.py (1808:1822) - megatron_patch/model/qwen1_5_megablocks/transformer.py (1671:1685) duplicated block id: 3222 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (1008:1025) - megatron_patch/model/llama2/transformer.py (1138:1155) duplicated block id: 3223 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (1621:1656) - megatron_patch/model/falcon/transformer.py (1263:1296) duplicated block id: 3224 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/language_model.py (316:330) - megatron_patch/model/qwen/language_model.py (301:315) duplicated block id: 3225 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/language_model.py (311:325) - megatron_patch/model/bloom/language_model.py (319:333) duplicated block id: 3226 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/language_model.py (357:371) - megatron_patch/model/chatglm/language_model.py (317:331) duplicated block id: 3227 size: 14 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/chatglm/checkpoint_reshaping_and_interoperability.py (298:323) - toolkits/model_checkpoints_convertor/falcon/checkpoint_reshaping_and_interoperability.py (308:333) duplicated block id: 3228 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/language_model.py (317:331) - megatron_patch/model/llava/language_model.py (304:318) duplicated block id: 3229 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (657:672) - megatron_patch/model/glm130b/transformer.py (758:773) duplicated block id: 3230 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/llama3_1/layer_specs.py (40:53) - megatron_patch/model/mixtral_bak/layer_specs.py (40:66) duplicated block id: 3231 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/language_model.py (547:567) - megatron_patch/model/qwen1_5_megablocks/language_model.py (521:541) duplicated block id: 3232 size: 14 cleaned lines of code in 2 files: - megatron_patch/generation/api.py (26:39) - megatron_patch/generation/api.py (106:119) duplicated block id: 3233 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/llama/language_model.py (301:315) - megatron_patch/model/qwen/language_model.py (301:315) duplicated block id: 3234 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/galactica/language_model.py (332:346) - megatron_patch/model/mistral/language_model.py (302:316) duplicated block id: 3235 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/language_model.py (510:530) - megatron_patch/model/mistral/language_model.py (534:554) duplicated block id: 3236 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/llama/language_model.py (301:315) - megatron_patch/model/starcoder/language_model.py (302:316) duplicated block id: 3237 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/falcon/transformer.py (1263:1296) - megatron_patch/model/llama3/transformer_legacy.py (1711:1746) duplicated block id: 3238 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/llama/language_model.py (560:575) - megatron_patch/model/qwen1_5_megablocks/language_model.py (498:513) duplicated block id: 3239 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/qwen1_5/moe/experts.py (20:48) - megatron_patch/model/qwen2/moe/experts.py (35:68) duplicated block id: 3240 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (1677:1691) - megatron_patch/model/qwen_vl/transformer.py (1808:1822) duplicated block id: 3241 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (1363:1377) - megatron_patch/model/glm130b/transformer.py (941:955) duplicated block id: 3242 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/transformer/attention.py (72:97) - megatron_patch/model/mixtral_bak/transformer/attention.py (24:45) duplicated block id: 3243 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/language_model.py (357:371) - megatron_patch/model/falcon40b/language_model.py (301:315) duplicated block id: 3244 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/falcon/language_model.py (301:315) - megatron_patch/model/qwen_vl/language_model.py (302:316) duplicated block id: 3245 size: 14 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_mha.py (228:243) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_mha_to_moe.py (144:159) duplicated block id: 3246 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/llama/transformer.py (564:580) - megatron_patch/model/mistral/transformer.py (772:788) duplicated block id: 3247 size: 14 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (901:917) - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (811:827) duplicated block id: 3248 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/bloom/language_model.py (481:501) - megatron_patch/model/qwen_vl/language_model.py (551:571) duplicated block id: 3249 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/language_model.py (357:371) - megatron_patch/model/galactica/language_model.py (332:346) duplicated block id: 3250 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/llama/language_model.py (579:599) - megatron_patch/model/qwen1_5_megablocks/language_model.py (521:541) duplicated block id: 3251 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/llava/language_model.py (182:200) - megatron_patch/model/starcoder/language_model.py (178:196) duplicated block id: 3252 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (440:456) - megatron_patch/model/galactica/transformer.py (393:409) duplicated block id: 3253 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/language_model.py (301:315) - megatron_patch/model/qwen/language_model.py (301:315) duplicated block id: 3254 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (1751:1786) - megatron_patch/model/llama/transformer.py (1057:1090) duplicated block id: 3255 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/language_model.py (317:331) - megatron_patch/model/qwen_vl/language_model.py (302:316) duplicated block id: 3256 size: 14 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_mha.py (268:281) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_vl.py (618:631) duplicated block id: 3257 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/language_model.py (301:315) - megatron_patch/model/starcoder/language_model.py (302:316) duplicated block id: 3258 size: 14 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_mha.py (268:281) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_vl.py (675:688) duplicated block id: 3259 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/language_model.py (568:588) - megatron_patch/model/qwen_vl/language_model.py (551:571) duplicated block id: 3260 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/falcon/language_model.py (301:315) - megatron_patch/model/llama2/language_model.py (301:315) duplicated block id: 3261 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/mixtral_bak/moe/experts.py (20:49) - megatron_patch/model/qwen2/moe/experts.py (35:68) duplicated block id: 3262 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/qwen/language_model.py (179:197) - megatron_patch/model/starcoder/language_model.py (178:196) duplicated block id: 3263 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/language_model.py (599:619) - megatron_patch/model/falcon/language_model.py (568:588) duplicated block id: 3264 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/language_model.py (357:371) - megatron_patch/model/falcon/language_model.py (301:315) duplicated block id: 3265 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (1751:1786) - megatron_patch/model/bloom/transformer.py (1165:1198) duplicated block id: 3266 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (646:660) - megatron_patch/model/falcon40b/transformer.py (583:597) duplicated block id: 3267 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/bloom/language_model.py (194:212) - megatron_patch/model/glm130b/language_model.py (192:210) duplicated block id: 3268 size: 14 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/deepseek/hf2mcore_deepseek_v2_moe.py (499:515) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_dense_and_moe_gqa.py (764:779) duplicated block id: 3269 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer_legacy.py (1767:1781) - megatron_patch/model/llava/transformer.py (1808:1822) duplicated block id: 3270 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/galactica/transformer.py (177:197) - megatron_patch/model/glm130b/transformer.py (245:264) duplicated block id: 3271 size: 14 cleaned lines of code in 2 files: - megatron_patch/model/falcon/language_model.py (568:588) - megatron_patch/model/llama3/language_model.py (499:519) duplicated block id: 3272 size: 13 cleaned lines of code in 2 files: - megatron_patch/model/llama/transformer.py (627:648) - megatron_patch/model/qwen/transformer.py (821:841) duplicated block id: 3273 size: 13 cleaned lines of code in 2 files: - megatron_patch/model/falcon/transformer.py (904:920) - megatron_patch/model/glm130b/transformer.py (737:753) duplicated block id: 3274 size: 13 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon/checkpoint_reshaping_and_interoperability.py (90:102) - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (90:104) duplicated block id: 3275 size: 13 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2.5_vl.py (140:152) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_dense_and_moe_gqa.py (241:253) duplicated block id: 3276 size: 13 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (146:164) - megatron_patch/model/qwen1_5_megablocks/transformer.py (148:166) duplicated block id: 3277 size: 13 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/transformer.py (247:270) - megatron_patch/model/glm130b/transformer.py (317:341) duplicated block id: 3278 size: 13 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (89:101) - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (90:104) duplicated block id: 3279 size: 13 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (230:246) - toolkits/model_checkpoints_convertor/glm/checkpoint_reshaping_and_interoperability.py (171:187) duplicated block id: 3280 size: 13 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/language_model.py (537:553) - megatron_patch/model/mistral/language_model.py (563:579) duplicated block id: 3281 size: 13 cleaned lines of code in 2 files: - megatron_patch/model/mixtral_bak/transformer/attention.py (165:177) - megatron_patch/model/qwen2_vl/attention_vision.py (209:221) duplicated block id: 3282 size: 13 cleaned lines of code in 2 files: - megatron_patch/model/llama2/transformer.py (419:444) - megatron_patch/model/starcoder/transformer.py (324:349) duplicated block id: 3283 size: 13 cleaned lines of code in 2 files: - megatron_patch/model/qwen1_5/transformer/attention.py (191:203) - megatron_patch/model/qwen2_vl/attention.py (210:222) duplicated block id: 3284 size: 13 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/language_model.py (378:394) - megatron_patch/model/falcon/language_model.py (366:382) duplicated block id: 3285 size: 13 cleaned lines of code in 2 files: - toolkits/pretrain_data_preprocessing/clean_raw_text.py (63:78) - toolkits/pretrain_data_preprocessing/preprocess_wudao2.py (69:84) duplicated block id: 3286 size: 13 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer_legacy.py (418:443) - megatron_patch/model/starcoder/transformer.py (324:349) duplicated block id: 3287 size: 13 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/reward_model_to_megatron.py (200:216) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (262:278) duplicated block id: 3288 size: 13 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon40b/checkpoint_reshaping_and_interoperability.py (816:832) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (1132:1148) duplicated block id: 3289 size: 13 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/language_model.py (573:587) - megatron_patch/model/qwen1_5_megablocks/language_model.py (481:495) duplicated block id: 3290 size: 13 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/language_model.py (537:553) - megatron_patch/model/llama3/language_model.py (528:544) duplicated block id: 3291 size: 13 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_moe.py (157:169) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_dense_and_moe_gqa.py (241:253) duplicated block id: 3292 size: 13 cleaned lines of code in 2 files: - megatron_patch/model/mistral/transformer.py (420:445) - megatron_patch/model/starcoder/transformer.py (324:349) duplicated block id: 3293 size: 13 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (1094:1112) - megatron_patch/model/baichuan2/transformer.py (1227:1245) duplicated block id: 3294 size: 13 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (387:405) - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (337:355) duplicated block id: 3295 size: 13 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon/checkpoint_reshaping_and_interoperability.py (840:855) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (1132:1148) duplicated block id: 3296 size: 13 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/transformer.py (688:709) - megatron_patch/model/mistral/transformer.py (873:893) duplicated block id: 3297 size: 13 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (1027:1041) - megatron_patch/model/mistral/transformer.py (1154:1168) duplicated block id: 3298 size: 13 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/language_model.py (378:394) - megatron_patch/model/llama/language_model.py (366:382) duplicated block id: 3299 size: 13 cleaned lines of code in 2 files: - megatron_patch/model/llama/transformer.py (627:648) - megatron_patch/model/qwen_vl/transformer.py (875:895) duplicated block id: 3300 size: 13 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/language_model.py (628:644) - megatron_patch/model/glm130b/language_model.py (537:553) duplicated block id: 3301 size: 13 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (230:246) - toolkits/model_checkpoints_convertor/starcoder/checkpoint_reshaping_and_interoperability.py (222:238) duplicated block id: 3302 size: 13 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (156:171) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (185:200) duplicated block id: 3303 size: 13 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (1094:1112) - megatron_patch/model/qwen1_5_megablocks/transformer.py (1111:1129) duplicated block id: 3304 size: 13 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (371:390) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (324:342) duplicated block id: 3305 size: 13 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (389:401) - megatron_patch/model/glm130b/transformer.py (422:434) duplicated block id: 3306 size: 13 cleaned lines of code in 2 files: - megatron_patch/model/llava/language_model.py (542:556) - megatron_patch/model/qwen/language_model.py (478:492) duplicated block id: 3307 size: 13 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/glm130b/checkpoint_reshaping_and_interoperability.py (171:187) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (189:205) duplicated block id: 3308 size: 13 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/transformer.py (972:995) - megatron_patch/model/galactica/transformer.py (863:886) duplicated block id: 3309 size: 13 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (723:743) - megatron_patch/model/starcoder/transformer.py (731:751) duplicated block id: 3310 size: 13 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/transformer.py (737:753) - megatron_patch/model/llama/transformer.py (676:692) duplicated block id: 3311 size: 13 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/reward_model_to_megatron.py (318:330) - toolkits/model_checkpoints_convertor/falcon/checkpoint_reshaping_and_interoperability.py (615:627) duplicated block id: 3312 size: 13 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/reward_model_to_megatron.py (200:216) - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (272:288) duplicated block id: 3313 size: 13 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (146:164) - megatron_patch/model/qwen_vl/transformer.py (150:168) duplicated block id: 3314 size: 13 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/checkpoint_reshaping_and_interoperability.py (375:393) - toolkits/model_checkpoints_convertor/falcon/checkpoint_reshaping_and_interoperability.py (693:711) duplicated block id: 3315 size: 13 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (570:587) - toolkits/model_checkpoints_convertor/yi/checkpoint_reshaping_and_interoperability.py (562:579) duplicated block id: 3316 size: 13 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/layer_specs.py (132:146) - megatron_patch/model/qwen3_moe/gpt_layer_specs.py (253:267) duplicated block id: 3317 size: 13 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/transformer.py (855:878) - megatron_patch/model/falcon/transformer.py (1233:1256) duplicated block id: 3318 size: 13 cleaned lines of code in 2 files: - megatron_patch/model/llama3_1/model.py (82:98) - megatron_patch/model/mixtral/model.py (65:81) duplicated block id: 3319 size: 13 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (913:928) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (1132:1148) duplicated block id: 3320 size: 13 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/hf2te.py (452:468) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (624:640) duplicated block id: 3321 size: 13 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (723:743) - megatron_patch/model/falcon/transformer.py (757:777) duplicated block id: 3322 size: 13 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/galactica/checkpoint_reshaping_and_interoperability.py (217:233) - toolkits/model_checkpoints_convertor/starcoder/checkpoint_reshaping_and_interoperability.py (222:238) duplicated block id: 3323 size: 13 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/transformer.py (301:325) - megatron_patch/model/glm130b/transformer.py (317:341) duplicated block id: 3324 size: 13 cleaned lines of code in 2 files: - megatron_patch/model/llama2/transformer.py (732:752) - megatron_patch/model/starcoder/transformer.py (731:751) duplicated block id: 3325 size: 13 cleaned lines of code in 2 files: - megatron_patch/model/qwen_vl/transformer.py (731:751) - megatron_patch/model/starcoder/transformer.py (731:751) duplicated block id: 3326 size: 13 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (370:389) - toolkits/model_checkpoints_convertor/yi/checkpoint_reshaping_and_interoperability.py (374:392) duplicated block id: 3327 size: 13 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/gpt_model.py (49:65) - megatron_patch/model/starcoder/gpt_model.py (52:69) duplicated block id: 3328 size: 13 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (89:101) - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (100:114) duplicated block id: 3329 size: 13 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/language_model.py (573:587) - megatron_patch/model/llava/language_model.py (542:556) duplicated block id: 3330 size: 13 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (146:164) - megatron_patch/model/llama2/transformer.py (150:168) duplicated block id: 3331 size: 13 cleaned lines of code in 2 files: - megatron_patch/model/llama3/model.py (67:83) - megatron_patch/model/mixtral/model.py (65:81) duplicated block id: 3332 size: 13 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (612:632) - megatron_patch/model/starcoder/transformer.py (731:751) duplicated block id: 3333 size: 13 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (100:114) - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (113:125) duplicated block id: 3334 size: 13 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/glm/checkpoint_reshaping_and_interoperability.py (171:187) - toolkits/model_checkpoints_convertor/yi/checkpoint_reshaping_and_interoperability.py (215:231) duplicated block id: 3335 size: 13 cleaned lines of code in 2 files: - megatron_patch/model/qwen1_5_megablocks/transformer.py (310:335) - megatron_patch/model/starcoder/transformer.py (324:349) duplicated block id: 3336 size: 13 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/deepseek/hf2mcore_deepseek_v2_moe.py (338:356) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_dense_and_moe_gqa.py (484:502) duplicated block id: 3337 size: 13 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore.py (329:341) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_dense_and_moe_gqa.py (241:253) duplicated block id: 3338 size: 13 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (939:954) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (1132:1148) duplicated block id: 3339 size: 13 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (266:282) - toolkits/model_checkpoints_convertor/starcoder/checkpoint_reshaping_and_interoperability.py (222:238) duplicated block id: 3340 size: 13 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/hf2te.py (306:321) - toolkits/model_checkpoints_convertor/falcon/checkpoint_reshaping_and_interoperability.py (444:459) duplicated block id: 3341 size: 13 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (1094:1112) - megatron_patch/model/qwen/transformer.py (1169:1187) duplicated block id: 3342 size: 13 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (612:632) - megatron_patch/model/falcon/transformer.py (757:777) duplicated block id: 3343 size: 13 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (230:246) - toolkits/model_checkpoints_convertor/bloom/checkpoint_reshaping_and_interoperability.py (200:216) duplicated block id: 3344 size: 13 cleaned lines of code in 2 files: - megatron_patch/model/qwen1_5_megablocks/transformer.py (623:643) - megatron_patch/model/starcoder/transformer.py (731:751) duplicated block id: 3345 size: 13 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/mistral/hf2mcore.py (154:168) - toolkits/model_checkpoints_convertor/qwen/hf2megablocks_qwen1.5.py (208:222) duplicated block id: 3346 size: 13 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/reward_model_to_megatron.py (768:783) - toolkits/model_checkpoints_convertor/glm/checkpoint_reshaping_and_interoperability.py (519:534) duplicated block id: 3347 size: 13 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (498:510) - megatron_patch/model/starcoder/transformer.py (656:668) duplicated block id: 3348 size: 13 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/transformer.py (1002:1033) - megatron_patch/model/qwen1_5_megablocks/transformer.py (1614:1646) duplicated block id: 3349 size: 13 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/checkpoint_reshaping_and_interoperability.py (200:216) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (262:278) duplicated block id: 3350 size: 13 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (371:390) - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (337:355) duplicated block id: 3351 size: 13 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (228:244) - toolkits/model_checkpoints_convertor/starcoder/checkpoint_reshaping_and_interoperability.py (222:238) duplicated block id: 3352 size: 13 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (498:510) - megatron_patch/model/falcon/transformer.py (690:702) duplicated block id: 3353 size: 13 cleaned lines of code in 2 files: - megatron_patch/model/falcon/transformer.py (1014:1036) - megatron_patch/model/llama/transformer.py (770:792) duplicated block id: 3354 size: 13 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/glm/checkpoint_reshaping_and_interoperability.py (171:187) - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (266:282) duplicated block id: 3355 size: 13 cleaned lines of code in 2 files: - megatron_patch/model/llama/transformer.py (1057:1088) - megatron_patch/model/qwen1_5_megablocks/transformer.py (1614:1646) duplicated block id: 3356 size: 13 cleaned lines of code in 2 files: - megatron_patch/model/llava/transformer.py (731:751) - megatron_patch/model/starcoder/transformer.py (731:751) duplicated block id: 3357 size: 13 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/transformer/attention.py (394:407) - megatron_patch/model/qwen2_vl/attention.py (386:399) duplicated block id: 3358 size: 13 cleaned lines of code in 2 files: - megatron_patch/model/qwen3_moe/moe/router.py (95:107) - megatron_patch/model/qwen3_moe/moe/router.py (156:168) duplicated block id: 3359 size: 13 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/reward_model_to_megatron.py (374:392) - toolkits/model_checkpoints_convertor/falcon40b/checkpoint_reshaping_and_interoperability.py (678:696) duplicated block id: 3360 size: 13 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/language_model.py (378:394) - megatron_patch/model/galactica/language_model.py (397:413) duplicated block id: 3361 size: 13 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/transformer.py (534:550) - megatron_patch/model/glm130b/transformer.py (737:753) duplicated block id: 3362 size: 13 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/transformer.py (688:709) - megatron_patch/model/qwen_vl/transformer.py (875:895) duplicated block id: 3363 size: 13 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/language_model.py (537:553) - megatron_patch/model/qwen1_5_megablocks/language_model.py (550:566) duplicated block id: 3364 size: 13 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (678:696) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (635:653) duplicated block id: 3365 size: 13 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/glm/checkpoint_reshaping_and_interoperability.py (171:187) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (262:278) duplicated block id: 3366 size: 13 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/reward_model_to_megatron.py (200:216) - toolkits/model_checkpoints_convertor/galactica/checkpoint_reshaping_and_interoperability.py (217:233) duplicated block id: 3367 size: 13 cleaned lines of code in 2 files: - megatron_patch/model/falcon/transformer.py (757:777) - megatron_patch/model/llama3/transformer_legacy.py (710:730) duplicated block id: 3368 size: 13 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/glm/checkpoint_reshaping_and_interoperability.py (171:187) - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (272:288) duplicated block id: 3369 size: 13 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/transformer_block.py (259:274) - megatron_patch/model/qwen2/transformer_block.py (206:221) duplicated block id: 3370 size: 13 cleaned lines of code in 2 files: - megatron_patch/model/falcon/transformer.py (855:875) - megatron_patch/model/glm130b/transformer.py (688:709) duplicated block id: 3371 size: 13 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_mha.py (120:132) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_dense_and_moe_gqa.py (241:253) duplicated block id: 3372 size: 13 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2.5_vl.py (189:201) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_dense_and_moe_gqa.py (241:253) duplicated block id: 3373 size: 13 cleaned lines of code in 2 files: - megatron_patch/model/qwen_vl/transformer.py (420:445) - megatron_patch/model/starcoder/transformer.py (324:349) duplicated block id: 3374 size: 13 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/language_model.py (537:553) - megatron_patch/model/llava/language_model.py (611:627) duplicated block id: 3375 size: 13 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (1027:1041) - megatron_patch/model/qwen_vl/transformer.py (1156:1170) duplicated block id: 3376 size: 13 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/reward_model_to_megatron.py (200:216) - toolkits/model_checkpoints_convertor/chatglm/checkpoint_reshaping_and_interoperability.py (200:216) duplicated block id: 3377 size: 13 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_mha.py (209:226) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_moe.py (272:289) duplicated block id: 3378 size: 13 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/hf2te.py (306:321) - toolkits/model_checkpoints_convertor/falcon40b/checkpoint_reshaping_and_interoperability.py (433:448) duplicated block id: 3379 size: 13 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/language_model.py (378:394) - megatron_patch/model/falcon40b/language_model.py (366:382) duplicated block id: 3380 size: 13 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (1094:1112) - megatron_patch/model/qwen_vl/transformer.py (1226:1244) duplicated block id: 3381 size: 13 cleaned lines of code in 2 files: - megatron_patch/model/qwen/transformer.py (705:725) - megatron_patch/model/starcoder/transformer.py (731:751) duplicated block id: 3382 size: 13 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/galactica/checkpoint_reshaping_and_interoperability.py (89:101) - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (90:104) duplicated block id: 3383 size: 13 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/checkpoint_reshaping_and_interoperability.py (200:216) - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (272:288) duplicated block id: 3384 size: 13 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/hf2te.py (339:354) - toolkits/model_checkpoints_convertor/chatglm/checkpoint_reshaping_and_interoperability.py (447:462) duplicated block id: 3385 size: 13 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/transformer.py (688:709) - megatron_patch/model/qwen/transformer.py (821:841) duplicated block id: 3386 size: 13 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (1015:1031) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (1132:1148) duplicated block id: 3387 size: 13 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/checkpoint_reshaping_and_interoperability.py (200:216) - toolkits/model_checkpoints_convertor/chatglm/checkpoint_reshaping_and_interoperability.py (200:216) duplicated block id: 3388 size: 13 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (272:288) - toolkits/model_checkpoints_convertor/starcoder/checkpoint_reshaping_and_interoperability.py (222:238) duplicated block id: 3389 size: 13 cleaned lines of code in 2 files: - megatron_patch/model/llama/transformer.py (627:648) - megatron_patch/model/llava/transformer.py (875:895) duplicated block id: 3390 size: 13 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/language_model.py (573:587) - megatron_patch/model/llama2/language_model.py (482:496) duplicated block id: 3391 size: 13 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/chatglm/checkpoint_reshaping_and_interoperability.py (200:216) - toolkits/model_checkpoints_convertor/starcoder/checkpoint_reshaping_and_interoperability.py (222:238) duplicated block id: 3392 size: 13 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (146:164) - megatron_patch/model/llava/transformer.py (150:168) duplicated block id: 3393 size: 13 cleaned lines of code in 2 files: - megatron_patch/model/falcon/transformer.py (757:777) - megatron_patch/model/qwen/transformer.py (705:725) duplicated block id: 3394 size: 13 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (387:405) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (324:342) duplicated block id: 3395 size: 13 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/transformer.py (881:913) - megatron_patch/model/starcoder/transformer.py (1309:1341) duplicated block id: 3396 size: 13 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (230:246) - toolkits/model_checkpoints_convertor/bloom/reward_model_to_megatron.py (200:216) duplicated block id: 3397 size: 13 cleaned lines of code in 2 files: - megatron_patch/model/falcon/transformer.py (1301:1316) - megatron_patch/model/llama/transformer.py (1095:1110) duplicated block id: 3398 size: 13 cleaned lines of code in 2 files: - megatron_patch/model/qwen1_5/transformer/attention.py (369:382) - megatron_patch/model/qwen2_vl/attention.py (505:518) duplicated block id: 3399 size: 13 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (728:748) - megatron_patch/model/glm130b/transformer.py (688:709) duplicated block id: 3400 size: 13 cleaned lines of code in 2 files: - megatron_patch/model/qwen3_moe/moe/router.py (52:64) - megatron_patch/model/qwen3_moe/moe/router.py (156:168) duplicated block id: 3401 size: 13 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (728:748) - megatron_patch/model/llama/transformer.py (627:648) duplicated block id: 3402 size: 13 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/reward_model_to_megatron.py (318:330) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (719:731) duplicated block id: 3403 size: 13 cleaned lines of code in 2 files: - megatron_patch/model/qwen1_5/moe/token_dispatcher.py (126:141) - megatron_patch/model/qwen2/moe/token_dispatcher.py (125:140) duplicated block id: 3404 size: 13 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (589:604) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (593:608) duplicated block id: 3405 size: 13 cleaned lines of code in 2 files: - megatron_patch/model/bloom/language_model.py (376:392) - megatron_patch/model/chatglm/language_model.py (378:394) duplicated block id: 3406 size: 13 cleaned lines of code in 2 files: - megatron_patch/model/qwen1_5/transformer/attention.py (369:382) - megatron_patch/model/qwen2_vl/attention_vision.py (504:517) duplicated block id: 3407 size: 13 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (228:244) - toolkits/model_checkpoints_convertor/bloom/reward_model_to_megatron.py (200:216) duplicated block id: 3408 size: 13 cleaned lines of code in 2 files: - megatron_patch/model/falcon/transformer.py (757:777) - megatron_patch/model/llama2/transformer.py (732:752) duplicated block id: 3409 size: 13 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/checkpoint_reshaping_and_interoperability.py (375:393) - toolkits/model_checkpoints_convertor/falcon40b/checkpoint_reshaping_and_interoperability.py (678:696) duplicated block id: 3410 size: 13 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/transformer.py (317:341) - megatron_patch/model/llama/transformer.py (289:313) duplicated block id: 3411 size: 13 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/glm130b/checkpoint_reshaping_and_interoperability.py (171:187) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (285:301) duplicated block id: 3412 size: 13 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/checkpoint_reshaping_and_interoperability.py (769:784) - toolkits/model_checkpoints_convertor/glm/checkpoint_reshaping_and_interoperability.py (519:534) duplicated block id: 3413 size: 13 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (210:226) - toolkits/model_checkpoints_convertor/starcoder/checkpoint_reshaping_and_interoperability.py (222:238) duplicated block id: 3414 size: 13 cleaned lines of code in 2 files: - megatron_patch/model/galactica/transformer.py (889:921) - megatron_patch/model/starcoder/transformer.py (1309:1341) duplicated block id: 3415 size: 13 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/glm130b/checkpoint_reshaping_and_interoperability.py (171:187) - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (210:226) duplicated block id: 3416 size: 13 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (1094:1112) - megatron_patch/model/llama3/transformer_legacy.py (1185:1203) duplicated block id: 3417 size: 13 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (230:246) - toolkits/model_checkpoints_convertor/glm130b/checkpoint_reshaping_and_interoperability.py (171:187) duplicated block id: 3418 size: 13 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/glm130b/checkpoint_reshaping_and_interoperability.py (171:187) - toolkits/model_checkpoints_convertor/yi/checkpoint_reshaping_and_interoperability.py (215:231) duplicated block id: 3419 size: 13 cleaned lines of code in 2 files: - megatron_patch/model/qwen1_5/transformer/attention.py (191:203) - megatron_patch/model/qwen2_vl/attention_vision.py (209:221) duplicated block id: 3420 size: 13 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (228:244) - toolkits/model_checkpoints_convertor/glm130b/checkpoint_reshaping_and_interoperability.py (171:187) duplicated block id: 3421 size: 13 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/language_model.py (489:503) - megatron_patch/model/llava/language_model.py (542:556) duplicated block id: 3422 size: 13 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (146:164) - megatron_patch/model/mistral/transformer.py (150:168) duplicated block id: 3423 size: 13 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (100:114) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (118:130) duplicated block id: 3424 size: 13 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/transformer/attention.py (504:517) - megatron_patch/model/qwen1_5/transformer/attention.py (369:382) duplicated block id: 3425 size: 13 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/transformer.py (688:709) - megatron_patch/model/llava/transformer.py (875:895) duplicated block id: 3426 size: 13 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (693:708) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (593:608) duplicated block id: 3427 size: 13 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (1027:1041) - megatron_patch/model/llama3/transformer_legacy.py (1106:1120) duplicated block id: 3428 size: 13 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer/attention.py (369:382) - megatron_patch/model/qwen2_vl/attention.py (505:518) duplicated block id: 3429 size: 13 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (1027:1041) - megatron_patch/model/llama2/transformer.py (1157:1171) duplicated block id: 3430 size: 13 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (1094:1112) - megatron_patch/model/llama2/transformer.py (1227:1245) duplicated block id: 3431 size: 13 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/transformer.py (1040:1055) - megatron_patch/model/llama/transformer.py (1095:1110) duplicated block id: 3432 size: 13 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/checkpoint_reshaping_and_interoperability.py (200:216) - toolkits/model_checkpoints_convertor/galactica/checkpoint_reshaping_and_interoperability.py (217:233) duplicated block id: 3433 size: 13 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (262:278) - toolkits/model_checkpoints_convertor/starcoder/checkpoint_reshaping_and_interoperability.py (222:238) duplicated block id: 3434 size: 13 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (1094:1112) - megatron_patch/model/llava/transformer.py (1226:1244) duplicated block id: 3435 size: 13 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (536:549) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (454:468) duplicated block id: 3436 size: 13 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/hf2te.py (424:440) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (624:640) duplicated block id: 3437 size: 13 cleaned lines of code in 2 files: - megatron_patch/model/mixtral_bak/transformer/attention.py (165:177) - megatron_patch/model/qwen2_vl/attention.py (210:222) duplicated block id: 3438 size: 13 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/reward_model_to_megatron.py (200:216) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (285:301) duplicated block id: 3439 size: 13 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (285:301) - toolkits/model_checkpoints_convertor/starcoder/checkpoint_reshaping_and_interoperability.py (222:238) duplicated block id: 3440 size: 13 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (589:607) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (635:653) duplicated block id: 3441 size: 13 cleaned lines of code in 2 files: - megatron_patch/model/bloom/language_model.py (449:461) - megatron_patch/model/falcon40b/language_model.py (526:538) duplicated block id: 3442 size: 13 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (228:244) - toolkits/model_checkpoints_convertor/glm/checkpoint_reshaping_and_interoperability.py (171:187) duplicated block id: 3443 size: 13 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (1027:1041) - megatron_patch/model/llava/transformer.py (1156:1170) duplicated block id: 3444 size: 13 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (839:859) - megatron_patch/model/glm130b/transformer.py (688:709) duplicated block id: 3445 size: 13 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (805:821) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (740:756) duplicated block id: 3446 size: 13 cleaned lines of code in 2 files: - megatron_patch/model/llava/transformer.py (420:445) - megatron_patch/model/starcoder/transformer.py (324:349) duplicated block id: 3447 size: 13 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/moe/shared_experts.py (158:170) - megatron_patch/model/mixtral/moe/experts.py (646:658) duplicated block id: 3448 size: 13 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/galactica/checkpoint_reshaping_and_interoperability.py (217:233) - toolkits/model_checkpoints_convertor/glm130b/checkpoint_reshaping_and_interoperability.py (171:187) duplicated block id: 3449 size: 13 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/hf2te.py (33:45) - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (43:56) duplicated block id: 3450 size: 13 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (431:456) - megatron_patch/model/starcoder/transformer.py (324:349) duplicated block id: 3451 size: 13 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon40b/checkpoint_reshaping_and_interoperability.py (90:102) - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (90:104) duplicated block id: 3452 size: 13 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/language_model.py (573:587) - megatron_patch/model/mistral/language_model.py (484:498) duplicated block id: 3453 size: 13 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/language_model.py (544:560) - megatron_patch/model/glm130b/language_model.py (537:553) duplicated block id: 3454 size: 13 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llava/hf2mcore_llava.py (436:450) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_moe.py (406:420) duplicated block id: 3455 size: 13 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/hf2te.py (367:382) - toolkits/model_checkpoints_convertor/chatglm/checkpoint_reshaping_and_interoperability.py (447:462) duplicated block id: 3456 size: 13 cleaned lines of code in 2 files: - megatron_patch/model/bloom/language_model.py (449:461) - megatron_patch/model/falcon/language_model.py (526:538) duplicated block id: 3457 size: 13 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (228:244) - toolkits/model_checkpoints_convertor/bloom/checkpoint_reshaping_and_interoperability.py (200:216) duplicated block id: 3458 size: 13 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (146:164) - megatron_patch/model/baichuan2/transformer.py (145:163) duplicated block id: 3459 size: 13 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/language_model.py (537:553) - megatron_patch/model/llama2/language_model.py (551:567) duplicated block id: 3460 size: 13 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (322:347) - megatron_patch/model/starcoder/transformer.py (324:349) duplicated block id: 3461 size: 13 cleaned lines of code in 2 files: - megatron_patch/model/falcon/transformer.py (618:632) - megatron_patch/model/starcoder/transformer.py (567:580) duplicated block id: 3462 size: 13 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/transformer.py (855:878) - megatron_patch/model/falcon40b/transformer.py (972:995) duplicated block id: 3463 size: 13 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/transformer.py (193:209) - megatron_patch/model/starcoder/transformer.py (198:214) duplicated block id: 3464 size: 13 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer/attention.py (191:203) - megatron_patch/model/qwen2_vl/attention_vision.py (209:221) duplicated block id: 3465 size: 13 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (100:114) - toolkits/model_checkpoints_convertor/yi/checkpoint_reshaping_and_interoperability.py (90:102) duplicated block id: 3466 size: 13 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (90:104) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (118:130) duplicated block id: 3467 size: 13 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (189:205) - toolkits/model_checkpoints_convertor/starcoder/checkpoint_reshaping_and_interoperability.py (222:238) duplicated block id: 3468 size: 13 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/reward_model_to_megatron.py (318:330) - toolkits/model_checkpoints_convertor/starcoder/checkpoint_reshaping_and_interoperability.py (340:352) duplicated block id: 3469 size: 13 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/transformer.py (688:709) - megatron_patch/model/llama2/transformer.py (876:896) duplicated block id: 3470 size: 13 cleaned lines of code in 2 files: - megatron_patch/model/falcon/transformer.py (690:702) - megatron_patch/model/llama/transformer.py (497:509) duplicated block id: 3471 size: 13 cleaned lines of code in 2 files: - megatron_patch/model/llama/language_model.py (684:699) - megatron_patch/model/llama3/language_model.py (602:617) duplicated block id: 3472 size: 13 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/glm/checkpoint_reshaping_and_interoperability.py (171:187) - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (210:226) duplicated block id: 3473 size: 13 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (589:604) - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (693:708) duplicated block id: 3474 size: 13 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (542:559) - toolkits/model_checkpoints_convertor/yi/checkpoint_reshaping_and_interoperability.py (562:579) duplicated block id: 3475 size: 13 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/chatglm/checkpoint_reshaping_and_interoperability.py (200:216) - toolkits/model_checkpoints_convertor/glm130b/checkpoint_reshaping_and_interoperability.py (171:187) duplicated block id: 3476 size: 13 cleaned lines of code in 2 files: - megatron_patch/model/galactica/language_model.py (702:717) - megatron_patch/model/llama3/language_model.py (602:617) duplicated block id: 3477 size: 13 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/galactica/checkpoint_reshaping_and_interoperability.py (89:101) - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (100:114) duplicated block id: 3478 size: 13 cleaned lines of code in 2 files: - megatron_patch/model/galactica/transformer.py (93:108) - megatron_patch/model/llama/transformer.py (127:142) duplicated block id: 3479 size: 13 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (100:114) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (117:129) duplicated block id: 3480 size: 13 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (553:571) - megatron_patch/model/starcoder/transformer.py (750:769) duplicated block id: 3481 size: 13 cleaned lines of code in 2 files: - megatron_patch/model/falcon/transformer.py (855:875) - megatron_patch/model/llama/transformer.py (627:648) duplicated block id: 3482 size: 13 cleaned lines of code in 2 files: - megatron_patch/model/falcon/transformer.py (757:777) - megatron_patch/model/mistral/transformer.py (730:750) duplicated block id: 3483 size: 13 cleaned lines of code in 2 files: - megatron_patch/model/bloom/language_model.py (449:461) - megatron_patch/model/galactica/language_model.py (552:564) duplicated block id: 3484 size: 13 cleaned lines of code in 2 files: - megatron_patch/model/qwen2/transformer/attention.py (187:199) - megatron_patch/model/qwen2_vl/attention_vision.py (209:221) duplicated block id: 3485 size: 13 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/language_model.py (503:515) - megatron_patch/model/galactica/language_model.py (552:564) duplicated block id: 3486 size: 13 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/glm/checkpoint_reshaping_and_interoperability.py (171:187) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (189:205) duplicated block id: 3487 size: 13 cleaned lines of code in 2 files: - megatron_patch/model/qwen2/transformer/attention.py (187:199) - megatron_patch/model/qwen2_vl/attention.py (210:222) duplicated block id: 3488 size: 13 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/chatglm/checkpoint_reshaping_and_interoperability.py (79:91) - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (100:114) duplicated block id: 3489 size: 13 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (146:164) - megatron_patch/model/llama3/transformer_legacy.py (154:172) duplicated block id: 3490 size: 13 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/chatglm/checkpoint_reshaping_and_interoperability.py (79:91) - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (90:104) duplicated block id: 3491 size: 13 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (306:329) - megatron_patch/model/glm130b/transformer.py (317:341) duplicated block id: 3492 size: 13 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/reward_model_to_megatron.py (200:216) - toolkits/model_checkpoints_convertor/yi/checkpoint_reshaping_and_interoperability.py (215:231) duplicated block id: 3493 size: 13 cleaned lines of code in 2 files: - megatron_patch/model/falcon/transformer.py (1233:1256) - megatron_patch/model/galactica/transformer.py (863:886) duplicated block id: 3494 size: 13 cleaned lines of code in 2 files: - megatron_patch/model/bloom/language_model.py (449:461) - megatron_patch/model/llama/language_model.py (526:538) duplicated block id: 3495 size: 13 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/hf2te.py (334:349) - toolkits/model_checkpoints_convertor/falcon40b/checkpoint_reshaping_and_interoperability.py (433:448) duplicated block id: 3496 size: 13 cleaned lines of code in 2 files: - megatron_patch/model/mixtral_bak/moe/token_dispatcher.py (24:64) - megatron_patch/model/qwen1_5/moe/token_dispatcher.py (25:65) duplicated block id: 3497 size: 13 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (146:164) - megatron_patch/model/qwen/transformer.py (144:162) duplicated block id: 3498 size: 13 cleaned lines of code in 2 files: - megatron_patch/model/falcon/transformer.py (757:777) - megatron_patch/model/qwen1_5_megablocks/transformer.py (623:643) duplicated block id: 3499 size: 13 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore.py (289:301) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_dense_and_moe_gqa.py (241:253) duplicated block id: 3500 size: 13 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (1027:1041) - megatron_patch/model/baichuan2/transformer.py (1159:1173) duplicated block id: 3501 size: 13 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/layer_specs.py (147:162) - megatron_patch/model/qwen2/layer_specs.py (112:126) duplicated block id: 3502 size: 13 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_dense_and_moe_gqa.py (241:253) - toolkits/model_checkpoints_convertor/qwen/hf2megablocks_qwen1.5.py (283:295) duplicated block id: 3503 size: 13 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_dense_and_moe_gqa.py (241:253) - toolkits/model_checkpoints_convertor/qwen/hf2megablocks_qwen1.5.py (322:334) duplicated block id: 3504 size: 13 cleaned lines of code in 2 files: - megatron_patch/model/mixtral_bak/transformer/mlp.py (72:88) - megatron_patch/model/qwen2/transformer/mlp.py (109:125) duplicated block id: 3505 size: 13 cleaned lines of code in 2 files: - megatron_patch/model/falcon/transformer.py (150:165) - megatron_patch/model/llama/transformer.py (127:142) duplicated block id: 3506 size: 13 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/model.py (65:81) - megatron_patch/model/qwen2/model.py (68:84) duplicated block id: 3507 size: 13 cleaned lines of code in 2 files: - megatron_patch/model/llama/transformer.py (627:648) - megatron_patch/model/llama2/transformer.py (876:896) duplicated block id: 3508 size: 13 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/chatglm/checkpoint_reshaping_and_interoperability.py (200:216) - toolkits/model_checkpoints_convertor/glm/checkpoint_reshaping_and_interoperability.py (171:187) duplicated block id: 3509 size: 13 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/reward_model_to_megatron.py (200:216) - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (210:226) duplicated block id: 3510 size: 13 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/hf2te.py (34:46) - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (43:56) duplicated block id: 3511 size: 13 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/hf2te.py (334:349) - toolkits/model_checkpoints_convertor/falcon/checkpoint_reshaping_and_interoperability.py (444:459) duplicated block id: 3512 size: 13 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/layer_specs.py (128:142) - megatron_patch/model/qwen1_5/layer_specs.py (85:100) duplicated block id: 3513 size: 13 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer_legacy.py (710:730) - megatron_patch/model/starcoder/transformer.py (731:751) duplicated block id: 3514 size: 13 cleaned lines of code in 2 files: - megatron_patch/model/qwen/transformer.py (413:438) - megatron_patch/model/starcoder/transformer.py (324:349) duplicated block id: 3515 size: 13 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/deepseek/hf2mcore_deepseek_v2_moe.py (497:510) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_moe.py (646:659) duplicated block id: 3516 size: 13 cleaned lines of code in 2 files: - megatron_patch/model/galactica/transformer.py (251:275) - megatron_patch/model/glm130b/transformer.py (317:341) duplicated block id: 3517 size: 13 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/glm130b/checkpoint_reshaping_and_interoperability.py (171:187) - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (266:282) duplicated block id: 3518 size: 13 cleaned lines of code in 2 files: - megatron_patch/model/mixtral_bak/transformer/mlp.py (72:88) - megatron_patch/model/qwen1_5/transformer/mlp.py (110:126) duplicated block id: 3519 size: 13 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/language_model.py (503:515) - megatron_patch/model/falcon40b/language_model.py (526:538) duplicated block id: 3520 size: 13 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/transformer.py (688:709) - megatron_patch/model/qwen1_5_megablocks/transformer.py (767:787) duplicated block id: 3521 size: 13 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llava/hf2mcore_llava.py (436:450) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_mha.py (297:311) duplicated block id: 3522 size: 13 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/checkpoint_reshaping_and_interoperability.py (200:216) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (285:301) duplicated block id: 3523 size: 13 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/checkpoint_reshaping_and_interoperability.py (200:216) - toolkits/model_checkpoints_convertor/yi/checkpoint_reshaping_and_interoperability.py (215:231) duplicated block id: 3524 size: 13 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/transformer/attention.py (394:407) - megatron_patch/model/qwen2_vl/attention_vision.py (385:398) duplicated block id: 3525 size: 13 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (89:101) - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (100:114) duplicated block id: 3526 size: 13 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon40b/checkpoint_reshaping_and_interoperability.py (90:102) - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (100:114) duplicated block id: 3527 size: 13 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_dense_and_moe_gqa.py (241:253) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_vl.py (189:201) duplicated block id: 3528 size: 13 cleaned lines of code in 2 files: - megatron_patch/model/falcon/transformer.py (1263:1294) - megatron_patch/model/qwen1_5_megablocks/transformer.py (1614:1646) duplicated block id: 3529 size: 13 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/hf2te.py (68:81) - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (90:104) duplicated block id: 3530 size: 13 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (1027:1041) - megatron_patch/model/qwen/transformer.py (1102:1116) duplicated block id: 3531 size: 13 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer/attention.py (369:382) - megatron_patch/model/mixtral/transformer/attention.py (504:517) duplicated block id: 3532 size: 13 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/reward_model_to_megatron.py (200:216) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (189:205) duplicated block id: 3533 size: 13 cleaned lines of code in 2 files: - megatron_patch/model/llama3/language_model.py (473:487) - megatron_patch/model/llava/language_model.py (542:556) duplicated block id: 3534 size: 13 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/transformer.py (674:690) - megatron_patch/model/glm130b/transformer.py (737:753) duplicated block id: 3535 size: 13 cleaned lines of code in 2 files: - megatron_patch/model/falcon/transformer.py (757:777) - megatron_patch/model/qwen_vl/transformer.py (731:751) duplicated block id: 3536 size: 13 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (89:101) - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (90:104) duplicated block id: 3537 size: 13 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/reward_model_to_megatron.py (200:216) - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (266:282) duplicated block id: 3538 size: 13 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/transformer.py (143:158) - megatron_patch/model/llama/transformer.py (127:142) duplicated block id: 3539 size: 13 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer/attention.py (191:203) - megatron_patch/model/qwen2_vl/attention.py (210:222) duplicated block id: 3540 size: 13 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/checkpoint_reshaping_and_interoperability.py (200:216) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (189:205) duplicated block id: 3541 size: 13 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/glm/checkpoint_reshaping_and_interoperability.py (171:187) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (285:301) duplicated block id: 3542 size: 13 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (1165:1196) - megatron_patch/model/qwen1_5_megablocks/transformer.py (1614:1646) duplicated block id: 3543 size: 13 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/language_model.py (503:515) - megatron_patch/model/llama/language_model.py (526:538) duplicated block id: 3544 size: 13 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (116:131) - megatron_patch/model/llama/transformer.py (127:142) duplicated block id: 3545 size: 13 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (662:678) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (740:756) duplicated block id: 3546 size: 13 cleaned lines of code in 2 files: - megatron_patch/model/mistral/transformer.py (730:750) - megatron_patch/model/starcoder/transformer.py (731:751) duplicated block id: 3547 size: 13 cleaned lines of code in 2 files: - megatron_patch/model/llama/transformer.py (627:648) - megatron_patch/model/qwen1_5_megablocks/transformer.py (767:787) duplicated block id: 3548 size: 13 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/model.py (65:81) - megatron_patch/model/qwen1_5/model.py (67:83) duplicated block id: 3549 size: 13 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (561:579) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (635:653) duplicated block id: 3550 size: 13 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_mha.py (209:226) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_mha_to_moe.py (123:140) duplicated block id: 3551 size: 13 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (113:125) - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (90:104) duplicated block id: 3552 size: 13 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/reward_model_to_megatron.py (318:330) - toolkits/model_checkpoints_convertor/falcon40b/checkpoint_reshaping_and_interoperability.py (618:630) duplicated block id: 3553 size: 13 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (839:859) - megatron_patch/model/llama/transformer.py (627:648) duplicated block id: 3554 size: 13 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/language_model.py (537:553) - megatron_patch/model/qwen_vl/language_model.py (580:596) duplicated block id: 3555 size: 13 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/transformer.py (546:558) - megatron_patch/model/starcoder/transformer.py (656:668) duplicated block id: 3556 size: 13 cleaned lines of code in 2 files: - megatron_patch/model/llama/transformer.py (627:648) - megatron_patch/model/mistral/transformer.py (873:893) duplicated block id: 3557 size: 13 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (1094:1112) - megatron_patch/model/mistral/transformer.py (1224:1242) duplicated block id: 3558 size: 13 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (90:104) - toolkits/model_checkpoints_convertor/yi/checkpoint_reshaping_and_interoperability.py (90:102) duplicated block id: 3559 size: 13 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/chatglm/checkpoint_reshaping_and_interoperability.py (298:322) - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (384:408) duplicated block id: 3560 size: 13 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/galactica/checkpoint_reshaping_and_interoperability.py (217:233) - toolkits/model_checkpoints_convertor/glm/checkpoint_reshaping_and_interoperability.py (171:187) duplicated block id: 3561 size: 13 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (441:460) - toolkits/model_checkpoints_convertor/yi/checkpoint_reshaping_and_interoperability.py (374:392) duplicated block id: 3562 size: 13 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/layer_specs.py (132:146) - megatron_patch/model/qwen2_moe/layer_specs.py (232:246) duplicated block id: 3563 size: 13 cleaned lines of code in 2 files: - megatron_patch/model/bloom/language_model.py (139:165) - megatron_patch/model/galactica/language_model.py (166:192) duplicated block id: 3564 size: 13 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (457:476) - toolkits/model_checkpoints_convertor/yi/checkpoint_reshaping_and_interoperability.py (374:392) duplicated block id: 3565 size: 13 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (386:404) - toolkits/model_checkpoints_convertor/yi/checkpoint_reshaping_and_interoperability.py (374:392) duplicated block id: 3566 size: 13 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (1027:1041) - megatron_patch/model/qwen1_5_megablocks/transformer.py (1041:1055) duplicated block id: 3567 size: 13 cleaned lines of code in 2 files: - megatron_patch/model/falcon/transformer.py (757:777) - megatron_patch/model/llava/transformer.py (731:751) duplicated block id: 3568 size: 13 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/language_model.py (503:515) - megatron_patch/model/falcon/language_model.py (526:538) duplicated block id: 3569 size: 13 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/checkpoint_reshaping_and_interoperability.py (200:216) - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (210:226) duplicated block id: 3570 size: 13 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon/checkpoint_reshaping_and_interoperability.py (90:102) - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (100:114) duplicated block id: 3571 size: 13 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (635:653) - toolkits/model_checkpoints_convertor/yi/checkpoint_reshaping_and_interoperability.py (581:600) duplicated block id: 3572 size: 13 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/glm130b/checkpoint_reshaping_and_interoperability.py (171:187) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (262:278) duplicated block id: 3573 size: 13 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/moe/experts.py (645:657) - megatron_patch/model/deepseek_v2/moe/shared_experts.py (158:170) duplicated block id: 3574 size: 13 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/language_model.py (537:553) - megatron_patch/model/qwen/language_model.py (533:549) duplicated block id: 3575 size: 13 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/reward_model_to_megatron.py (374:392) - toolkits/model_checkpoints_convertor/falcon/checkpoint_reshaping_and_interoperability.py (693:711) duplicated block id: 3576 size: 13 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (90:104) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (117:129) duplicated block id: 3577 size: 13 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/starcoder/checkpoint_reshaping_and_interoperability.py (222:238) - toolkits/model_checkpoints_convertor/yi/checkpoint_reshaping_and_interoperability.py (215:231) duplicated block id: 3578 size: 13 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (635:653) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (673:691) duplicated block id: 3579 size: 13 cleaned lines of code in 2 files: - megatron_patch/model/falcon/transformer.py (690:702) - megatron_patch/model/falcon40b/transformer.py (546:558) duplicated block id: 3580 size: 13 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer/attention.py (369:382) - megatron_patch/model/qwen2_vl/attention_vision.py (504:517) duplicated block id: 3581 size: 13 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/hf2te.py (69:82) - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (90:104) duplicated block id: 3582 size: 13 cleaned lines of code in 2 files: - megatron_patch/model/galactica/transformer.py (566:582) - megatron_patch/model/glm130b/transformer.py (737:753) duplicated block id: 3583 size: 13 cleaned lines of code in 2 files: - megatron_patch/model/llama/transformer.py (497:509) - megatron_patch/model/starcoder/transformer.py (656:668) duplicated block id: 3584 size: 13 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_dense_and_moe_gqa.py (241:253) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_vl.py (140:152) duplicated block id: 3585 size: 13 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/checkpoint_reshaping_and_interoperability.py (200:216) - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (266:282) duplicated block id: 3586 size: 13 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/glm130b/checkpoint_reshaping_and_interoperability.py (171:187) - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (272:288) duplicated block id: 3587 size: 13 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/chatglm/checkpoint_reshaping_and_interoperability.py (518:536) - toolkits/model_checkpoints_convertor/falcon/checkpoint_reshaping_and_interoperability.py (530:546) duplicated block id: 3588 size: 12 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/language_model.py (75:88) - megatron_patch/model/baichuan2/language_model.py (32:45) duplicated block id: 3589 size: 12 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (243:256) - toolkits/model_checkpoints_convertor/starcoder/checkpoint_reshaping_and_interoperability.py (254:267) duplicated block id: 3590 size: 12 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon40b/checkpoint_reshaping_and_interoperability.py (757:768) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (889:902) duplicated block id: 3591 size: 12 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/galactica/checkpoint_reshaping_and_interoperability.py (139:151) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (90:103) duplicated block id: 3592 size: 12 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (554:566) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (680:692) duplicated block id: 3593 size: 12 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/layer_specs.py (105:119) - megatron_patch/model/mixtral/layer_specs.py (138:153) duplicated block id: 3594 size: 12 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon40b/checkpoint_reshaping_and_interoperability.py (139:150) - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (101:113) duplicated block id: 3595 size: 12 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/glm/checkpoint_reshaping_and_interoperability.py (499:514) - toolkits/model_checkpoints_convertor/starcoder/checkpoint_reshaping_and_interoperability.py (746:761) duplicated block id: 3596 size: 12 cleaned lines of code in 2 files: - megatron_patch/model/falcon/language_model.py (394:408) - megatron_patch/model/llama2/language_model.py (413:427) duplicated block id: 3597 size: 12 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/reward_model_to_megatron.py (232:245) - toolkits/model_checkpoints_convertor/galactica/checkpoint_reshaping_and_interoperability.py (250:263) duplicated block id: 3598 size: 12 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/transformer.py (298:322) - megatron_patch/model/mistral/transformer.py (424:448) duplicated block id: 3599 size: 12 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/checkpoint_reshaping_and_interoperability.py (232:245) - toolkits/model_checkpoints_convertor/falcon/checkpoint_reshaping_and_interoperability.py (234:247) duplicated block id: 3600 size: 12 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/transformer.py (881:911) - megatron_patch/model/glm130b/transformer.py (1199:1229) duplicated block id: 3601 size: 12 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/deepseek/hf2mcore_deepseek_v2_moe.py (68:82) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2.5_vl.py (72:86) duplicated block id: 3602 size: 12 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/checkpoint_reshaping_and_interoperability.py (354:365) - toolkits/model_checkpoints_convertor/starcoder/checkpoint_reshaping_and_interoperability.py (377:388) duplicated block id: 3603 size: 12 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (1134:1155) - megatron_patch/model/falcon40b/transformer.py (970:992) duplicated block id: 3604 size: 12 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (1425:1436) - megatron_patch/model/llama3/transformer_legacy.py (1522:1533) duplicated block id: 3605 size: 12 cleaned lines of code in 2 files: - toolkits/pretrain_data_preprocessing/clean_raw_text.py (12:25) - toolkits/pretrain_data_preprocessing/preprocess_wudao2.py (11:24) duplicated block id: 3606 size: 12 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (901:913) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (833:845) duplicated block id: 3607 size: 12 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (111:123) - toolkits/model_checkpoints_convertor/starcoder/checkpoint_reshaping_and_interoperability.py (117:128) duplicated block id: 3608 size: 12 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/transformer_layer.py (377:400) - megatron_patch/model/qwen2/transformer_layer.py (253:265) duplicated block id: 3609 size: 12 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (957:977) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (981:1001) duplicated block id: 3610 size: 12 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/reward_model_to_megatron.py (524:536) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (1116:1128) duplicated block id: 3611 size: 12 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (748:760) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (554:566) duplicated block id: 3612 size: 12 cleaned lines of code in 2 files: - megatron_patch/model/llama/language_model.py (394:408) - megatron_patch/model/mistral/language_model.py (415:429) duplicated block id: 3613 size: 12 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (139:151) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (90:103) duplicated block id: 3614 size: 12 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/multi_latent_attention.py (98:109) - megatron_patch/model/llama3/transformer/attention.py (104:115) duplicated block id: 3615 size: 12 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (139:151) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (90:103) duplicated block id: 3616 size: 12 cleaned lines of code in 2 files: - megatron_patch/model/bloom/language_model.py (518:533) - megatron_patch/model/falcon40b/language_model.py (605:620) duplicated block id: 3617 size: 12 cleaned lines of code in 2 files: - megatron_patch/model/falcon/language_model.py (394:408) - megatron_patch/model/qwen1_5_megablocks/language_model.py (412:426) duplicated block id: 3618 size: 12 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore.py (603:615) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (833:845) duplicated block id: 3619 size: 12 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (816:829) - toolkits/model_checkpoints_convertor/falcon/checkpoint_reshaping_and_interoperability.py (773:784) duplicated block id: 3620 size: 12 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (295:308) - toolkits/model_checkpoints_convertor/starcoder/checkpoint_reshaping_and_interoperability.py (254:267) duplicated block id: 3621 size: 12 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/checkpoint_reshaping_and_interoperability.py (232:245) - toolkits/model_checkpoints_convertor/galactica/checkpoint_reshaping_and_interoperability.py (250:263) duplicated block id: 3622 size: 12 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer/mlp.py (118:131) - megatron_patch/model/qwen2/moe/experts.py (290:303) duplicated block id: 3623 size: 12 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/model.py (22:69) - megatron_patch/model/llama3/model.py (34:63) duplicated block id: 3624 size: 12 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/language_model.py (394:408) - megatron_patch/model/qwen1_5_megablocks/language_model.py (412:426) duplicated block id: 3625 size: 12 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/moe/shared_experts.py (167:180) - megatron_patch/model/llama3/transformer/mlp.py (118:131) duplicated block id: 3626 size: 12 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/transformer_config.py (13:24) - megatron_patch/model/qwen2_moe/transformer_config.py (33:44) duplicated block id: 3627 size: 12 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/hf2te.py (508:520) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (554:566) duplicated block id: 3628 size: 12 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (1003:1015) - toolkits/model_checkpoints_convertor/starcoder/checkpoint_reshaping_and_interoperability.py (540:552) duplicated block id: 3629 size: 12 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (261:274) - toolkits/model_checkpoints_convertor/starcoder/checkpoint_reshaping_and_interoperability.py (254:267) duplicated block id: 3630 size: 12 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/checkpoint_reshaping_and_interoperability.py (232:245) - toolkits/model_checkpoints_convertor/falcon40b/checkpoint_reshaping_and_interoperability.py (232:245) duplicated block id: 3631 size: 12 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (101:113) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (167:178) duplicated block id: 3632 size: 12 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/glm130b/checkpoint_reshaping_and_interoperability.py (203:216) - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (305:318) duplicated block id: 3633 size: 12 cleaned lines of code in 2 files: - megatron_patch/model/llama/transformer.py (697:710) - megatron_patch/model/starcoder/transformer.py (865:878) duplicated block id: 3634 size: 12 cleaned lines of code in 2 files: - megatron_patch/model/llama3_1/model.py (46:75) - megatron_patch/model/mixtral/model.py (31:60) duplicated block id: 3635 size: 12 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/galactica/checkpoint_reshaping_and_interoperability.py (138:149) - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (101:113) duplicated block id: 3636 size: 12 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (127:143) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (103:119) duplicated block id: 3637 size: 12 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (263:276) - toolkits/model_checkpoints_convertor/bloom/checkpoint_reshaping_and_interoperability.py (232:245) duplicated block id: 3638 size: 12 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/reward_model_to_megatron.py (354:365) - toolkits/model_checkpoints_convertor/starcoder/checkpoint_reshaping_and_interoperability.py (377:388) duplicated block id: 3639 size: 12 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon/checkpoint_reshaping_and_interoperability.py (139:150) - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (111:123) duplicated block id: 3640 size: 12 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2.5_vl.py (72:86) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_moe.py (136:150) duplicated block id: 3641 size: 12 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/moe/experts.py (207:225) - megatron_patch/model/qwen2/moe/experts.py (384:397) duplicated block id: 3642 size: 12 cleaned lines of code in 2 files: - megatron_patch/model/llama3/model.py (34:63) - megatron_patch/model/qwen2/model.py (33:62) duplicated block id: 3643 size: 12 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore.py (233:248) - toolkits/model_checkpoints_convertor/qwen/hf2megablocks_qwen1.5.py (228:242) duplicated block id: 3644 size: 12 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/checkpoint_reshaping_and_interoperability.py (749:764) - toolkits/model_checkpoints_convertor/starcoder/checkpoint_reshaping_and_interoperability.py (746:761) duplicated block id: 3645 size: 12 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/hf2te.py (106:118) - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (101:113) duplicated block id: 3646 size: 12 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/reward_model_to_megatron.py (232:245) - toolkits/model_checkpoints_convertor/falcon40b/checkpoint_reshaping_and_interoperability.py (232:245) duplicated block id: 3647 size: 12 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/deepseek/hf2mcore_deepseek_v2_moe.py (68:82) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_vl.py (72:86) duplicated block id: 3648 size: 12 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (1624:1656) - megatron_patch/model/galactica/transformer.py (889:919) duplicated block id: 3649 size: 12 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (807:821) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (828:842) duplicated block id: 3650 size: 12 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/chatglm/checkpoint_reshaping_and_interoperability.py (121:132) - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (111:123) duplicated block id: 3651 size: 12 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/reward_model_to_megatron.py (524:536) - toolkits/model_checkpoints_convertor/llama/hf2mcore.py (603:615) duplicated block id: 3652 size: 12 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (138:149) - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (111:123) duplicated block id: 3653 size: 12 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/language_model.py (510:523) - megatron_patch/model/glm130b/language_model.py (478:491) duplicated block id: 3654 size: 12 cleaned lines of code in 2 files: - megatron_patch/data/utils.py (119:132) - megatron_patch/data/utils.py (283:296) duplicated block id: 3655 size: 12 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/moe/experts.py (206:224) - megatron_patch/model/deepseek_v2/moe/experts.py (789:802) duplicated block id: 3656 size: 12 cleaned lines of code in 2 files: - megatron_patch/data/utils.py (119:132) - megatron_patch/data/utils.py (257:269) duplicated block id: 3657 size: 12 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/moe/shared_experts.py (167:180) - megatron_patch/model/qwen1_5/transformer/mlp.py (134:147) duplicated block id: 3658 size: 12 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore.py (603:615) - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (811:823) duplicated block id: 3659 size: 12 cleaned lines of code in 2 files: - megatron_patch/model/qwen1_5/model.py (34:63) - megatron_patch/model/qwen2/model.py (33:62) duplicated block id: 3660 size: 12 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/multi_latent_attention.py (98:109) - megatron_patch/model/qwen1_5/transformer/attention.py (104:115) duplicated block id: 3661 size: 12 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/hf2te.py (107:119) - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (101:113) duplicated block id: 3662 size: 12 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/transformer.py (881:911) - megatron_patch/model/qwen/transformer.py (1696:1728) duplicated block id: 3663 size: 12 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/model.py (22:69) - megatron_patch/model/mixtral/model.py (31:60) duplicated block id: 3664 size: 12 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (337:354) - toolkits/model_checkpoints_convertor/yi/checkpoint_reshaping_and_interoperability.py (375:392) duplicated block id: 3665 size: 12 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (435:459) - megatron_patch/model/glm130b/transformer.py (298:322) duplicated block id: 3666 size: 12 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/language_model.py (31:44) - megatron_patch/model/falcon40b/language_model.py (29:42) duplicated block id: 3667 size: 12 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (263:276) - toolkits/model_checkpoints_convertor/glm/checkpoint_reshaping_and_interoperability.py (203:216) duplicated block id: 3668 size: 12 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (101:113) - toolkits/model_checkpoints_convertor/starcoder/checkpoint_reshaping_and_interoperability.py (117:128) duplicated block id: 3669 size: 12 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/reward_model_to_megatron.py (748:763) - toolkits/model_checkpoints_convertor/starcoder/checkpoint_reshaping_and_interoperability.py (746:761) duplicated block id: 3670 size: 12 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (305:318) - toolkits/model_checkpoints_convertor/starcoder/checkpoint_reshaping_and_interoperability.py (254:267) duplicated block id: 3671 size: 12 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/language_model.py (471:485) - megatron_patch/model/llama/language_model.py (394:408) duplicated block id: 3672 size: 12 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llava/hf2mcore_llava.py (453:465) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_vl.py (220:232) duplicated block id: 3673 size: 12 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/reward_model_to_megatron.py (232:245) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (318:331) duplicated block id: 3674 size: 12 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (606:618) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (554:566) duplicated block id: 3675 size: 12 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (901:913) - toolkits/model_checkpoints_convertor/qwen/hf2megablocks_qwen1.5.py (613:625) duplicated block id: 3676 size: 12 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/checkpoint_reshaping_and_interoperability.py (232:245) - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (243:256) duplicated block id: 3677 size: 12 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/glm/checkpoint_reshaping_and_interoperability.py (203:216) - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (243:256) duplicated block id: 3678 size: 12 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (90:103) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (168:180) duplicated block id: 3679 size: 12 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/checkpoint_reshaping_and_interoperability.py (581:600) - toolkits/model_checkpoints_convertor/starcoder/checkpoint_reshaping_and_interoperability.py (596:614) duplicated block id: 3680 size: 12 cleaned lines of code in 2 files: - megatron_patch/model/falcon/transformer.py (855:872) - megatron_patch/model/llama3/transformer_legacy.py (827:844) duplicated block id: 3681 size: 12 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon/checkpoint_reshaping_and_interoperability.py (234:247) - toolkits/model_checkpoints_convertor/starcoder/checkpoint_reshaping_and_interoperability.py (254:267) duplicated block id: 3682 size: 12 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/transformer.py (298:322) - megatron_patch/model/qwen1_5_megablocks/transformer.py (314:338) duplicated block id: 3683 size: 12 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/hf2te.py (480:492) - toolkits/model_checkpoints_convertor/galactica/checkpoint_reshaping_and_interoperability.py (631:643) duplicated block id: 3684 size: 12 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/reward_model_to_megatron.py (232:245) - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (243:256) duplicated block id: 3685 size: 12 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2megablocks_qwen1.5.py (613:625) - toolkits/model_checkpoints_convertor/starcoder/checkpoint_reshaping_and_interoperability.py (540:552) duplicated block id: 3686 size: 12 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (651:664) - megatron_patch/model/llama/transformer.py (697:710) duplicated block id: 3687 size: 12 cleaned lines of code in 2 files: - megatron_patch/model/falcon/transformer.py (926:951) - megatron_patch/model/starcoder/transformer.py (865:878) duplicated block id: 3688 size: 12 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (111:123) - toolkits/model_checkpoints_convertor/yi/checkpoint_reshaping_and_interoperability.py (139:150) duplicated block id: 3689 size: 12 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/transformer_config.py (45:82) - megatron_patch/model/mixtral_bak/transformer_config.py (78:91) duplicated block id: 3690 size: 12 cleaned lines of code in 2 files: - megatron_patch/model/bloom/language_model.py (31:44) - megatron_patch/model/chatglm/language_model.py (31:44) duplicated block id: 3691 size: 12 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon/checkpoint_reshaping_and_interoperability.py (234:247) - toolkits/model_checkpoints_convertor/glm130b/checkpoint_reshaping_and_interoperability.py (203:216) duplicated block id: 3692 size: 12 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/transformer.py (548:561) - megatron_patch/model/falcon40b/transformer.py (694:719) duplicated block id: 3693 size: 12 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon/checkpoint_reshaping_and_interoperability.py (774:786) - toolkits/model_checkpoints_convertor/starcoder/checkpoint_reshaping_and_interoperability.py (491:503) duplicated block id: 3694 size: 12 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (927:939) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (833:845) duplicated block id: 3695 size: 12 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (664:678) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (828:842) duplicated block id: 3696 size: 12 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/chatglm/checkpoint_reshaping_and_interoperability.py (277:296) - toolkits/model_checkpoints_convertor/falcon/checkpoint_reshaping_and_interoperability.py (282:301) duplicated block id: 3697 size: 12 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon/checkpoint_reshaping_and_interoperability.py (828:840) - toolkits/model_checkpoints_convertor/qwen/hf2megablocks_qwen1.5.py (613:625) duplicated block id: 3698 size: 12 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/deepseek/hf2mcore_deepseek_v2_moe.py (25:48) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_moe.py (29:53) duplicated block id: 3699 size: 12 cleaned lines of code in 2 files: - megatron_patch/model/galactica/transformer.py (889:919) - megatron_patch/model/glm130b/transformer.py (1199:1229) duplicated block id: 3700 size: 12 cleaned lines of code in 2 files: - megatron_patch/model/llama/language_model.py (394:408) - megatron_patch/model/qwen1_5_megablocks/language_model.py (412:426) duplicated block id: 3701 size: 12 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon/checkpoint_reshaping_and_interoperability.py (828:840) - toolkits/model_checkpoints_convertor/llama/hf2mcore.py (603:615) duplicated block id: 3702 size: 12 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/checkpoint_reshaping_and_interoperability.py (117:128) - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (101:113) duplicated block id: 3703 size: 12 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/checkpoint_reshaping_and_interoperability.py (232:245) - toolkits/model_checkpoints_convertor/yi/checkpoint_reshaping_and_interoperability.py (248:261) duplicated block id: 3704 size: 12 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/language_model.py (394:408) - megatron_patch/model/mistral/language_model.py (415:429) duplicated block id: 3705 size: 12 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/moe/moe_layer.py (111:122) - megatron_patch/model/qwen3_moe/moe/moe_layer.py (36:47) duplicated block id: 3706 size: 12 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/glm/checkpoint_reshaping_and_interoperability.py (203:216) - toolkits/model_checkpoints_convertor/yi/checkpoint_reshaping_and_interoperability.py (248:261) duplicated block id: 3707 size: 12 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (138:149) - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (101:113) duplicated block id: 3708 size: 12 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/deepseek/hf2mcore_deepseek_v3_moe.py (103:117) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_vl.py (54:68) duplicated block id: 3709 size: 12 cleaned lines of code in 2 files: - megatron_patch/model/qwen2/model.py (33:62) - megatron_patch/model/qwen2_vl/gpt_model.py (22:51) duplicated block id: 3710 size: 12 cleaned lines of code in 2 files: - megatron_patch/model/llama3/model.py (34:63) - megatron_patch/model/qwen2_vl/gpt_model.py (22:51) duplicated block id: 3711 size: 12 cleaned lines of code in 2 files: - megatron_patch/model/falcon/language_model.py (319:341) - megatron_patch/model/starcoder/language_model.py (320:343) duplicated block id: 3712 size: 12 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (833:845) - toolkits/model_checkpoints_convertor/qwen/hf2megablocks_qwen1.5.py (613:625) duplicated block id: 3713 size: 12 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_moe.py (136:150) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_vl.py (72:86) duplicated block id: 3714 size: 12 cleaned lines of code in 2 files: - megatron_patch/model/falcon/transformer.py (568:579) - megatron_patch/model/glm130b/transformer.py (423:434) duplicated block id: 3715 size: 12 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/checkpoint_reshaping_and_interoperability.py (232:245) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (222:235) duplicated block id: 3716 size: 12 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/moe/moe_layer.py (110:121) - megatron_patch/model/qwen3_moe/moe/moe_layer.py (36:47) duplicated block id: 3717 size: 12 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon40b/checkpoint_reshaping_and_interoperability.py (804:816) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (1116:1128) duplicated block id: 3718 size: 12 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon40b/checkpoint_reshaping_and_interoperability.py (232:245) - toolkits/model_checkpoints_convertor/starcoder/checkpoint_reshaping_and_interoperability.py (254:267) duplicated block id: 3719 size: 12 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/hf2te.py (107:119) - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (111:123) duplicated block id: 3720 size: 12 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (1092:1108) - megatron_patch/model/bloom/transformer.py (795:811) duplicated block id: 3721 size: 12 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/hf2te.py (106:118) - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (111:123) duplicated block id: 3722 size: 12 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (1754:1786) - megatron_patch/model/chatglm/transformer.py (881:911) duplicated block id: 3723 size: 12 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon40b/checkpoint_reshaping_and_interoperability.py (232:245) - toolkits/model_checkpoints_convertor/glm/checkpoint_reshaping_and_interoperability.py (203:216) duplicated block id: 3724 size: 12 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/transformer.py (298:322) - megatron_patch/model/llama3/transformer_legacy.py (422:446) duplicated block id: 3725 size: 12 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon40b/checkpoint_reshaping_and_interoperability.py (758:770) - toolkits/model_checkpoints_convertor/starcoder/checkpoint_reshaping_and_interoperability.py (491:503) duplicated block id: 3726 size: 12 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/model.py (81:97) - megatron_patch/model/qwen1_5/model.py (71:87) duplicated block id: 3727 size: 12 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/hf2te.py (480:492) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (554:566) duplicated block id: 3728 size: 12 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/model.py (81:97) - megatron_patch/model/qwen2/model.py (72:88) duplicated block id: 3729 size: 12 cleaned lines of code in 2 files: - megatron_patch/model/galactica/language_model.py (425:439) - megatron_patch/model/qwen/language_model.py (410:424) duplicated block id: 3730 size: 12 cleaned lines of code in 2 files: - megatron_patch/model/llama3_1/model.py (46:75) - megatron_patch/model/mixtral_bak/model.py (31:60) duplicated block id: 3731 size: 12 cleaned lines of code in 2 files: - megatron_patch/model/qwen1_5/model.py (34:63) - megatron_patch/model/qwen2_vl/gpt_model.py (22:51) duplicated block id: 3732 size: 12 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (101:113) - toolkits/model_checkpoints_convertor/yi/checkpoint_reshaping_and_interoperability.py (139:150) duplicated block id: 3733 size: 12 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/checkpoint_reshaping_and_interoperability.py (525:537) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (1116:1128) duplicated block id: 3734 size: 12 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/glm/checkpoint_reshaping_and_interoperability.py (203:216) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (318:331) duplicated block id: 3735 size: 12 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/hf2te.py (508:520) - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (551:563) duplicated block id: 3736 size: 12 cleaned lines of code in 2 files: - megatron_patch/model/llama3/model.py (34:63) - megatron_patch/model/llama3_1/model.py (46:75) duplicated block id: 3737 size: 12 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore_llama3_1.py (647:659) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_vl.py (220:232) duplicated block id: 3738 size: 12 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/starcoder/checkpoint_reshaping_and_interoperability.py (254:267) - toolkits/model_checkpoints_convertor/yi/checkpoint_reshaping_and_interoperability.py (248:261) duplicated block id: 3739 size: 12 cleaned lines of code in 2 files: - megatron_patch/model/llama/language_model.py (394:408) - megatron_patch/model/llama2/language_model.py (413:427) duplicated block id: 3740 size: 12 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/hf2te.py (508:520) - toolkits/model_checkpoints_convertor/galactica/checkpoint_reshaping_and_interoperability.py (631:643) duplicated block id: 3741 size: 12 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (261:274) - toolkits/model_checkpoints_convertor/glm130b/checkpoint_reshaping_and_interoperability.py (203:216) duplicated block id: 3742 size: 12 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (833:845) - toolkits/model_checkpoints_convertor/starcoder/checkpoint_reshaping_and_interoperability.py (540:552) duplicated block id: 3743 size: 12 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (634:646) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (554:566) duplicated block id: 3744 size: 12 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (651:664) - megatron_patch/model/galactica/transformer.py (587:605) duplicated block id: 3745 size: 12 cleaned lines of code in 2 files: - megatron_patch/model/llama3/layer_specs.py (76:87) - megatron_patch/model/qwen1_5/layer_specs.py (76:87) duplicated block id: 3746 size: 12 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (222:235) - toolkits/model_checkpoints_convertor/starcoder/checkpoint_reshaping_and_interoperability.py (254:267) duplicated block id: 3747 size: 12 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/hf2te.py (33:44) - toolkits/model_checkpoints_convertor/chatglm/checkpoint_reshaping_and_interoperability.py (38:49) duplicated block id: 3748 size: 12 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (162:173) - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (101:113) duplicated block id: 3749 size: 12 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (111:123) - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (162:173) duplicated block id: 3750 size: 12 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore.py (603:615) - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (1108:1120) duplicated block id: 3751 size: 12 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/glm/checkpoint_reshaping_and_interoperability.py (203:216) - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (299:312) duplicated block id: 3752 size: 12 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/multi_latent_attention.py (98:109) - megatron_patch/model/qwen2_vl/attention_vision.py (118:129) duplicated block id: 3753 size: 12 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/language_model.py (32:45) - megatron_patch/model/qwen/language_model.py (30:43) duplicated block id: 3754 size: 12 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/transformer.py (881:911) - megatron_patch/model/llama2/transformer.py (1755:1787) duplicated block id: 3755 size: 12 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/hf2te.py (149:162) - toolkits/model_checkpoints_convertor/bloom/reward_model_to_megatron.py (232:245) duplicated block id: 3756 size: 12 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/transformer.py (298:322) - megatron_patch/model/qwen_vl/transformer.py (424:448) duplicated block id: 3757 size: 12 cleaned lines of code in 2 files: - megatron_patch/model/galactica/language_model.py (425:439) - megatron_patch/model/mistral/language_model.py (415:429) duplicated block id: 3758 size: 12 cleaned lines of code in 2 files: - megatron_patch/model/llama3_1/layer_specs.py (76:87) - megatron_patch/model/qwen1_5/layer_specs.py (76:87) duplicated block id: 3759 size: 12 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/chatglm/checkpoint_reshaping_and_interoperability.py (121:132) - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (101:113) duplicated block id: 3760 size: 12 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore_llama3_1.py (647:659) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2.5_vl.py (220:232) duplicated block id: 3761 size: 12 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/checkpoint_reshaping_and_interoperability.py (525:537) - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (1108:1120) duplicated block id: 3762 size: 12 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/glm/checkpoint_reshaping_and_interoperability.py (203:216) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (222:235) duplicated block id: 3763 size: 12 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore.py (603:615) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (1003:1015) duplicated block id: 3764 size: 12 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/reward_model_to_megatron.py (232:245) - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (305:318) duplicated block id: 3765 size: 12 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (138:149) - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (111:123) duplicated block id: 3766 size: 12 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (691:705) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (742:756) duplicated block id: 3767 size: 12 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (477:490) - megatron_patch/model/glm130b/transformer.py (523:536) duplicated block id: 3768 size: 12 cleaned lines of code in 2 files: - megatron_patch/model/galactica/transformer.py (889:919) - megatron_patch/model/llama2/transformer.py (1755:1787) duplicated block id: 3769 size: 12 cleaned lines of code in 2 files: - megatron_patch/model/qwen1_5/moe/token_dispatcher.py (207:220) - megatron_patch/model/qwen2/moe/token_dispatcher.py (215:228) duplicated block id: 3770 size: 12 cleaned lines of code in 2 files: - megatron_patch/model/llama3_1/model.py (46:75) - megatron_patch/model/qwen1_5/model.py (34:63) duplicated block id: 3771 size: 12 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/glm/checkpoint_reshaping_and_interoperability.py (203:216) - toolkits/model_checkpoints_convertor/qwen/hf2megablocks_qwen1.5.py (406:419) duplicated block id: 3772 size: 12 cleaned lines of code in 2 files: - megatron_patch/data/utils.py (257:269) - megatron_patch/data/utils.py (283:296) duplicated block id: 3773 size: 12 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llava/hf2mcore_llava.py (453:465) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_dense_and_moe_gqa.py (350:363) duplicated block id: 3774 size: 12 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/hf2te.py (108:121) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (90:103) duplicated block id: 3775 size: 12 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (927:939) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (1116:1128) duplicated block id: 3776 size: 12 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/checkpoint_reshaping_and_interoperability.py (232:245) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (318:331) duplicated block id: 3777 size: 12 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2megablocks_qwen1.5.py (613:625) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (1116:1128) duplicated block id: 3778 size: 12 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/reward_model_to_megatron.py (232:245) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (222:235) duplicated block id: 3779 size: 12 cleaned lines of code in 2 files: - megatron_patch/model/llava_mcore/transformer_config.py (38:49) - megatron_patch/model/llava_mcore/transformer_config.py (51:62) duplicated block id: 3780 size: 12 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/model.py (22:69) - megatron_patch/model/llama3_1/model.py (46:75) duplicated block id: 3781 size: 12 cleaned lines of code in 2 files: - megatron_patch/model/mixtral_bak/moe/token_dispatcher.py (309:322) - megatron_patch/model/qwen1_5/moe/token_dispatcher.py (272:285) duplicated block id: 3782 size: 12 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/hf2te.py (149:162) - toolkits/model_checkpoints_convertor/glm/checkpoint_reshaping_and_interoperability.py (203:216) duplicated block id: 3783 size: 12 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon/checkpoint_reshaping_and_interoperability.py (828:840) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (1116:1128) duplicated block id: 3784 size: 12 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (263:276) - toolkits/model_checkpoints_convertor/glm130b/checkpoint_reshaping_and_interoperability.py (203:216) duplicated block id: 3785 size: 12 cleaned lines of code in 2 files: - megatron_patch/model/galactica/transformer.py (889:919) - megatron_patch/model/qwen_vl/transformer.py (1754:1786) duplicated block id: 3786 size: 12 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/glm130b/checkpoint_reshaping_and_interoperability.py (203:216) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (295:308) duplicated block id: 3787 size: 12 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/model.py (22:69) - megatron_patch/model/qwen1_5/model.py (34:63) duplicated block id: 3788 size: 12 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (261:274) - toolkits/model_checkpoints_convertor/bloom/checkpoint_reshaping_and_interoperability.py (232:245) duplicated block id: 3789 size: 12 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/moe/experts.py (206:224) - megatron_patch/model/mixtral/moe/experts.py (790:803) duplicated block id: 3790 size: 12 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (1754:1786) - megatron_patch/model/galactica/transformer.py (889:919) duplicated block id: 3791 size: 12 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon40b/checkpoint_reshaping_and_interoperability.py (804:816) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (833:845) duplicated block id: 3792 size: 12 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (299:312) - toolkits/model_checkpoints_convertor/starcoder/checkpoint_reshaping_and_interoperability.py (254:267) duplicated block id: 3793 size: 12 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon/checkpoint_reshaping_and_interoperability.py (234:247) - toolkits/model_checkpoints_convertor/glm/checkpoint_reshaping_and_interoperability.py (203:216) duplicated block id: 3794 size: 12 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (890:905) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (957:972) duplicated block id: 3795 size: 12 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/language_model.py (412:431) - megatron_patch/model/starcoder/language_model.py (393:412) duplicated block id: 3796 size: 12 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/language_model.py (394:408) - megatron_patch/model/llama3/language_model.py (404:418) duplicated block id: 3797 size: 12 cleaned lines of code in 2 files: - megatron_patch/model/falcon/language_model.py (394:408) - megatron_patch/model/qwen/language_model.py (410:424) duplicated block id: 3798 size: 12 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon40b/checkpoint_reshaping_and_interoperability.py (580:592) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (554:566) duplicated block id: 3799 size: 12 cleaned lines of code in 2 files: - megatron_patch/model/galactica/language_model.py (350:372) - megatron_patch/model/starcoder/language_model.py (320:343) duplicated block id: 3800 size: 12 cleaned lines of code in 2 files: - megatron_patch/model/galactica/language_model.py (581:594) - megatron_patch/model/llava/language_model.py (559:572) duplicated block id: 3801 size: 12 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/transformer.py (688:706) - megatron_patch/model/llama3/transformer_legacy.py (827:844) duplicated block id: 3802 size: 12 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/model.py (81:97) - megatron_patch/model/qwen2_vl/gpt_model.py (61:76) duplicated block id: 3803 size: 12 cleaned lines of code in 2 files: - megatron_patch/model/mixtral_bak/moe/router.py (49:78) - megatron_patch/model/qwen1_5/moe/router.py (59:88) duplicated block id: 3804 size: 12 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/reward_model_to_megatron.py (524:536) - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (1108:1120) duplicated block id: 3805 size: 12 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon/checkpoint_reshaping_and_interoperability.py (773:784) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (889:902) duplicated block id: 3806 size: 12 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/model.py (22:69) - megatron_patch/model/mixtral_bak/model.py (31:60) duplicated block id: 3807 size: 12 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/model.py (31:60) - megatron_patch/model/qwen2_vl/gpt_model.py (22:51) duplicated block id: 3808 size: 12 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (901:913) - toolkits/model_checkpoints_convertor/llama/hf2mcore.py (603:615) duplicated block id: 3809 size: 12 cleaned lines of code in 2 files: - megatron_patch/model/galactica/language_model.py (581:594) - megatron_patch/model/qwen1_5_megablocks/language_model.py (498:511) duplicated block id: 3810 size: 12 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/transformer/attention.py (191:203) - megatron_patch/model/qwen2_vl/attention_vision.py (179:191) duplicated block id: 3811 size: 12 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (90:103) - toolkits/model_checkpoints_convertor/starcoder/checkpoint_reshaping_and_interoperability.py (118:130) duplicated block id: 3812 size: 12 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (789:800) - toolkits/model_checkpoints_convertor/falcon/checkpoint_reshaping_and_interoperability.py (773:784) duplicated block id: 3813 size: 12 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/hf2te.py (508:520) - toolkits/model_checkpoints_convertor/chatglm/checkpoint_reshaping_and_interoperability.py (561:573) duplicated block id: 3814 size: 12 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/transformer.py (548:561) - megatron_patch/model/galactica/transformer.py (587:605) duplicated block id: 3815 size: 12 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (811:823) - toolkits/model_checkpoints_convertor/starcoder/checkpoint_reshaping_and_interoperability.py (540:552) duplicated block id: 3816 size: 12 cleaned lines of code in 2 files: - megatron_patch/model/llama/transformer.py (627:645) - megatron_patch/model/llama3/transformer_legacy.py (827:844) duplicated block id: 3817 size: 12 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (111:123) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (166:177) duplicated block id: 3818 size: 12 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/language_model.py (471:485) - megatron_patch/model/falcon/language_model.py (394:408) duplicated block id: 3819 size: 12 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/galactica/checkpoint_reshaping_and_interoperability.py (138:149) - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (111:123) duplicated block id: 3820 size: 12 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_moe.py (56:69) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_dense_and_moe_gqa.py (48:61) duplicated block id: 3821 size: 12 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon40b/checkpoint_reshaping_and_interoperability.py (140:152) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (90:103) duplicated block id: 3822 size: 12 cleaned lines of code in 2 files: - megatron_patch/model/galactica/transformer.py (889:919) - megatron_patch/model/mistral/transformer.py (1752:1784) duplicated block id: 3823 size: 12 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/galactica/checkpoint_reshaping_and_interoperability.py (250:263) - toolkits/model_checkpoints_convertor/starcoder/checkpoint_reshaping_and_interoperability.py (254:267) duplicated block id: 3824 size: 12 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/reward_model_to_megatron.py (524:536) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (1003:1015) duplicated block id: 3825 size: 12 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/checkpoint_reshaping_and_interoperability.py (525:537) - toolkits/model_checkpoints_convertor/qwen/hf2megablocks_qwen1.5.py (613:625) duplicated block id: 3826 size: 12 cleaned lines of code in 2 files: - megatron_patch/model/mixtral_bak/moe/router.py (49:78) - megatron_patch/model/qwen2/moe/router.py (134:163) duplicated block id: 3827 size: 12 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/model.py (31:60) - megatron_patch/model/qwen2/model.py (33:62) duplicated block id: 3828 size: 12 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/checkpoint_reshaping_and_interoperability.py (525:537) - toolkits/model_checkpoints_convertor/llama/hf2mcore.py (603:615) duplicated block id: 3829 size: 12 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/language_model.py (31:44) - megatron_patch/model/llama/language_model.py (29:42) duplicated block id: 3830 size: 12 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/hf2te.py (107:120) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (90:103) duplicated block id: 3831 size: 12 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon/checkpoint_reshaping_and_interoperability.py (139:150) - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (101:113) duplicated block id: 3832 size: 12 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (554:566) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (763:775) duplicated block id: 3833 size: 12 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore.py (603:615) - toolkits/model_checkpoints_convertor/starcoder/checkpoint_reshaping_and_interoperability.py (540:552) duplicated block id: 3834 size: 12 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/moe/experts.py (207:225) - megatron_patch/model/mixtral/moe/experts.py (790:803) duplicated block id: 3835 size: 12 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (261:274) - toolkits/model_checkpoints_convertor/glm/checkpoint_reshaping_and_interoperability.py (203:216) duplicated block id: 3836 size: 12 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (1624:1656) - megatron_patch/model/chatglm/transformer.py (881:911) duplicated block id: 3837 size: 12 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/multi_latent_attention.py (98:109) - megatron_patch/model/mixtral_bak/transformer/attention.py (78:89) duplicated block id: 3838 size: 12 cleaned lines of code in 2 files: - megatron_patch/model/llama3/layer_specs.py (76:87) - megatron_patch/model/qwen2/layer_specs.py (86:97) duplicated block id: 3839 size: 12 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/reward_model_to_megatron.py (232:245) - toolkits/model_checkpoints_convertor/yi/checkpoint_reshaping_and_interoperability.py (248:261) duplicated block id: 3840 size: 12 cleaned lines of code in 2 files: - megatron_patch/model/bloom/language_model.py (518:533) - megatron_patch/model/falcon/language_model.py (605:620) duplicated block id: 3841 size: 12 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_moe.py (112:126) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_vl.py (54:68) duplicated block id: 3842 size: 12 cleaned lines of code in 2 files: - megatron_patch/model/galactica/language_model.py (425:439) - megatron_patch/model/llama2/language_model.py (413:427) duplicated block id: 3843 size: 12 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/glm130b/checkpoint_reshaping_and_interoperability.py (456:471) - toolkits/model_checkpoints_convertor/starcoder/checkpoint_reshaping_and_interoperability.py (746:761) duplicated block id: 3844 size: 12 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/glm130b/checkpoint_reshaping_and_interoperability.py (203:216) - toolkits/model_checkpoints_convertor/yi/checkpoint_reshaping_and_interoperability.py (248:261) duplicated block id: 3845 size: 12 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (554:566) - toolkits/model_checkpoints_convertor/yi/checkpoint_reshaping_and_interoperability.py (648:660) duplicated block id: 3846 size: 12 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/transformer_config.py (173:235) - megatron_patch/model/mixtral_bak/transformer_config.py (110:125) duplicated block id: 3847 size: 12 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon40b/checkpoint_reshaping_and_interoperability.py (804:816) - toolkits/model_checkpoints_convertor/llama/hf2mcore.py (603:615) duplicated block id: 3848 size: 12 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (318:331) - toolkits/model_checkpoints_convertor/starcoder/checkpoint_reshaping_and_interoperability.py (254:267) duplicated block id: 3849 size: 12 cleaned lines of code in 2 files: - megatron_patch/model/falcon/language_model.py (394:408) - megatron_patch/model/mistral/language_model.py (415:429) duplicated block id: 3850 size: 12 cleaned lines of code in 2 files: - megatron_patch/model/llava_mcore/llava_spec.py (68:79) - megatron_patch/model/qwen2_moe/layer_specs.py (213:225) duplicated block id: 3851 size: 12 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/transformer.py (881:911) - megatron_patch/model/qwen_vl/transformer.py (1754:1786) duplicated block id: 3852 size: 12 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/transformer.py (881:911) - megatron_patch/model/llava/transformer.py (1754:1786) duplicated block id: 3853 size: 12 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/language_model.py (334:356) - megatron_patch/model/starcoder/language_model.py (320:343) duplicated block id: 3854 size: 12 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (444:457) - megatron_patch/model/chatglm/transformer.py (377:390) duplicated block id: 3855 size: 12 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon40b/checkpoint_reshaping_and_interoperability.py (232:245) - toolkits/model_checkpoints_convertor/glm130b/checkpoint_reshaping_and_interoperability.py (203:216) duplicated block id: 3856 size: 12 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (901:913) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (1116:1128) duplicated block id: 3857 size: 12 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (90:103) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (167:179) duplicated block id: 3858 size: 12 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (101:113) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (166:177) duplicated block id: 3859 size: 12 cleaned lines of code in 2 files: - megatron_patch/model/llama/language_model.py (394:408) - megatron_patch/model/llama3/language_model.py (404:418) duplicated block id: 3860 size: 12 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore.py (603:615) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (1116:1128) duplicated block id: 3861 size: 12 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (651:664) - megatron_patch/model/falcon/transformer.py (926:951) duplicated block id: 3862 size: 12 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (624:641) - megatron_patch/model/glm130b/transformer.py (724:741) duplicated block id: 3863 size: 12 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (263:276) - toolkits/model_checkpoints_convertor/starcoder/checkpoint_reshaping_and_interoperability.py (254:267) duplicated block id: 3864 size: 12 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/checkpoint_reshaping_and_interoperability.py (117:128) - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (111:123) duplicated block id: 3865 size: 12 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/reward_model_to_megatron.py (524:536) - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (811:823) duplicated block id: 3866 size: 12 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon/checkpoint_reshaping_and_interoperability.py (577:589) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (554:566) duplicated block id: 3867 size: 12 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/reward_model_to_megatron.py (232:245) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (295:308) duplicated block id: 3868 size: 12 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/transformer.py (298:322) - megatron_patch/model/llava/transformer.py (424:448) duplicated block id: 3869 size: 12 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/hf2te.py (149:162) - toolkits/model_checkpoints_convertor/glm130b/checkpoint_reshaping_and_interoperability.py (203:216) duplicated block id: 3870 size: 12 cleaned lines of code in 2 files: - megatron_patch/model/mixtral_bak/model.py (31:60) - megatron_patch/model/qwen2/model.py (33:62) duplicated block id: 3871 size: 12 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/hf2te.py (149:162) - toolkits/model_checkpoints_convertor/starcoder/checkpoint_reshaping_and_interoperability.py (254:267) duplicated block id: 3872 size: 12 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/galactica/checkpoint_reshaping_and_interoperability.py (631:643) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (554:566) duplicated block id: 3873 size: 12 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/language_model.py (32:45) - megatron_patch/model/llava/language_model.py (33:46) duplicated block id: 3874 size: 12 cleaned lines of code in 2 files: - megatron_patch/model/galactica/language_model.py (581:594) - megatron_patch/model/qwen_vl/language_model.py (528:541) duplicated block id: 3875 size: 12 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/reward_model_to_megatron.py (117:128) - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (111:123) duplicated block id: 3876 size: 12 cleaned lines of code in 2 files: - megatron_patch/model/mixtral_bak/model.py (31:60) - megatron_patch/model/qwen2_vl/gpt_model.py (22:51) duplicated block id: 3877 size: 12 cleaned lines of code in 2 files: - megatron_patch/model/galactica/transformer.py (587:605) - megatron_patch/model/starcoder/transformer.py (865:878) duplicated block id: 3878 size: 12 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/moe/experts.py (789:802) - megatron_patch/model/mixtral/moe/experts.py (207:225) duplicated block id: 3879 size: 12 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/language_model.py (31:44) - megatron_patch/model/falcon/language_model.py (29:42) duplicated block id: 3880 size: 12 cleaned lines of code in 2 files: - megatron_patch/model/falcon/language_model.py (394:408) - megatron_patch/model/llama3/language_model.py (404:418) duplicated block id: 3881 size: 12 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/transformer.py (908:920) - megatron_patch/model/starcoder/transformer.py (1144:1156) duplicated block id: 3882 size: 12 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/galactica/checkpoint_reshaping_and_interoperability.py (250:263) - toolkits/model_checkpoints_convertor/glm130b/checkpoint_reshaping_and_interoperability.py (203:216) duplicated block id: 3883 size: 12 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon40b/checkpoint_reshaping_and_interoperability.py (757:768) - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (932:945) duplicated block id: 3884 size: 12 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/language_model.py (420:434) - megatron_patch/model/galactica/language_model.py (425:439) duplicated block id: 3885 size: 12 cleaned lines of code in 2 files: - megatron_patch/model/bloom/language_model.py (456:469) - megatron_patch/model/glm130b/language_model.py (478:491) duplicated block id: 3886 size: 12 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/glm/checkpoint_reshaping_and_interoperability.py (203:216) - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (305:318) duplicated block id: 3887 size: 12 cleaned lines of code in 2 files: - megatron_patch/model/galactica/transformer.py (889:919) - megatron_patch/model/llama3/transformer_legacy.py (1714:1746) duplicated block id: 3888 size: 12 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/language_model.py (31:44) - megatron_patch/model/glm130b/language_model.py (30:43) duplicated block id: 3889 size: 12 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/checkpoint_reshaping_and_interoperability.py (525:537) - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (811:823) duplicated block id: 3890 size: 12 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/hf2te.py (480:492) - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (551:563) duplicated block id: 3891 size: 12 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/language_model.py (32:45) - megatron_patch/model/qwen_vl/language_model.py (31:44) duplicated block id: 3892 size: 12 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/language_model.py (32:45) - megatron_patch/model/llama2/language_model.py (30:43) duplicated block id: 3893 size: 12 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/checkpoint_reshaping_and_interoperability.py (232:245) - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (305:318) duplicated block id: 3894 size: 12 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/model.py (81:97) - megatron_patch/model/llama3_1/model.py (86:102) duplicated block id: 3895 size: 12 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/transformer.py (298:322) - megatron_patch/model/llama2/transformer.py (423:447) duplicated block id: 3896 size: 12 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/transformer.py (694:719) - megatron_patch/model/starcoder/transformer.py (865:878) duplicated block id: 3897 size: 12 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/reward_model_to_megatron.py (524:536) - toolkits/model_checkpoints_convertor/qwen/hf2megablocks_qwen1.5.py (613:625) duplicated block id: 3898 size: 12 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/deepseek/hf2mcore_deepseek_v3_moe.py (103:117) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2.5_vl.py (54:68) duplicated block id: 3899 size: 12 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/transformer.py (423:434) - megatron_patch/model/llama/transformer.py (378:389) duplicated block id: 3900 size: 12 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/moe/experts.py (206:224) - megatron_patch/model/qwen2/moe/experts.py (384:397) duplicated block id: 3901 size: 12 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/model.py (22:69) - megatron_patch/model/qwen2/model.py (33:62) duplicated block id: 3902 size: 12 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/reward_model_to_megatron.py (118:130) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (90:103) duplicated block id: 3903 size: 12 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/checkpoint_reshaping_and_interoperability.py (525:537) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (1003:1015) duplicated block id: 3904 size: 12 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/moe/shared_experts.py (167:180) - megatron_patch/model/mixtral_bak/transformer/mlp.py (93:106) duplicated block id: 3905 size: 12 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/model.py (66:81) - megatron_patch/model/qwen2_vl/gpt_model.py (58:72) duplicated block id: 3906 size: 12 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/checkpoint_reshaping_and_interoperability.py (118:130) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (90:103) duplicated block id: 3907 size: 12 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/language_model.py (32:45) - megatron_patch/model/qwen1_5_megablocks/language_model.py (29:42) duplicated block id: 3908 size: 12 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/checkpoint_reshaping_and_interoperability.py (525:537) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (833:845) duplicated block id: 3909 size: 12 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_mha.py (54:67) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_dense_and_moe_gqa.py (48:61) duplicated block id: 3910 size: 12 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (651:664) - megatron_patch/model/falcon40b/transformer.py (694:719) duplicated block id: 3911 size: 12 cleaned lines of code in 2 files: - megatron_patch/model/qwen1_5/transformer/mlp.py (134:147) - megatron_patch/model/qwen2/moe/experts.py (290:303) duplicated block id: 3912 size: 12 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (326:350) - megatron_patch/model/glm130b/transformer.py (298:322) duplicated block id: 3913 size: 12 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (163:175) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (90:103) duplicated block id: 3914 size: 12 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/multi_latent_attention.py (98:109) - megatron_patch/model/qwen2_vl/attention.py (119:130) duplicated block id: 3915 size: 12 cleaned lines of code in 2 files: - megatron_patch/model/llama3_1/model.py (46:75) - megatron_patch/model/qwen2_vl/gpt_model.py (22:51) duplicated block id: 3916 size: 12 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/reward_model_to_megatron.py (232:245) - toolkits/model_checkpoints_convertor/qwen/hf2megablocks_qwen1.5.py (406:419) duplicated block id: 3917 size: 12 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon/checkpoint_reshaping_and_interoperability.py (828:840) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (833:845) duplicated block id: 3918 size: 12 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (324:341) - toolkits/model_checkpoints_convertor/yi/checkpoint_reshaping_and_interoperability.py (375:392) duplicated block id: 3919 size: 12 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (634:646) - toolkits/model_checkpoints_convertor/baichuan/hf2te.py (508:520) duplicated block id: 3920 size: 12 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/reward_model_to_megatron.py (232:245) - toolkits/model_checkpoints_convertor/falcon/checkpoint_reshaping_and_interoperability.py (234:247) duplicated block id: 3921 size: 12 cleaned lines of code in 2 files: - megatron_patch/model/mixtral_bak/transformer/mlp.py (93:106) - megatron_patch/model/qwen2/moe/experts.py (290:303) duplicated block id: 3922 size: 12 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (261:274) - toolkits/model_checkpoints_convertor/bloom/reward_model_to_megatron.py (232:245) duplicated block id: 3923 size: 12 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/glm130b/checkpoint_reshaping_and_interoperability.py (203:216) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (318:331) duplicated block id: 3924 size: 12 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/reward_model_to_megatron.py (232:245) - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (299:312) duplicated block id: 3925 size: 12 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (138:149) - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (101:113) duplicated block id: 3926 size: 12 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/language_model.py (394:408) - megatron_patch/model/llama2/language_model.py (413:427) duplicated block id: 3927 size: 12 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (1108:1120) - toolkits/model_checkpoints_convertor/qwen/hf2megablocks_qwen1.5.py (613:625) duplicated block id: 3928 size: 12 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon40b/checkpoint_reshaping_and_interoperability.py (804:816) - toolkits/model_checkpoints_convertor/qwen/hf2megablocks_qwen1.5.py (613:625) duplicated block id: 3929 size: 12 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (1092:1108) - megatron_patch/model/starcoder/transformer.py (940:956) duplicated block id: 3930 size: 12 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/language_model.py (394:408) - megatron_patch/model/qwen/language_model.py (410:424) duplicated block id: 3931 size: 12 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon40b/checkpoint_reshaping_and_interoperability.py (139:150) - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (111:123) duplicated block id: 3932 size: 12 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (1116:1128) - toolkits/model_checkpoints_convertor/starcoder/checkpoint_reshaping_and_interoperability.py (540:552) duplicated block id: 3933 size: 12 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/transformer.py (548:561) - megatron_patch/model/falcon/transformer.py (926:951) duplicated block id: 3934 size: 12 cleaned lines of code in 2 files: - megatron_patch/model/qwen1_5/layer_specs.py (76:87) - megatron_patch/model/qwen2/layer_specs.py (86:97) duplicated block id: 3935 size: 12 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/transformer.py (332:343) - megatron_patch/model/glm130b/transformer.py (423:434) duplicated block id: 3936 size: 12 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/hf2te.py (480:492) - toolkits/model_checkpoints_convertor/chatglm/checkpoint_reshaping_and_interoperability.py (561:573) duplicated block id: 3937 size: 12 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/language_model.py (319:341) - megatron_patch/model/starcoder/language_model.py (320:343) duplicated block id: 3938 size: 12 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2megablocks_qwen1.5.py (613:625) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (1003:1015) duplicated block id: 3939 size: 12 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/glm/checkpoint_reshaping_and_interoperability.py (203:216) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (295:308) duplicated block id: 3940 size: 12 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (263:276) - toolkits/model_checkpoints_convertor/bloom/reward_model_to_megatron.py (232:245) duplicated block id: 3941 size: 12 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/language_model.py (420:434) - megatron_patch/model/falcon40b/language_model.py (394:408) duplicated block id: 3942 size: 12 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/transformer/attention.py (191:203) - megatron_patch/model/qwen2_vl/attention.py (180:192) duplicated block id: 3943 size: 12 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/model.py (81:97) - megatron_patch/model/llama3/model.py (71:87) duplicated block id: 3944 size: 12 cleaned lines of code in 2 files: - megatron_patch/model/mixtral_bak/moe/token_dispatcher.py (309:322) - megatron_patch/model/qwen2/moe/token_dispatcher.py (278:291) duplicated block id: 3945 size: 12 cleaned lines of code in 2 files: - megatron_patch/model/bloom/language_model.py (518:533) - megatron_patch/model/llama/language_model.py (616:631) duplicated block id: 3946 size: 12 cleaned lines of code in 2 files: - megatron_patch/model/galactica/transformer.py (889:919) - megatron_patch/model/llava/transformer.py (1754:1786) duplicated block id: 3947 size: 12 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/hf2te.py (34:45) - toolkits/model_checkpoints_convertor/chatglm/checkpoint_reshaping_and_interoperability.py (38:49) duplicated block id: 3948 size: 12 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (691:705) - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (807:821) duplicated block id: 3949 size: 12 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore_llama3_1.py (834:848) - toolkits/model_checkpoints_convertor/llava/hf2mcore_llava.py (787:801) duplicated block id: 3950 size: 12 cleaned lines of code in 2 files: - megatron_patch/model/galactica/transformer.py (340:351) - megatron_patch/model/glm130b/transformer.py (423:434) duplicated block id: 3951 size: 12 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/language_model.py (32:45) - megatron_patch/model/mistral/language_model.py (31:44) duplicated block id: 3952 size: 12 cleaned lines of code in 2 files: - megatron_patch/model/galactica/language_model.py (425:439) - megatron_patch/model/llama3/language_model.py (404:418) duplicated block id: 3953 size: 12 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (1108:1120) - toolkits/model_checkpoints_convertor/starcoder/checkpoint_reshaping_and_interoperability.py (540:552) duplicated block id: 3954 size: 12 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/chatglm/checkpoint_reshaping_and_interoperability.py (122:134) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (90:103) duplicated block id: 3955 size: 12 cleaned lines of code in 2 files: - megatron_patch/model/galactica/language_model.py (425:439) - megatron_patch/model/qwen1_5_megablocks/language_model.py (412:426) duplicated block id: 3956 size: 12 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/transformer.py (391:402) - megatron_patch/model/glm130b/transformer.py (423:434) duplicated block id: 3957 size: 12 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/language_model.py (471:485) - megatron_patch/model/falcon40b/language_model.py (394:408) duplicated block id: 3958 size: 12 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (742:756) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (828:842) duplicated block id: 3959 size: 12 cleaned lines of code in 2 files: - megatron_patch/model/llama/language_model.py (394:408) - megatron_patch/model/qwen/language_model.py (410:424) duplicated block id: 3960 size: 12 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/checkpoint_reshaping_and_interoperability.py (232:245) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (295:308) duplicated block id: 3961 size: 12 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/transformer.py (881:911) - megatron_patch/model/llama3/transformer_legacy.py (1714:1746) duplicated block id: 3962 size: 12 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (847:863) - megatron_patch/model/chatglm/transformer.py (625:641) duplicated block id: 3963 size: 12 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/chatglm/checkpoint_reshaping_and_interoperability.py (561:573) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (554:566) duplicated block id: 3964 size: 12 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore_llama3_1.py (647:659) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_dense_and_moe_gqa.py (350:363) duplicated block id: 3965 size: 12 cleaned lines of code in 2 files: - megatron_patch/model/mixtral_bak/model.py (65:80) - megatron_patch/model/qwen2_vl/gpt_model.py (58:72) duplicated block id: 3966 size: 12 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (789:800) - toolkits/model_checkpoints_convertor/falcon40b/checkpoint_reshaping_and_interoperability.py (757:768) duplicated block id: 3967 size: 12 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/transformer.py (298:322) - megatron_patch/model/qwen/transformer.py (417:441) duplicated block id: 3968 size: 12 cleaned lines of code in 2 files: - megatron_patch/model/galactica/transformer.py (889:919) - megatron_patch/model/qwen/transformer.py (1696:1728) duplicated block id: 3969 size: 12 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (927:939) - toolkits/model_checkpoints_convertor/llama/hf2mcore.py (603:615) duplicated block id: 3970 size: 12 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2megablocks_qwen1.5.py (406:419) - toolkits/model_checkpoints_convertor/starcoder/checkpoint_reshaping_and_interoperability.py (254:267) duplicated block id: 3971 size: 12 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llava/hf2mcore_llava.py (453:465) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2.5_vl.py (220:232) duplicated block id: 3972 size: 12 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/reward_model_to_megatron.py (117:128) - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (101:113) duplicated block id: 3973 size: 12 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/multi_latent_attention.py (98:109) - megatron_patch/model/qwen2/transformer/attention.py (100:111) duplicated block id: 3974 size: 12 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/language_model.py (420:434) - megatron_patch/model/falcon/language_model.py (394:408) duplicated block id: 3975 size: 12 cleaned lines of code in 2 files: - megatron_patch/model/bloom/language_model.py (518:533) - megatron_patch/model/galactica/language_model.py (634:649) duplicated block id: 3976 size: 12 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/glm130b/checkpoint_reshaping_and_interoperability.py (203:216) - toolkits/model_checkpoints_convertor/qwen/hf2megablocks_qwen1.5.py (406:419) duplicated block id: 3977 size: 12 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon/checkpoint_reshaping_and_interoperability.py (140:152) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (90:103) duplicated block id: 3978 size: 12 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/model.py (22:69) - megatron_patch/model/qwen2_vl/gpt_model.py (22:51) duplicated block id: 3979 size: 12 cleaned lines of code in 2 files: - megatron_patch/model/galactica/language_model.py (581:594) - megatron_patch/model/llama2/language_model.py (499:512) duplicated block id: 3980 size: 12 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/glm130b/checkpoint_reshaping_and_interoperability.py (203:216) - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (243:256) duplicated block id: 3981 size: 12 cleaned lines of code in 2 files: - megatron_patch/model/llama3_1/layer_specs.py (76:87) - megatron_patch/model/qwen2/layer_specs.py (86:97) duplicated block id: 3982 size: 12 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (90:103) - toolkits/model_checkpoints_convertor/yi/checkpoint_reshaping_and_interoperability.py (140:152) duplicated block id: 3983 size: 12 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/reward_model_to_megatron.py (524:536) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (833:845) duplicated block id: 3984 size: 12 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/language_model.py (31:44) - megatron_patch/model/galactica/language_model.py (29:42) duplicated block id: 3985 size: 12 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (1134:1155) - megatron_patch/model/falcon/transformer.py (1231:1253) duplicated block id: 3986 size: 12 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/transformer.py (548:561) - megatron_patch/model/llama/transformer.py (697:710) duplicated block id: 3987 size: 12 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_dense_and_moe_gqa.py (12:36) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_moe.py (29:53) duplicated block id: 3988 size: 12 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/galactica/checkpoint_reshaping_and_interoperability.py (250:263) - toolkits/model_checkpoints_convertor/glm/checkpoint_reshaping_and_interoperability.py (203:216) duplicated block id: 3989 size: 12 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (111:123) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (167:178) duplicated block id: 3990 size: 12 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/language_model.py (420:434) - megatron_patch/model/llama/language_model.py (394:408) duplicated block id: 3991 size: 12 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (816:829) - toolkits/model_checkpoints_convertor/falcon40b/checkpoint_reshaping_and_interoperability.py (757:768) duplicated block id: 3992 size: 12 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/checkpoint_reshaping_and_interoperability.py (232:245) - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (299:312) duplicated block id: 3993 size: 12 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (811:823) - toolkits/model_checkpoints_convertor/qwen/hf2megablocks_qwen1.5.py (613:625) duplicated block id: 3994 size: 12 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/multi_latent_attention.py (98:109) - megatron_patch/model/mixtral/transformer/attention.py (132:143) duplicated block id: 3995 size: 12 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/glm130b/checkpoint_reshaping_and_interoperability.py (203:216) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (222:235) duplicated block id: 3996 size: 12 cleaned lines of code in 2 files: - megatron_patch/model/galactica/language_model.py (581:594) - megatron_patch/model/llama/language_model.py (560:573) duplicated block id: 3997 size: 12 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2.5_vl.py (54:68) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_moe.py (112:126) duplicated block id: 3998 size: 12 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (634:646) - toolkits/model_checkpoints_convertor/baichuan2/hf2te.py (480:492) duplicated block id: 3999 size: 12 cleaned lines of code in 2 files: - megatron_patch/model/llama/language_model.py (319:341) - megatron_patch/model/starcoder/language_model.py (320:343) duplicated block id: 4000 size: 12 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (691:705) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (828:842) duplicated block id: 4001 size: 12 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon/checkpoint_reshaping_and_interoperability.py (773:784) - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (932:945) duplicated block id: 4002 size: 12 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/glm130b/checkpoint_reshaping_and_interoperability.py (203:216) - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (299:312) duplicated block id: 4003 size: 12 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/checkpoint_reshaping_and_interoperability.py (232:245) - toolkits/model_checkpoints_convertor/qwen/hf2megablocks_qwen1.5.py (406:419) duplicated block id: 4004 size: 12 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/hf2te.py (149:162) - toolkits/model_checkpoints_convertor/bloom/checkpoint_reshaping_and_interoperability.py (232:245) duplicated block id: 4005 size: 12 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/language_model.py (471:485) - megatron_patch/model/galactica/language_model.py (425:439) duplicated block id: 4006 size: 12 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/language_model.py (32:45) - megatron_patch/model/llama3/language_model.py (19:32) duplicated block id: 4007 size: 12 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/transformer.py (881:911) - megatron_patch/model/mistral/transformer.py (1752:1784) duplicated block id: 4008 size: 12 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (927:939) - toolkits/model_checkpoints_convertor/qwen/hf2megablocks_qwen1.5.py (613:625) duplicated block id: 4009 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/starcoder/transformer.py (302:325) - megatron_patch/model/starcoder/transformer.py (425:448) duplicated block id: 4010 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (756:767) - megatron_patch/model/glm130b/transformer.py (612:623) duplicated block id: 4011 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/multi_latent_attention.py (111:121) - megatron_patch/model/qwen2_vl/attention_vision.py (333:343) duplicated block id: 4012 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer_legacy.py (51:83) - megatron_patch/model/starcoder/transformer.py (47:80) duplicated block id: 4013 size: 11 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/hf2te.py (214:229) - toolkits/model_checkpoints_convertor/galactica/checkpoint_reshaping_and_interoperability.py (343:358) duplicated block id: 4014 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/llama/language_model.py (433:443) - megatron_patch/model/llava/language_model.py (459:469) duplicated block id: 4015 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/language_model.py (608:620) - megatron_patch/model/starcoder/language_model.py (495:505) duplicated block id: 4016 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/language_model.py (431:448) - megatron_patch/model/mistral/language_model.py (371:388) duplicated block id: 4017 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (1725:1745) - megatron_patch/model/llava/transformer.py (1724:1744) duplicated block id: 4018 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/galactica/language_model.py (255:273) - megatron_patch/model/starcoder/language_model.py (229:246) duplicated block id: 4019 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/bloom/language_model.py (363:373) - megatron_patch/model/llama/language_model.py (351:361) duplicated block id: 4020 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/mixtral_bak/moe/experts.py (180:208) - megatron_patch/model/qwen2/moe/experts.py (354:367) duplicated block id: 4021 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/falcon/language_model.py (433:443) - megatron_patch/model/llava/language_model.py (459:469) duplicated block id: 4022 size: 11 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/hf2te.py (35:45) - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (72:85) duplicated block id: 4023 size: 11 cleaned lines of code in 2 files: - toolkits/pretrain_data_preprocessing/preprocess_data.py (146:156) - toolkits/sft_data_preprocessing/build_idxmap_sft_dataset.py (244:256) duplicated block id: 4024 size: 11 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/deepseek/hf2mcore_deepseek_v2_moe.py (34:50) - toolkits/model_checkpoints_convertor/llama/hf2mcore_llama3_1.py (24:39) duplicated block id: 4025 size: 11 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/checkpoint_reshaping_and_interoperability.py (477:487) - toolkits/model_checkpoints_convertor/falcon40b/checkpoint_reshaping_and_interoperability.py (758:768) duplicated block id: 4026 size: 11 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (780:791) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (665:676) duplicated block id: 4027 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/llama2/transformer.py (1725:1745) - megatron_patch/model/llama3/transformer_legacy.py (1685:1705) duplicated block id: 4028 size: 11 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/mistral/hf2mcore.py (426:436) - toolkits/model_checkpoints_convertor/mistral/hf2mcore.py (461:471) duplicated block id: 4029 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/language_model.py (326:341) - megatron_patch/model/starcoder/language_model.py (271:286) duplicated block id: 4030 size: 11 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (72:85) - toolkits/model_checkpoints_convertor/yi/checkpoint_reshaping_and_interoperability.py (45:55) duplicated block id: 4031 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (1595:1615) - megatron_patch/model/qwen1_5_megablocks/transformer.py (1587:1607) duplicated block id: 4032 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/llama/language_model.py (433:443) - megatron_patch/model/qwen/language_model.py (448:458) duplicated block id: 4033 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/qwen1_5_megablocks/transformer.py (45:77) - megatron_patch/model/starcoder/transformer.py (47:80) duplicated block id: 4034 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/llama3/model.py (89:100) - megatron_patch/model/mixtral_bak/model.py (82:93) duplicated block id: 4035 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/qwen_vl/language_model.py (271:286) - megatron_patch/model/starcoder/language_model.py (271:286) duplicated block id: 4036 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/falcon/language_model.py (129:154) - megatron_patch/model/qwen1_5_megablocks/language_model.py (121:147) duplicated block id: 4037 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/bloom/language_model.py (404:416) - megatron_patch/model/llama3/language_model.py (404:416) duplicated block id: 4038 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/language_model.py (403:413) - megatron_patch/model/qwen_vl/language_model.py (348:358) duplicated block id: 4039 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/falcon/transformer.py (289:310) - megatron_patch/model/qwen_vl/transformer.py (424:445) duplicated block id: 4040 size: 11 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore.py (315:326) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_moe.py (143:154) duplicated block id: 4041 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/moe/router.py (66:97) - megatron_patch/model/qwen1_5/moe/router.py (75:105) duplicated block id: 4042 size: 11 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (72:85) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (72:82) duplicated block id: 4043 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/language_model.py (132:158) - megatron_patch/model/falcon40b/language_model.py (129:154) duplicated block id: 4044 size: 11 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore.py (267:277) - toolkits/model_checkpoints_convertor/qwen/hf2megablocks_qwen1.5.py (261:271) duplicated block id: 4045 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/transformer.py (45:76) - megatron_patch/model/mistral/transformer.py (47:79) duplicated block id: 4046 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/bloom/language_model.py (363:373) - megatron_patch/model/falcon/language_model.py (351:361) duplicated block id: 4047 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/language_model.py (131:156) - megatron_patch/model/llava/language_model.py (125:151) duplicated block id: 4048 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/llama3_1/layer_specs.py (89:101) - megatron_patch/model/qwen2/layer_specs.py (99:112) duplicated block id: 4049 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/falcon/transformer.py (289:310) - megatron_patch/model/llava/transformer.py (424:445) duplicated block id: 4050 size: 11 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/checkpoint_reshaping_and_interoperability.py (477:487) - toolkits/model_checkpoints_convertor/falcon/checkpoint_reshaping_and_interoperability.py (774:784) duplicated block id: 4051 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/language_model.py (433:443) - megatron_patch/model/llava/language_model.py (459:469) duplicated block id: 4052 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/language_model.py (433:443) - megatron_patch/model/qwen_vl/language_model.py (451:461) duplicated block id: 4053 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/language_model.py (471:483) - megatron_patch/model/glm130b/language_model.py (423:435) duplicated block id: 4054 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/language_model.py (433:443) - megatron_patch/model/qwen/language_model.py (448:458) duplicated block id: 4055 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/transformer.py (298:319) - megatron_patch/model/starcoder/transformer.py (328:349) duplicated block id: 4056 size: 11 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/hf2te.py (110:121) - toolkits/model_checkpoints_convertor/mistral/hf2mcore.py (105:116) duplicated block id: 4057 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/llava_mcore/vision/vit_layer_specs.py (65:75) - megatron_patch/model/qwen2_moe/layer_specs.py (215:225) duplicated block id: 4058 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer/attention.py (158:172) - megatron_patch/model/qwen2_vl/attention_vision.py (167:181) duplicated block id: 4059 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/language_model.py (34:46) - megatron_patch/model/starcoder/language_model.py (30:42) duplicated block id: 4060 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/qwen2_moe/layer_specs.py (16:36) - megatron_patch/model/qwen3_moe/gpt_layer_specs.py (26:41) duplicated block id: 4061 size: 11 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/chatglm/checkpoint_reshaping_and_interoperability.py (123:134) - toolkits/model_checkpoints_convertor/mistral/hf2mcore.py (105:116) duplicated block id: 4062 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/mlp.py (96:110) - megatron_patch/model/qwen1_5/transformer/mlp.py (118:132) duplicated block id: 4063 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/mixtral_bak/model.py (82:93) - megatron_patch/model/qwen1_5/model.py (89:100) duplicated block id: 4064 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/bloom/language_model.py (404:416) - megatron_patch/model/qwen1_5_megablocks/language_model.py (412:424) duplicated block id: 4065 size: 11 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore.py (317:328) - toolkits/model_checkpoints_convertor/mistral/hf2mcore.py (257:268) duplicated block id: 4066 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/qwen/transformer.py (42:74) - megatron_patch/model/starcoder/transformer.py (47:80) duplicated block id: 4067 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/language_model.py (458:468) - megatron_patch/model/chatglm/language_model.py (439:449) duplicated block id: 4068 size: 11 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (60:72) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2.5_vl.py (56:68) duplicated block id: 4069 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/llama3/language_model.py (342:352) - megatron_patch/model/qwen_vl/language_model.py (348:358) duplicated block id: 4070 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/llama/language_model.py (433:443) - megatron_patch/model/qwen_vl/language_model.py (451:461) duplicated block id: 4071 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/transformer.py (45:76) - megatron_patch/model/qwen/transformer.py (42:74) duplicated block id: 4072 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/llama/language_model.py (351:361) - megatron_patch/model/starcoder/language_model.py (347:357) duplicated block id: 4073 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer/attention.py (507:527) - megatron_patch/model/qwen2_vl/attention_vision.py (655:675) duplicated block id: 4074 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer/attention.py (158:172) - megatron_patch/model/qwen2_vl/attention.py (168:182) duplicated block id: 4075 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/bloom/language_model.py (131:156) - megatron_patch/model/llama3/language_model.py (111:137) duplicated block id: 4076 size: 11 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/deepseek/hf2mcore_deepseek_v3_moe.py (105:117) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (60:72) duplicated block id: 4077 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/language_model.py (509:519) - megatron_patch/model/chatglm/language_model.py (439:449) duplicated block id: 4078 size: 11 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (164:175) - toolkits/model_checkpoints_convertor/qwen/hf2megablocks_qwen1.5.py (131:142) duplicated block id: 4079 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/language_model.py (132:158) - megatron_patch/model/llama/language_model.py (129:154) duplicated block id: 4080 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/llava/transformer.py (1724:1744) - megatron_patch/model/qwen/transformer.py (1667:1687) duplicated block id: 4081 size: 11 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/reward_model_to_megatron.py (119:130) - toolkits/model_checkpoints_convertor/llama/hf2mcore.py (137:148) duplicated block id: 4082 size: 11 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/mistral/hf2mcore.py (257:268) - toolkits/model_checkpoints_convertor/qwen/hf2megablocks_qwen1.5.py (310:321) duplicated block id: 4083 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/falcon/language_model.py (433:443) - megatron_patch/model/llama2/language_model.py (451:461) duplicated block id: 4084 size: 11 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore.py (137:148) - toolkits/model_checkpoints_convertor/starcoder/checkpoint_reshaping_and_interoperability.py (119:130) duplicated block id: 4085 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (1725:1745) - megatron_patch/model/mistral/transformer.py (1722:1742) duplicated block id: 4086 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/llava_mcore/llava_model.py (132:142) - megatron_patch/model/qwen2_5_vl/model.py (106:116) duplicated block id: 4087 size: 11 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/hf2te.py (214:229) - toolkits/model_checkpoints_convertor/falcon40b/checkpoint_reshaping_and_interoperability.py (314:329) duplicated block id: 4088 size: 11 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/hf2te.py (109:120) - toolkits/model_checkpoints_convertor/qwen/hf2megablocks_qwen1.5.py (131:142) duplicated block id: 4089 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/llama/language_model.py (129:154) - megatron_patch/model/llama2/language_model.py (122:148) duplicated block id: 4090 size: 11 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (72:85) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (73:83) duplicated block id: 4091 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/language_model.py (131:156) - megatron_patch/model/mistral/language_model.py (123:149) duplicated block id: 4092 size: 11 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/checkpoint_reshaping_and_interoperability.py (119:130) - toolkits/model_checkpoints_convertor/qwen/hf2megablocks_qwen1.5.py (131:142) duplicated block id: 4093 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (947:957) - megatron_patch/model/llama3/transformer_legacy.py (934:944) duplicated block id: 4094 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/language_model.py (458:468) - megatron_patch/model/bloom/language_model.py (435:445) duplicated block id: 4095 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/language_model.py (46:60) - megatron_patch/model/glm130b/language_model.py (46:60) duplicated block id: 4096 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/bloom/language_model.py (435:445) - megatron_patch/model/qwen_vl/language_model.py (451:461) duplicated block id: 4097 size: 11 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/deepseek/hf2mcore_deepseek_v2_moe.py (52:64) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (60:72) duplicated block id: 4098 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/language_model.py (439:449) - megatron_patch/model/mistral/language_model.py (453:463) duplicated block id: 4099 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/language_model.py (335:356) - megatron_patch/model/starcoder/language_model.py (320:342) duplicated block id: 4100 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/llama3_1/layer_specs.py (89:101) - megatron_patch/model/qwen1_5/layer_specs.py (89:102) duplicated block id: 4101 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/transformer.py (45:76) - megatron_patch/model/llama2/transformer.py (47:79) duplicated block id: 4102 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/language_model.py (131:156) - megatron_patch/model/qwen/language_model.py (122:148) duplicated block id: 4103 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/qwen2/transformer/attention.py (154:168) - megatron_patch/model/qwen2_vl/attention_vision.py (167:181) duplicated block id: 4104 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/language_model.py (129:154) - megatron_patch/model/llama2/language_model.py (122:148) duplicated block id: 4105 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/language_model.py (129:154) - megatron_patch/model/llava/language_model.py (125:151) duplicated block id: 4106 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (756:767) - megatron_patch/model/bloom/transformer.py (560:571) duplicated block id: 4107 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/qwen1_5/transformer/attention.py (507:527) - megatron_patch/model/qwen2_vl/attention.py (656:676) duplicated block id: 4108 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/llama3/language_model.py (404:416) - megatron_patch/model/starcoder/language_model.py (389:401) duplicated block id: 4109 size: 11 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (790:800) - toolkits/model_checkpoints_convertor/bloom/reward_model_to_megatron.py (476:486) duplicated block id: 4110 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/falcon/language_model.py (129:154) - megatron_patch/model/llama3/language_model.py (111:137) duplicated block id: 4111 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/language_model.py (423:435) - megatron_patch/model/mistral/language_model.py (415:427) duplicated block id: 4112 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/llama2/transformer.py (47:79) - megatron_patch/model/starcoder/transformer.py (47:80) duplicated block id: 4113 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/gpt_model.py (43:58) - megatron_patch/model/llama/gpt_model.py (44:59) duplicated block id: 4114 size: 11 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore_llama3_1.py (24:39) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_dense_and_moe_gqa.py (22:38) duplicated block id: 4115 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/mistral/language_model.py (348:358) - megatron_patch/model/qwen_vl/language_model.py (348:358) duplicated block id: 4116 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/bloom/language_model.py (435:445) - megatron_patch/model/qwen/language_model.py (448:458) duplicated block id: 4117 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/llama3/rms_norm.py (4:18) - megatron_patch/model/llama3_1/rms_norm.py (4:20) duplicated block id: 4118 size: 11 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/reward_model_to_megatron.py (476:486) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (957:969) duplicated block id: 4119 size: 11 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/reward_model_to_megatron.py (476:486) - toolkits/model_checkpoints_convertor/falcon40b/checkpoint_reshaping_and_interoperability.py (758:768) duplicated block id: 4120 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/language_model.py (454:464) - megatron_patch/model/qwen_vl/language_model.py (451:461) duplicated block id: 4121 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer_legacy.py (1685:1705) - megatron_patch/model/llava/transformer.py (1724:1744) duplicated block id: 4122 size: 11 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (44:54) - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (72:85) duplicated block id: 4123 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/language_model.py (458:468) - megatron_patch/model/glm130b/language_model.py (454:464) duplicated block id: 4124 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/bloom/language_model.py (404:416) - megatron_patch/model/qwen/language_model.py (410:422) duplicated block id: 4125 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/bloom/language_model.py (131:156) - megatron_patch/model/qwen/language_model.py (122:148) duplicated block id: 4126 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/llama/language_model.py (129:154) - megatron_patch/model/qwen_vl/language_model.py (123:149) duplicated block id: 4127 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/language_model.py (132:158) - megatron_patch/model/glm130b/language_model.py (131:156) duplicated block id: 4128 size: 11 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon40b/checkpoint_reshaping_and_interoperability.py (45:55) - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (72:85) duplicated block id: 4129 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/language_model.py (131:156) - megatron_patch/model/llama2/language_model.py (122:148) duplicated block id: 4130 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/llama/language_model.py (129:154) - megatron_patch/model/qwen1_5_megablocks/language_model.py (121:147) duplicated block id: 4131 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/transformer/attention.py (655:675) - megatron_patch/model/qwen1_5/transformer/attention.py (507:527) duplicated block id: 4132 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/language_model.py (46:60) - megatron_patch/model/falcon/language_model.py (45:59) duplicated block id: 4133 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/language_model.py (509:519) - megatron_patch/model/glm130b/language_model.py (454:464) duplicated block id: 4134 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/galactica/language_model.py (464:474) - megatron_patch/model/qwen_vl/language_model.py (451:461) duplicated block id: 4135 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/language_model.py (423:435) - megatron_patch/model/qwen1_5_megablocks/language_model.py (412:424) duplicated block id: 4136 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/language_model.py (420:432) - megatron_patch/model/starcoder/language_model.py (389:401) duplicated block id: 4137 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/transformer.py (377:389) - megatron_patch/model/llama/transformer.py (434:446) duplicated block id: 4138 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/falcon/language_model.py (129:154) - megatron_patch/model/qwen_vl/language_model.py (123:149) duplicated block id: 4139 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/transformer.py (364:375) - megatron_patch/model/starcoder/transformer.py (569:580) duplicated block id: 4140 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/language_model.py (439:449) - megatron_patch/model/llama3/language_model.py (442:452) duplicated block id: 4141 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer/mlp.py (177:187) - megatron_patch/model/mixtral_bak/transformer/mlp.py (150:160) duplicated block id: 4142 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/qwen/language_model.py (347:357) - megatron_patch/model/qwen_vl/language_model.py (348:358) duplicated block id: 4143 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/llava/language_model.py (273:288) - megatron_patch/model/starcoder/language_model.py (271:286) duplicated block id: 4144 size: 11 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/mistral/hf2mcore.py (105:116) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (169:180) duplicated block id: 4145 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (947:957) - megatron_patch/model/qwen1_5_megablocks/transformer.py (868:878) duplicated block id: 4146 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/falcon/language_model.py (351:361) - megatron_patch/model/starcoder/language_model.py (347:357) duplicated block id: 4147 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/mistral/language_model.py (415:427) - megatron_patch/model/starcoder/language_model.py (389:401) duplicated block id: 4148 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/language_model.py (167:197) - megatron_patch/model/falcon40b/language_model.py (129:154) duplicated block id: 4149 size: 11 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (72:85) - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (68:78) duplicated block id: 4150 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/llama3/language_model.py (265:280) - megatron_patch/model/starcoder/language_model.py (271:286) duplicated block id: 4151 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/language_model.py (131:156) - megatron_patch/model/qwen1_5_megablocks/language_model.py (121:147) duplicated block id: 4152 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/llama2/language_model.py (270:285) - megatron_patch/model/starcoder/language_model.py (271:286) duplicated block id: 4153 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/qwen/transformer.py (1667:1687) - megatron_patch/model/qwen1_5_megablocks/transformer.py (1587:1607) duplicated block id: 4154 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/language_model.py (365:375) - megatron_patch/model/starcoder/language_model.py (347:357) duplicated block id: 4155 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/language_model.py (167:197) - megatron_patch/model/falcon/language_model.py (129:154) duplicated block id: 4156 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/language_model.py (132:158) - megatron_patch/model/falcon/language_model.py (129:154) duplicated block id: 4157 size: 11 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (933:945) - toolkits/model_checkpoints_convertor/starcoder/checkpoint_reshaping_and_interoperability.py (491:501) duplicated block id: 4158 size: 11 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore.py (137:148) - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (164:175) duplicated block id: 4159 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (1595:1615) - megatron_patch/model/llama2/transformer.py (1725:1745) duplicated block id: 4160 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/llava/language_model.py (350:360) - megatron_patch/model/qwen1_5_megablocks/language_model.py (346:356) duplicated block id: 4161 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/language_model.py (454:464) - megatron_patch/model/starcoder/language_model.py (418:428) duplicated block id: 4162 size: 11 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/mistral/hf2mcore.py (105:116) - toolkits/model_checkpoints_convertor/starcoder/checkpoint_reshaping_and_interoperability.py (119:130) duplicated block id: 4163 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/bloom/language_model.py (131:156) - megatron_patch/model/qwen1_5_megablocks/language_model.py (121:147) duplicated block id: 4164 size: 11 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (140:151) - toolkits/model_checkpoints_convertor/llama/hf2mcore.py (137:148) duplicated block id: 4165 size: 11 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/checkpoint_reshaping_and_interoperability.py (701:712) - toolkits/model_checkpoints_convertor/glm/checkpoint_reshaping_and_interoperability.py (428:439) duplicated block id: 4166 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/mistral/transformer.py (1722:1742) - megatron_patch/model/qwen/transformer.py (1667:1687) duplicated block id: 4167 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/language_model.py (362:372) - megatron_patch/model/starcoder/language_model.py (347:357) duplicated block id: 4168 size: 11 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/hf2te.py (36:46) - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (72:85) duplicated block id: 4169 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/llava/transformer.py (47:79) - megatron_patch/model/starcoder/transformer.py (47:80) duplicated block id: 4170 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/language_model.py (129:154) - megatron_patch/model/qwen_vl/language_model.py (123:149) duplicated block id: 4171 size: 11 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/glm130b/checkpoint_reshaping_and_interoperability.py (409:420) - toolkits/model_checkpoints_convertor/starcoder/checkpoint_reshaping_and_interoperability.py (701:712) duplicated block id: 4172 size: 11 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (817:829) - toolkits/model_checkpoints_convertor/bloom/checkpoint_reshaping_and_interoperability.py (477:487) duplicated block id: 4173 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/mistral/language_model.py (271:286) - megatron_patch/model/starcoder/language_model.py (271:286) duplicated block id: 4174 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/qwen1_5/transformer/attention.py (158:172) - megatron_patch/model/qwen2_vl/attention.py (168:182) duplicated block id: 4175 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/mixtral_bak/transformer/mlp.py (150:160) - megatron_patch/model/qwen1_5/transformer/mlp.py (193:203) duplicated block id: 4176 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (1725:1745) - megatron_patch/model/llama2/transformer.py (1725:1745) duplicated block id: 4177 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/language_model.py (357:367) - megatron_patch/model/llava/language_model.py (350:360) duplicated block id: 4178 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/llava_mcore/vision/vit_layer_specs.py (65:75) - megatron_patch/model/qwen3_moe/gpt_layer_specs.py (236:246) duplicated block id: 4179 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/llama/language_model.py (129:154) - megatron_patch/model/qwen/language_model.py (122:148) duplicated block id: 4180 size: 11 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/chatglm/checkpoint_reshaping_and_interoperability.py (123:134) - toolkits/model_checkpoints_convertor/qwen/hf2megablocks_qwen1.5.py (131:142) duplicated block id: 4181 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/language_model.py (454:464) - megatron_patch/model/qwen/language_model.py (448:458) duplicated block id: 4182 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (645:656) - megatron_patch/model/bloom/transformer.py (560:571) duplicated block id: 4183 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/language_model.py (167:197) - megatron_patch/model/llama/language_model.py (129:154) duplicated block id: 4184 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/falcon/language_model.py (433:443) - megatron_patch/model/mistral/language_model.py (453:463) duplicated block id: 4185 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/moe/experts.py (65:77) - megatron_patch/model/qwen2/moe/experts.py (67:79) duplicated block id: 4186 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/llava_mcore/llava_model.py (75:85) - megatron_patch/model/qwen2_5_vl/model.py (52:63) duplicated block id: 4187 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/transformer_block.py (494:504) - megatron_patch/model/qwen2_5_vl/transformer_block.py (560:570) duplicated block id: 4188 size: 11 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/deepspeed_to_megatron.py (171:181) - toolkits/model_checkpoints_convertor/bloom/deepspeed_to_megatron_ori.py (171:181) duplicated block id: 4189 size: 11 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/checkpoint_reshaping_and_interoperability.py (701:712) - toolkits/model_checkpoints_convertor/glm130b/checkpoint_reshaping_and_interoperability.py (409:420) duplicated block id: 4190 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/language_model.py (284:299) - megatron_patch/model/starcoder/language_model.py (271:286) duplicated block id: 4191 size: 11 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/hf2te.py (214:229) - toolkits/model_checkpoints_convertor/chatglm/checkpoint_reshaping_and_interoperability.py (306:321) duplicated block id: 4192 size: 11 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/reward_model_to_megatron.py (700:711) - toolkits/model_checkpoints_convertor/glm/checkpoint_reshaping_and_interoperability.py (428:439) duplicated block id: 4193 size: 11 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore.py (554:564) - toolkits/model_checkpoints_convertor/mistral/hf2mcore.py (426:436) duplicated block id: 4194 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/model.py (181:192) - megatron_patch/model/llama3_1/model.py (221:232) duplicated block id: 4195 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/llama2/language_model.py (347:357) - megatron_patch/model/llava/language_model.py (350:360) duplicated block id: 4196 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/bloom/language_model.py (47:61) - megatron_patch/model/chatglm/language_model.py (46:60) duplicated block id: 4197 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/qwen/language_model.py (410:422) - megatron_patch/model/starcoder/language_model.py (389:401) duplicated block id: 4198 size: 11 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon40b/checkpoint_reshaping_and_interoperability.py (141:152) - toolkits/model_checkpoints_convertor/llama/hf2mcore.py (137:148) duplicated block id: 4199 size: 11 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/deepseek/hf2mcore_deepseek_v2_moe.py (167:177) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_dense_and_moe_gqa.py (281:291) duplicated block id: 4200 size: 11 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/hf2te.py (110:121) - toolkits/model_checkpoints_convertor/llama/hf2mcore.py (137:148) duplicated block id: 4201 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/language_model.py (458:468) - megatron_patch/model/llama/language_model.py (433:443) duplicated block id: 4202 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/language_model.py (362:372) - megatron_patch/model/glm130b/language_model.py (365:375) duplicated block id: 4203 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/llama2/transformer.py (1725:1745) - megatron_patch/model/qwen/transformer.py (1667:1687) duplicated block id: 4204 size: 11 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_mha.py (76:89) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_moe.py (85:98) duplicated block id: 4205 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer_legacy.py (1685:1705) - megatron_patch/model/qwen1_5_megablocks/transformer.py (1587:1607) duplicated block id: 4206 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/language_model.py (132:158) - megatron_patch/model/bloom/language_model.py (131:156) duplicated block id: 4207 size: 11 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2megablocks_qwen1.5.py (131:142) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (168:179) duplicated block id: 4208 size: 11 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/reward_model_to_megatron.py (119:130) - toolkits/model_checkpoints_convertor/mistral/hf2mcore.py (105:116) duplicated block id: 4209 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/moe/moe_layer.py (127:139) - megatron_patch/model/qwen2/moe/moe_layer.py (128:140) duplicated block id: 4210 size: 11 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon40b/checkpoint_reshaping_and_interoperability.py (758:768) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (957:969) duplicated block id: 4211 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/starcoder/transformer.py (277:289) - megatron_patch/model/starcoder/transformer.py (400:412) duplicated block id: 4212 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/falcon/transformer.py (289:310) - megatron_patch/model/starcoder/transformer.py (328:349) duplicated block id: 4213 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/galactica/language_model.py (382:392) - megatron_patch/model/starcoder/language_model.py (347:357) duplicated block id: 4214 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/falcon/language_model.py (433:443) - megatron_patch/model/starcoder/language_model.py (418:428) duplicated block id: 4215 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/transformer.py (377:389) - megatron_patch/model/galactica/transformer.py (397:409) duplicated block id: 4216 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/llava_mcore/llava_spec.py (69:79) - megatron_patch/model/qwen3_moe/gpt_layer_specs.py (236:246) duplicated block id: 4217 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/language_model.py (433:443) - megatron_patch/model/llama3/language_model.py (442:452) duplicated block id: 4218 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/llama/language_model.py (433:443) - megatron_patch/model/llama2/language_model.py (451:461) duplicated block id: 4219 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/language_model.py (167:197) - megatron_patch/model/bloom/language_model.py (131:156) duplicated block id: 4220 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/language_model.py (454:464) - megatron_patch/model/llama2/language_model.py (451:461) duplicated block id: 4221 size: 11 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (102:113) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (90:101) duplicated block id: 4222 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/language_model.py (423:435) - megatron_patch/model/llama2/language_model.py (413:425) duplicated block id: 4223 size: 11 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore.py (137:148) - toolkits/model_checkpoints_convertor/yi/checkpoint_reshaping_and_interoperability.py (141:152) duplicated block id: 4224 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/transformer.py (364:375) - megatron_patch/model/falcon/transformer.py (621:632) duplicated block id: 4225 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/language_model.py (241:259) - megatron_patch/model/galactica/language_model.py (255:273) duplicated block id: 4226 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/falcon/language_model.py (608:620) - megatron_patch/model/starcoder/language_model.py (495:505) duplicated block id: 4227 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/galactica/transformer.py (889:917) - megatron_patch/model/qwen1_5_megablocks/transformer.py (1617:1646) duplicated block id: 4228 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/moe/experts.py (64:76) - megatron_patch/model/qwen2/moe/experts.py (67:79) duplicated block id: 4229 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/llama2/language_model.py (413:425) - megatron_patch/model/starcoder/language_model.py (389:401) duplicated block id: 4230 size: 11 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (72:85) - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (45:56) duplicated block id: 4231 size: 11 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon/checkpoint_reshaping_and_interoperability.py (45:55) - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (72:85) duplicated block id: 4232 size: 11 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/galactica/checkpoint_reshaping_and_interoperability.py (140:151) - toolkits/model_checkpoints_convertor/llama/hf2mcore.py (137:148) duplicated block id: 4233 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer/attention.py (507:527) - megatron_patch/model/mixtral/transformer/attention.py (655:675) duplicated block id: 4234 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/bloom/language_model.py (131:156) - megatron_patch/model/mistral/language_model.py (123:149) duplicated block id: 4235 size: 11 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/checkpoint_reshaping_and_interoperability.py (477:487) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (957:969) duplicated block id: 4236 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/falcon/transformer.py (289:310) - megatron_patch/model/qwen1_5_megablocks/transformer.py (314:335) duplicated block id: 4237 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/bloom/language_model.py (435:445) - megatron_patch/model/llama2/language_model.py (451:461) duplicated block id: 4238 size: 11 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (834:847) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (907:920) duplicated block id: 4239 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/galactica/language_model.py (464:474) - megatron_patch/model/qwen1_5_megablocks/language_model.py (450:460) duplicated block id: 4240 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/language_model.py (362:372) - megatron_patch/model/falcon/language_model.py (351:361) duplicated block id: 4241 size: 11 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (553:565) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (471:483) duplicated block id: 4242 size: 11 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_dense_and_moe_gqa.py (687:697) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_dense_and_moe_gqa.py (764:774) duplicated block id: 4243 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/falcon/language_model.py (129:154) - megatron_patch/model/qwen/language_model.py (122:148) duplicated block id: 4244 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/llama3/layer_specs.py (89:101) - megatron_patch/model/mixtral_bak/layer_specs.py (110:123) duplicated block id: 4245 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/language_model.py (439:449) - megatron_patch/model/qwen1_5_megablocks/language_model.py (450:460) duplicated block id: 4246 size: 11 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (354:369) - toolkits/model_checkpoints_convertor/baichuan/hf2te.py (214:229) duplicated block id: 4247 size: 11 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/checkpoint_reshaping_and_interoperability.py (404:414) - toolkits/model_checkpoints_convertor/starcoder/checkpoint_reshaping_and_interoperability.py (418:428) duplicated block id: 4248 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/mixtral_bak/transformer/attention.py (132:146) - megatron_patch/model/qwen2_vl/attention.py (168:182) duplicated block id: 4249 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/language_model.py (509:519) - megatron_patch/model/falcon40b/language_model.py (433:443) duplicated block id: 4250 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/transformer.py (45:76) - megatron_patch/model/qwen_vl/transformer.py (47:79) duplicated block id: 4251 size: 11 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (817:829) - toolkits/model_checkpoints_convertor/starcoder/checkpoint_reshaping_and_interoperability.py (491:501) duplicated block id: 4252 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/galactica/language_model.py (464:474) - megatron_patch/model/llama2/language_model.py (451:461) duplicated block id: 4253 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/falcon/transformer.py (429:452) - megatron_patch/model/starcoder/transformer.py (302:325) duplicated block id: 4254 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/language_model.py (458:468) - megatron_patch/model/galactica/language_model.py (464:474) duplicated block id: 4255 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/bloom/language_model.py (363:373) - megatron_patch/model/galactica/language_model.py (382:392) duplicated block id: 4256 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/language_model.py (129:154) - megatron_patch/model/llama3/language_model.py (111:137) duplicated block id: 4257 size: 11 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/reward_model_to_megatron.py (403:413) - toolkits/model_checkpoints_convertor/falcon/checkpoint_reshaping_and_interoperability.py (714:724) duplicated block id: 4258 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/language_model.py (420:432) - megatron_patch/model/glm130b/language_model.py (423:435) duplicated block id: 4259 size: 11 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (790:800) - toolkits/model_checkpoints_convertor/starcoder/checkpoint_reshaping_and_interoperability.py (491:501) duplicated block id: 4260 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/falcon/language_model.py (433:443) - megatron_patch/model/qwen/language_model.py (448:458) duplicated block id: 4261 size: 11 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/checkpoint_reshaping_and_interoperability.py (477:487) - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (933:945) duplicated block id: 4262 size: 11 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore.py (519:529) - toolkits/model_checkpoints_convertor/llama/hf2mcore.py (554:564) duplicated block id: 4263 size: 11 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (140:151) - toolkits/model_checkpoints_convertor/mistral/hf2mcore.py (105:116) duplicated block id: 4264 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/bloom/language_model.py (435:445) - megatron_patch/model/mistral/language_model.py (453:463) duplicated block id: 4265 size: 11 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore.py (137:148) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (168:179) duplicated block id: 4266 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/qwen1_5_megablocks/language_model.py (346:356) - megatron_patch/model/qwen_vl/language_model.py (348:358) duplicated block id: 4267 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/transformer.py (45:76) - megatron_patch/model/llava/transformer.py (47:79) duplicated block id: 4268 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/language_model.py (458:468) - megatron_patch/model/falcon40b/language_model.py (433:443) duplicated block id: 4269 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (46:78) - megatron_patch/model/starcoder/transformer.py (47:80) duplicated block id: 4270 size: 11 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon/checkpoint_reshaping_and_interoperability.py (774:784) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (957:969) duplicated block id: 4271 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/language_model.py (471:483) - megatron_patch/model/starcoder/language_model.py (389:401) duplicated block id: 4272 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/falcon/language_model.py (129:154) - megatron_patch/model/llama2/language_model.py (122:148) duplicated block id: 4273 size: 11 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore.py (137:148) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (92:103) duplicated block id: 4274 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/language_model.py (509:519) - megatron_patch/model/galactica/language_model.py (464:474) duplicated block id: 4275 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (1595:1615) - megatron_patch/model/llava/transformer.py (1724:1744) duplicated block id: 4276 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/language_model.py (458:468) - megatron_patch/model/falcon/language_model.py (433:443) duplicated block id: 4277 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/llama/language_model.py (433:443) - megatron_patch/model/qwen1_5_megablocks/language_model.py (450:460) duplicated block id: 4278 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/language_model.py (423:435) - megatron_patch/model/qwen/language_model.py (410:422) duplicated block id: 4279 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/llama/language_model.py (619:631) - megatron_patch/model/starcoder/language_model.py (495:505) duplicated block id: 4280 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/language_model.py (454:464) - megatron_patch/model/llava/language_model.py (459:469) duplicated block id: 4281 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/qwen2_moe/layer_specs.py (202:213) - megatron_patch/model/qwen3_moe/gpt_layer_specs.py (196:207) duplicated block id: 4282 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/transformer.py (45:76) - megatron_patch/model/llama3/transformer_legacy.py (51:83) duplicated block id: 4283 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/bloom/language_model.py (435:445) - megatron_patch/model/qwen1_5_megablocks/language_model.py (450:460) duplicated block id: 4284 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/llava_mcore/llava_model.py (132:142) - megatron_patch/model/qwen2_vl/model.py (106:116) duplicated block id: 4285 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/bloom/gpt_model.py (43:58) - megatron_patch/model/llama/gpt_model.py (44:59) duplicated block id: 4286 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (1725:1745) - megatron_patch/model/qwen1_5_megablocks/transformer.py (1587:1607) duplicated block id: 4287 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/galactica/language_model.py (464:474) - megatron_patch/model/qwen/language_model.py (448:458) duplicated block id: 4288 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/language_model.py (362:372) - megatron_patch/model/falcon40b/language_model.py (351:361) duplicated block id: 4289 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/llama2/language_model.py (347:357) - megatron_patch/model/qwen_vl/language_model.py (348:358) duplicated block id: 4290 size: 11 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_dense_and_moe_gqa.py (110:121) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_vl.py (113:124) duplicated block id: 4291 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/language_model.py (280:295) - megatron_patch/model/starcoder/language_model.py (271:286) duplicated block id: 4292 size: 11 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/galactica/checkpoint_reshaping_and_interoperability.py (140:151) - toolkits/model_checkpoints_convertor/mistral/hf2mcore.py (105:116) duplicated block id: 4293 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/llama/language_model.py (129:154) - megatron_patch/model/mistral/language_model.py (123:149) duplicated block id: 4294 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/language_model.py (439:449) - megatron_patch/model/qwen_vl/language_model.py (451:461) duplicated block id: 4295 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/language_model.py (509:519) - megatron_patch/model/falcon/language_model.py (433:443) duplicated block id: 4296 size: 11 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/glm/checkpoint_reshaping_and_interoperability.py (428:439) - toolkits/model_checkpoints_convertor/starcoder/checkpoint_reshaping_and_interoperability.py (701:712) duplicated block id: 4297 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/llama/language_model.py (129:154) - megatron_patch/model/llama3/language_model.py (111:137) duplicated block id: 4298 size: 11 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/reward_model_to_megatron.py (476:486) - toolkits/model_checkpoints_convertor/falcon/checkpoint_reshaping_and_interoperability.py (774:784) duplicated block id: 4299 size: 11 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/hf2te.py (214:229) - toolkits/model_checkpoints_convertor/falcon/checkpoint_reshaping_and_interoperability.py (316:331) duplicated block id: 4300 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/language_model.py (454:464) - megatron_patch/model/llama3/language_model.py (442:452) duplicated block id: 4301 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/falcon/language_model.py (129:154) - megatron_patch/model/mistral/language_model.py (123:149) duplicated block id: 4302 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/language_model.py (46:60) - megatron_patch/model/galactica/language_model.py (45:59) duplicated block id: 4303 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/bloom/language_model.py (435:445) - megatron_patch/model/llava/language_model.py (459:469) duplicated block id: 4304 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/falcon/transformer.py (289:310) - megatron_patch/model/llama2/transformer.py (423:444) duplicated block id: 4305 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/transformer_block.py (507:523) - megatron_patch/model/qwen2_5_vl/transformer_block.py (575:591) duplicated block id: 4306 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer_legacy.py (1685:1705) - megatron_patch/model/qwen_vl/transformer.py (1724:1744) duplicated block id: 4307 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/falcon/transformer.py (289:310) - megatron_patch/model/mistral/transformer.py (424:445) duplicated block id: 4308 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/bloom/language_model.py (435:445) - megatron_patch/model/starcoder/language_model.py (418:428) duplicated block id: 4309 size: 11 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (484:496) - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (553:565) duplicated block id: 4310 size: 11 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/mistral/hf2mcore.py (172:186) - toolkits/model_checkpoints_convertor/qwen/hf2megablocks_qwen1.5.py (229:242) duplicated block id: 4311 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/falcon/language_model.py (433:443) - megatron_patch/model/llama3/language_model.py (442:452) duplicated block id: 4312 size: 11 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (60:72) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_vl.py (56:68) duplicated block id: 4313 size: 11 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore.py (219:231) - toolkits/model_checkpoints_convertor/mistral/hf2mcore.py (158:170) duplicated block id: 4314 size: 11 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (44:54) - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (72:85) duplicated block id: 4315 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (1595:1615) - megatron_patch/model/qwen_vl/transformer.py (1724:1744) duplicated block id: 4316 size: 11 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/galactica/checkpoint_reshaping_and_interoperability.py (140:151) - toolkits/model_checkpoints_convertor/qwen/hf2megablocks_qwen1.5.py (131:142) duplicated block id: 4317 size: 11 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/reward_model_to_megatron.py (476:486) - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (933:945) duplicated block id: 4318 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/language_model.py (433:443) - megatron_patch/model/mistral/language_model.py (453:463) duplicated block id: 4319 size: 11 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/hf2te.py (214:229) - toolkits/model_checkpoints_convertor/yi/checkpoint_reshaping_and_interoperability.py (341:356) duplicated block id: 4320 size: 11 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore.py (623:635) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_moe.py (433:445) duplicated block id: 4321 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/language_model.py (403:413) - megatron_patch/model/llava/language_model.py (350:360) duplicated block id: 4322 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/falcon/transformer.py (289:310) - megatron_patch/model/qwen/transformer.py (417:438) duplicated block id: 4323 size: 11 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2megablocks_qwen1.5.py (131:142) - toolkits/model_checkpoints_convertor/starcoder/checkpoint_reshaping_and_interoperability.py (119:130) duplicated block id: 4324 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (560:571) - megatron_patch/model/llama3/transformer_legacy.py (743:754) duplicated block id: 4325 size: 11 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/reward_model_to_megatron.py (476:486) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (890:902) duplicated block id: 4326 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/bloom/language_model.py (131:156) - megatron_patch/model/llava/language_model.py (125:151) duplicated block id: 4327 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/qwen/language_model.py (270:285) - megatron_patch/model/starcoder/language_model.py (271:286) duplicated block id: 4328 size: 11 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (140:151) - toolkits/model_checkpoints_convertor/qwen/hf2megablocks_qwen1.5.py (131:142) duplicated block id: 4329 size: 11 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/mistral/hf2mcore.py (255:266) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_dense_and_moe_gqa.py (225:236) duplicated block id: 4330 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/bloom/language_model.py (363:373) - megatron_patch/model/chatglm/language_model.py (362:372) duplicated block id: 4331 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (326:347) - megatron_patch/model/falcon/transformer.py (289:310) duplicated block id: 4332 size: 11 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/mistral/hf2mcore.py (105:116) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (92:103) duplicated block id: 4333 size: 11 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (112:123) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (90:101) duplicated block id: 4334 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/language_model.py (46:60) - megatron_patch/model/llama/language_model.py (45:59) duplicated block id: 4335 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/llama/language_model.py (433:443) - megatron_patch/model/mistral/language_model.py (453:463) duplicated block id: 4336 size: 11 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/reward_model_to_megatron.py (403:413) - toolkits/model_checkpoints_convertor/starcoder/checkpoint_reshaping_and_interoperability.py (418:428) duplicated block id: 4337 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/language_model.py (433:443) - megatron_patch/model/qwen1_5_megablocks/language_model.py (450:460) duplicated block id: 4338 size: 11 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2.5_vl.py (113:124) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_dense_and_moe_gqa.py (110:121) duplicated block id: 4339 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/language_model.py (362:372) - megatron_patch/model/galactica/language_model.py (382:392) duplicated block id: 4340 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/llama/language_model.py (433:443) - megatron_patch/model/llama3/language_model.py (442:452) duplicated block id: 4341 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/qwen2/transformer/attention.py (154:168) - megatron_patch/model/qwen2_vl/attention.py (168:182) duplicated block id: 4342 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/language_model.py (471:483) - megatron_patch/model/bloom/language_model.py (404:416) duplicated block id: 4343 size: 11 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/hf2te.py (110:121) - toolkits/model_checkpoints_convertor/qwen/hf2megablocks_qwen1.5.py (131:142) duplicated block id: 4344 size: 11 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/hf2te.py (371:382) - toolkits/model_checkpoints_convertor/yi/checkpoint_reshaping_and_interoperability.py (500:511) duplicated block id: 4345 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/bloom/language_model.py (363:373) - megatron_patch/model/falcon40b/language_model.py (351:361) duplicated block id: 4346 size: 11 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_moe.py (143:154) - toolkits/model_checkpoints_convertor/qwen/hf2megablocks_qwen1.5.py (308:319) duplicated block id: 4347 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/transformer.py (881:909) - megatron_patch/model/qwen1_5_megablocks/transformer.py (1617:1646) duplicated block id: 4348 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/language_model.py (357:367) - megatron_patch/model/qwen_vl/language_model.py (348:358) duplicated block id: 4349 size: 11 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/mistral/hf2mcore.py (105:116) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (168:179) duplicated block id: 4350 size: 11 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/checkpoint_reshaping_and_interoperability.py (404:414) - toolkits/model_checkpoints_convertor/falcon/checkpoint_reshaping_and_interoperability.py (714:724) duplicated block id: 4351 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/language_model.py (509:519) - megatron_patch/model/llama/language_model.py (433:443) duplicated block id: 4352 size: 11 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/reward_model_to_megatron.py (119:130) - toolkits/model_checkpoints_convertor/qwen/hf2megablocks_qwen1.5.py (131:142) duplicated block id: 4353 size: 11 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon/checkpoint_reshaping_and_interoperability.py (141:152) - toolkits/model_checkpoints_convertor/mistral/hf2mcore.py (105:116) duplicated block id: 4354 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/bloom/language_model.py (404:416) - megatron_patch/model/mistral/language_model.py (415:427) duplicated block id: 4355 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/llava/language_model.py (350:360) - megatron_patch/model/mistral/language_model.py (348:358) duplicated block id: 4356 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/bloom/language_model.py (131:156) - megatron_patch/model/qwen_vl/language_model.py (123:149) duplicated block id: 4357 size: 11 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (957:969) - toolkits/model_checkpoints_convertor/starcoder/checkpoint_reshaping_and_interoperability.py (491:501) duplicated block id: 4358 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/llama/language_model.py (129:154) - megatron_patch/model/llava/language_model.py (125:151) duplicated block id: 4359 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/gpt_model.py (43:58) - megatron_patch/model/llama/gpt_model.py (44:59) duplicated block id: 4360 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/moe/moe_layer.py (134:146) - megatron_patch/model/qwen2/moe/moe_layer.py (128:140) duplicated block id: 4361 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/falcon/language_model.py (129:154) - megatron_patch/model/llava/language_model.py (125:151) duplicated block id: 4362 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/language_model.py (433:443) - megatron_patch/model/starcoder/language_model.py (418:428) duplicated block id: 4363 size: 11 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (890:902) - toolkits/model_checkpoints_convertor/starcoder/checkpoint_reshaping_and_interoperability.py (491:501) duplicated block id: 4364 size: 11 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/hf2te.py (109:120) - toolkits/model_checkpoints_convertor/llama/hf2mcore.py (137:148) duplicated block id: 4365 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/falcon/language_model.py (433:443) - megatron_patch/model/qwen_vl/language_model.py (451:461) duplicated block id: 4366 size: 11 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (60:72) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_moe.py (114:126) duplicated block id: 4367 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/falcon/language_model.py (433:443) - megatron_patch/model/qwen1_5_megablocks/language_model.py (450:460) duplicated block id: 4368 size: 11 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/hf2te.py (109:120) - toolkits/model_checkpoints_convertor/mistral/hf2mcore.py (105:116) duplicated block id: 4369 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/language_model.py (454:464) - megatron_patch/model/qwen1_5_megablocks/language_model.py (450:460) duplicated block id: 4370 size: 11 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (140:151) - toolkits/model_checkpoints_convertor/mistral/hf2mcore.py (105:116) duplicated block id: 4371 size: 11 cleaned lines of code in 2 files: - megatron_patch/tokenizer/tokenization_baichuan.py (55:65) - megatron_patch/tokenizer/tokenization_yi.py (73:83) duplicated block id: 4372 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/qwen1_5/transformer/attention.py (158:172) - megatron_patch/model/qwen2_vl/attention_vision.py (167:181) duplicated block id: 4373 size: 11 cleaned lines of code in 2 files: - megatron_patch/template/helper.py (47:58) - megatron_patch/template/helper.py (93:104) duplicated block id: 4374 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/language_model.py (351:361) - megatron_patch/model/starcoder/language_model.py (347:357) duplicated block id: 4375 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/language_model.py (131:156) - megatron_patch/model/qwen_vl/language_model.py (123:149) duplicated block id: 4376 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/transformer.py (45:76) - megatron_patch/model/qwen1_5_megablocks/transformer.py (45:77) duplicated block id: 4377 size: 11 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (60:72) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_dense_and_moe_gqa.py (40:52) duplicated block id: 4378 size: 11 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_dense_and_moe_gqa.py (764:774) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_moe.py (648:659) duplicated block id: 4379 size: 11 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/reward_model_to_megatron.py (700:711) - toolkits/model_checkpoints_convertor/glm130b/checkpoint_reshaping_and_interoperability.py (409:420) duplicated block id: 4380 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/language_model.py (439:449) - megatron_patch/model/llava/language_model.py (459:469) duplicated block id: 4381 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/llama3_1/layer_specs.py (89:101) - megatron_patch/model/mixtral_bak/layer_specs.py (110:123) duplicated block id: 4382 size: 11 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (164:175) - toolkits/model_checkpoints_convertor/mistral/hf2mcore.py (105:116) duplicated block id: 4383 size: 11 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_dense_and_moe_gqa.py (687:697) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_moe.py (648:659) duplicated block id: 4384 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer/attention.py (507:527) - megatron_patch/model/qwen2_vl/attention.py (656:676) duplicated block id: 4385 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/mixtral_bak/transformer/attention.py (132:146) - megatron_patch/model/qwen2_vl/attention_vision.py (167:181) duplicated block id: 4386 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/qwen1_5/transformer/attention.py (507:527) - megatron_patch/model/qwen2_vl/attention_vision.py (655:675) duplicated block id: 4387 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/language_model.py (509:519) - megatron_patch/model/bloom/language_model.py (435:445) duplicated block id: 4388 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (1725:1745) - megatron_patch/model/qwen_vl/transformer.py (1724:1744) duplicated block id: 4389 size: 11 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/checkpoint_reshaping_and_interoperability.py (404:414) - toolkits/model_checkpoints_convertor/falcon40b/checkpoint_reshaping_and_interoperability.py (699:709) duplicated block id: 4390 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/llama/language_model.py (433:443) - megatron_patch/model/starcoder/language_model.py (418:428) duplicated block id: 4391 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/language_model.py (129:154) - megatron_patch/model/qwen/language_model.py (122:148) duplicated block id: 4392 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/qwen2_vl/rope_utils.py (55:65) - megatron_patch/model/qwen2_vl/rotary_pos_embedding.py (46:56) duplicated block id: 4393 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/bloom/language_model.py (363:373) - megatron_patch/model/glm130b/language_model.py (365:375) duplicated block id: 4394 size: 11 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore.py (519:529) - toolkits/model_checkpoints_convertor/mistral/hf2mcore.py (461:471) duplicated block id: 4395 size: 11 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (92:103) - toolkits/model_checkpoints_convertor/qwen/hf2megablocks_qwen1.5.py (131:142) duplicated block id: 4396 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/language_model.py (100:129) - megatron_patch/model/starcoder/language_model.py (87:118) duplicated block id: 4397 size: 11 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/chatglm/checkpoint_reshaping_and_interoperability.py (123:134) - toolkits/model_checkpoints_convertor/llama/hf2mcore.py (137:148) duplicated block id: 4398 size: 11 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (140:151) - toolkits/model_checkpoints_convertor/qwen/hf2megablocks_qwen1.5.py (131:142) duplicated block id: 4399 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/mistral/transformer.py (47:79) - megatron_patch/model/starcoder/transformer.py (47:80) duplicated block id: 4400 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/moe/router.py (66:97) - megatron_patch/model/qwen2/moe/router.py (150:180) duplicated block id: 4401 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/language_model.py (439:449) - megatron_patch/model/llama2/language_model.py (451:461) duplicated block id: 4402 size: 11 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (341:356) - toolkits/model_checkpoints_convertor/baichuan2/hf2te.py (198:213) duplicated block id: 4403 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/falcon/transformer.py (289:310) - megatron_patch/model/llama3/transformer_legacy.py (422:443) duplicated block id: 4404 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/language_model.py (46:60) - megatron_patch/model/falcon40b/language_model.py (45:59) duplicated block id: 4405 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/galactica/language_model.py (637:649) - megatron_patch/model/starcoder/language_model.py (495:505) duplicated block id: 4406 size: 11 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/galactica/checkpoint_reshaping_and_interoperability.py (44:54) - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (72:85) duplicated block id: 4407 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/bloom/language_model.py (435:445) - megatron_patch/model/llama3/language_model.py (442:452) duplicated block id: 4408 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/galactica/language_model.py (464:474) - megatron_patch/model/llava/language_model.py (459:469) duplicated block id: 4409 size: 11 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (817:829) - toolkits/model_checkpoints_convertor/bloom/reward_model_to_megatron.py (476:486) duplicated block id: 4410 size: 11 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon/checkpoint_reshaping_and_interoperability.py (141:152) - toolkits/model_checkpoints_convertor/qwen/hf2megablocks_qwen1.5.py (131:142) duplicated block id: 4411 size: 11 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/chatglm/checkpoint_reshaping_and_interoperability.py (278:296) - toolkits/model_checkpoints_convertor/falcon40b/checkpoint_reshaping_and_interoperability.py (281:299) duplicated block id: 4412 size: 11 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/mistral/hf2mcore.py (105:116) - toolkits/model_checkpoints_convertor/yi/checkpoint_reshaping_and_interoperability.py (141:152) duplicated block id: 4413 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (1595:1615) - megatron_patch/model/mistral/transformer.py (1722:1742) duplicated block id: 4414 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (42:74) - megatron_patch/model/starcoder/transformer.py (47:80) duplicated block id: 4415 size: 11 cleaned lines of code in 2 files: - toolkits/pretrain_data_preprocessing/preprocess_data.py (146:156) - toolkits/pretrain_data_preprocessing/preprocess_data_megatron.py (251:263) duplicated block id: 4416 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/qwen_vl/transformer.py (47:79) - megatron_patch/model/starcoder/transformer.py (47:80) duplicated block id: 4417 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/transformer.py (612:623) - megatron_patch/model/starcoder/transformer.py (758:769) duplicated block id: 4418 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/mixtral_bak/transformer/mlp.py (150:160) - megatron_patch/model/qwen2/transformer/mlp.py (196:206) duplicated block id: 4419 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/language_model.py (129:154) - megatron_patch/model/qwen1_5_megablocks/language_model.py (121:147) duplicated block id: 4420 size: 11 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/checkpoint_reshaping_and_interoperability.py (477:487) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (890:902) duplicated block id: 4421 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/mlp.py (96:110) - megatron_patch/model/llama3/transformer/mlp.py (102:116) duplicated block id: 4422 size: 11 cleaned lines of code in 2 files: - megatron_patch/generation/generation.py (226:239) - megatron_patch/generation/generation.py (399:412) duplicated block id: 4423 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/transformer.py (612:623) - megatron_patch/model/llama3/transformer_legacy.py (743:754) duplicated block id: 4424 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/language_model.py (167:197) - megatron_patch/model/glm130b/language_model.py (131:156) duplicated block id: 4425 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/llava/language_model.py (350:360) - megatron_patch/model/qwen/language_model.py (347:357) duplicated block id: 4426 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (46:78) - megatron_patch/model/glm130b/transformer.py (45:76) duplicated block id: 4427 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/galactica/language_model.py (464:474) - megatron_patch/model/llama3/language_model.py (442:452) duplicated block id: 4428 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/qwen/transformer.py (1667:1687) - megatron_patch/model/qwen_vl/transformer.py (1724:1744) duplicated block id: 4429 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (435:456) - megatron_patch/model/falcon/transformer.py (289:310) duplicated block id: 4430 size: 11 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/deepseek/hf2mcore_deepseek_v2_moe.py (234:249) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_moe.py (365:381) duplicated block id: 4431 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer_legacy.py (1685:1705) - megatron_patch/model/mistral/transformer.py (1722:1742) duplicated block id: 4432 size: 11 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon40b/checkpoint_reshaping_and_interoperability.py (141:152) - toolkits/model_checkpoints_convertor/mistral/hf2mcore.py (105:116) duplicated block id: 4433 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/falcon/transformer.py (403:415) - megatron_patch/model/starcoder/transformer.py (277:289) duplicated block id: 4434 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/multi_latent_attention.py (111:121) - megatron_patch/model/qwen2_vl/attention.py (334:344) duplicated block id: 4435 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/language_model.py (362:372) - megatron_patch/model/llama/language_model.py (351:361) duplicated block id: 4436 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (42:74) - megatron_patch/model/glm130b/transformer.py (45:76) duplicated block id: 4437 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (560:571) - megatron_patch/model/qwen/transformer.py (738:749) duplicated block id: 4438 size: 11 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (805:818) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (907:920) duplicated block id: 4439 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/bloom/language_model.py (404:416) - megatron_patch/model/llama2/language_model.py (413:425) duplicated block id: 4440 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/galactica/language_model.py (464:474) - megatron_patch/model/mistral/language_model.py (453:463) duplicated block id: 4441 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/language_model.py (131:156) - megatron_patch/model/llama3/language_model.py (111:137) duplicated block id: 4442 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/llama3/layer_specs.py (89:101) - megatron_patch/model/qwen2/layer_specs.py (99:112) duplicated block id: 4443 size: 11 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon40b/checkpoint_reshaping_and_interoperability.py (141:152) - toolkits/model_checkpoints_convertor/qwen/hf2megablocks_qwen1.5.py (131:142) duplicated block id: 4444 size: 11 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon/checkpoint_reshaping_and_interoperability.py (141:152) - toolkits/model_checkpoints_convertor/llama/hf2mcore.py (137:148) duplicated block id: 4445 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/qwen1_5_megablocks/language_model.py (269:284) - megatron_patch/model/starcoder/language_model.py (271:286) duplicated block id: 4446 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/language_model.py (439:449) - megatron_patch/model/qwen/language_model.py (448:458) duplicated block id: 4447 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/language_model.py (129:154) - megatron_patch/model/mistral/language_model.py (123:149) duplicated block id: 4448 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/language_model.py (439:449) - megatron_patch/model/starcoder/language_model.py (418:428) duplicated block id: 4449 size: 11 cleaned lines of code in 2 files: - toolkits/distributed_checkpoints_convertor/impl/general/m2h_synchronizer.py (188:200) - toolkits/distributed_checkpoints_convertor/impl/general/m2h_synchronizer.py (214:226) duplicated block id: 4450 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/llama3_1/model.py (221:232) - megatron_patch/model/mixtral/model.py (141:152) duplicated block id: 4451 size: 11 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/hf2te.py (343:354) - toolkits/model_checkpoints_convertor/yi/checkpoint_reshaping_and_interoperability.py (500:511) duplicated block id: 4452 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/language_model.py (420:432) - megatron_patch/model/bloom/language_model.py (404:416) duplicated block id: 4453 size: 11 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore.py (137:148) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (169:180) duplicated block id: 4454 size: 11 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2megablocks_qwen1.5.py (131:142) - toolkits/model_checkpoints_convertor/yi/checkpoint_reshaping_and_interoperability.py (141:152) duplicated block id: 4455 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/bloom/language_model.py (131:156) - megatron_patch/model/llama2/language_model.py (122:148) duplicated block id: 4456 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/qwen2/moe/moe_layer.py (128:140) - megatron_patch/model/qwen3_moe/moe/moe_layer.py (66:78) duplicated block id: 4457 size: 11 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/reward_model_to_megatron.py (403:413) - toolkits/model_checkpoints_convertor/falcon40b/checkpoint_reshaping_and_interoperability.py (699:709) duplicated block id: 4458 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/llava_mcore/llava_model.py (75:85) - megatron_patch/model/qwen2_vl/model.py (52:63) duplicated block id: 4459 size: 11 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/checkpoint_reshaping_and_interoperability.py (119:130) - toolkits/model_checkpoints_convertor/mistral/hf2mcore.py (105:116) duplicated block id: 4460 size: 11 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (790:800) - toolkits/model_checkpoints_convertor/bloom/checkpoint_reshaping_and_interoperability.py (477:487) duplicated block id: 4461 size: 11 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (140:151) - toolkits/model_checkpoints_convertor/llama/hf2mcore.py (137:148) duplicated block id: 4462 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/language_model.py (454:464) - megatron_patch/model/mistral/language_model.py (453:463) duplicated block id: 4463 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (645:656) - megatron_patch/model/glm130b/transformer.py (612:623) duplicated block id: 4464 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/galactica/language_model.py (464:474) - megatron_patch/model/starcoder/language_model.py (418:428) duplicated block id: 4465 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/transformer.py (612:623) - megatron_patch/model/qwen/transformer.py (738:749) duplicated block id: 4466 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/language_model.py (423:435) - megatron_patch/model/llama3/language_model.py (404:416) duplicated block id: 4467 size: 11 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2megablocks_qwen1.5.py (131:142) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (169:180) duplicated block id: 4468 size: 11 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/hf2te.py (214:229) - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (392:407) duplicated block id: 4469 size: 11 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/checkpoint_reshaping_and_interoperability.py (119:130) - toolkits/model_checkpoints_convertor/llama/hf2mcore.py (137:148) duplicated block id: 4470 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/qwen1_5_megablocks/language_model.py (412:424) - megatron_patch/model/starcoder/language_model.py (389:401) duplicated block id: 4471 size: 11 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/deepseek/hf2mcore_deepseek_v2_moe.py (499:510) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_dense_and_moe_gqa.py (687:697) duplicated block id: 4472 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/language_model.py (433:443) - megatron_patch/model/llama2/language_model.py (451:461) duplicated block id: 4473 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/llama3/language_model.py (342:352) - megatron_patch/model/llava/language_model.py (350:360) duplicated block id: 4474 size: 11 cleaned lines of code in 2 files: - megatron_patch/model/llama3/layer_specs.py (89:101) - megatron_patch/model/qwen1_5/layer_specs.py (89:102) duplicated block id: 4475 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/galactica/language_model.py (706:717) - megatron_patch/model/llava/language_model.py (692:703) duplicated block id: 4476 size: 10 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/hf2te.py (178:194) - toolkits/model_checkpoints_convertor/falcon/checkpoint_reshaping_and_interoperability.py (289:305) duplicated block id: 4477 size: 10 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (916:925) - toolkits/model_checkpoints_convertor/starcoder/checkpoint_reshaping_and_interoperability.py (827:836) duplicated block id: 4478 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/qwen1_5/moe/token_dispatcher.py (475:488) - megatron_patch/model/qwen2/moe/token_dispatcher.py (521:535) duplicated block id: 4479 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (498:507) - megatron_patch/model/glm130b/transformer.py (545:554) duplicated block id: 4480 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/language_model.py (100:128) - megatron_patch/model/mistral/language_model.py (90:120) duplicated block id: 4481 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/falcon/language_model.py (197:210) - megatron_patch/model/llama2/language_model.py (184:197) duplicated block id: 4482 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/galactica/language_model.py (581:590) - megatron_patch/model/mistral/language_model.py (501:510) duplicated block id: 4483 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/falcon/transformer.py (992:1002) - megatron_patch/model/qwen/transformer.py (1177:1187) duplicated block id: 4484 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/llama3/language_model.py (19:30) - megatron_patch/model/starcoder/language_model.py (29:40) duplicated block id: 4485 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/transformer.py (90:100) - megatron_patch/model/llama/transformer.py (63:73) duplicated block id: 4486 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/qwen2/layer_specs.py (53:62) - megatron_patch/model/qwen2_moe/layer_specs.py (159:168) duplicated block id: 4487 size: 10 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (1088:1105) - toolkits/model_checkpoints_convertor/starcoder/checkpoint_reshaping_and_interoperability.py (520:537) duplicated block id: 4488 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/bloom/language_model.py (100:128) - megatron_patch/model/llama2/language_model.py (89:119) duplicated block id: 4489 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/language_model.py (132:155) - megatron_patch/model/mistral/language_model.py (125:149) duplicated block id: 4490 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/bloom/language_model.py (263:276) - megatron_patch/model/qwen/language_model.py (247:260) duplicated block id: 4491 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/language_model.py (132:155) - megatron_patch/model/llama2/language_model.py (124:148) duplicated block id: 4492 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/bloom/language_model.py (100:128) - megatron_patch/model/qwen/language_model.py (89:119) duplicated block id: 4493 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (561:571) - megatron_patch/model/falcon/transformer.py (797:808) duplicated block id: 4494 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/model.py (112:121) - megatron_patch/model/qwen1_5/model.py (130:139) duplicated block id: 4495 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/language_model.py (245:258) - megatron_patch/model/mistral/language_model.py (248:261) duplicated block id: 4496 size: 10 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/deepseek/hf2mcore_deepseek_v3_moe.py (519:530) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_moe.py (519:532) duplicated block id: 4497 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/mistral/language_model.py (31:42) - megatron_patch/model/starcoder/language_model.py (29:40) duplicated block id: 4498 size: 10 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (114:123) - toolkits/model_checkpoints_convertor/mistral/hf2mcore.py (105:114) duplicated block id: 4499 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (799:810) - megatron_patch/model/llama3/transformer_legacy.py (787:798) duplicated block id: 4500 size: 10 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_moe.py (352:361) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_vl.py (143:152) duplicated block id: 4501 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/qwen2/layer_specs.py (58:67) - megatron_patch/model/qwen2_vl/layer_specs.py (49:58) duplicated block id: 4502 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/transformer/attention.py (170:181) - megatron_patch/model/qwen2_vl/attention.py (159:170) duplicated block id: 4503 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/llama/transformer.py (187:198) - megatron_patch/model/starcoder/transformer.py (217:228) duplicated block id: 4504 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/falcon/transformer.py (1166:1182) - megatron_patch/model/starcoder/transformer.py (1226:1238) duplicated block id: 4505 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/bloom/language_model.py (100:128) - megatron_patch/model/chatglm/language_model.py (100:128) duplicated block id: 4506 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/llama3_1/layer_specs.py (49:58) - megatron_patch/model/qwen2_vl/layer_specs.py (49:58) duplicated block id: 4507 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/llava_mcore/vision/vit_layer_specs.py (69:78) - megatron_patch/model/mixtral_bak/layer_specs.py (99:108) duplicated block id: 4508 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/bloom/language_model.py (263:276) - megatron_patch/model/llava/language_model.py (250:263) duplicated block id: 4509 size: 10 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/hf2te.py (178:194) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (384:400) duplicated block id: 4510 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/qwen1_5/transformer/attention.py (338:361) - megatron_patch/model/qwen2_vl/attention_vision.py (471:494) duplicated block id: 4511 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/galactica/language_model.py (276:289) - megatron_patch/model/llava/language_model.py (250:263) duplicated block id: 4512 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/transformer.py (831:841) - megatron_patch/model/starcoder/transformer.py (1145:1155) duplicated block id: 4513 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/language_model.py (397:409) - megatron_patch/model/glm130b/language_model.py (413:425) duplicated block id: 4514 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/transformer.py (831:841) - megatron_patch/model/glm130b/transformer.py (909:919) duplicated block id: 4515 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/model.py (81:93) - megatron_patch/model/mixtral_bak/model.py (68:80) duplicated block id: 4516 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/language_model.py (99:127) - megatron_patch/model/starcoder/language_model.py (87:117) duplicated block id: 4517 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/language_model.py (197:210) - megatron_patch/model/qwen/language_model.py (184:197) duplicated block id: 4518 size: 10 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (286:306) - toolkits/model_checkpoints_convertor/falcon40b/checkpoint_reshaping_and_interoperability.py (255:275) duplicated block id: 4519 size: 10 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/mistral/hf2mcore.py (375:388) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_vl.py (574:588) duplicated block id: 4520 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (797:811) - megatron_patch/model/qwen1_5_megablocks/transformer.py (1111:1125) duplicated block id: 4521 size: 10 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/reward_model_to_megatron.py (504:521) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (983:1000) duplicated block id: 4522 size: 10 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/hf2te.py (281:292) - toolkits/model_checkpoints_convertor/chatglm/checkpoint_reshaping_and_interoperability.py (362:373) duplicated block id: 4523 size: 10 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/mistral/hf2mcore.py (274:283) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_moe.py (352:361) duplicated block id: 4524 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/language_model.py (169:197) - megatron_patch/model/chatglm/language_model.py (132:155) duplicated block id: 4525 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/language_model.py (99:127) - megatron_patch/model/llava/language_model.py (92:122) duplicated block id: 4526 size: 10 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/mistral/hf2mcore.py (204:213) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_dense_and_moe_gqa.py (204:213) duplicated block id: 4527 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/model.py (81:93) - megatron_patch/model/mixtral/model.py (69:81) duplicated block id: 4528 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/language_model.py (99:127) - megatron_patch/model/mistral/language_model.py (90:120) duplicated block id: 4529 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/language_model.py (397:409) - megatron_patch/model/starcoder/language_model.py (379:391) duplicated block id: 4530 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/galactica/transformer.py (431:442) - megatron_patch/model/glm130b/transformer.py (523:534) duplicated block id: 4531 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/bloom/language_model.py (100:128) - megatron_patch/model/llava/language_model.py (92:122) duplicated block id: 4532 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (797:811) - megatron_patch/model/llama2/transformer.py (1227:1241) duplicated block id: 4533 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/language_model.py (257:270) - megatron_patch/model/llama/language_model.py (245:258) duplicated block id: 4534 size: 10 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/hf2te.py (309:320) - toolkits/model_checkpoints_convertor/chatglm/checkpoint_reshaping_and_interoperability.py (362:373) duplicated block id: 4535 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/llama3_1/model.py (188:197) - megatron_patch/model/mixtral/model.py (112:121) duplicated block id: 4536 size: 10 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/hf2te.py (424:435) - toolkits/model_checkpoints_convertor/falcon40b/checkpoint_reshaping_and_interoperability.py (547:558) duplicated block id: 4537 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/llama/transformer.py (570:580) - megatron_patch/model/starcoder/transformer.py (759:769) duplicated block id: 4538 size: 10 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/reward_model_to_megatron.py (817:826) - toolkits/model_checkpoints_convertor/falcon/checkpoint_reshaping_and_interoperability.py (843:852) duplicated block id: 4539 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/llama/language_model.py (98:126) - megatron_patch/model/qwen/language_model.py (89:119) duplicated block id: 4540 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/galactica/language_model.py (98:126) - megatron_patch/model/qwen1_5_megablocks/language_model.py (88:118) duplicated block id: 4541 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (1136:1155) - megatron_patch/model/galactica/transformer.py (863:883) duplicated block id: 4542 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/mixtral_bak/model.py (94:107) - megatron_patch/model/qwen1_5/model.py (102:115) duplicated block id: 4543 size: 10 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (668:678) - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (781:791) duplicated block id: 4544 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/transformer.py (715:725) - megatron_patch/model/llama/transformer.py (886:896) duplicated block id: 4545 size: 10 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/checkpoint_reshaping_and_interoperability.py (818:827) - toolkits/model_checkpoints_convertor/falcon/checkpoint_reshaping_and_interoperability.py (843:852) duplicated block id: 4546 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/transformer.py (129:140) - megatron_patch/model/starcoder/transformer.py (217:228) duplicated block id: 4547 size: 10 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/hf2te.py (178:194) - toolkits/model_checkpoints_convertor/yi/checkpoint_reshaping_and_interoperability.py (314:330) duplicated block id: 4548 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (1235:1245) - megatron_patch/model/llama/transformer.py (751:761) duplicated block id: 4549 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/transformer.py (711:724) - megatron_patch/model/qwen1_5_megablocks/transformer.py (790:803) duplicated block id: 4550 size: 10 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (942:951) - toolkits/model_checkpoints_convertor/starcoder/checkpoint_reshaping_and_interoperability.py (827:836) duplicated block id: 4551 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/falcon/language_model.py (197:210) - megatron_patch/model/starcoder/language_model.py (183:196) duplicated block id: 4552 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/language_model.py (257:270) - megatron_patch/model/falcon/language_model.py (245:258) duplicated block id: 4553 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/falcon/language_model.py (197:210) - megatron_patch/model/mistral/language_model.py (185:198) duplicated block id: 4554 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (1502:1512) - megatron_patch/model/glm130b/transformer.py (945:955) duplicated block id: 4555 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/llama/transformer.py (570:580) - megatron_patch/model/qwen/transformer.py (739:749) duplicated block id: 4556 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/bloom/language_model.py (199:212) - megatron_patch/model/mistral/language_model.py (185:198) duplicated block id: 4557 size: 10 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/chatglm/checkpoint_reshaping_and_interoperability.py (362:373) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (485:496) duplicated block id: 4558 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/language_model.py (303:316) - megatron_patch/model/falcon/language_model.py (245:258) duplicated block id: 4559 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (1077:1089) - megatron_patch/model/galactica/transformer.py (794:810) duplicated block id: 4560 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/transformer.py (502:518) - megatron_patch/model/starcoder/transformer.py (806:821) duplicated block id: 4561 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/qwen1_5_megablocks/transformer.py (1111:1125) - megatron_patch/model/starcoder/transformer.py (942:956) duplicated block id: 4562 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/language_model.py (257:270) - megatron_patch/model/bloom/language_model.py (263:276) duplicated block id: 4563 size: 10 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (614:626) - toolkits/model_checkpoints_convertor/galactica/checkpoint_reshaping_and_interoperability.py (616:628) duplicated block id: 4564 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/mixtral_bak/moe/router.py (27:46) - megatron_patch/model/qwen1_5/moe/router.py (34:55) duplicated block id: 4565 size: 10 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (299:319) - toolkits/model_checkpoints_convertor/falcon40b/checkpoint_reshaping_and_interoperability.py (255:275) duplicated block id: 4566 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/llava/transformer.py (1226:1240) - megatron_patch/model/starcoder/transformer.py (942:956) duplicated block id: 4567 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/language_model.py (198:211) - megatron_patch/model/starcoder/language_model.py (183:196) duplicated block id: 4568 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/llama3_1/rms_norm.py (5:20) - megatron_patch/model/qwen2/rms_norm.py (5:18) duplicated block id: 4569 size: 10 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore_llama3_1.py (28:39) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2.5_vl.py (43:54) duplicated block id: 4570 size: 10 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/reward_model_to_megatron.py (817:826) - toolkits/model_checkpoints_convertor/falcon40b/checkpoint_reshaping_and_interoperability.py (819:828) duplicated block id: 4571 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/language_model.py (100:128) - megatron_patch/model/falcon40b/language_model.py (98:126) duplicated block id: 4572 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/language_model.py (709:720) - megatron_patch/model/llama/language_model.py (688:699) duplicated block id: 4573 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/llama/transformer.py (650:663) - megatron_patch/model/qwen/transformer.py (844:857) duplicated block id: 4574 size: 10 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (648:662) - toolkits/model_checkpoints_convertor/yi/checkpoint_reshaping_and_interoperability.py (557:571) duplicated block id: 4575 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/galactica/language_model.py (228:241) - megatron_patch/model/llava/language_model.py (187:200) duplicated block id: 4576 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/transformer.py (711:724) - megatron_patch/model/qwen/transformer.py (844:857) duplicated block id: 4577 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer_legacy.py (1185:1199) - megatron_patch/model/starcoder/transformer.py (942:956) duplicated block id: 4578 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/transformer.py (790:802) - megatron_patch/model/starcoder/transformer.py (1226:1238) duplicated block id: 4579 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/language_model.py (194:207) - megatron_patch/model/galactica/language_model.py (228:241) duplicated block id: 4580 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/falcon/language_model.py (98:126) - megatron_patch/model/qwen1_5_megablocks/language_model.py (88:118) duplicated block id: 4581 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/galactica/language_model.py (228:241) - megatron_patch/model/qwen_vl/language_model.py (185:198) duplicated block id: 4582 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/llama/language_model.py (197:210) - megatron_patch/model/mistral/language_model.py (185:198) duplicated block id: 4583 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/transformer_block.py (87:131) - megatron_patch/model/qwen2/transformer_block.py (48:78) duplicated block id: 4584 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/falcon/language_model.py (197:210) - megatron_patch/model/qwen1_5_megablocks/language_model.py (183:196) duplicated block id: 4585 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/galactica/language_model.py (706:717) - megatron_patch/model/mistral/language_model.py (644:655) duplicated block id: 4586 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/llama/language_model.py (245:258) - megatron_patch/model/mistral/language_model.py (248:261) duplicated block id: 4587 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/llama3/layer_specs.py (43:52) - megatron_patch/model/qwen3_moe/gpt_layer_specs.py (145:154) duplicated block id: 4588 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/llama2/language_model.py (30:41) - megatron_patch/model/starcoder/language_model.py (29:40) duplicated block id: 4589 size: 10 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_moe.py (352:361) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_vl.py (192:201) duplicated block id: 4590 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/galactica/language_model.py (228:241) - megatron_patch/model/llama2/language_model.py (184:197) duplicated block id: 4591 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/transformer.py (583:593) - megatron_patch/model/glm130b/transformer.py (613:623) duplicated block id: 4592 size: 10 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore.py (266:275) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_moe.py (123:132) duplicated block id: 4593 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/qwen1_5/layer_specs.py (48:57) - megatron_patch/model/qwen2_vl/layer_specs.py (49:58) duplicated block id: 4594 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/falcon/language_model.py (245:258) - megatron_patch/model/llama2/language_model.py (247:260) duplicated block id: 4595 size: 10 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/checkpoint_reshaping_and_interoperability.py (818:827) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (1135:1144) duplicated block id: 4596 size: 10 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (1018:1027) - toolkits/model_checkpoints_convertor/starcoder/checkpoint_reshaping_and_interoperability.py (827:836) duplicated block id: 4597 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/language_model.py (132:155) - megatron_patch/model/llava/language_model.py (127:151) duplicated block id: 4598 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/language_model.py (198:211) - megatron_patch/model/llava/language_model.py (187:200) duplicated block id: 4599 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/language_model.py (709:720) - megatron_patch/model/galactica/language_model.py (706:717) duplicated block id: 4600 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/language_model.py (194:207) - megatron_patch/model/falcon40b/language_model.py (197:210) duplicated block id: 4601 size: 10 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/checkpoint_reshaping_and_interoperability.py (505:522) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (983:1000) duplicated block id: 4602 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/llama/language_model.py (688:699) - megatron_patch/model/mistral/language_model.py (644:655) duplicated block id: 4603 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/language_model.py (245:258) - megatron_patch/model/llama3/language_model.py (242:255) duplicated block id: 4604 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/galactica/language_model.py (228:241) - megatron_patch/model/qwen/language_model.py (184:197) duplicated block id: 4605 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/language_model.py (99:127) - megatron_patch/model/qwen1_5_megablocks/language_model.py (88:118) duplicated block id: 4606 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/transformer.py (419:429) - megatron_patch/model/llama/transformer.py (427:437) duplicated block id: 4607 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/language_model.py (197:210) - megatron_patch/model/qwen_vl/language_model.py (185:198) duplicated block id: 4608 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/language_model.py (100:128) - megatron_patch/model/llama/language_model.py (98:126) duplicated block id: 4609 size: 10 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore.py (583:594) - toolkits/model_checkpoints_convertor/qwen/hf2megablocks_qwen1.5.py (593:604) duplicated block id: 4610 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/llama/transformer.py (751:761) - megatron_patch/model/llama2/transformer.py (1235:1245) duplicated block id: 4611 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/language_model.py (134:164) - megatron_patch/model/galactica/language_model.py (98:126) duplicated block id: 4612 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/bloom/language_model.py (100:128) - megatron_patch/model/starcoder/language_model.py (87:117) duplicated block id: 4613 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/language_model.py (194:207) - megatron_patch/model/llama/language_model.py (197:210) duplicated block id: 4614 size: 10 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore.py (137:146) - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (114:123) duplicated block id: 4615 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/qwen2/transformer/attention.py (334:357) - megatron_patch/model/qwen2_vl/attention.py (472:495) duplicated block id: 4616 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/llama/transformer.py (751:761) - megatron_patch/model/qwen1_5_megablocks/transformer.py (1119:1129) duplicated block id: 4617 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/llama/language_model.py (98:126) - megatron_patch/model/qwen_vl/language_model.py (90:120) duplicated block id: 4618 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/galactica/language_model.py (276:289) - megatron_patch/model/mistral/language_model.py (248:261) duplicated block id: 4619 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/language_model.py (75:86) - megatron_patch/model/starcoder/language_model.py (29:40) duplicated block id: 4620 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/bloom/language_model.py (199:212) - megatron_patch/model/llava/language_model.py (187:200) duplicated block id: 4621 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/llama/transformer.py (650:663) - megatron_patch/model/llava/transformer.py (898:911) duplicated block id: 4622 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/layer_specs.py (147:158) - megatron_patch/model/mixtral_bak/layer_specs.py (123:145) duplicated block id: 4623 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (984:994) - megatron_patch/model/falcon40b/transformer.py (831:841) duplicated block id: 4624 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/qwen/transformer.py (1169:1183) - megatron_patch/model/starcoder/transformer.py (942:956) duplicated block id: 4625 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/language_model.py (611:623) - megatron_patch/model/llama3/language_model.py (602:614) duplicated block id: 4626 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/qwen1_5/layer_specs.py (43:52) - megatron_patch/model/qwen2_moe/layer_specs.py (159:168) duplicated block id: 4627 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/language_model.py (235:248) - megatron_patch/model/bloom/language_model.py (199:212) duplicated block id: 4628 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/language_model.py (397:409) - megatron_patch/model/llama/language_model.py (384:396) duplicated block id: 4629 size: 10 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_dense_and_moe_gqa.py (56:68) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_moe.py (142:154) duplicated block id: 4630 size: 10 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/chatglm/checkpoint_reshaping_and_interoperability.py (362:373) - toolkits/model_checkpoints_convertor/yi/checkpoint_reshaping_and_interoperability.py (430:441) duplicated block id: 4631 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/falcon/language_model.py (98:126) - megatron_patch/model/qwen/language_model.py (89:119) duplicated block id: 4632 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/llama3/language_model.py (110:135) - megatron_patch/model/starcoder/language_model.py (119:142) duplicated block id: 4633 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/transformer.py (347:360) - megatron_patch/model/glm130b/transformer.py (438:451) duplicated block id: 4634 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/transformer.py (909:919) - megatron_patch/model/llama/transformer.py (886:896) duplicated block id: 4635 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/transformer.py (715:725) - megatron_patch/model/falcon40b/transformer.py (831:841) duplicated block id: 4636 size: 10 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_mha.py (123:132) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_moe.py (352:361) duplicated block id: 4637 size: 10 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_dense_and_moe_gqa.py (244:253) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_moe.py (352:361) duplicated block id: 4638 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (63:73) - megatron_patch/model/glm130b/transformer.py (90:100) duplicated block id: 4639 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/galactica/language_model.py (228:241) - megatron_patch/model/mistral/language_model.py (185:198) duplicated block id: 4640 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/llama/language_model.py (245:258) - megatron_patch/model/llama2/language_model.py (247:260) duplicated block id: 4641 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/mixtral_bak/layer_specs.py (57:66) - megatron_patch/model/qwen3_moe/gpt_layer_specs.py (145:154) duplicated block id: 4642 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/llama/transformer.py (650:663) - megatron_patch/model/llama2/transformer.py (899:912) duplicated block id: 4643 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/falcon/language_model.py (245:258) - megatron_patch/model/qwen_vl/language_model.py (248:261) duplicated block id: 4644 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/llama/language_model.py (98:126) - megatron_patch/model/llava/language_model.py (92:122) duplicated block id: 4645 size: 10 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/reward_model_to_megatron.py (504:521) - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (1088:1105) duplicated block id: 4646 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/language_model.py (197:210) - megatron_patch/model/qwen1_5_megablocks/language_model.py (183:196) duplicated block id: 4647 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/mixtral_bak/moe/token_dispatcher.py (122:157) - megatron_patch/model/qwen1_5/moe/token_dispatcher.py (96:126) duplicated block id: 4648 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/language_model.py (198:211) - megatron_patch/model/qwen1_5_megablocks/language_model.py (183:196) duplicated block id: 4649 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/language_model.py (99:129) - megatron_patch/model/glm130b/language_model.py (99:127) duplicated block id: 4650 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/language_model.py (100:128) - megatron_patch/model/llama2/language_model.py (89:119) duplicated block id: 4651 size: 10 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (916:925) - toolkits/model_checkpoints_convertor/bloom/reward_model_to_megatron.py (817:826) duplicated block id: 4652 size: 10 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/hf2te.py (452:463) - toolkits/model_checkpoints_convertor/falcon40b/checkpoint_reshaping_and_interoperability.py (547:558) duplicated block id: 4653 size: 10 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/checkpoint_reshaping_and_interoperability.py (818:827) - toolkits/model_checkpoints_convertor/falcon40b/checkpoint_reshaping_and_interoperability.py (819:828) duplicated block id: 4654 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/moe/router.py (20:42) - megatron_patch/model/qwen1_5/moe/router.py (36:59) duplicated block id: 4655 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/language_model.py (625:636) - megatron_patch/model/llama/language_model.py (688:699) duplicated block id: 4656 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/mixtral_bak/moe/token_dispatcher.py (24:51) - megatron_patch/model/qwen2/moe/token_dispatcher.py (26:53) duplicated block id: 4657 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/llava/transformer.py (620:634) - megatron_patch/model/mistral/transformer.py (620:634) duplicated block id: 4658 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (832:844) - megatron_patch/model/llama3/transformer_legacy.py (915:927) duplicated block id: 4659 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (419:428) - megatron_patch/model/glm130b/transformer.py (453:462) duplicated block id: 4660 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (1227:1241) - megatron_patch/model/starcoder/transformer.py (942:956) duplicated block id: 4661 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/language_model.py (397:409) - megatron_patch/model/falcon40b/language_model.py (384:396) duplicated block id: 4662 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/language_model.py (303:316) - megatron_patch/model/bloom/language_model.py (263:276) duplicated block id: 4663 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/falcon/language_model.py (98:126) - megatron_patch/model/llama3/language_model.py (78:108) duplicated block id: 4664 size: 10 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/chatglm/checkpoint_reshaping_and_interoperability.py (362:373) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (515:526) duplicated block id: 4665 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/bloom/language_model.py (100:128) - megatron_patch/model/llama3/language_model.py (78:108) duplicated block id: 4666 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/language_model.py (166:195) - megatron_patch/model/starcoder/language_model.py (119:142) duplicated block id: 4667 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/galactica/transformer.py (794:810) - megatron_patch/model/starcoder/transformer.py (1226:1238) duplicated block id: 4668 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/transformer.py (945:955) - megatron_patch/model/llava/transformer.py (1501:1511) duplicated block id: 4669 size: 10 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon40b/checkpoint_reshaping_and_interoperability.py (784:801) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (983:1000) duplicated block id: 4670 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/llama/language_model.py (197:210) - megatron_patch/model/llama2/language_model.py (184:197) duplicated block id: 4671 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/falcon/language_model.py (98:126) - megatron_patch/model/qwen_vl/language_model.py (90:120) duplicated block id: 4672 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/llama/transformer.py (570:580) - megatron_patch/model/llama3/transformer_legacy.py (744:754) duplicated block id: 4673 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/mistral/transformer.py (620:634) - megatron_patch/model/qwen_vl/transformer.py (620:634) duplicated block id: 4674 size: 10 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (114:123) - toolkits/model_checkpoints_convertor/qwen/hf2megablocks_qwen1.5.py (131:140) duplicated block id: 4675 size: 10 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/chatglm/checkpoint_reshaping_and_interoperability.py (256:276) - toolkits/model_checkpoints_convertor/falcon40b/checkpoint_reshaping_and_interoperability.py (255:275) duplicated block id: 4676 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (984:994) - megatron_patch/model/falcon/transformer.py (1092:1102) duplicated block id: 4677 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/galactica/language_model.py (98:126) - megatron_patch/model/llama3/language_model.py (78:108) duplicated block id: 4678 size: 10 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore.py (332:341) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_moe.py (352:361) duplicated block id: 4679 size: 10 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (286:306) - toolkits/model_checkpoints_convertor/falcon/checkpoint_reshaping_and_interoperability.py (257:277) duplicated block id: 4680 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/mixtral_bak/layer_specs.py (57:66) - megatron_patch/model/qwen2_moe/layer_specs.py (159:168) duplicated block id: 4681 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer/attention.py (338:361) - megatron_patch/model/qwen2_vl/attention_vision.py (471:494) duplicated block id: 4682 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/llama/language_model.py (245:258) - megatron_patch/model/qwen1_5_megablocks/language_model.py (246:259) duplicated block id: 4683 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/language_model.py (198:211) - megatron_patch/model/mistral/language_model.py (185:198) duplicated block id: 4684 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/falcon/transformer.py (992:1002) - megatron_patch/model/llama2/transformer.py (1235:1245) duplicated block id: 4685 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/language_model.py (303:316) - megatron_patch/model/galactica/language_model.py (276:289) duplicated block id: 4686 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/bloom/language_model.py (100:128) - megatron_patch/model/qwen1_5_megablocks/language_model.py (88:118) duplicated block id: 4687 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (178:189) - megatron_patch/model/starcoder/transformer.py (217:228) duplicated block id: 4688 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/falcon/language_model.py (197:210) - megatron_patch/model/llava/language_model.py (187:200) duplicated block id: 4689 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (1227:1241) - megatron_patch/model/bloom/transformer.py (797:811) duplicated block id: 4690 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/language_model.py (194:207) - megatron_patch/model/bloom/language_model.py (199:212) duplicated block id: 4691 size: 10 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore_llama3_1.py (24:37) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_moe.py (36:53) duplicated block id: 4692 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/transformer.py (59:69) - megatron_patch/model/glm130b/transformer.py (90:100) duplicated block id: 4693 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/llava_mcore/llava_spec.py (46:55) - megatron_patch/model/qwen2_vl/layer_specs.py (49:58) duplicated block id: 4694 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (797:811) - megatron_patch/model/mistral/transformer.py (1224:1238) duplicated block id: 4695 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/galactica/transformer.py (720:730) - megatron_patch/model/glm130b/transformer.py (909:919) duplicated block id: 4696 size: 10 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon/checkpoint_reshaping_and_interoperability.py (808:825) - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (1088:1105) duplicated block id: 4697 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/galactica/language_model.py (98:126) - megatron_patch/model/starcoder/language_model.py (87:117) duplicated block id: 4698 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/qwen2/transformer_config.py (21:39) - megatron_patch/model/qwen2_moe/transformer_config.py (75:93) duplicated block id: 4699 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/llama3_1/model.py (264:274) - megatron_patch/model/qwen2_vl/gpt_model.py (190:200) duplicated block id: 4700 size: 10 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/hf2te.py (194:210) - toolkits/model_checkpoints_convertor/yi/checkpoint_reshaping_and_interoperability.py (314:330) duplicated block id: 4701 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/language_model.py (99:129) - megatron_patch/model/chatglm/language_model.py (100:128) duplicated block id: 4702 size: 10 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (942:951) - toolkits/model_checkpoints_convertor/bloom/reward_model_to_megatron.py (817:826) duplicated block id: 4703 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/transformer.py (570:580) - megatron_patch/model/glm130b/transformer.py (777:787) duplicated block id: 4704 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/falcon/transformer.py (63:73) - megatron_patch/model/glm130b/transformer.py (90:100) duplicated block id: 4705 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/layer_specs.py (147:158) - megatron_patch/model/qwen1_5/layer_specs.py (102:112) duplicated block id: 4706 size: 10 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/chatglm/checkpoint_reshaping_and_interoperability.py (362:373) - toolkits/model_checkpoints_convertor/galactica/checkpoint_reshaping_and_interoperability.py (454:465) duplicated block id: 4707 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/transformer.py (523:534) - megatron_patch/model/llama/transformer.py (476:487) duplicated block id: 4708 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (797:811) - megatron_patch/model/qwen/transformer.py (1169:1183) duplicated block id: 4709 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/falcon/language_model.py (98:126) - megatron_patch/model/llava/language_model.py (92:122) duplicated block id: 4710 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/transformer.py (945:955) - megatron_patch/model/qwen_vl/transformer.py (1501:1511) duplicated block id: 4711 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/transformer.py (777:787) - megatron_patch/model/starcoder/transformer.py (888:898) duplicated block id: 4712 size: 10 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/checkpoint_reshaping_and_interoperability.py (505:522) - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (1088:1105) duplicated block id: 4713 size: 10 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon40b/checkpoint_reshaping_and_interoperability.py (819:828) - toolkits/model_checkpoints_convertor/starcoder/checkpoint_reshaping_and_interoperability.py (827:836) duplicated block id: 4714 size: 10 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2.5_vl.py (143:152) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_moe.py (352:361) duplicated block id: 4715 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/llama/transformer.py (751:761) - megatron_patch/model/mistral/transformer.py (1232:1242) duplicated block id: 4716 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/falcon/language_model.py (245:258) - megatron_patch/model/qwen/language_model.py (247:260) duplicated block id: 4717 size: 10 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (314:330) - toolkits/model_checkpoints_convertor/baichuan2/hf2te.py (178:194) duplicated block id: 4718 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/qwen_vl/language_model.py (122:147) - megatron_patch/model/starcoder/language_model.py (119:142) duplicated block id: 4719 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/language_model.py (100:128) - megatron_patch/model/qwen/language_model.py (89:119) duplicated block id: 4720 size: 10 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/chatglm/checkpoint_reshaping_and_interoperability.py (362:373) - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (508:519) duplicated block id: 4721 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/llama/positional_embeddings.py (23:34) - megatron_patch/model/llama2/rotary_pos_embedding.py (27:38) duplicated block id: 4722 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/llama/language_model.py (560:569) - megatron_patch/model/mistral/language_model.py (501:510) duplicated block id: 4723 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (561:571) - megatron_patch/model/qwen1_5_megablocks/transformer.py (671:681) duplicated block id: 4724 size: 10 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_dense_and_moe_gqa.py (183:192) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_moe.py (352:361) duplicated block id: 4725 size: 10 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (614:626) - toolkits/model_checkpoints_convertor/chatglm/checkpoint_reshaping_and_interoperability.py (546:558) duplicated block id: 4726 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/bloom/language_model.py (100:128) - megatron_patch/model/mistral/language_model.py (90:120) duplicated block id: 4727 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/mistral/language_model.py (122:147) - megatron_patch/model/starcoder/language_model.py (119:142) duplicated block id: 4728 size: 10 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon/checkpoint_reshaping_and_interoperability.py (543:554) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (624:635) duplicated block id: 4729 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/qwen1_5/moe/router.py (267:291) - megatron_patch/model/qwen2/moe/router.py (359:380) duplicated block id: 4730 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/llama/transformer.py (650:663) - megatron_patch/model/qwen_vl/transformer.py (898:911) duplicated block id: 4731 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/falcon/language_model.py (245:258) - megatron_patch/model/llava/language_model.py (250:263) duplicated block id: 4732 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/llama/language_model.py (98:126) - megatron_patch/model/qwen1_5_megablocks/language_model.py (88:118) duplicated block id: 4733 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/falcon/transformer.py (992:1002) - megatron_patch/model/llama3/transformer_legacy.py (1193:1203) duplicated block id: 4734 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/galactica/language_model.py (98:126) - megatron_patch/model/llava/language_model.py (92:122) duplicated block id: 4735 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/language_model.py (303:316) - megatron_patch/model/llama/language_model.py (245:258) duplicated block id: 4736 size: 10 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_moe.py (160:169) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_moe.py (352:361) duplicated block id: 4737 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/language_model.py (245:258) - megatron_patch/model/qwen1_5_megablocks/language_model.py (246:259) duplicated block id: 4738 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/galactica/language_model.py (276:289) - megatron_patch/model/llama2/language_model.py (247:260) duplicated block id: 4739 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/language_model.py (397:409) - megatron_patch/model/galactica/language_model.py (415:427) duplicated block id: 4740 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/transformer.py (711:724) - megatron_patch/model/llava/transformer.py (898:911) duplicated block id: 4741 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/mixtral_bak/model.py (108:117) - megatron_patch/model/qwen1_5/model.py (130:139) duplicated block id: 4742 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/transformer_block.py (305:314) - megatron_patch/model/qwen2/transformer_block.py (257:266) duplicated block id: 4743 size: 10 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon/checkpoint_reshaping_and_interoperability.py (808:825) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (983:1000) duplicated block id: 4744 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/language_model.py (197:210) - megatron_patch/model/llama2/language_model.py (184:197) duplicated block id: 4745 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/bloom/language_model.py (263:276) - megatron_patch/model/qwen_vl/language_model.py (248:261) duplicated block id: 4746 size: 10 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (299:319) - toolkits/model_checkpoints_convertor/falcon/checkpoint_reshaping_and_interoperability.py (257:277) duplicated block id: 4747 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/qwen2/transformer/attention.py (334:357) - megatron_patch/model/qwen2_vl/attention_vision.py (471:494) duplicated block id: 4748 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/transformer_layer.py (274:311) - megatron_patch/model/qwen2/transformer_layer.py (176:192) duplicated block id: 4749 size: 10 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore_llama3_1.py (623:632) - toolkits/model_checkpoints_convertor/llava/hf2mcore_llava.py (344:353) duplicated block id: 4750 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/language_model.py (99:127) - megatron_patch/model/llama3/language_model.py (78:108) duplicated block id: 4751 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/language_model.py (99:129) - megatron_patch/model/falcon40b/language_model.py (98:126) duplicated block id: 4752 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/mixtral_bak/transformer/attention.py (311:334) - megatron_patch/model/qwen2_vl/attention.py (472:495) duplicated block id: 4753 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/language_model.py (235:248) - megatron_patch/model/galactica/language_model.py (228:241) duplicated block id: 4754 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/language_model.py (134:158) - megatron_patch/model/chatglm/language_model.py (132:155) duplicated block id: 4755 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/galactica/language_model.py (706:717) - megatron_patch/model/qwen1_5_megablocks/language_model.py (631:642) duplicated block id: 4756 size: 10 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/hf2te.py (178:194) - toolkits/model_checkpoints_convertor/falcon40b/checkpoint_reshaping_and_interoperability.py (287:303) duplicated block id: 4757 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/transformer/attention.py (170:181) - megatron_patch/model/qwen2_vl/attention_vision.py (158:169) duplicated block id: 4758 size: 10 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/reward_model_to_megatron.py (504:521) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (1096:1113) duplicated block id: 4759 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/galactica/language_model.py (706:717) - megatron_patch/model/qwen_vl/language_model.py (661:672) duplicated block id: 4760 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/transformer.py (711:724) - megatron_patch/model/mistral/transformer.py (896:909) duplicated block id: 4761 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/multi_latent_attention.py (199:208) - megatron_patch/model/qwen2_vl/attention_vision.py (493:502) duplicated block id: 4762 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/transformer.py (945:955) - megatron_patch/model/mistral/transformer.py (1499:1509) duplicated block id: 4763 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/llama/positional_embeddings.py (23:34) - megatron_patch/model/qwen1_5_megablocks/rotary_pos_embedding.py (27:38) duplicated block id: 4764 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/transformer.py (613:623) - megatron_patch/model/mistral/transformer.py (778:788) duplicated block id: 4765 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/language_model.py (245:258) - megatron_patch/model/llava/language_model.py (250:263) duplicated block id: 4766 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/bloom/language_model.py (199:212) - megatron_patch/model/qwen/language_model.py (184:197) duplicated block id: 4767 size: 10 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/hf2te.py (194:210) - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (365:381) duplicated block id: 4768 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/language_model.py (134:164) - megatron_patch/model/falcon/language_model.py (98:126) duplicated block id: 4769 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/bloom/language_model.py (263:276) - megatron_patch/model/qwen1_5_megablocks/language_model.py (246:259) duplicated block id: 4770 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/llama/transformer.py (960:974) - megatron_patch/model/starcoder/transformer.py (1226:1238) duplicated block id: 4771 size: 10 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore.py (137:146) - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (104:113) duplicated block id: 4772 size: 10 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon40b/checkpoint_reshaping_and_interoperability.py (547:558) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (624:635) duplicated block id: 4773 size: 10 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (916:925) - toolkits/model_checkpoints_convertor/bloom/checkpoint_reshaping_and_interoperability.py (818:827) duplicated block id: 4774 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/llama3_1/layer_specs.py (44:53) - megatron_patch/model/qwen2_moe/layer_specs.py (159:168) duplicated block id: 4775 size: 10 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (455:466) - toolkits/model_checkpoints_convertor/chatglm/checkpoint_reshaping_and_interoperability.py (362:373) duplicated block id: 4776 size: 10 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/chatglm/checkpoint_reshaping_and_interoperability.py (362:373) - toolkits/model_checkpoints_convertor/falcon/checkpoint_reshaping_and_interoperability.py (417:428) duplicated block id: 4777 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (1102:1112) - megatron_patch/model/llama/transformer.py (751:761) duplicated block id: 4778 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/llama/transformer.py (650:663) - megatron_patch/model/qwen1_5_megablocks/transformer.py (790:803) duplicated block id: 4779 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/llama3_1/layer_specs.py (44:53) - megatron_patch/model/qwen3_moe/gpt_layer_specs.py (145:154) duplicated block id: 4780 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/llava_mcore/vision/vit_layer_specs.py (69:78) - megatron_patch/model/mixtral/layer_specs.py (121:130) duplicated block id: 4781 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/language_model.py (134:164) - megatron_patch/model/llama/language_model.py (98:126) duplicated block id: 4782 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/llava_mcore/llava_spec.py (41:50) - megatron_patch/model/qwen2_moe/layer_specs.py (159:168) duplicated block id: 4783 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/llava/language_model.py (124:149) - megatron_patch/model/starcoder/language_model.py (119:142) duplicated block id: 4784 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/llama3/model.py (130:139) - megatron_patch/model/mixtral_bak/model.py (108:117) duplicated block id: 4785 size: 10 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/hf2te.py (452:463) - toolkits/model_checkpoints_convertor/falcon/checkpoint_reshaping_and_interoperability.py (543:554) duplicated block id: 4786 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/bloom/language_model.py (199:212) - megatron_patch/model/qwen_vl/language_model.py (185:198) duplicated block id: 4787 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/mistral/transformer.py (1224:1238) - megatron_patch/model/starcoder/transformer.py (942:956) duplicated block id: 4788 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/llama/language_model.py (684:696) - megatron_patch/model/starcoder/language_model.py (551:563) duplicated block id: 4789 size: 10 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/hf2te.py (194:210) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (362:378) duplicated block id: 4790 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/language_model.py (98:126) - megatron_patch/model/llama2/language_model.py (89:119) duplicated block id: 4791 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/llava_mcore/llava_spec.py (73:82) - megatron_patch/model/mixtral_bak/layer_specs.py (99:108) duplicated block id: 4792 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (862:875) - megatron_patch/model/llama/transformer.py (650:663) duplicated block id: 4793 size: 10 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (104:113) - toolkits/model_checkpoints_convertor/qwen/hf2megablocks_qwen1.5.py (131:140) duplicated block id: 4794 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/falcon/transformer.py (992:1002) - megatron_patch/model/llava/transformer.py (1234:1244) duplicated block id: 4795 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/falcon/language_model.py (98:126) - megatron_patch/model/starcoder/language_model.py (87:117) duplicated block id: 4796 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/transformer.py (613:623) - megatron_patch/model/llama2/transformer.py (780:790) duplicated block id: 4797 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/llama/language_model.py (245:258) - megatron_patch/model/qwen_vl/language_model.py (248:261) duplicated block id: 4798 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/falcon/transformer.py (814:824) - megatron_patch/model/llama/transformer.py (586:596) duplicated block id: 4799 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (984:994) - megatron_patch/model/llama/transformer.py (886:896) duplicated block id: 4800 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/language_model.py (100:128) - megatron_patch/model/falcon/language_model.py (98:126) duplicated block id: 4801 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/galactica/language_model.py (228:241) - megatron_patch/model/starcoder/language_model.py (183:196) duplicated block id: 4802 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/galactica/language_model.py (98:126) - megatron_patch/model/qwen_vl/language_model.py (90:120) duplicated block id: 4803 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/qwen/language_model.py (121:146) - megatron_patch/model/starcoder/language_model.py (119:142) duplicated block id: 4804 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/language_model.py (235:248) - megatron_patch/model/llama/language_model.py (197:210) duplicated block id: 4805 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/qwen2/layer_specs.py (53:62) - megatron_patch/model/qwen3_moe/gpt_layer_specs.py (145:154) duplicated block id: 4806 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/language_model.py (235:248) - megatron_patch/model/falcon40b/language_model.py (197:210) duplicated block id: 4807 size: 10 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/hf2te.py (194:210) - toolkits/model_checkpoints_convertor/galactica/checkpoint_reshaping_and_interoperability.py (316:332) duplicated block id: 4808 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/language_model.py (132:155) - megatron_patch/model/llama3/language_model.py (113:137) duplicated block id: 4809 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/llama/language_model.py (197:210) - megatron_patch/model/starcoder/language_model.py (183:196) duplicated block id: 4810 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (1077:1089) - megatron_patch/model/falcon40b/transformer.py (905:921) duplicated block id: 4811 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/language_model.py (98:126) - megatron_patch/model/llava/language_model.py (92:122) duplicated block id: 4812 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/falcon/language_model.py (197:210) - megatron_patch/model/qwen_vl/language_model.py (185:198) duplicated block id: 4813 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/qwen/language_model.py (30:41) - megatron_patch/model/starcoder/language_model.py (29:40) duplicated block id: 4814 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (1077:1089) - megatron_patch/model/falcon/transformer.py (1166:1182) duplicated block id: 4815 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (961:974) - megatron_patch/model/bloom/transformer.py (707:722) duplicated block id: 4816 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/llama/language_model.py (688:699) - megatron_patch/model/qwen/language_model.py (614:625) duplicated block id: 4817 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/transformer.py (715:725) - megatron_patch/model/falcon/transformer.py (1092:1102) duplicated block id: 4818 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/falcon/transformer.py (992:1002) - megatron_patch/model/mistral/transformer.py (1232:1242) duplicated block id: 4819 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (1077:1089) - megatron_patch/model/llama/transformer.py (960:974) duplicated block id: 4820 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/layer_specs.py (79:88) - megatron_patch/model/qwen2_moe/layer_specs.py (159:168) duplicated block id: 4821 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/language_model.py (245:258) - megatron_patch/model/qwen_vl/language_model.py (248:261) duplicated block id: 4822 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/language_model.py (100:128) - megatron_patch/model/qwen_vl/language_model.py (90:120) duplicated block id: 4823 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/llama/language_model.py (98:126) - megatron_patch/model/llama2/language_model.py (89:119) duplicated block id: 4824 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/language_model.py (98:126) - megatron_patch/model/qwen/language_model.py (89:119) duplicated block id: 4825 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/multi_latent_attention.py (199:208) - megatron_patch/model/mixtral/transformer/attention.py (493:502) duplicated block id: 4826 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/mixtral_bak/model.py (94:107) - megatron_patch/model/qwen2/model.py (103:116) duplicated block id: 4827 size: 10 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/hf2te.py (194:210) - toolkits/model_checkpoints_convertor/falcon/checkpoint_reshaping_and_interoperability.py (289:305) duplicated block id: 4828 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/language_model.py (303:316) - megatron_patch/model/falcon40b/language_model.py (245:258) duplicated block id: 4829 size: 10 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/mistral/hf2mcore.py (375:388) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2.5_vl.py (567:581) duplicated block id: 4830 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/falcon/language_model.py (245:258) - megatron_patch/model/qwen1_5_megablocks/language_model.py (246:259) duplicated block id: 4831 size: 10 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_moe.py (290:303) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_dense_and_moe_gqa.py (500:512) duplicated block id: 4832 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/language_model.py (132:155) - megatron_patch/model/qwen_vl/language_model.py (125:149) duplicated block id: 4833 size: 10 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/hf2te.py (178:194) - toolkits/model_checkpoints_convertor/galactica/checkpoint_reshaping_and_interoperability.py (316:332) duplicated block id: 4834 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/multi_latent_attention.py (199:208) - megatron_patch/model/qwen2_vl/attention.py (494:503) duplicated block id: 4835 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/transformer.py (905:921) - megatron_patch/model/starcoder/transformer.py (1226:1238) duplicated block id: 4836 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/transformer.py (613:623) - megatron_patch/model/llava/transformer.py (779:789) duplicated block id: 4837 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/llama3/language_model.py (168:179) - megatron_patch/model/starcoder/language_model.py (178:189) duplicated block id: 4838 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/language_model.py (99:127) - megatron_patch/model/qwen_vl/language_model.py (90:120) duplicated block id: 4839 size: 10 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (333:354) - toolkits/model_checkpoints_convertor/chatglm/checkpoint_reshaping_and_interoperability.py (298:319) duplicated block id: 4840 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/model.py (112:121) - megatron_patch/model/qwen2/model.py (131:140) duplicated block id: 4841 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/language_model.py (99:127) - megatron_patch/model/qwen/language_model.py (89:119) duplicated block id: 4842 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/llama/language_model.py (245:258) - megatron_patch/model/llava/language_model.py (250:263) duplicated block id: 4843 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/language_model.py (397:409) - megatron_patch/model/falcon/language_model.py (384:396) duplicated block id: 4844 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/llama3/model.py (130:139) - megatron_patch/model/mixtral/model.py (112:121) duplicated block id: 4845 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/galactica/language_model.py (706:717) - megatron_patch/model/llama2/language_model.py (632:643) duplicated block id: 4846 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/falcon/language_model.py (98:126) - megatron_patch/model/mistral/language_model.py (90:120) duplicated block id: 4847 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (405:418) - megatron_patch/model/glm130b/transformer.py (438:451) duplicated block id: 4848 size: 10 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore_llama3_1.py (28:39) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_vl.py (43:54) duplicated block id: 4849 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/language_model.py (99:129) - megatron_patch/model/falcon/language_model.py (98:126) duplicated block id: 4850 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer/attention.py (338:361) - megatron_patch/model/qwen2_vl/attention.py (472:495) duplicated block id: 4851 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/language_model.py (98:126) - megatron_patch/model/qwen_vl/language_model.py (90:120) duplicated block id: 4852 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/language_model.py (625:636) - megatron_patch/model/galactica/language_model.py (706:717) duplicated block id: 4853 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/transformer_layer.py (20:57) - megatron_patch/model/qwen2/transformer_layer.py (30:41) duplicated block id: 4854 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/language_model.py (134:164) - megatron_patch/model/falcon40b/language_model.py (98:126) duplicated block id: 4855 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/qwen2_5_vl/transformer_config.py (81:99) - megatron_patch/model/qwen2_vl/transformer_config.py (70:87) duplicated block id: 4856 size: 10 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (327:343) - toolkits/model_checkpoints_convertor/baichuan/hf2te.py (194:210) duplicated block id: 4857 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (477:488) - megatron_patch/model/llama/transformer.py (476:487) duplicated block id: 4858 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/language_model.py (99:129) - megatron_patch/model/bloom/language_model.py (100:128) duplicated block id: 4859 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (862:875) - megatron_patch/model/glm130b/transformer.py (711:724) duplicated block id: 4860 size: 10 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/checkpoint_reshaping_and_interoperability.py (786:799) - toolkits/model_checkpoints_convertor/glm/checkpoint_reshaping_and_interoperability.py (537:550) duplicated block id: 4861 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/qwen2/transformer_block.py (257:266) - megatron_patch/model/qwen2_5_vl/transformer_block.py (352:361) duplicated block id: 4862 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/falcon/transformer.py (206:217) - megatron_patch/model/starcoder/transformer.py (217:228) duplicated block id: 4863 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (1136:1155) - megatron_patch/model/chatglm/transformer.py (855:875) duplicated block id: 4864 size: 10 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon40b/checkpoint_reshaping_and_interoperability.py (784:801) - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (1088:1105) duplicated block id: 4865 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/transformer.py (715:725) - megatron_patch/model/starcoder/transformer.py (1145:1155) duplicated block id: 4866 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/bloom/language_model.py (199:212) - megatron_patch/model/starcoder/language_model.py (183:196) duplicated block id: 4867 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/galactica/language_model.py (276:289) - megatron_patch/model/qwen/language_model.py (247:260) duplicated block id: 4868 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (477:488) - megatron_patch/model/chatglm/transformer.py (415:426) duplicated block id: 4869 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (757:767) - megatron_patch/model/llama/transformer.py (570:580) duplicated block id: 4870 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/falcon/language_model.py (245:258) - megatron_patch/model/llama3/language_model.py (242:255) duplicated block id: 4871 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (561:571) - megatron_patch/model/mistral/transformer.py (778:788) duplicated block id: 4872 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/llama/transformer.py (751:761) - megatron_patch/model/qwen_vl/transformer.py (1234:1244) duplicated block id: 4873 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/language_model.py (132:155) - megatron_patch/model/qwen/language_model.py (124:148) duplicated block id: 4874 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/llama/language_model.py (197:210) - megatron_patch/model/qwen1_5_megablocks/language_model.py (183:196) duplicated block id: 4875 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/language_model.py (32:43) - megatron_patch/model/starcoder/language_model.py (29:40) duplicated block id: 4876 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/transformer.py (945:955) - megatron_patch/model/llama3/transformer_legacy.py (1460:1470) duplicated block id: 4877 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/language_model.py (99:129) - megatron_patch/model/llama/language_model.py (98:126) duplicated block id: 4878 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/falcon/transformer.py (1092:1102) - megatron_patch/model/starcoder/transformer.py (1145:1155) duplicated block id: 4879 size: 10 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/chatglm/checkpoint_reshaping_and_interoperability.py (362:373) - toolkits/model_checkpoints_convertor/falcon40b/checkpoint_reshaping_and_interoperability.py (406:417) duplicated block id: 4880 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/llava_mcore/llava_spec.py (41:50) - megatron_patch/model/qwen3_moe/gpt_layer_specs.py (145:154) duplicated block id: 4881 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/falcon/transformer.py (690:699) - megatron_patch/model/glm130b/transformer.py (545:554) duplicated block id: 4882 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/transformer.py (945:955) - megatron_patch/model/qwen/transformer.py (1444:1454) duplicated block id: 4883 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/bloom/language_model.py (263:276) - megatron_patch/model/llama2/language_model.py (247:260) duplicated block id: 4884 size: 10 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (327:343) - toolkits/model_checkpoints_convertor/baichuan2/hf2te.py (178:194) duplicated block id: 4885 size: 10 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/deepseek/hf2mcore_deepseek_v2_moe.py (74:86) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_dense_and_moe_gqa.py (56:68) duplicated block id: 4886 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/transformer.py (415:426) - megatron_patch/model/glm130b/transformer.py (523:534) duplicated block id: 4887 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (561:571) - megatron_patch/model/falcon40b/transformer.py (583:593) duplicated block id: 4888 size: 10 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (1096:1113) - toolkits/model_checkpoints_convertor/starcoder/checkpoint_reshaping_and_interoperability.py (520:537) duplicated block id: 4889 size: 10 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/reward_model_to_megatron.py (817:826) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (1018:1027) duplicated block id: 4890 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/qwen1_5/layer_specs.py (43:52) - megatron_patch/model/qwen3_moe/gpt_layer_specs.py (145:154) duplicated block id: 4891 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/transformer.py (945:955) - megatron_patch/model/llama2/transformer.py (1502:1512) duplicated block id: 4892 size: 10 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llava/hf2mcore_llava.py (344:353) - toolkits/model_checkpoints_convertor/llava/hf2mcore_llava.py (392:401) duplicated block id: 4893 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/bloom/language_model.py (263:276) - megatron_patch/model/llama3/language_model.py (242:255) duplicated block id: 4894 size: 10 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore.py (292:301) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_moe.py (352:361) duplicated block id: 4895 size: 10 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore_llama3_1.py (575:584) - toolkits/model_checkpoints_convertor/llama/hf2mcore_llama3_1.py (623:632) duplicated block id: 4896 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/mixtral_bak/model.py (108:117) - megatron_patch/model/qwen2_vl/gpt_model.py (121:130) duplicated block id: 4897 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/galactica/transformer.py (720:730) - megatron_patch/model/starcoder/transformer.py (1145:1155) duplicated block id: 4898 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/layer_specs.py (79:88) - megatron_patch/model/qwen3_moe/gpt_layer_specs.py (145:154) duplicated block id: 4899 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/language_model.py (100:128) - megatron_patch/model/qwen1_5_megablocks/language_model.py (88:118) duplicated block id: 4900 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/language_model.py (134:164) - megatron_patch/model/chatglm/language_model.py (100:128) duplicated block id: 4901 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/language_model.py (98:126) - megatron_patch/model/llama3/language_model.py (78:108) duplicated block id: 4902 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/language_model.py (198:211) - megatron_patch/model/qwen/language_model.py (184:197) duplicated block id: 4903 size: 10 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/reward_model_to_megatron.py (817:826) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (1135:1144) duplicated block id: 4904 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/llava_mcore/llava_spec.py (73:82) - megatron_patch/model/mixtral/layer_specs.py (121:130) duplicated block id: 4905 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/language_model.py (235:248) - megatron_patch/model/chatglm/language_model.py (198:211) duplicated block id: 4906 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/transformer.py (546:555) - megatron_patch/model/glm130b/transformer.py (545:554) duplicated block id: 4907 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/llava/language_model.py (33:44) - megatron_patch/model/starcoder/language_model.py (29:40) duplicated block id: 4908 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/galactica/language_model.py (702:714) - megatron_patch/model/starcoder/language_model.py (551:563) duplicated block id: 4909 size: 10 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2.5_vl.py (695:707) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_moe.py (661:674) duplicated block id: 4910 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/transformer.py (212:223) - megatron_patch/model/starcoder/transformer.py (217:228) duplicated block id: 4911 size: 10 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/hf2te.py (194:210) - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (314:330) duplicated block id: 4912 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/mixtral_bak/transformer/mlp.py (57:67) - megatron_patch/model/qwen2/transformer/mlp.py (86:96) duplicated block id: 4913 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/llama/language_model.py (245:258) - megatron_patch/model/qwen/language_model.py (247:260) duplicated block id: 4914 size: 10 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/chatglm/checkpoint_reshaping_and_interoperability.py (256:276) - toolkits/model_checkpoints_convertor/falcon/checkpoint_reshaping_and_interoperability.py (257:277) duplicated block id: 4915 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/transformer.py (545:554) - megatron_patch/model/starcoder/transformer.py (656:665) duplicated block id: 4916 size: 10 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_mha_to_moe.py (141:154) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_dense_and_moe_gqa.py (500:512) duplicated block id: 4917 size: 10 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_moe.py (352:361) - toolkits/model_checkpoints_convertor/qwen/hf2megablocks_qwen1.5.py (286:295) duplicated block id: 4918 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/llama/transformer.py (886:896) - megatron_patch/model/starcoder/transformer.py (1145:1155) duplicated block id: 4919 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (561:571) - megatron_patch/model/llama2/transformer.py (780:790) duplicated block id: 4920 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/falcon/transformer.py (992:1002) - megatron_patch/model/qwen_vl/transformer.py (1234:1244) duplicated block id: 4921 size: 10 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_moe.py (352:361) - toolkits/model_checkpoints_convertor/qwen/hf2megablocks_qwen1.5.py (325:334) duplicated block id: 4922 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/llama2/language_model.py (121:146) - megatron_patch/model/starcoder/language_model.py (119:142) duplicated block id: 4923 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/galactica/language_model.py (98:126) - megatron_patch/model/qwen/language_model.py (89:119) duplicated block id: 4924 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/language_model.py (257:270) - megatron_patch/model/galactica/language_model.py (276:289) duplicated block id: 4925 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/llama/language_model.py (245:258) - megatron_patch/model/llama3/language_model.py (242:255) duplicated block id: 4926 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (797:811) - megatron_patch/model/llama3/transformer_legacy.py (1185:1199) duplicated block id: 4927 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/language_model.py (198:211) - megatron_patch/model/qwen_vl/language_model.py (185:198) duplicated block id: 4928 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/language_model.py (99:129) - megatron_patch/model/galactica/language_model.py (98:126) duplicated block id: 4929 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/transformer.py (715:725) - megatron_patch/model/glm130b/transformer.py (909:919) duplicated block id: 4930 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (379:390) - megatron_patch/model/starcoder/transformer.py (266:278) duplicated block id: 4931 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/falcon/language_model.py (245:258) - megatron_patch/model/mistral/language_model.py (248:261) duplicated block id: 4932 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/language_model.py (99:127) - megatron_patch/model/llama2/language_model.py (89:119) duplicated block id: 4933 size: 10 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_dense_and_moe_gqa.py (138:147) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_moe.py (352:361) duplicated block id: 4934 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/llama/transformer.py (650:663) - megatron_patch/model/mistral/transformer.py (896:909) duplicated block id: 4935 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/galactica/language_model.py (228:241) - megatron_patch/model/qwen1_5_megablocks/language_model.py (183:196) duplicated block id: 4936 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (1077:1089) - megatron_patch/model/chatglm/transformer.py (790:802) duplicated block id: 4937 size: 10 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/hf2te.py (424:435) - toolkits/model_checkpoints_convertor/falcon/checkpoint_reshaping_and_interoperability.py (543:554) duplicated block id: 4938 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (477:488) - megatron_patch/model/galactica/transformer.py (431:442) duplicated block id: 4939 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/language_model.py (257:270) - megatron_patch/model/falcon40b/language_model.py (245:258) duplicated block id: 4940 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/llama/language_model.py (688:699) - megatron_patch/model/qwen1_5_megablocks/language_model.py (631:642) duplicated block id: 4941 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/qwen2_moe/layer_specs.py (299:311) - megatron_patch/model/qwen3_moe/gpt_layer_specs.py (319:331) duplicated block id: 4942 size: 10 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/mistral/hf2mcore.py (232:241) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_moe.py (352:361) duplicated block id: 4943 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/transformer.py (711:724) - megatron_patch/model/llama2/transformer.py (899:912) duplicated block id: 4944 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/transformer.py (545:554) - megatron_patch/model/llama/transformer.py (497:506) duplicated block id: 4945 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/llama2/rotary_pos_embedding.py (27:38) - megatron_patch/model/llava/rotary_pos_embedding.py (26:37) duplicated block id: 4946 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/language_model.py (235:248) - megatron_patch/model/falcon/language_model.py (197:210) duplicated block id: 4947 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/qwen_vl/language_model.py (31:42) - megatron_patch/model/starcoder/language_model.py (29:40) duplicated block id: 4948 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/language_model.py (131:156) - megatron_patch/model/starcoder/language_model.py (119:142) duplicated block id: 4949 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/llava/rotary_pos_embedding.py (26:37) - megatron_patch/model/qwen1_5_megablocks/rotary_pos_embedding.py (27:38) duplicated block id: 4950 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/model.py (112:121) - megatron_patch/model/qwen2_vl/gpt_model.py (121:130) duplicated block id: 4951 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/mixtral_bak/transformer/mlp.py (57:67) - megatron_patch/model/qwen1_5/transformer/mlp.py (87:97) duplicated block id: 4952 size: 10 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/hf2te.py (194:210) - toolkits/model_checkpoints_convertor/falcon40b/checkpoint_reshaping_and_interoperability.py (287:303) duplicated block id: 4953 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/language_model.py (98:126) - megatron_patch/model/qwen1_5_megablocks/language_model.py (88:118) duplicated block id: 4954 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/falcon/transformer.py (1092:1102) - megatron_patch/model/glm130b/transformer.py (909:919) duplicated block id: 4955 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (797:811) - megatron_patch/model/llava/transformer.py (1226:1240) duplicated block id: 4956 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/qwen1_5/moe/token_dispatcher.py (25:52) - megatron_patch/model/qwen2/moe/token_dispatcher.py (26:53) duplicated block id: 4957 size: 10 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (942:951) - toolkits/model_checkpoints_convertor/bloom/checkpoint_reshaping_and_interoperability.py (818:827) duplicated block id: 4958 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/language_model.py (100:128) - megatron_patch/model/glm130b/language_model.py (99:127) duplicated block id: 4959 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/transformer.py (84:96) - megatron_patch/model/starcoder/transformer.py (117:129) duplicated block id: 4960 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/language_model.py (134:164) - megatron_patch/model/bloom/language_model.py (100:128) duplicated block id: 4961 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/transformer.py (715:725) - megatron_patch/model/galactica/transformer.py (720:730) duplicated block id: 4962 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/galactica/language_model.py (706:717) - megatron_patch/model/qwen/language_model.py (614:625) duplicated block id: 4963 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/llama/transformer.py (751:761) - megatron_patch/model/llava/transformer.py (1234:1244) duplicated block id: 4964 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/qwen_vl/transformer.py (1226:1240) - megatron_patch/model/starcoder/transformer.py (942:956) duplicated block id: 4965 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/qwen1_5/moe/router.py (22:34) - megatron_patch/model/qwen2/moe/router.py (21:33) duplicated block id: 4966 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/galactica/transformer.py (149:160) - megatron_patch/model/starcoder/transformer.py (217:228) duplicated block id: 4967 size: 10 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/checkpoint_reshaping_and_interoperability.py (505:522) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (1096:1113) duplicated block id: 4968 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/llama/transformer.py (751:761) - megatron_patch/model/qwen/transformer.py (1177:1187) duplicated block id: 4969 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/bloom/language_model.py (394:406) - megatron_patch/model/chatglm/language_model.py (397:409) duplicated block id: 4970 size: 10 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon40b/checkpoint_reshaping_and_interoperability.py (784:801) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (1096:1113) duplicated block id: 4971 size: 10 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/checkpoint_reshaping_and_interoperability.py (818:827) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (1018:1027) duplicated block id: 4972 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/model.py (145:154) - megatron_patch/model/mixtral/model.py (112:121) duplicated block id: 4973 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/language_model.py (197:210) - megatron_patch/model/starcoder/language_model.py (183:196) duplicated block id: 4974 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/llama3/rms_norm.py (5:18) - megatron_patch/model/qwen2/rms_norm.py (5:18) duplicated block id: 4975 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/llama/language_model.py (197:210) - megatron_patch/model/llava/language_model.py (187:200) duplicated block id: 4976 size: 10 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/chatglm/checkpoint_reshaping_and_interoperability.py (40:49) - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (72:82) duplicated block id: 4977 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/galactica/language_model.py (276:289) - megatron_patch/model/qwen1_5_megablocks/language_model.py (246:259) duplicated block id: 4978 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (984:994) - megatron_patch/model/chatglm/transformer.py (715:725) duplicated block id: 4979 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/transformer.py (637:653) - megatron_patch/model/starcoder/transformer.py (806:821) duplicated block id: 4980 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/language_model.py (134:164) - megatron_patch/model/glm130b/language_model.py (99:127) duplicated block id: 4981 size: 10 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (642:656) - toolkits/model_checkpoints_convertor/yi/checkpoint_reshaping_and_interoperability.py (557:571) duplicated block id: 4982 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/llama2/transformer.py (1227:1241) - megatron_patch/model/starcoder/transformer.py (942:956) duplicated block id: 4983 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (751:764) - megatron_patch/model/llama/transformer.py (650:663) duplicated block id: 4984 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/qwen2_moe/layer_specs.py (42:59) - megatron_patch/model/qwen3_moe/gpt_layer_specs.py (44:65) duplicated block id: 4985 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (561:571) - megatron_patch/model/qwen_vl/transformer.py (779:789) duplicated block id: 4986 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (1235:1245) - megatron_patch/model/falcon/transformer.py (992:1002) duplicated block id: 4987 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/qwen1_5/moe/token_dispatcher.py (305:314) - megatron_patch/model/qwen2/moe/token_dispatcher.py (313:322) duplicated block id: 4988 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/llama/language_model.py (98:126) - megatron_patch/model/mistral/language_model.py (90:120) duplicated block id: 4989 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (970:982) - megatron_patch/model/chatglm/transformer.py (700:712) duplicated block id: 4990 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer_legacy.py (787:798) - megatron_patch/model/qwen/transformer.py (781:792) duplicated block id: 4991 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/language_model.py (98:126) - megatron_patch/model/starcoder/language_model.py (87:117) duplicated block id: 4992 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/language_model.py (197:210) - megatron_patch/model/mistral/language_model.py (185:198) duplicated block id: 4993 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/mixtral_bak/transformer/attention.py (311:334) - megatron_patch/model/qwen2_vl/attention_vision.py (471:494) duplicated block id: 4994 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/bloom/language_model.py (263:276) - megatron_patch/model/mistral/language_model.py (248:261) duplicated block id: 4995 size: 10 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (1135:1144) - toolkits/model_checkpoints_convertor/starcoder/checkpoint_reshaping_and_interoperability.py (827:836) duplicated block id: 4996 size: 10 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_moe.py (200:210) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_moe.py (290:300) duplicated block id: 4997 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/language_model.py (245:258) - megatron_patch/model/qwen/language_model.py (247:260) duplicated block id: 4998 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/transformer.py (199:210) - megatron_patch/model/starcoder/transformer.py (217:228) duplicated block id: 4999 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (605:621) - megatron_patch/model/starcoder/transformer.py (806:821) duplicated block id: 5000 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/transformer.py (613:623) - megatron_patch/model/qwen_vl/transformer.py (779:789) duplicated block id: 5001 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/language_model.py (198:211) - megatron_patch/model/llama2/language_model.py (184:197) duplicated block id: 5002 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/qwen1_5_megablocks/language_model.py (120:145) - megatron_patch/model/starcoder/language_model.py (119:142) duplicated block id: 5003 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/falcon/language_model.py (197:210) - megatron_patch/model/qwen/language_model.py (184:197) duplicated block id: 5004 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/transformer.py (711:724) - megatron_patch/model/qwen_vl/transformer.py (898:911) duplicated block id: 5005 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/llama/language_model.py (688:699) - megatron_patch/model/qwen_vl/language_model.py (661:672) duplicated block id: 5006 size: 10 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/hf2te.py (194:210) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (384:400) duplicated block id: 5007 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/llama/language_model.py (688:699) - megatron_patch/model/llama2/language_model.py (632:643) duplicated block id: 5008 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/mixtral_bak/model.py (108:117) - megatron_patch/model/qwen2/model.py (131:140) duplicated block id: 5009 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/falcon/language_model.py (98:126) - megatron_patch/model/llama2/language_model.py (89:119) duplicated block id: 5010 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/llama3/layer_specs.py (43:52) - megatron_patch/model/qwen2_moe/layer_specs.py (159:168) duplicated block id: 5011 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/llama/language_model.py (688:699) - megatron_patch/model/llava/language_model.py (692:703) duplicated block id: 5012 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/llama/transformer.py (751:761) - megatron_patch/model/llama3/transformer_legacy.py (1193:1203) duplicated block id: 5013 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/language_model.py (197:210) - megatron_patch/model/starcoder/language_model.py (183:196) duplicated block id: 5014 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/model.py (145:154) - megatron_patch/model/mixtral_bak/model.py (108:117) duplicated block id: 5015 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (984:994) - megatron_patch/model/galactica/transformer.py (720:730) duplicated block id: 5016 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/galactica/language_model.py (98:126) - megatron_patch/model/llama2/language_model.py (89:119) duplicated block id: 5017 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/galactica/language_model.py (276:289) - megatron_patch/model/qwen_vl/language_model.py (248:261) duplicated block id: 5018 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/language_model.py (194:207) - megatron_patch/model/falcon/language_model.py (197:210) duplicated block id: 5019 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/falcon/transformer.py (797:808) - megatron_patch/model/glm130b/transformer.py (613:623) duplicated block id: 5020 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/llama3/layer_specs.py (48:57) - megatron_patch/model/qwen2_vl/layer_specs.py (49:58) duplicated block id: 5021 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (751:764) - megatron_patch/model/glm130b/transformer.py (711:724) duplicated block id: 5022 size: 10 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (983:1000) - toolkits/model_checkpoints_convertor/starcoder/checkpoint_reshaping_and_interoperability.py (520:537) duplicated block id: 5023 size: 10 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon/checkpoint_reshaping_and_interoperability.py (174:185) - toolkits/model_checkpoints_convertor/falcon40b/checkpoint_reshaping_and_interoperability.py (174:185) duplicated block id: 5024 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/language_model.py (100:128) - megatron_patch/model/llava/language_model.py (92:122) duplicated block id: 5025 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (561:571) - megatron_patch/model/llava/transformer.py (779:789) duplicated block id: 5026 size: 10 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon/checkpoint_reshaping_and_interoperability.py (843:852) - toolkits/model_checkpoints_convertor/starcoder/checkpoint_reshaping_and_interoperability.py (827:836) duplicated block id: 5027 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/llama/language_model.py (197:210) - megatron_patch/model/qwen_vl/language_model.py (185:198) duplicated block id: 5028 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/language_model.py (197:210) - megatron_patch/model/llava/language_model.py (187:200) duplicated block id: 5029 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (1102:1112) - megatron_patch/model/falcon/transformer.py (992:1002) duplicated block id: 5030 size: 10 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/deepseek/hf2mcore_deepseek_v3_moe.py (139:151) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_dense_and_moe_gqa.py (56:68) duplicated block id: 5031 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/llama/language_model.py (98:126) - megatron_patch/model/starcoder/language_model.py (87:117) duplicated block id: 5032 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/llama3/model.py (102:115) - megatron_patch/model/mixtral_bak/model.py (94:107) duplicated block id: 5033 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/language_model.py (132:155) - megatron_patch/model/qwen1_5_megablocks/language_model.py (123:147) duplicated block id: 5034 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/galactica/language_model.py (98:126) - megatron_patch/model/mistral/language_model.py (90:120) duplicated block id: 5035 size: 10 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/hf2te.py (178:194) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (362:378) duplicated block id: 5036 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/qwen1_5_megablocks/language_model.py (29:40) - megatron_patch/model/starcoder/language_model.py (29:40) duplicated block id: 5037 size: 10 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2.5_vl.py (192:201) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_moe.py (352:361) duplicated block id: 5038 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/bloom/language_model.py (581:593) - megatron_patch/model/llama3/language_model.py (602:614) duplicated block id: 5039 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (675:685) - megatron_patch/model/glm130b/transformer.py (777:787) duplicated block id: 5040 size: 10 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/hf2te.py (178:194) - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (365:381) duplicated block id: 5041 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (1136:1155) - megatron_patch/model/llama/transformer.py (1027:1047) duplicated block id: 5042 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/language_model.py (245:258) - megatron_patch/model/llama2/language_model.py (247:260) duplicated block id: 5043 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/llama/language_model.py (197:210) - megatron_patch/model/qwen/language_model.py (184:197) duplicated block id: 5044 size: 10 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/mistral/hf2mcore.py (105:114) - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (104:113) duplicated block id: 5045 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/llama3_1/model.py (188:197) - megatron_patch/model/mixtral_bak/model.py (108:117) duplicated block id: 5046 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/transformer.py (613:623) - megatron_patch/model/qwen1_5_megablocks/transformer.py (671:681) duplicated block id: 5047 size: 10 cleaned lines of code in 2 files: - toolkits/pretrain_data_preprocessing/preprocess_data_megatron.py (217:226) - toolkits/sft_data_preprocessing/build_idxmap_sft_dataset.py (208:217) duplicated block id: 5048 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (797:811) - megatron_patch/model/qwen_vl/transformer.py (1226:1240) duplicated block id: 5049 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/bloom/language_model.py (199:212) - megatron_patch/model/llama2/language_model.py (184:197) duplicated block id: 5050 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/language_model.py (194:207) - megatron_patch/model/chatglm/language_model.py (198:211) duplicated block id: 5051 size: 10 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore_llama3_1.py (575:584) - toolkits/model_checkpoints_convertor/llava/hf2mcore_llava.py (392:401) duplicated block id: 5052 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/language_model.py (100:128) - megatron_patch/model/galactica/language_model.py (98:126) duplicated block id: 5053 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/bloom/language_model.py (199:212) - megatron_patch/model/qwen1_5_megablocks/language_model.py (183:196) duplicated block id: 5054 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/qwen1_5/transformer/attention.py (338:361) - megatron_patch/model/qwen2_vl/attention.py (472:495) duplicated block id: 5055 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/bloom/language_model.py (100:128) - megatron_patch/model/qwen_vl/language_model.py (90:120) duplicated block id: 5056 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/language_model.py (647:659) - megatron_patch/model/llama3/language_model.py (602:614) duplicated block id: 5057 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/falcon/transformer.py (992:1002) - megatron_patch/model/qwen1_5_megablocks/transformer.py (1119:1129) duplicated block id: 5058 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (646:656) - megatron_patch/model/llama/transformer.py (570:580) duplicated block id: 5059 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/galactica/language_model.py (276:289) - megatron_patch/model/llama3/language_model.py (242:255) duplicated block id: 5060 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (688:699) - megatron_patch/model/llama3/transformer_legacy.py (787:798) duplicated block id: 5061 size: 10 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (427:438) - toolkits/model_checkpoints_convertor/chatglm/checkpoint_reshaping_and_interoperability.py (362:373) duplicated block id: 5062 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/llama/language_model.py (98:126) - megatron_patch/model/llama3/language_model.py (78:108) duplicated block id: 5063 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/language_model.py (100:128) - megatron_patch/model/llama3/language_model.py (78:108) duplicated block id: 5064 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/qwen2_5_vl/transformer_config.py (49:58) - megatron_patch/model/qwen2_vl/transformer_config.py (41:50) duplicated block id: 5065 size: 10 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/language_model.py (98:126) - megatron_patch/model/mistral/language_model.py (90:120) duplicated block id: 5066 size: 10 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/deepseek/hf2mcore_deepseek_v2_moe.py (147:159) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_moe.py (200:210) duplicated block id: 5067 size: 10 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon/checkpoint_reshaping_and_interoperability.py (808:825) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (1096:1113) duplicated block id: 5068 size: 9 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore_llama3_1.py (634:642) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_dense_and_moe_gqa.py (689:697) duplicated block id: 5069 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/llama/language_model.py (32:42) - megatron_patch/model/llama3/language_model.py (20:30) duplicated block id: 5070 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/language_model.py (244:259) - megatron_patch/model/llama2/language_model.py (231:245) duplicated block id: 5071 size: 9 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore_llama3_1.py (634:642) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_dense_and_moe_gqa.py (766:774) duplicated block id: 5072 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/bloom/language_model.py (213:226) - megatron_patch/model/llama2/language_model.py (199:212) duplicated block id: 5073 size: 9 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/hf2te.py (356:370) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (564:578) duplicated block id: 5074 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (1307:1316) - megatron_patch/model/llama2/transformer.py (1440:1449) duplicated block id: 5075 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/language_model.py (349:357) - megatron_patch/model/qwen1_5_megablocks/language_model.py (333:341) duplicated block id: 5076 size: 9 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/galactica/checkpoint_reshaping_and_interoperability.py (542:556) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (414:428) duplicated block id: 5077 size: 9 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/checkpoint_reshaping_and_interoperability.py (714:728) - toolkits/model_checkpoints_convertor/chatglm/checkpoint_reshaping_and_interoperability.py (465:480) duplicated block id: 5078 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/language_model.py (262:274) - megatron_patch/model/qwen/language_model.py (248:260) duplicated block id: 5079 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/transformer.py (572:581) - megatron_patch/model/falcon40b/transformer.py (599:607) duplicated block id: 5080 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/galactica/language_model.py (365:373) - megatron_patch/model/qwen/language_model.py (334:342) duplicated block id: 5081 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/language_model.py (227:242) - megatron_patch/model/qwen_vl/language_model.py (232:246) duplicated block id: 5082 size: 9 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/checkpoint_reshaping_and_interoperability.py (714:728) - toolkits/model_checkpoints_convertor/falcon40b/checkpoint_reshaping_and_interoperability.py (485:500) duplicated block id: 5083 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/mlp.py (72:80) - megatron_patch/model/mixtral/transformer/mlp.py (96:104) duplicated block id: 5084 size: 9 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/reward_model_to_megatron.py (713:727) - toolkits/model_checkpoints_convertor/falcon40b/checkpoint_reshaping_and_interoperability.py (485:500) duplicated block id: 5085 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/language_model.py (33:43) - megatron_patch/model/qwen_vl/language_model.py (32:42) duplicated block id: 5086 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/qwen1_5/layer_specs.py (89:100) - megatron_patch/model/qwen3_moe/gpt_layer_specs.py (253:263) duplicated block id: 5087 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (1307:1316) - megatron_patch/model/qwen_vl/transformer.py (1439:1448) duplicated block id: 5088 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/llama2/transformer.py (834:843) - megatron_patch/model/qwen_vl/transformer.py (768:777) duplicated block id: 5089 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/falcon/language_model.py (334:342) - megatron_patch/model/qwen1_5_megablocks/language_model.py (333:341) duplicated block id: 5090 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (970:981) - megatron_patch/model/galactica/transformer.py (707:718) duplicated block id: 5091 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/bloom/language_model.py (213:226) - megatron_patch/model/mistral/language_model.py (200:213) duplicated block id: 5092 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/language_model.py (33:43) - megatron_patch/model/starcoder/language_model.py (30:40) duplicated block id: 5093 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/qwen/transformer.py (385:411) - megatron_patch/model/starcoder/transformer.py (295:322) duplicated block id: 5094 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/llama/language_model.py (32:42) - megatron_patch/model/qwen/language_model.py (31:41) duplicated block id: 5095 size: 9 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore.py (330:338) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_dense_and_moe_gqa.py (320:328) duplicated block id: 5096 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/qwen2_vl/rope_utils.py (68:84) - megatron_patch/model/qwen2_vl/rotary_pos_embedding.py (172:188) duplicated block id: 5097 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/llava_mcore/llava_model.py (107:117) - megatron_patch/model/qwen2_5_vl/model.py (78:88) duplicated block id: 5098 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/bloom/gpt_model.py (28:41) - megatron_patch/model/llama/gpt_model.py (28:41) duplicated block id: 5099 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer/attention.py (434:442) - megatron_patch/model/qwen2_vl/attention.py (599:607) duplicated block id: 5100 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/language_model.py (34:44) - megatron_patch/model/llama3/language_model.py (20:30) duplicated block id: 5101 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/language_model.py (33:43) - megatron_patch/model/llama2/language_model.py (31:41) duplicated block id: 5102 size: 9 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/galactica/checkpoint_reshaping_and_interoperability.py (296:313) - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (173:190) duplicated block id: 5103 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/layer_specs.py (30:44) - megatron_patch/model/qwen3_moe/gpt_layer_specs.py (43:59) duplicated block id: 5104 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/llama2/transformer.py (834:843) - megatron_patch/model/llava/transformer.py (768:777) duplicated block id: 5105 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/llama3_1/model.py (278:291) - megatron_patch/model/qwen1_5/model.py (215:228) duplicated block id: 5106 size: 9 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/hf2te.py (356:370) - toolkits/model_checkpoints_convertor/chatglm/checkpoint_reshaping_and_interoperability.py (464:478) duplicated block id: 5107 size: 9 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (883:897) - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (742:756) duplicated block id: 5108 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/mistral/language_model.py (471:481) - megatron_patch/model/starcoder/language_model.py (433:443) duplicated block id: 5109 size: 9 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/hf2te.py (198:211) - toolkits/model_checkpoints_convertor/falcon/checkpoint_reshaping_and_interoperability.py (316:329) duplicated block id: 5110 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/qwen2/transformer/attention.py (457:465) - megatron_patch/model/qwen2_vl/attention.py (580:588) duplicated block id: 5111 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/language_model.py (262:274) - megatron_patch/model/llava/language_model.py (251:263) duplicated block id: 5112 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/language_model.py (33:43) - megatron_patch/model/llava/language_model.py (34:44) duplicated block id: 5113 size: 9 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (173:190) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (342:359) duplicated block id: 5114 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (333:348) - megatron_patch/model/starcoder/transformer.py (489:503) duplicated block id: 5115 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/falcon/transformer.py (192:203) - megatron_patch/model/starcoder/transformer.py (203:214) duplicated block id: 5116 size: 9 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon40b/checkpoint_reshaping_and_interoperability.py (689:703) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (819:834) duplicated block id: 5117 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/qwen2_moe/layer_specs.py (283:292) - megatron_patch/model/qwen3_moe/gpt_layer_specs.py (304:313) duplicated block id: 5118 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (627:642) - megatron_patch/model/glm130b/transformer.py (594:609) duplicated block id: 5119 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/transformer.py (572:581) - megatron_patch/model/llama/transformer.py (587:596) duplicated block id: 5120 size: 9 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon40b/checkpoint_reshaping_and_interoperability.py (484:498) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (414:428) duplicated block id: 5121 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (489:501) - megatron_patch/model/falcon/transformer.py (507:519) duplicated block id: 5122 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/language_model.py (628:636) - megatron_patch/model/falcon/language_model.py (680:688) duplicated block id: 5123 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/language_model.py (344:352) - megatron_patch/model/glm130b/language_model.py (349:357) duplicated block id: 5124 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (380:392) - megatron_patch/model/falcon/transformer.py (507:519) duplicated block id: 5125 size: 9 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/reward_model_to_megatron.py (670:680) - toolkits/model_checkpoints_convertor/starcoder/checkpoint_reshaping_and_interoperability.py (675:685) duplicated block id: 5126 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/language_model.py (244:259) - megatron_patch/model/llava/language_model.py (234:248) duplicated block id: 5127 size: 9 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (565:580) - toolkits/model_checkpoints_convertor/starcoder/checkpoint_reshaping_and_interoperability.py (714:728) duplicated block id: 5128 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/llama3_1/layer_specs.py (89:99) - megatron_patch/model/qwen2_moe/layer_specs.py (232:242) duplicated block id: 5129 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/language_model.py (76:86) - megatron_patch/model/chatglm/language_model.py (34:44) duplicated block id: 5130 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/language_model.py (34:44) - megatron_patch/model/qwen/language_model.py (31:41) duplicated block id: 5131 size: 9 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/galactica/checkpoint_reshaping_and_interoperability.py (543:558) - toolkits/model_checkpoints_convertor/glm130b/checkpoint_reshaping_and_interoperability.py (422:436) duplicated block id: 5132 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/qwen2/transformer/attention.py (438:446) - megatron_patch/model/qwen2_vl/attention_vision.py (598:606) duplicated block id: 5133 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/transformer_layer.py (129:143) - megatron_patch/model/qwen2/transformer_layer.py (96:110) duplicated block id: 5134 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/bloom/language_model.py (245:260) - megatron_patch/model/llava/language_model.py (234:248) duplicated block id: 5135 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/llama/language_model.py (334:342) - megatron_patch/model/llama2/language_model.py (334:342) duplicated block id: 5136 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/language_model.py (334:342) - megatron_patch/model/qwen1_5_megablocks/language_model.py (333:341) duplicated block id: 5137 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/llama/language_model.py (32:42) - megatron_patch/model/qwen_vl/language_model.py (32:42) duplicated block id: 5138 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/language_model.py (244:259) - megatron_patch/model/mistral/language_model.py (232:246) duplicated block id: 5139 size: 9 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/mistral/hf2mcore.py (230:238) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_dense_and_moe_gqa.py (320:328) duplicated block id: 5140 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/falcon/language_model.py (227:242) - megatron_patch/model/llama2/language_model.py (231:245) duplicated block id: 5141 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/qwen2/layer_specs.py (99:110) - megatron_patch/model/qwen3_moe/gpt_layer_specs.py (253:263) duplicated block id: 5142 size: 9 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (354:367) - toolkits/model_checkpoints_convertor/baichuan2/hf2te.py (198:211) duplicated block id: 5143 size: 9 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/deepseek/hf2mcore_deepseek_v2_moe.py (515:526) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_vl.py (702:715) duplicated block id: 5144 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/language_model.py (241:255) - megatron_patch/model/starcoder/language_model.py (232:246) duplicated block id: 5145 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/llama/language_model.py (32:42) - megatron_patch/model/llava/language_model.py (34:44) duplicated block id: 5146 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (1307:1316) - megatron_patch/model/qwen1_5_megablocks/transformer.py (1325:1334) duplicated block id: 5147 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/language_model.py (33:43) - megatron_patch/model/llama/language_model.py (32:42) duplicated block id: 5148 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (643:654) - megatron_patch/model/falcon40b/transformer.py (680:691) duplicated block id: 5149 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/llava_mcore/layer_specs.py (20:34) - megatron_patch/model/mixtral/layer_specs.py (37:51) duplicated block id: 5150 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/transformer.py (115:126) - megatron_patch/model/starcoder/transformer.py (203:214) duplicated block id: 5151 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/llama3/layer_specs.py (89:99) - megatron_patch/model/qwen2_moe/layer_specs.py (232:242) duplicated block id: 5152 size: 9 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_moe.py (145:154) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_dense_and_moe_gqa.py (227:236) duplicated block id: 5153 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/qwen2/transformer/attention.py (438:446) - megatron_patch/model/qwen2/transformer/attention.py (457:465) duplicated block id: 5154 size: 9 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/checkpoint_reshaping_and_interoperability.py (671:681) - toolkits/model_checkpoints_convertor/starcoder/checkpoint_reshaping_and_interoperability.py (675:685) duplicated block id: 5155 size: 9 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore.py (259:267) - toolkits/model_checkpoints_convertor/mistral/hf2mcore.py (197:205) duplicated block id: 5156 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/language_model.py (334:342) - megatron_patch/model/llama3/language_model.py (329:337) duplicated block id: 5157 size: 9 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/hf2te.py (198:211) - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (392:405) duplicated block id: 5158 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/language_model.py (33:43) - megatron_patch/model/galactica/language_model.py (32:42) duplicated block id: 5159 size: 9 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon40b/checkpoint_reshaping_and_interoperability.py (485:500) - toolkits/model_checkpoints_convertor/glm130b/checkpoint_reshaping_and_interoperability.py (422:436) duplicated block id: 5160 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/qwen1_5/transformer/attention.py (455:463) - megatron_patch/model/qwen2_vl/attention_vision.py (579:587) duplicated block id: 5161 size: 9 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_mha.py (280:290) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_moe.py (388:399) duplicated block id: 5162 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/multi_latent_attention.py (111:119) - megatron_patch/model/mixtral/transformer/attention.py (344:352) duplicated block id: 5163 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/transformer.py (680:691) - megatron_patch/model/starcoder/transformer.py (857:868) duplicated block id: 5164 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/bloom/language_model.py (34:44) - megatron_patch/model/llava/language_model.py (34:44) duplicated block id: 5165 size: 9 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/galactica/checkpoint_reshaping_and_interoperability.py (296:313) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (152:169) duplicated block id: 5166 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/transformer.py (936:951) - megatron_patch/model/galactica/transformer.py (826:841) duplicated block id: 5167 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/galactica/language_model.py (258:273) - megatron_patch/model/llava/language_model.py (234:248) duplicated block id: 5168 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer/attention.py (434:442) - megatron_patch/model/mixtral/transformer/attention.py (598:606) duplicated block id: 5169 size: 9 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/checkpoint_reshaping_and_interoperability.py (714:728) - toolkits/model_checkpoints_convertor/galactica/checkpoint_reshaping_and_interoperability.py (543:558) duplicated block id: 5170 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/qwen2/transformer/attention.py (317:325) - megatron_patch/model/qwen2_vl/attention_vision.py (453:461) duplicated block id: 5171 size: 9 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2.5_vl.py (43:52) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_moe.py (41:53) duplicated block id: 5172 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/mlp.py (72:80) - megatron_patch/model/llama3/transformer/mlp.py (94:102) duplicated block id: 5173 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/gpt_model.py (28:41) - megatron_patch/model/llama/gpt_model.py (28:41) duplicated block id: 5174 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/transformer.py (251:268) - megatron_patch/model/mistral/transformer.py (391:408) duplicated block id: 5175 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/transformer/attention.py (598:606) - megatron_patch/model/qwen2/transformer/attention.py (438:446) duplicated block id: 5176 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/llama/language_model.py (227:242) - megatron_patch/model/qwen_vl/language_model.py (232:246) duplicated block id: 5177 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/falcon/language_model.py (227:242) - megatron_patch/model/mistral/language_model.py (232:246) duplicated block id: 5178 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/transformer.py (654:667) - megatron_patch/model/qwen_vl/transformer.py (845:858) duplicated block id: 5179 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/language_model.py (33:43) - megatron_patch/model/falcon/language_model.py (32:42) duplicated block id: 5180 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/bloom/language_model.py (245:260) - megatron_patch/model/llama3/language_model.py (221:235) duplicated block id: 5181 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (164:175) - megatron_patch/model/starcoder/transformer.py (203:214) duplicated block id: 5182 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/transformer/attention.py (579:587) - megatron_patch/model/qwen2_vl/attention.py (599:607) duplicated block id: 5183 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/mixtral_bak/layer_specs.py (136:145) - megatron_patch/model/qwen3_moe/gpt_layer_specs.py (298:307) duplicated block id: 5184 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/bloom/language_model.py (34:44) - megatron_patch/model/llama3/language_model.py (20:30) duplicated block id: 5185 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/layer_specs.py (132:142) - megatron_patch/model/qwen2/layer_specs.py (99:110) duplicated block id: 5186 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/transformer.py (594:609) - megatron_patch/model/qwen/transformer.py (720:735) duplicated block id: 5187 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/mistral/transformer.py (767:776) - megatron_patch/model/mistral/transformer.py (831:840) duplicated block id: 5188 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/transformer/attention.py (674:682) - megatron_patch/model/qwen2_vl/attention_vision.py (493:501) duplicated block id: 5189 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/language_model.py (349:357) - megatron_patch/model/llama2/language_model.py (334:342) duplicated block id: 5190 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/transformer/attention.py (598:606) - megatron_patch/model/qwen2_vl/attention_vision.py (579:587) duplicated block id: 5191 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/qwen2/layer_specs.py (65:76) - megatron_patch/model/qwen3_moe/gpt_layer_specs.py (157:167) duplicated block id: 5192 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (1585:1593) - megatron_patch/model/qwen1_5_megablocks/transformer.py (1577:1585) duplicated block id: 5193 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/falcon/language_model.py (227:242) - megatron_patch/model/llava/language_model.py (234:248) duplicated block id: 5194 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/qwen1_5_megablocks/transformer.py (725:734) - megatron_patch/model/qwen_vl/transformer.py (768:777) duplicated block id: 5195 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/llama/transformer.py (317:332) - megatron_patch/model/starcoder/transformer.py (489:503) duplicated block id: 5196 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/language_model.py (262:274) - megatron_patch/model/qwen_vl/language_model.py (249:261) duplicated block id: 5197 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/language_model.py (390:398) - megatron_patch/model/llama/language_model.py (334:342) duplicated block id: 5198 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/falcon/transformer.py (786:795) - megatron_patch/model/llama/transformer.py (587:596) duplicated block id: 5199 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer_legacy.py (1675:1683) - megatron_patch/model/llava/transformer.py (1714:1722) duplicated block id: 5200 size: 9 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/deepseek/hf2mcore_deepseek_v2_moe.py (66:76) - toolkits/model_checkpoints_convertor/deepseek/hf2mcore_deepseek_v3_moe.py (125:135) duplicated block id: 5201 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer/attention.py (321:329) - megatron_patch/model/qwen2_vl/attention_vision.py (453:461) duplicated block id: 5202 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer/attention.py (434:442) - megatron_patch/model/qwen2_vl/attention_vision.py (598:606) duplicated block id: 5203 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/qwen/language_model.py (231:245) - megatron_patch/model/starcoder/language_model.py (232:246) duplicated block id: 5204 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/bloom/language_model.py (34:44) - megatron_patch/model/mistral/language_model.py (32:42) duplicated block id: 5205 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/falcon/language_model.py (32:42) - megatron_patch/model/qwen_vl/language_model.py (32:42) duplicated block id: 5206 size: 9 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_dense_and_moe_gqa.py (896:906) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_vl.py (706:716) duplicated block id: 5207 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/llama3/layer_specs.py (89:99) - megatron_patch/model/qwen3_moe/gpt_layer_specs.py (253:263) duplicated block id: 5208 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/qwen1_5/transformer/attention.py (321:329) - megatron_patch/model/qwen2_vl/attention.py (454:462) duplicated block id: 5209 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/galactica/language_model.py (365:373) - megatron_patch/model/qwen1_5_megablocks/language_model.py (333:341) duplicated block id: 5210 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/falcon/language_model.py (32:42) - megatron_patch/model/llama2/language_model.py (31:41) duplicated block id: 5211 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/qwen2/layer_specs.py (65:76) - megatron_patch/model/qwen2_moe/layer_specs.py (171:181) duplicated block id: 5212 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/mlp.py (88:96) - megatron_patch/model/mixtral_bak/transformer/mlp.py (57:65) duplicated block id: 5213 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/qwen2_vl/attention.py (580:588) - megatron_patch/model/qwen2_vl/attention_vision.py (598:606) duplicated block id: 5214 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/transformer.py (818:833) - megatron_patch/model/falcon/transformer.py (1197:1212) duplicated block id: 5215 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/llama/transformer.py (239:256) - megatron_patch/model/qwen_vl/transformer.py (391:408) duplicated block id: 5216 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/falcon/transformer.py (1197:1212) - megatron_patch/model/llama/transformer.py (990:1005) duplicated block id: 5217 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/language_model.py (712:720) - megatron_patch/model/falcon/language_model.py (680:688) duplicated block id: 5218 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/galactica/language_model.py (258:273) - megatron_patch/model/llama3/language_model.py (221:235) duplicated block id: 5219 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/bloom/language_model.py (131:154) - megatron_patch/model/starcoder/language_model.py (120:142) duplicated block id: 5220 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/transformer/attention.py (493:501) - megatron_patch/model/mixtral/transformer/attention.py (674:682) duplicated block id: 5221 size: 9 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/hf2te.py (214:227) - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (341:354) duplicated block id: 5222 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/falcon/language_model.py (227:242) - megatron_patch/model/qwen1_5_megablocks/language_model.py (230:244) duplicated block id: 5223 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/llama/language_model.py (334:342) - megatron_patch/model/qwen1_5_megablocks/language_model.py (333:341) duplicated block id: 5224 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/qwen/transformer.py (712:725) - megatron_patch/model/starcoder/transformer.py (715:728) duplicated block id: 5225 size: 9 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/deepseek/hf2mcore_deepseek_v2_moe.py (118:126) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_mha.py (98:107) duplicated block id: 5226 size: 9 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/reward_model_to_megatron.py (713:727) - toolkits/model_checkpoints_convertor/galactica/checkpoint_reshaping_and_interoperability.py (543:558) duplicated block id: 5227 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/language_model.py (76:86) - megatron_patch/model/glm130b/language_model.py (33:43) duplicated block id: 5228 size: 9 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/mistral/hf2mcore.py (257:266) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_moe.py (145:154) duplicated block id: 5229 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/falcon/language_model.py (680:688) - megatron_patch/model/llama2/language_model.py (635:643) duplicated block id: 5230 size: 9 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (742:756) - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (742:756) duplicated block id: 5231 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/gpt_model.py (67:75) - megatron_patch/model/llama/gpt_model.py (71:79) duplicated block id: 5232 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/language_model.py (32:42) - megatron_patch/model/qwen/language_model.py (31:41) duplicated block id: 5233 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/llama3_1/model.py (278:291) - megatron_patch/model/qwen2_vl/gpt_model.py (203:216) duplicated block id: 5234 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/galactica/language_model.py (32:42) - megatron_patch/model/llava/language_model.py (34:44) duplicated block id: 5235 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/falcon/language_model.py (680:688) - megatron_patch/model/qwen_vl/language_model.py (664:672) duplicated block id: 5236 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/falcon/transformer.py (786:795) - megatron_patch/model/falcon/transformer.py (815:824) duplicated block id: 5237 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/bloom/language_model.py (67:75) - megatron_patch/model/chatglm/language_model.py (65:73) duplicated block id: 5238 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/language_model.py (334:342) - megatron_patch/model/qwen/language_model.py (334:342) duplicated block id: 5239 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/llama/transformer.py (173:184) - megatron_patch/model/starcoder/transformer.py (203:214) duplicated block id: 5240 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/language_model.py (32:42) - megatron_patch/model/llama2/language_model.py (31:41) duplicated block id: 5241 size: 9 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/chatglm/checkpoint_reshaping_and_interoperability.py (298:316) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (381:399) duplicated block id: 5242 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/falcon/transformer.py (507:519) - megatron_patch/model/llava/transformer.py (478:490) duplicated block id: 5243 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/bloom/gpt_model.py (67:75) - megatron_patch/model/falcon40b/gpt_model.py (72:80) duplicated block id: 5244 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/bloom/language_model.py (245:260) - megatron_patch/model/mistral/language_model.py (232:246) duplicated block id: 5245 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/language_model.py (680:688) - megatron_patch/model/llama3/language_model.py (609:617) duplicated block id: 5246 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/language_model.py (33:43) - megatron_patch/model/llama3/language_model.py (20:30) duplicated block id: 5247 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/language_model.py (344:352) - megatron_patch/model/galactica/language_model.py (365:373) duplicated block id: 5248 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/falcon/transformer.py (507:519) - megatron_patch/model/qwen/transformer.py (471:483) duplicated block id: 5249 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/transformer.py (251:268) - megatron_patch/model/qwen_vl/transformer.py (391:408) duplicated block id: 5250 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/llama2/transformer.py (769:778) - megatron_patch/model/qwen_vl/transformer.py (833:842) duplicated block id: 5251 size: 9 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/hf2te.py (198:211) - toolkits/model_checkpoints_convertor/chatglm/checkpoint_reshaping_and_interoperability.py (306:319) duplicated block id: 5252 size: 9 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon/checkpoint_reshaping_and_interoperability.py (496:511) - toolkits/model_checkpoints_convertor/glm130b/checkpoint_reshaping_and_interoperability.py (422:436) duplicated block id: 5253 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/language_model.py (244:259) - megatron_patch/model/glm130b/language_model.py (243:258) duplicated block id: 5254 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/layer_specs.py (26:35) - megatron_patch/model/qwen2/layer_specs.py (22:31) duplicated block id: 5255 size: 9 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore.py (317:326) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_dense_and_moe_gqa.py (227:236) duplicated block id: 5256 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/bloom/language_model.py (213:226) - megatron_patch/model/starcoder/language_model.py (198:211) duplicated block id: 5257 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/language_model.py (344:352) - megatron_patch/model/falcon40b/language_model.py (334:342) duplicated block id: 5258 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/falcon/language_model.py (680:688) - megatron_patch/model/llama3/language_model.py (609:617) duplicated block id: 5259 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer/attention.py (434:442) - megatron_patch/model/llama3/transformer/attention.py (455:463) duplicated block id: 5260 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/language_model.py (304:316) - megatron_patch/model/chatglm/language_model.py (262:274) duplicated block id: 5261 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/galactica/language_model.py (258:273) - megatron_patch/model/qwen_vl/language_model.py (232:246) duplicated block id: 5262 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/galactica/language_model.py (258:273) - megatron_patch/model/qwen/language_model.py (231:245) duplicated block id: 5263 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/llama/transformer.py (239:256) - megatron_patch/model/llava/transformer.py (391:408) duplicated block id: 5264 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/language_model.py (65:73) - megatron_patch/model/llama/language_model.py (65:73) duplicated block id: 5265 size: 9 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (152:169) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (342:359) duplicated block id: 5266 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/mixtral_bak/transformer/attention.py (294:302) - megatron_patch/model/qwen2_vl/attention.py (454:462) duplicated block id: 5267 size: 9 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (294:311) - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (173:190) duplicated block id: 5268 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/falcon/language_model.py (227:242) - megatron_patch/model/qwen/language_model.py (231:245) duplicated block id: 5269 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer_legacy.py (1148:1158) - megatron_patch/model/qwen/transformer.py (1142:1152) duplicated block id: 5270 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/language_model.py (262:274) - megatron_patch/model/llama3/language_model.py (243:255) duplicated block id: 5271 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/transformer.py (936:951) - megatron_patch/model/llama/transformer.py (990:1005) duplicated block id: 5272 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/language_model.py (212:225) - megatron_patch/model/llava/language_model.py (202:215) duplicated block id: 5273 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/language_model.py (476:486) - megatron_patch/model/starcoder/language_model.py (433:443) duplicated block id: 5274 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/falcon/transformer.py (786:795) - megatron_patch/model/falcon40b/transformer.py (599:607) duplicated block id: 5275 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/llama3_1/layer_specs.py (89:99) - megatron_patch/model/qwen3_moe/gpt_layer_specs.py (253:263) duplicated block id: 5276 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/transformer_layer.py (103:114) - megatron_patch/model/qwen2/transformer_layer.py (79:91) duplicated block id: 5277 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/transformer/attention.py (598:606) - megatron_patch/model/qwen1_5/transformer/attention.py (434:442) duplicated block id: 5278 size: 9 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llava/hf2mcore_llava.py (453:462) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_moe.py (378:388) duplicated block id: 5279 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/transformer/attention.py (579:587) - megatron_patch/model/mixtral/transformer/attention.py (598:606) duplicated block id: 5280 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/galactica/language_model.py (32:42) - megatron_patch/model/qwen_vl/language_model.py (32:42) duplicated block id: 5281 size: 9 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (184:195) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (188:199) duplicated block id: 5282 size: 9 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/reward_model_to_megatron.py (713:727) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (565:580) duplicated block id: 5283 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/bloom/language_model.py (34:44) - megatron_patch/model/qwen/language_model.py (31:41) duplicated block id: 5284 size: 9 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/hf2te.py (198:211) - toolkits/model_checkpoints_convertor/yi/checkpoint_reshaping_and_interoperability.py (341:354) duplicated block id: 5285 size: 9 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (769:783) - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (628:642) duplicated block id: 5286 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/llama/transformer.py (682:693) - megatron_patch/model/starcoder/transformer.py (857:868) duplicated block id: 5287 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/language_model.py (212:225) - megatron_patch/model/llama2/language_model.py (199:212) duplicated block id: 5288 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/language_model.py (209:222) - megatron_patch/model/bloom/language_model.py (213:226) duplicated block id: 5289 size: 9 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore_llama3_1.py (647:656) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_moe.py (378:388) duplicated block id: 5290 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/transformer.py (224:242) - megatron_patch/model/falcon40b/transformer.py (278:296) duplicated block id: 5291 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/falcon/transformer.py (507:519) - megatron_patch/model/llama3/transformer_legacy.py (476:488) duplicated block id: 5292 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/mistral/language_model.py (376:389) - megatron_patch/model/qwen1_5_megablocks/language_model.py (373:386) duplicated block id: 5293 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/language_model.py (349:357) - megatron_patch/model/llama3/language_model.py (329:337) duplicated block id: 5294 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/language_model.py (76:86) - megatron_patch/model/bloom/language_model.py (34:44) duplicated block id: 5295 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/qwen2/transformer/attention.py (317:325) - megatron_patch/model/qwen2_vl/attention.py (454:462) duplicated block id: 5296 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/qwen1_5/transformer/attention.py (434:442) - megatron_patch/model/qwen2_vl/attention.py (599:607) duplicated block id: 5297 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/mixtral_bak/layer_specs.py (110:121) - megatron_patch/model/qwen3_moe/gpt_layer_specs.py (253:263) duplicated block id: 5298 size: 9 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_mha.py (54:64) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_moe.py (122:132) duplicated block id: 5299 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/language_model.py (680:688) - megatron_patch/model/qwen1_5_megablocks/language_model.py (634:642) duplicated block id: 5300 size: 9 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore.py (290:298) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_dense_and_moe_gqa.py (320:328) duplicated block id: 5301 size: 9 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_dense_and_moe_gqa.py (320:328) - toolkits/model_checkpoints_convertor/qwen/hf2megablocks_qwen1.5.py (284:292) duplicated block id: 5302 size: 9 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_dense_and_moe_gqa.py (320:328) - toolkits/model_checkpoints_convertor/qwen/hf2megablocks_qwen1.5.py (323:331) duplicated block id: 5303 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/qwen1_5/layer_specs.py (55:66) - megatron_patch/model/qwen3_moe/gpt_layer_specs.py (157:167) duplicated block id: 5304 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/transformer/attention.py (454:462) - megatron_patch/model/qwen2_vl/attention.py (454:462) duplicated block id: 5305 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/layer_specs.py (37:51) - megatron_patch/model/qwen3_moe/gpt_layer_specs.py (43:59) duplicated block id: 5306 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/qwen/language_model.py (466:476) - megatron_patch/model/starcoder/language_model.py (433:443) duplicated block id: 5307 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/galactica/language_model.py (183:195) - megatron_patch/model/starcoder/language_model.py (143:155) duplicated block id: 5308 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/llama2/transformer.py (391:417) - megatron_patch/model/starcoder/transformer.py (295:322) duplicated block id: 5309 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (1199:1209) - megatron_patch/model/llama3/transformer_legacy.py (1148:1158) duplicated block id: 5310 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/qwen2/layer_specs.py (99:110) - megatron_patch/model/qwen2_moe/layer_specs.py (232:242) duplicated block id: 5311 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/language_model.py (287:301) - megatron_patch/model/bloom/language_model.py (245:260) duplicated block id: 5312 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/qwen2_vl/attention.py (580:588) - megatron_patch/model/qwen2_vl/attention.py (599:607) duplicated block id: 5313 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/falcon/transformer.py (815:824) - megatron_patch/model/llama/transformer.py (559:568) duplicated block id: 5314 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/galactica/language_model.py (32:42) - megatron_patch/model/mistral/language_model.py (32:42) duplicated block id: 5315 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/layer_specs.py (115:124) - megatron_patch/model/qwen2_moe/layer_specs.py (277:286) duplicated block id: 5316 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/llama2/transformer.py (769:778) - megatron_patch/model/qwen1_5_megablocks/transformer.py (725:734) duplicated block id: 5317 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/transformer/attention.py (598:606) - megatron_patch/model/qwen2_vl/attention.py (580:588) duplicated block id: 5318 size: 9 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/chatglm/checkpoint_reshaping_and_interoperability.py (465:480) - toolkits/model_checkpoints_convertor/starcoder/checkpoint_reshaping_and_interoperability.py (714:728) duplicated block id: 5319 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/language_model.py (241:255) - megatron_patch/model/falcon40b/language_model.py (227:242) duplicated block id: 5320 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/bloom/language_model.py (213:226) - megatron_patch/model/qwen1_5_megablocks/language_model.py (198:211) duplicated block id: 5321 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/falcon/language_model.py (154:166) - megatron_patch/model/starcoder/language_model.py (143:155) duplicated block id: 5322 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/falcon/language_model.py (32:42) - megatron_patch/model/qwen/language_model.py (31:41) duplicated block id: 5323 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/language_model.py (287:301) - megatron_patch/model/falcon/language_model.py (227:242) duplicated block id: 5324 size: 9 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/hf2te.py (384:398) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (564:578) duplicated block id: 5325 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (1585:1593) - megatron_patch/model/qwen_vl/transformer.py (1714:1722) duplicated block id: 5326 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/language_model.py (212:225) - megatron_patch/model/qwen/language_model.py (199:212) duplicated block id: 5327 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/falcon/language_model.py (680:688) - megatron_patch/model/llava/language_model.py (695:703) duplicated block id: 5328 size: 9 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/deepseek/hf2mcore_deepseek_v3_moe.py (137:147) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2.5_vl.py (76:86) duplicated block id: 5329 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/transformer.py (185:196) - megatron_patch/model/starcoder/transformer.py (203:214) duplicated block id: 5330 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/language_model.py (212:225) - megatron_patch/model/glm130b/language_model.py (212:225) duplicated block id: 5331 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/language_model.py (712:720) - megatron_patch/model/falcon40b/language_model.py (680:688) duplicated block id: 5332 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/language_model.py (65:73) - megatron_patch/model/galactica/language_model.py (65:73) duplicated block id: 5333 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/language_model.py (227:242) - megatron_patch/model/llama3/language_model.py (221:235) duplicated block id: 5334 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (1076:1089) - megatron_patch/model/glm130b/transformer.py (807:819) duplicated block id: 5335 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (1585:1593) - megatron_patch/model/llava/transformer.py (1714:1722) duplicated block id: 5336 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/language_model.py (241:255) - megatron_patch/model/galactica/language_model.py (258:273) duplicated block id: 5337 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/llama3_1/layer_specs.py (56:66) - megatron_patch/model/qwen3_moe/gpt_layer_specs.py (157:167) duplicated block id: 5338 size: 9 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore.py (481:489) - toolkits/model_checkpoints_convertor/mistral/hf2mcore.py (388:396) duplicated block id: 5339 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/llama/language_model.py (227:242) - megatron_patch/model/llava/language_model.py (234:248) duplicated block id: 5340 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/galactica/transformer.py (536:550) - megatron_patch/model/starcoder/transformer.py (808:821) duplicated block id: 5341 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/galactica/gpt_model.py (72:80) - megatron_patch/model/glm130b/gpt_model.py (67:75) duplicated block id: 5342 size: 9 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore_llama3_1.py (634:642) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_moe.py (651:659) duplicated block id: 5343 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/falcon/language_model.py (32:42) - megatron_patch/model/llama3/language_model.py (20:30) duplicated block id: 5344 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/transformer.py (224:242) - megatron_patch/model/llama/transformer.py (266:284) duplicated block id: 5345 size: 9 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (427:441) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (564:578) duplicated block id: 5346 size: 9 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/galactica/checkpoint_reshaping_and_interoperability.py (543:558) - toolkits/model_checkpoints_convertor/starcoder/checkpoint_reshaping_and_interoperability.py (714:728) duplicated block id: 5347 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/language_model.py (287:301) - megatron_patch/model/llama/language_model.py (227:242) duplicated block id: 5348 size: 9 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (632:646) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (905:919) duplicated block id: 5349 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/layer_specs.py (149:158) - megatron_patch/model/qwen3_moe/gpt_layer_specs.py (298:307) duplicated block id: 5350 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/language_model.py (33:43) - megatron_patch/model/qwen/language_model.py (31:41) duplicated block id: 5351 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/mistral/language_model.py (375:388) - megatron_patch/model/qwen/language_model.py (374:387) duplicated block id: 5352 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/falcon/language_model.py (334:342) - megatron_patch/model/mistral/language_model.py (335:343) duplicated block id: 5353 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/bloom/language_model.py (34:44) - megatron_patch/model/starcoder/language_model.py (30:40) duplicated block id: 5354 size: 9 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore_llama3_1.py (477:486) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_dense_and_moe_gqa.py (877:886) duplicated block id: 5355 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/falcon/language_model.py (680:688) - megatron_patch/model/qwen1_5_megablocks/language_model.py (634:642) duplicated block id: 5356 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/language_model.py (349:357) - megatron_patch/model/mistral/language_model.py (335:343) duplicated block id: 5357 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/language_model.py (32:42) - megatron_patch/model/qwen_vl/language_model.py (32:42) duplicated block id: 5358 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/language_model.py (241:255) - megatron_patch/model/chatglm/language_model.py (244:259) duplicated block id: 5359 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/transformer.py (594:609) - megatron_patch/model/mistral/transformer.py (745:760) duplicated block id: 5360 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/llama/language_model.py (227:242) - megatron_patch/model/llama2/language_model.py (231:245) duplicated block id: 5361 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/galactica/language_model.py (32:42) - megatron_patch/model/llama2/language_model.py (31:41) duplicated block id: 5362 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/mistral/transformer.py (737:750) - megatron_patch/model/starcoder/transformer.py (715:728) duplicated block id: 5363 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/falcon/language_model.py (32:42) - megatron_patch/model/mistral/language_model.py (32:42) duplicated block id: 5364 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/falcon/language_model.py (334:342) - megatron_patch/model/llama3/language_model.py (329:337) duplicated block id: 5365 size: 9 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/chatglm/checkpoint_reshaping_and_interoperability.py (264:281) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (152:169) duplicated block id: 5366 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/llama/language_model.py (32:42) - megatron_patch/model/mistral/language_model.py (32:42) duplicated block id: 5367 size: 9 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/chatglm/checkpoint_reshaping_and_interoperability.py (464:478) - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (427:441) duplicated block id: 5368 size: 9 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/hf2te.py (384:398) - toolkits/model_checkpoints_convertor/chatglm/checkpoint_reshaping_and_interoperability.py (464:478) duplicated block id: 5369 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/bloom/language_model.py (213:226) - megatron_patch/model/llava/language_model.py (202:215) duplicated block id: 5370 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/language_model.py (390:398) - megatron_patch/model/glm130b/language_model.py (349:357) duplicated block id: 5371 size: 9 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_mha.py (121:129) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_dense_and_moe_gqa.py (320:328) duplicated block id: 5372 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (1307:1316) - megatron_patch/model/llama3/transformer_legacy.py (1398:1407) duplicated block id: 5373 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (1715:1723) - megatron_patch/model/llama2/transformer.py (1715:1723) duplicated block id: 5374 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/language_model.py (33:43) - megatron_patch/model/falcon40b/language_model.py (32:42) duplicated block id: 5375 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/mixtral_bak/layer_specs.py (110:121) - megatron_patch/model/qwen2_moe/layer_specs.py (232:242) duplicated block id: 5376 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/transformer.py (654:667) - megatron_patch/model/llama2/transformer.py (846:859) duplicated block id: 5377 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/qwen2_vl/attention_vision.py (493:501) - megatron_patch/model/qwen2_vl/attention_vision.py (674:682) duplicated block id: 5378 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/language_model.py (34:44) - megatron_patch/model/qwen1_5_megablocks/language_model.py (30:40) duplicated block id: 5379 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/llama2/language_model.py (469:479) - megatron_patch/model/starcoder/language_model.py (433:443) duplicated block id: 5380 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/falcon/transformer.py (910:921) - megatron_patch/model/starcoder/transformer.py (857:868) duplicated block id: 5381 size: 9 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (152:169) - toolkits/model_checkpoints_convertor/yi/checkpoint_reshaping_and_interoperability.py (294:311) duplicated block id: 5382 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/llava_mcore/llava_model.py (107:117) - megatron_patch/model/qwen2_vl/model.py (78:88) duplicated block id: 5383 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/galactica/language_model.py (365:373) - megatron_patch/model/llama2/language_model.py (334:342) duplicated block id: 5384 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/qwen1_5_megablocks/transformer.py (282:308) - megatron_patch/model/starcoder/transformer.py (295:322) duplicated block id: 5385 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/language_model.py (244:259) - megatron_patch/model/llama3/language_model.py (221:235) duplicated block id: 5386 size: 9 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/hf2te.py (356:370) - toolkits/model_checkpoints_convertor/falcon40b/checkpoint_reshaping_and_interoperability.py (484:498) duplicated block id: 5387 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/language_model.py (32:42) - megatron_patch/model/llava/language_model.py (34:44) duplicated block id: 5388 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/llama2/transformer.py (769:778) - megatron_patch/model/llava/transformer.py (833:842) duplicated block id: 5389 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/qwen1_5_megablocks/language_model.py (230:244) - megatron_patch/model/starcoder/language_model.py (232:246) duplicated block id: 5390 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/transformer.py (599:607) - megatron_patch/model/llama/transformer.py (559:568) duplicated block id: 5391 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/falcon/transformer.py (815:824) - megatron_patch/model/falcon40b/transformer.py (572:581) duplicated block id: 5392 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/llama/language_model.py (334:342) - megatron_patch/model/qwen/language_model.py (334:342) duplicated block id: 5393 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (738:753) - megatron_patch/model/glm130b/transformer.py (594:609) duplicated block id: 5394 size: 9 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (160:171) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (188:199) duplicated block id: 5395 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/transformer_layer.py (195:206) - megatron_patch/model/qwen2/transformer_layer.py (149:160) duplicated block id: 5396 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/llama/transformer.py (559:568) - megatron_patch/model/llama/transformer.py (587:596) duplicated block id: 5397 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/qwen1_5/transformer/attention.py (455:463) - megatron_patch/model/qwen2/transformer/attention.py (438:446) duplicated block id: 5398 size: 9 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/deepseek/hf2mcore_deepseek_v2_moe.py (60:70) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_mha_to_moe.py (53:63) duplicated block id: 5399 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer_legacy.py (717:730) - megatron_patch/model/starcoder/transformer.py (715:728) duplicated block id: 5400 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/gpt_model.py (27:40) - megatron_patch/model/glm130b/gpt_model.py (28:41) duplicated block id: 5401 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/language_model.py (250:263) - megatron_patch/model/chatglm/language_model.py (212:225) duplicated block id: 5402 size: 9 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/checkpoint_reshaping_and_interoperability.py (714:728) - toolkits/model_checkpoints_convertor/falcon/checkpoint_reshaping_and_interoperability.py (496:511) duplicated block id: 5403 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/galactica/language_model.py (365:373) - megatron_patch/model/mistral/language_model.py (335:343) duplicated block id: 5404 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/language_model.py (680:688) - megatron_patch/model/qwen/language_model.py (617:625) duplicated block id: 5405 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/language_model.py (33:43) - megatron_patch/model/mistral/language_model.py (32:42) duplicated block id: 5406 size: 9 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_dense_and_moe_gqa.py (181:189) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_dense_and_moe_gqa.py (320:328) duplicated block id: 5407 size: 9 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (628:642) - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (883:897) duplicated block id: 5408 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/language_model.py (680:688) - megatron_patch/model/mistral/language_model.py (647:655) duplicated block id: 5409 size: 9 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon/checkpoint_reshaping_and_interoperability.py (704:718) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (819:834) duplicated block id: 5410 size: 9 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/glm130b/checkpoint_reshaping_and_interoperability.py (422:436) - toolkits/model_checkpoints_convertor/starcoder/checkpoint_reshaping_and_interoperability.py (714:728) duplicated block id: 5411 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/language_model.py (250:263) - megatron_patch/model/bloom/language_model.py (213:226) duplicated block id: 5412 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer_legacy.py (390:416) - megatron_patch/model/starcoder/transformer.py (295:322) duplicated block id: 5413 size: 9 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/hf2te.py (356:370) - toolkits/model_checkpoints_convertor/galactica/checkpoint_reshaping_and_interoperability.py (542:556) duplicated block id: 5414 size: 9 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (726:742) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (811:826) duplicated block id: 5415 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/language_model.py (227:242) - megatron_patch/model/llava/language_model.py (234:248) duplicated block id: 5416 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/falcon/language_model.py (129:152) - megatron_patch/model/starcoder/language_model.py (120:142) duplicated block id: 5417 size: 9 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_dense_and_moe_gqa.py (136:144) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_dense_and_moe_gqa.py (320:328) duplicated block id: 5418 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/qwen2/transformer/attention.py (457:465) - megatron_patch/model/qwen2_vl/attention_vision.py (579:587) duplicated block id: 5419 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer/attention.py (434:442) - megatron_patch/model/qwen2/transformer/attention.py (457:465) duplicated block id: 5420 size: 9 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_dense_and_moe_gqa.py (320:328) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_vl.py (190:198) duplicated block id: 5421 size: 9 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (294:311) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (152:169) duplicated block id: 5422 size: 9 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/mistral/hf2mcore.py (207:215) - toolkits/model_checkpoints_convertor/qwen/hf2megablocks_qwen1.5.py (263:271) duplicated block id: 5423 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/mlp.py (72:80) - megatron_patch/model/mixtral_bak/transformer/mlp.py (72:80) duplicated block id: 5424 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/multi_latent_attention.py (199:207) - megatron_patch/model/qwen2_vl/attention_vision.py (674:682) duplicated block id: 5425 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer_legacy.py (1675:1683) - megatron_patch/model/qwen_vl/transformer.py (1714:1722) duplicated block id: 5426 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/llava/transformer.py (1714:1722) - megatron_patch/model/qwen/transformer.py (1657:1665) duplicated block id: 5427 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/falcon/transformer.py (258:275) - megatron_patch/model/mistral/transformer.py (391:408) duplicated block id: 5428 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/llama/language_model.py (227:242) - megatron_patch/model/mistral/language_model.py (232:246) duplicated block id: 5429 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/transformer/attention.py (579:587) - megatron_patch/model/qwen2_vl/attention_vision.py (598:606) duplicated block id: 5430 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/qwen1_5_megablocks/language_model.py (468:478) - megatron_patch/model/starcoder/language_model.py (433:443) duplicated block id: 5431 size: 9 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon40b/checkpoint_reshaping_and_interoperability.py (484:498) - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (496:510) duplicated block id: 5432 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (1585:1593) - megatron_patch/model/mistral/transformer.py (1712:1720) duplicated block id: 5433 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/layer_specs.py (30:44) - megatron_patch/model/llava_mcore/layer_specs.py (20:34) duplicated block id: 5434 size: 9 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (819:834) - toolkits/model_checkpoints_convertor/starcoder/checkpoint_reshaping_and_interoperability.py (408:422) duplicated block id: 5435 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/qwen2_vl/attention.py (494:502) - megatron_patch/model/qwen2_vl/attention.py (675:683) duplicated block id: 5436 size: 9 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (139:149) - toolkits/model_checkpoints_convertor/yi/checkpoint_reshaping_and_interoperability.py (159:169) duplicated block id: 5437 size: 9 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/deepseek/hf2mcore_deepseek_v3_moe.py (137:147) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_vl.py (76:86) duplicated block id: 5438 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/llava/language_model.py (571:579) - megatron_patch/model/mistral/language_model.py (523:531) duplicated block id: 5439 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/qwen1_5_megablocks/transformer.py (660:669) - megatron_patch/model/qwen1_5_megablocks/transformer.py (725:734) duplicated block id: 5440 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/transformer.py (594:609) - megatron_patch/model/qwen_vl/transformer.py (746:761) duplicated block id: 5441 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/qwen_vl/transformer.py (768:777) - megatron_patch/model/qwen_vl/transformer.py (833:842) duplicated block id: 5442 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer/attention.py (455:463) - megatron_patch/model/qwen1_5/transformer/attention.py (434:442) duplicated block id: 5443 size: 9 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/deepseek/hf2mcore_deepseek_v2_moe.py (60:70) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_mha.py (54:64) duplicated block id: 5444 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (970:981) - megatron_patch/model/falcon/transformer.py (1079:1090) duplicated block id: 5445 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/language_model.py (32:42) - megatron_patch/model/llama3/language_model.py (20:30) duplicated block id: 5446 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/mixtral_bak/transformer/attention.py (294:302) - megatron_patch/model/qwen2_vl/attention_vision.py (453:461) duplicated block id: 5447 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (970:981) - megatron_patch/model/llama/transformer.py (873:884) duplicated block id: 5448 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/llama3_1/layer_specs.py (89:99) - megatron_patch/model/mixtral/layer_specs.py (132:142) duplicated block id: 5449 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/transformer/attention.py (579:587) - megatron_patch/model/qwen2/transformer/attention.py (457:465) duplicated block id: 5450 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/qwen2_vl/attention.py (599:607) - megatron_patch/model/qwen2_vl/attention_vision.py (579:587) duplicated block id: 5451 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/llama/language_model.py (334:342) - megatron_patch/model/mistral/language_model.py (335:343) duplicated block id: 5452 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/llava/transformer.py (833:842) - megatron_patch/model/qwen_vl/transformer.py (768:777) duplicated block id: 5453 size: 9 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/mistral/hf2mcore.py (272:280) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_dense_and_moe_gqa.py (320:328) duplicated block id: 5454 size: 9 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_dense_and_moe_gqa.py (242:250) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_dense_and_moe_gqa.py (320:328) duplicated block id: 5455 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer_legacy.py (1675:1683) - megatron_patch/model/qwen1_5_megablocks/transformer.py (1577:1585) duplicated block id: 5456 size: 9 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon40b/checkpoint_reshaping_and_interoperability.py (459:469) - toolkits/model_checkpoints_convertor/yi/checkpoint_reshaping_and_interoperability.py (484:494) duplicated block id: 5457 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/language_model.py (243:258) - megatron_patch/model/starcoder/language_model.py (232:246) duplicated block id: 5458 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/llava/transformer.py (833:842) - megatron_patch/model/qwen1_5_megablocks/transformer.py (660:669) duplicated block id: 5459 size: 9 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_mha_to_moe.py (53:63) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_moe.py (122:132) duplicated block id: 5460 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/llama3/language_model.py (221:235) - megatron_patch/model/starcoder/language_model.py (232:246) duplicated block id: 5461 size: 9 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_moe.py (378:388) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_vl.py (220:229) duplicated block id: 5462 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/language_model.py (32:42) - megatron_patch/model/qwen1_5_megablocks/language_model.py (30:40) duplicated block id: 5463 size: 9 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/glm/checkpoint_reshaping_and_interoperability.py (441:455) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (565:580) duplicated block id: 5464 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (1307:1316) - megatron_patch/model/llava/transformer.py (1439:1448) duplicated block id: 5465 size: 9 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/checkpoint_reshaping_and_interoperability.py (714:728) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (565:580) duplicated block id: 5466 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/language_model.py (212:225) - megatron_patch/model/qwen1_5_megablocks/language_model.py (198:211) duplicated block id: 5467 size: 9 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (811:826) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (862:878) duplicated block id: 5468 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/language_model.py (34:44) - megatron_patch/model/qwen_vl/language_model.py (32:42) duplicated block id: 5469 size: 9 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2megablocks_qwen1.5.py (529:537) - toolkits/model_checkpoints_convertor/qwen/hf2megablocks_qwen1.5.py (564:572) duplicated block id: 5470 size: 9 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/galactica/checkpoint_reshaping_and_interoperability.py (542:556) - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (496:510) duplicated block id: 5471 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/language_model.py (76:86) - megatron_patch/model/galactica/language_model.py (32:42) duplicated block id: 5472 size: 9 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/deepseek/hf2mcore_deepseek_v2_moe.py (343:354) - toolkits/model_checkpoints_convertor/deepseek/hf2mcore_deepseek_v3_moe.py (527:538) duplicated block id: 5473 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/language_model.py (334:342) - megatron_patch/model/llama2/language_model.py (334:342) duplicated block id: 5474 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/llama3_1/model.py (278:291) - megatron_patch/model/qwen2/model.py (216:229) duplicated block id: 5475 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/falcon/language_model.py (32:42) - megatron_patch/model/qwen1_5_megablocks/language_model.py (30:40) duplicated block id: 5476 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/llava/transformer.py (768:777) - megatron_patch/model/qwen_vl/transformer.py (833:842) duplicated block id: 5477 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/falcon/transformer.py (507:519) - megatron_patch/model/mistral/transformer.py (478:490) duplicated block id: 5478 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (1016:1025) - megatron_patch/model/starcoder/transformer.py (1174:1182) duplicated block id: 5479 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/language_model.py (76:86) - megatron_patch/model/falcon/language_model.py (32:42) duplicated block id: 5480 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/language_model.py (262:274) - megatron_patch/model/mistral/language_model.py (249:261) duplicated block id: 5481 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (1715:1723) - megatron_patch/model/qwen_vl/transformer.py (1714:1722) duplicated block id: 5482 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/llama/language_model.py (154:166) - megatron_patch/model/starcoder/language_model.py (143:155) duplicated block id: 5483 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/falcon/gpt_model.py (72:80) - megatron_patch/model/glm130b/gpt_model.py (67:75) duplicated block id: 5484 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/transformer/attention.py (454:462) - megatron_patch/model/qwen2_vl/attention_vision.py (453:461) duplicated block id: 5485 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/language_model.py (212:225) - megatron_patch/model/qwen_vl/language_model.py (200:213) duplicated block id: 5486 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer/attention.py (434:442) - megatron_patch/model/qwen1_5/transformer/attention.py (455:463) duplicated block id: 5487 size: 9 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (742:756) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (632:646) duplicated block id: 5488 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/language_model.py (33:43) - megatron_patch/model/qwen1_5_megablocks/language_model.py (30:40) duplicated block id: 5489 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/bloom/gpt_model.py (67:75) - megatron_patch/model/galactica/gpt_model.py (72:80) duplicated block id: 5490 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/transformer.py (172:183) - megatron_patch/model/glm130b/transformer.py (184:196) duplicated block id: 5491 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/falcon/language_model.py (227:242) - megatron_patch/model/llama3/language_model.py (221:235) duplicated block id: 5492 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/llama/language_model.py (334:342) - megatron_patch/model/llama3/language_model.py (329:337) duplicated block id: 5493 size: 9 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_moe.py (116:124) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_dense_and_moe_gqa.py (197:205) duplicated block id: 5494 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/transformer.py (594:609) - megatron_patch/model/qwen1_5_megablocks/transformer.py (638:653) duplicated block id: 5495 size: 9 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (496:510) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (564:578) duplicated block id: 5496 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/gpt_model.py (27:40) - megatron_patch/model/bloom/gpt_model.py (28:41) duplicated block id: 5497 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/bloom/language_model.py (213:226) - megatron_patch/model/glm130b/language_model.py (212:225) duplicated block id: 5498 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/falcon/transformer.py (428:449) - megatron_patch/model/llava/transformer.py (398:418) duplicated block id: 5499 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer_legacy.py (1675:1683) - megatron_patch/model/mistral/transformer.py (1712:1720) duplicated block id: 5500 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (643:654) - megatron_patch/model/falcon/transformer.py (910:921) duplicated block id: 5501 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/language_model.py (287:301) - megatron_patch/model/chatglm/language_model.py (244:259) duplicated block id: 5502 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/qwen2/transformer/attention.py (438:446) - megatron_patch/model/qwen2_vl/attention.py (599:607) duplicated block id: 5503 size: 9 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_mha.py (228:238) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_dense_and_moe_gqa.py (502:512) duplicated block id: 5504 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/multi_latent_attention.py (199:207) - megatron_patch/model/qwen2_vl/attention.py (675:683) duplicated block id: 5505 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/gpt_model.py (72:80) - megatron_patch/model/glm130b/gpt_model.py (67:75) duplicated block id: 5506 size: 9 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/deepseek/hf2mcore_deepseek_v2_moe.py (60:70) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_moe.py (56:66) duplicated block id: 5507 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/qwen2_vl/attention.py (494:502) - megatron_patch/model/qwen2_vl/attention_vision.py (674:682) duplicated block id: 5508 size: 9 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_moe.py (126:134) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_dense_and_moe_gqa.py (207:215) duplicated block id: 5509 size: 9 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (840:856) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (811:826) duplicated block id: 5510 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/language_model.py (34:44) - megatron_patch/model/llava/language_model.py (34:44) duplicated block id: 5511 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (1585:1593) - megatron_patch/model/llama2/transformer.py (1715:1723) duplicated block id: 5512 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/multi_latent_attention.py (199:207) - megatron_patch/model/mixtral/transformer/attention.py (674:682) duplicated block id: 5513 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/falcon/transformer.py (428:449) - megatron_patch/model/mistral/transformer.py (398:418) duplicated block id: 5514 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/language_model.py (32:42) - megatron_patch/model/mistral/language_model.py (32:42) duplicated block id: 5515 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/galactica/language_model.py (32:42) - megatron_patch/model/qwen1_5_megablocks/language_model.py (30:40) duplicated block id: 5516 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/gpt_model.py (27:40) - megatron_patch/model/chatglm/gpt_model.py (28:41) duplicated block id: 5517 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/language_model.py (241:255) - megatron_patch/model/falcon/language_model.py (227:242) duplicated block id: 5518 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/language_model.py (33:43) - megatron_patch/model/glm130b/language_model.py (33:43) duplicated block id: 5519 size: 9 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/deepseek/hf2mcore_deepseek_v3_moe.py (113:123) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_mha.py (54:64) duplicated block id: 5520 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/qwen1_5/layer_specs.py (55:66) - megatron_patch/model/qwen2_moe/layer_specs.py (171:181) duplicated block id: 5521 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/language_model.py (129:152) - megatron_patch/model/starcoder/language_model.py (120:142) duplicated block id: 5522 size: 9 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon/checkpoint_reshaping_and_interoperability.py (495:509) - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (427:441) duplicated block id: 5523 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/bloom/language_model.py (213:226) - megatron_patch/model/qwen/language_model.py (199:212) duplicated block id: 5524 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/falcon/language_model.py (32:42) - megatron_patch/model/starcoder/language_model.py (30:40) duplicated block id: 5525 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/language_model.py (680:688) - megatron_patch/model/llava/language_model.py (695:703) duplicated block id: 5526 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/bloom/language_model.py (34:44) - megatron_patch/model/qwen1_5_megablocks/language_model.py (30:40) duplicated block id: 5527 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/language_model.py (65:73) - megatron_patch/model/falcon40b/language_model.py (65:73) duplicated block id: 5528 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/language_model.py (227:242) - megatron_patch/model/mistral/language_model.py (232:246) duplicated block id: 5529 size: 9 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/deepseek/hf2mcore_deepseek_v3_moe.py (113:123) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_mha_to_moe.py (53:63) duplicated block id: 5530 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/language_model.py (349:357) - megatron_patch/model/qwen/language_model.py (334:342) duplicated block id: 5531 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/bloom/language_model.py (169:179) - megatron_patch/model/galactica/language_model.py (196:206) duplicated block id: 5532 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/language_model.py (287:301) - megatron_patch/model/starcoder/language_model.py (232:246) duplicated block id: 5533 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (518:530) - megatron_patch/model/chatglm/transformer.py (437:450) duplicated block id: 5534 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/llama3/layer_specs.py (89:99) - megatron_patch/model/mixtral/layer_specs.py (132:142) duplicated block id: 5535 size: 9 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (628:642) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (905:919) duplicated block id: 5536 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/transformer.py (274:289) - megatron_patch/model/starcoder/transformer.py (489:503) duplicated block id: 5537 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (643:654) - megatron_patch/model/galactica/transformer.py (572:583) duplicated block id: 5538 size: 9 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon/checkpoint_reshaping_and_interoperability.py (496:511) - toolkits/model_checkpoints_convertor/glm/checkpoint_reshaping_and_interoperability.py (441:455) duplicated block id: 5539 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/language_model.py (628:636) - megatron_patch/model/falcon40b/language_model.py (680:688) duplicated block id: 5540 size: 9 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon40b/checkpoint_reshaping_and_interoperability.py (484:498) - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (427:441) duplicated block id: 5541 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/transformer.py (594:609) - megatron_patch/model/llava/transformer.py (746:761) duplicated block id: 5542 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/qwen_vl/language_model.py (232:246) - megatron_patch/model/starcoder/language_model.py (232:246) duplicated block id: 5543 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/bloom/language_model.py (213:226) - megatron_patch/model/llama3/language_model.py (189:202) duplicated block id: 5544 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/qwen1_5/transformer/attention.py (434:442) - megatron_patch/model/qwen2/transformer/attention.py (457:465) duplicated block id: 5545 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/moe/token_dispatcher.py (88:101) - megatron_patch/model/qwen1_5/moe/token_dispatcher.py (75:88) duplicated block id: 5546 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer/attention.py (455:463) - megatron_patch/model/qwen2_vl/attention_vision.py (579:587) duplicated block id: 5547 size: 9 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (173:190) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (364:381) duplicated block id: 5548 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/qwen_vl/transformer.py (738:751) - megatron_patch/model/starcoder/transformer.py (715:728) duplicated block id: 5549 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/galactica/language_model.py (32:42) - megatron_patch/model/llama3/language_model.py (20:30) duplicated block id: 5550 size: 9 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/chatglm/checkpoint_reshaping_and_interoperability.py (464:478) - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (496:510) duplicated block id: 5551 size: 9 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_moe.py (56:66) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_moe.py (122:132) duplicated block id: 5552 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/galactica/language_model.py (258:273) - megatron_patch/model/llama2/language_model.py (231:245) duplicated block id: 5553 size: 9 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_dense_and_moe_gqa.py (281:289) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_moe.py (308:316) duplicated block id: 5554 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (94:112) - megatron_patch/model/starcoder/transformer.py (85:105) duplicated block id: 5555 size: 9 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/reward_model_to_megatron.py (713:727) - toolkits/model_checkpoints_convertor/falcon/checkpoint_reshaping_and_interoperability.py (496:511) duplicated block id: 5556 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/bloom/language_model.py (245:260) - megatron_patch/model/qwen1_5_megablocks/language_model.py (230:244) duplicated block id: 5557 size: 9 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_dense_and_moe_gqa.py (281:289) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_moe.py (218:226) duplicated block id: 5558 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (970:981) - megatron_patch/model/falcon40b/transformer.py (818:829) duplicated block id: 5559 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/llama/language_model.py (129:152) - megatron_patch/model/starcoder/language_model.py (120:142) duplicated block id: 5560 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/qwen1_5_megablocks/transformer.py (660:669) - megatron_patch/model/qwen_vl/transformer.py (833:842) duplicated block id: 5561 size: 9 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/reward_model_to_megatron.py (713:727) - toolkits/model_checkpoints_convertor/chatglm/checkpoint_reshaping_and_interoperability.py (465:480) duplicated block id: 5562 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/galactica/language_model.py (32:42) - megatron_patch/model/qwen/language_model.py (31:41) duplicated block id: 5563 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer/attention.py (455:463) - megatron_patch/model/qwen2_vl/attention.py (580:588) duplicated block id: 5564 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/language_model.py (244:259) - megatron_patch/model/qwen_vl/language_model.py (232:246) duplicated block id: 5565 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/transformer/attention.py (674:682) - megatron_patch/model/qwen2_vl/attention.py (494:502) duplicated block id: 5566 size: 9 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (769:783) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (632:646) duplicated block id: 5567 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/language_model.py (154:166) - megatron_patch/model/starcoder/language_model.py (143:155) duplicated block id: 5568 size: 9 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon40b/checkpoint_reshaping_and_interoperability.py (485:500) - toolkits/model_checkpoints_convertor/glm/checkpoint_reshaping_and_interoperability.py (441:455) duplicated block id: 5569 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/galactica/transformer.py (122:133) - megatron_patch/model/glm130b/transformer.py (184:196) duplicated block id: 5570 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (1715:1723) - megatron_patch/model/qwen1_5_megablocks/transformer.py (1577:1585) duplicated block id: 5571 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/llama2/transformer.py (739:752) - megatron_patch/model/starcoder/transformer.py (715:728) duplicated block id: 5572 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/language_model.py (209:222) - megatron_patch/model/chatglm/language_model.py (212:225) duplicated block id: 5573 size: 9 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (699:715) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (811:826) duplicated block id: 5574 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/transformer/attention.py (493:501) - megatron_patch/model/qwen2_vl/attention.py (675:683) duplicated block id: 5575 size: 9 cleaned lines of code in 2 files: - megatron_patch/data/utils.py (245:254) - megatron_patch/data/utils.py (427:436) duplicated block id: 5576 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/falcon/transformer.py (507:519) - megatron_patch/model/llama2/transformer.py (477:489) duplicated block id: 5577 size: 9 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/glm/checkpoint_reshaping_and_interoperability.py (441:455) - toolkits/model_checkpoints_convertor/starcoder/checkpoint_reshaping_and_interoperability.py (714:728) duplicated block id: 5578 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/language_model.py (334:342) - megatron_patch/model/mistral/language_model.py (335:343) duplicated block id: 5579 size: 9 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_mha.py (259:267) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_mha_to_moe.py (201:209) duplicated block id: 5580 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/qwen/transformer.py (1657:1665) - megatron_patch/model/qwen1_5_megablocks/transformer.py (1577:1585) duplicated block id: 5581 size: 9 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/hf2te.py (356:370) - toolkits/model_checkpoints_convertor/falcon/checkpoint_reshaping_and_interoperability.py (495:509) duplicated block id: 5582 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (1715:1723) - megatron_patch/model/mistral/transformer.py (1712:1720) duplicated block id: 5583 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/transformer.py (224:242) - megatron_patch/model/galactica/transformer.py (228:246) duplicated block id: 5584 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (730:743) - megatron_patch/model/starcoder/transformer.py (715:728) duplicated block id: 5585 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/transformer.py (639:653) - megatron_patch/model/galactica/transformer.py (536:550) duplicated block id: 5586 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (1307:1316) - megatron_patch/model/qwen/transformer.py (1382:1391) duplicated block id: 5587 size: 9 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (152:169) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (364:381) duplicated block id: 5588 size: 9 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_moe.py (158:166) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_dense_and_moe_gqa.py (320:328) duplicated block id: 5589 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/language_model.py (227:242) - megatron_patch/model/qwen/language_model.py (231:245) duplicated block id: 5590 size: 9 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (1011:1019) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (1041:1049) duplicated block id: 5591 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/layer_specs.py (33:42) - megatron_patch/model/qwen2/layer_specs.py (22:31) duplicated block id: 5592 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/llama/transformer.py (239:256) - megatron_patch/model/mistral/transformer.py (391:408) duplicated block id: 5593 size: 9 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_mha_to_moe.py (53:63) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_dense_and_moe_gqa.py (48:58) duplicated block id: 5594 size: 9 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon/checkpoint_reshaping_and_interoperability.py (495:509) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (414:428) duplicated block id: 5595 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/transformer.py (329:344) - megatron_patch/model/starcoder/transformer.py (489:503) duplicated block id: 5596 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/llama3_1/layer_specs.py (56:66) - megatron_patch/model/qwen2_moe/layer_specs.py (171:181) duplicated block id: 5597 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/language_model.py (212:225) - megatron_patch/model/mistral/language_model.py (200:213) duplicated block id: 5598 size: 9 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2.5_vl.py (141:149) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_dense_and_moe_gqa.py (320:328) duplicated block id: 5599 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/qwen1_5/layer_specs.py (105:114) - megatron_patch/model/qwen2_vl/layer_specs.py (106:114) duplicated block id: 5600 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/galactica/transformer.py (279:294) - megatron_patch/model/starcoder/transformer.py (489:503) duplicated block id: 5601 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/transformer/attention.py (493:501) - megatron_patch/model/qwen2_vl/attention_vision.py (674:682) duplicated block id: 5602 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/qwen2/transformer_block.py (206:215) - megatron_patch/model/qwen2_5_vl/transformer_block.py (298:307) duplicated block id: 5603 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/transformer.py (654:667) - megatron_patch/model/mistral/transformer.py (843:856) duplicated block id: 5604 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/language_model.py (155:167) - megatron_patch/model/starcoder/language_model.py (143:155) duplicated block id: 5605 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/transformer.py (654:667) - megatron_patch/model/qwen/transformer.py (791:804) duplicated block id: 5606 size: 9 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (883:897) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (632:646) duplicated block id: 5607 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/language_model.py (287:301) - megatron_patch/model/galactica/language_model.py (258:273) duplicated block id: 5608 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/galactica/language_model.py (32:42) - megatron_patch/model/starcoder/language_model.py (30:40) duplicated block id: 5609 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/llama2/transformer.py (1715:1723) - megatron_patch/model/qwen/transformer.py (1657:1665) duplicated block id: 5610 size: 9 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/deepseek/hf2mcore_deepseek_v2_moe.py (432:442) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_moe.py (599:608) duplicated block id: 5611 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/galactica/transformer.py (135:146) - megatron_patch/model/starcoder/transformer.py (203:214) duplicated block id: 5612 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/language_model.py (258:270) - megatron_patch/model/chatglm/language_model.py (262:274) duplicated block id: 5613 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/galactica/language_model.py (365:373) - megatron_patch/model/llama3/language_model.py (329:337) duplicated block id: 5614 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/llama/language_model.py (32:42) - megatron_patch/model/llama2/language_model.py (31:41) duplicated block id: 5615 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/falcon/language_model.py (334:342) - megatron_patch/model/qwen/language_model.py (334:342) duplicated block id: 5616 size: 9 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_dense_and_moe_gqa.py (320:328) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_vl.py (141:149) duplicated block id: 5617 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/mistral/transformer.py (1712:1720) - megatron_patch/model/qwen/transformer.py (1657:1665) duplicated block id: 5618 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/falcon/language_model.py (334:342) - megatron_patch/model/llama2/language_model.py (334:342) duplicated block id: 5619 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/llama/language_model.py (32:42) - megatron_patch/model/qwen1_5_megablocks/language_model.py (30:40) duplicated block id: 5620 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/language_model.py (156:168) - megatron_patch/model/starcoder/language_model.py (143:155) duplicated block id: 5621 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/galactica/language_model.py (258:273) - megatron_patch/model/qwen1_5_megablocks/language_model.py (230:244) duplicated block id: 5622 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/language_model.py (262:274) - megatron_patch/model/qwen1_5_megablocks/language_model.py (247:259) duplicated block id: 5623 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/qwen1_5/transformer/attention.py (455:463) - megatron_patch/model/qwen2_vl/attention.py (580:588) duplicated block id: 5624 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/language_model.py (212:225) - megatron_patch/model/starcoder/language_model.py (198:211) duplicated block id: 5625 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (1307:1316) - megatron_patch/model/baichuan2/transformer.py (1440:1449) duplicated block id: 5626 size: 9 cleaned lines of code in 2 files: - toolkits/pretrain_data_preprocessing/preprocess_data_megatron.py (401:413) - toolkits/sft_data_preprocessing/build_idxmap_sft_dataset.py (375:387) duplicated block id: 5627 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/falcon/transformer.py (428:449) - megatron_patch/model/qwen_vl/transformer.py (398:418) duplicated block id: 5628 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/language_model.py (34:44) - megatron_patch/model/llama2/language_model.py (31:41) duplicated block id: 5629 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/transformer.py (654:667) - megatron_patch/model/qwen1_5_megablocks/transformer.py (737:750) duplicated block id: 5630 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/language_model.py (287:301) - megatron_patch/model/falcon40b/language_model.py (227:242) duplicated block id: 5631 size: 9 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (307:324) - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (173:190) duplicated block id: 5632 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/language_model.py (76:86) - megatron_patch/model/falcon40b/language_model.py (32:42) duplicated block id: 5633 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/starcoder/transformer.py (715:728) - megatron_patch/model/starcoder/transformer.py (738:751) duplicated block id: 5634 size: 9 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/galactica/checkpoint_reshaping_and_interoperability.py (542:556) - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (427:441) duplicated block id: 5635 size: 9 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/hf2te.py (198:211) - toolkits/model_checkpoints_convertor/galactica/checkpoint_reshaping_and_interoperability.py (343:356) duplicated block id: 5636 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (698:711) - megatron_patch/model/glm130b/transformer.py (654:667) duplicated block id: 5637 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/language_model.py (241:255) - megatron_patch/model/llama/language_model.py (227:242) duplicated block id: 5638 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (619:632) - megatron_patch/model/starcoder/transformer.py (715:728) duplicated block id: 5639 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/language_model.py (227:242) - megatron_patch/model/llama2/language_model.py (231:245) duplicated block id: 5640 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/language_model.py (65:73) - megatron_patch/model/glm130b/language_model.py (66:74) duplicated block id: 5641 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (99:115) - megatron_patch/model/chatglm/transformer.py (62:78) duplicated block id: 5642 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/transformer.py (251:268) - megatron_patch/model/llava/transformer.py (391:408) duplicated block id: 5643 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/falcon/transformer.py (764:777) - megatron_patch/model/starcoder/transformer.py (715:728) duplicated block id: 5644 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/bloom/gpt_model.py (67:75) - megatron_patch/model/llama/gpt_model.py (71:79) duplicated block id: 5645 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/language_model.py (680:688) - megatron_patch/model/qwen_vl/language_model.py (664:672) duplicated block id: 5646 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (1307:1316) - megatron_patch/model/mistral/transformer.py (1437:1446) duplicated block id: 5647 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/qwen1_5/layer_specs.py (103:112) - megatron_patch/model/qwen3_moe/gpt_layer_specs.py (298:307) duplicated block id: 5648 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/llama3/model.py (215:228) - megatron_patch/model/llama3_1/model.py (278:291) duplicated block id: 5649 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/language_model.py (65:73) - megatron_patch/model/falcon/language_model.py (65:73) duplicated block id: 5650 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/language_model.py (33:43) - megatron_patch/model/bloom/language_model.py (34:44) duplicated block id: 5651 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/bloom/gpt_model.py (67:75) - megatron_patch/model/falcon/gpt_model.py (72:80) duplicated block id: 5652 size: 9 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (183:191) - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (201:209) duplicated block id: 5653 size: 9 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_moe.py (492:504) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_dense_and_moe_gqa.py (837:851) duplicated block id: 5654 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/falcon/language_model.py (32:42) - megatron_patch/model/llava/language_model.py (34:44) duplicated block id: 5655 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/qwen1_5_megablocks/transformer.py (630:643) - megatron_patch/model/starcoder/transformer.py (715:728) duplicated block id: 5656 size: 9 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/chatglm/checkpoint_reshaping_and_interoperability.py (464:478) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (414:428) duplicated block id: 5657 size: 9 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/chatglm/checkpoint_reshaping_and_interoperability.py (298:316) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (403:421) duplicated block id: 5658 size: 9 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_mha.py (259:267) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_moe.py (350:358) duplicated block id: 5659 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/language_model.py (344:352) - megatron_patch/model/llama/language_model.py (334:342) duplicated block id: 5660 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer/attention.py (455:463) - megatron_patch/model/mixtral/transformer/attention.py (579:587) duplicated block id: 5661 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/llama/language_model.py (227:242) - megatron_patch/model/qwen1_5_megablocks/language_model.py (230:244) duplicated block id: 5662 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/language_model.py (390:398) - megatron_patch/model/galactica/language_model.py (365:373) duplicated block id: 5663 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/falcon/transformer.py (258:275) - megatron_patch/model/qwen_vl/transformer.py (391:408) duplicated block id: 5664 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/qwen/transformer.py (1657:1665) - megatron_patch/model/qwen_vl/transformer.py (1714:1722) duplicated block id: 5665 size: 9 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/chatglm/checkpoint_reshaping_and_interoperability.py (465:480) - toolkits/model_checkpoints_convertor/glm130b/checkpoint_reshaping_and_interoperability.py (422:436) duplicated block id: 5666 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/bloom/language_model.py (213:226) - megatron_patch/model/qwen_vl/language_model.py (200:213) duplicated block id: 5667 size: 9 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon/checkpoint_reshaping_and_interoperability.py (495:509) - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (496:510) duplicated block id: 5668 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer/attention.py (321:329) - megatron_patch/model/qwen2_vl/attention.py (454:462) duplicated block id: 5669 size: 9 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_moe.py (124:132) - toolkits/model_checkpoints_convertor/qwen/hf2megablocks_qwen1.5.py (261:269) duplicated block id: 5670 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/llama3/language_model.py (460:470) - megatron_patch/model/starcoder/language_model.py (433:443) duplicated block id: 5671 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/qwen2_vl/attention_vision.py (579:587) - megatron_patch/model/qwen2_vl/attention_vision.py (598:606) duplicated block id: 5672 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/bloom/language_model.py (245:260) - megatron_patch/model/qwen/language_model.py (231:245) duplicated block id: 5673 size: 9 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (173:190) - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (345:362) duplicated block id: 5674 size: 9 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (173:190) - toolkits/model_checkpoints_convertor/yi/checkpoint_reshaping_and_interoperability.py (294:311) duplicated block id: 5675 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/language_model.py (537:547) - megatron_patch/model/starcoder/language_model.py (433:443) duplicated block id: 5676 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/mistral/language_model.py (232:246) - megatron_patch/model/starcoder/language_model.py (232:246) duplicated block id: 5677 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/falcon/language_model.py (680:688) - megatron_patch/model/qwen/language_model.py (617:625) duplicated block id: 5678 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/language_model.py (384:397) - megatron_patch/model/mistral/language_model.py (375:388) duplicated block id: 5679 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/language_model.py (244:259) - megatron_patch/model/qwen/language_model.py (231:245) duplicated block id: 5680 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/llava/language_model.py (234:248) - megatron_patch/model/starcoder/language_model.py (232:246) duplicated block id: 5681 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/transformer.py (594:609) - megatron_patch/model/llama3/transformer_legacy.py (725:740) duplicated block id: 5682 size: 9 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (414:428) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (564:578) duplicated block id: 5683 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/llava/transformer.py (768:777) - megatron_patch/model/llava/transformer.py (833:842) duplicated block id: 5684 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/language_model.py (262:274) - megatron_patch/model/llama2/language_model.py (248:260) duplicated block id: 5685 size: 9 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (742:756) - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (628:642) duplicated block id: 5686 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/qwen1_5/transformer/attention.py (321:329) - megatron_patch/model/qwen2_vl/attention_vision.py (453:461) duplicated block id: 5687 size: 9 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_moe.py (135:143) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_dense_and_moe_gqa.py (217:225) duplicated block id: 5688 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer/attention.py (455:463) - megatron_patch/model/qwen2/transformer/attention.py (438:446) duplicated block id: 5689 size: 9 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (742:756) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (905:919) duplicated block id: 5690 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/qwen1_5/layer_specs.py (89:100) - megatron_patch/model/qwen2_moe/layer_specs.py (232:242) duplicated block id: 5691 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/transformer.py (818:833) - megatron_patch/model/falcon40b/transformer.py (936:951) duplicated block id: 5692 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/transformer.py (594:609) - megatron_patch/model/llama2/transformer.py (747:762) duplicated block id: 5693 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/llama/language_model.py (32:42) - megatron_patch/model/starcoder/language_model.py (30:40) duplicated block id: 5694 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/falcon/language_model.py (680:688) - megatron_patch/model/mistral/language_model.py (647:655) duplicated block id: 5695 size: 9 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/deepseek/hf2mcore_deepseek_v3_moe.py (670:680) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2.5_vl.py (695:705) duplicated block id: 5696 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/language_model.py (131:154) - megatron_patch/model/starcoder/language_model.py (120:142) duplicated block id: 5697 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/bloom/language_model.py (34:44) - megatron_patch/model/qwen_vl/language_model.py (32:42) duplicated block id: 5698 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/moe/token_dispatcher.py (88:101) - megatron_patch/model/qwen2/moe/token_dispatcher.py (76:89) duplicated block id: 5699 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/qwen2_vl/attention.py (675:683) - megatron_patch/model/qwen2_vl/attention_vision.py (493:501) duplicated block id: 5700 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/llava/transformer.py (738:751) - megatron_patch/model/starcoder/transformer.py (715:728) duplicated block id: 5701 size: 9 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/chatglm/checkpoint_reshaping_and_interoperability.py (264:281) - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (173:190) duplicated block id: 5702 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/bloom/language_model.py (34:44) - megatron_patch/model/llama2/language_model.py (31:41) duplicated block id: 5703 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/transformer.py (654:667) - megatron_patch/model/llava/transformer.py (845:858) duplicated block id: 5704 size: 9 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/deepseek/hf2mcore_deepseek_v2_moe.py (502:510) - toolkits/model_checkpoints_convertor/llama/hf2mcore_llama3_1.py (634:642) duplicated block id: 5705 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/bloom/language_model.py (245:260) - megatron_patch/model/qwen_vl/language_model.py (232:246) duplicated block id: 5706 size: 9 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/galactica/checkpoint_reshaping_and_interoperability.py (543:558) - toolkits/model_checkpoints_convertor/glm/checkpoint_reshaping_and_interoperability.py (441:455) duplicated block id: 5707 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/llama3/layer_specs.py (55:66) - megatron_patch/model/qwen3_moe/gpt_layer_specs.py (157:167) duplicated block id: 5708 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (1715:1723) - megatron_patch/model/llava/transformer.py (1714:1722) duplicated block id: 5709 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/llama2/transformer.py (1715:1723) - megatron_patch/model/llama3/transformer_legacy.py (1675:1683) duplicated block id: 5710 size: 9 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon40b/checkpoint_reshaping_and_interoperability.py (485:500) - toolkits/model_checkpoints_convertor/starcoder/checkpoint_reshaping_and_interoperability.py (714:728) duplicated block id: 5711 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/gpt_model.py (28:41) - megatron_patch/model/llama/gpt_model.py (28:41) duplicated block id: 5712 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/llama/language_model.py (227:242) - megatron_patch/model/llama3/language_model.py (221:235) duplicated block id: 5713 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/language_model.py (76:86) - megatron_patch/model/llama/language_model.py (32:42) duplicated block id: 5714 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/falcon/transformer.py (507:519) - megatron_patch/model/qwen_vl/transformer.py (478:490) duplicated block id: 5715 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (294:320) - megatron_patch/model/starcoder/transformer.py (295:322) duplicated block id: 5716 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/galactica/language_model.py (258:273) - megatron_patch/model/mistral/language_model.py (232:246) duplicated block id: 5717 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/llama2/transformer.py (834:843) - megatron_patch/model/qwen1_5_megablocks/transformer.py (660:669) duplicated block id: 5718 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/llama2/language_model.py (231:245) - megatron_patch/model/starcoder/language_model.py (232:246) duplicated block id: 5719 size: 9 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/chatglm/checkpoint_reshaping_and_interoperability.py (465:480) - toolkits/model_checkpoints_convertor/glm/checkpoint_reshaping_and_interoperability.py (441:455) duplicated block id: 5720 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/language_model.py (34:44) - megatron_patch/model/mistral/language_model.py (32:42) duplicated block id: 5721 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/llama2/language_model.py (374:387) - megatron_patch/model/mistral/language_model.py (376:389) duplicated block id: 5722 size: 9 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/hf2te.py (384:398) - toolkits/model_checkpoints_convertor/galactica/checkpoint_reshaping_and_interoperability.py (542:556) duplicated block id: 5723 size: 9 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2.5_vl.py (220:229) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_moe.py (378:388) duplicated block id: 5724 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (643:654) - megatron_patch/model/llama/transformer.py (682:693) duplicated block id: 5725 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/llama/language_model.py (227:242) - megatron_patch/model/qwen/language_model.py (231:245) duplicated block id: 5726 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/transformer/attention.py (579:587) - megatron_patch/model/qwen1_5/transformer/attention.py (455:463) duplicated block id: 5727 size: 9 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/hf2te.py (384:398) - toolkits/model_checkpoints_convertor/falcon/checkpoint_reshaping_and_interoperability.py (495:509) duplicated block id: 5728 size: 9 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_dense_and_moe_gqa.py (565:573) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_dense_and_moe_gqa.py (602:610) duplicated block id: 5729 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/transformer.py (224:242) - megatron_patch/model/glm130b/transformer.py (294:312) duplicated block id: 5730 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/falcon/language_model.py (227:242) - megatron_patch/model/qwen_vl/language_model.py (232:246) duplicated block id: 5731 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/language_model.py (680:688) - megatron_patch/model/llama2/language_model.py (635:643) duplicated block id: 5732 size: 9 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore.py (219:229) - toolkits/model_checkpoints_convertor/qwen/hf2megablocks_qwen1.5.py (212:222) duplicated block id: 5733 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/transformer.py (224:242) - megatron_patch/model/falcon/transformer.py (285:303) duplicated block id: 5734 size: 9 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/hf2te.py (384:398) - toolkits/model_checkpoints_convertor/falcon40b/checkpoint_reshaping_and_interoperability.py (484:498) duplicated block id: 5735 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (518:530) - megatron_patch/model/galactica/transformer.py (452:465) duplicated block id: 5736 size: 9 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (345:362) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (152:169) duplicated block id: 5737 size: 9 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/deepseek/hf2mcore_deepseek_v3_moe.py (113:123) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_moe.py (56:66) duplicated block id: 5738 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/falcon/transformer.py (507:519) - megatron_patch/model/qwen1_5_megablocks/transformer.py (368:380) duplicated block id: 5739 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/language_model.py (227:242) - megatron_patch/model/qwen1_5_megablocks/language_model.py (230:244) duplicated block id: 5740 size: 9 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2.5_vl.py (190:198) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_dense_and_moe_gqa.py (320:328) duplicated block id: 5741 size: 9 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/hf2te.py (198:211) - toolkits/model_checkpoints_convertor/falcon40b/checkpoint_reshaping_and_interoperability.py (314:327) duplicated block id: 5742 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/language_model.py (344:352) - megatron_patch/model/falcon/language_model.py (334:342) duplicated block id: 5743 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/llama2/transformer.py (769:778) - megatron_patch/model/llama2/transformer.py (834:843) duplicated block id: 5744 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/language_model.py (212:225) - megatron_patch/model/llama3/language_model.py (189:202) duplicated block id: 5745 size: 9 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/deepseek/hf2mcore_deepseek_v2_moe.py (275:285) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_dense_and_moe_gqa.py (396:404) duplicated block id: 5746 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/bloom/language_model.py (245:260) - megatron_patch/model/llama2/language_model.py (231:245) duplicated block id: 5747 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/qwen1_5/transformer/attention.py (434:442) - megatron_patch/model/qwen1_5/transformer/attention.py (455:463) duplicated block id: 5748 size: 9 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon/checkpoint_reshaping_and_interoperability.py (496:511) - toolkits/model_checkpoints_convertor/starcoder/checkpoint_reshaping_and_interoperability.py (714:728) duplicated block id: 5749 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/language_model.py (32:42) - megatron_patch/model/starcoder/language_model.py (30:40) duplicated block id: 5750 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/falcon/transformer.py (1197:1212) - megatron_patch/model/galactica/transformer.py (826:841) duplicated block id: 5751 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/qwen2/layer_specs.py (113:122) - megatron_patch/model/qwen3_moe/gpt_layer_specs.py (298:307) duplicated block id: 5752 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/falcon/transformer.py (258:275) - megatron_patch/model/llava/transformer.py (391:408) duplicated block id: 5753 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/qwen1_5/transformer/attention.py (434:442) - megatron_patch/model/qwen2_vl/attention_vision.py (598:606) duplicated block id: 5754 size: 9 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_moe.py (41:53) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_vl.py (43:52) duplicated block id: 5755 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/language_model.py (390:398) - megatron_patch/model/falcon40b/language_model.py (334:342) duplicated block id: 5756 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/language_model.py (33:43) - megatron_patch/model/chatglm/language_model.py (34:44) duplicated block id: 5757 size: 9 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (769:783) - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (742:756) duplicated block id: 5758 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/qwen2_5_vl/transformer_config.py (21:38) - megatron_patch/model/qwen2_vl/transformer_config.py (21:38) duplicated block id: 5759 size: 9 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_dense_and_moe_gqa.py (227:236) - toolkits/model_checkpoints_convertor/qwen/hf2megablocks_qwen1.5.py (310:319) duplicated block id: 5760 size: 9 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon/checkpoint_reshaping_and_interoperability.py (470:480) - toolkits/model_checkpoints_convertor/yi/checkpoint_reshaping_and_interoperability.py (484:494) duplicated block id: 5761 size: 9 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/deepseek/hf2mcore_deepseek_v3_moe.py (527:538) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_dense_and_moe_gqa.py (489:500) duplicated block id: 5762 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/llava/transformer.py (768:777) - megatron_patch/model/qwen1_5_megablocks/transformer.py (725:734) duplicated block id: 5763 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/language_model.py (244:259) - megatron_patch/model/qwen1_5_megablocks/language_model.py (230:244) duplicated block id: 5764 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/mixtral_bak/layer_specs.py (138:147) - megatron_patch/model/qwen2_vl/layer_specs.py (106:114) duplicated block id: 5765 size: 9 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (307:324) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (152:169) duplicated block id: 5766 size: 9 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/glm130b/checkpoint_reshaping_and_interoperability.py (422:436) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (565:580) duplicated block id: 5767 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/language_model.py (262:274) - megatron_patch/model/glm130b/language_model.py (262:274) duplicated block id: 5768 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/language_model.py (241:255) - megatron_patch/model/bloom/language_model.py (245:260) duplicated block id: 5769 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/llama3/layer_specs.py (55:66) - megatron_patch/model/qwen2_moe/layer_specs.py (171:181) duplicated block id: 5770 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/galactica/transformer.py (572:583) - megatron_patch/model/starcoder/transformer.py (857:868) duplicated block id: 5771 size: 9 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/language_model.py (390:398) - megatron_patch/model/falcon/language_model.py (334:342) duplicated block id: 5772 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (145:154) - megatron_patch/model/starcoder/transformer.py (134:143) duplicated block id: 5773 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (506:518) - toolkits/model_checkpoints_convertor/baichuan2/hf2te.py (356:368) duplicated block id: 5774 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/language_model.py (579:589) - megatron_patch/model/mistral/language_model.py (569:579) duplicated block id: 5775 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/transformer.py (268:284) - megatron_patch/model/qwen/transformer.py (385:401) duplicated block id: 5776 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/llama/language_model.py (371:382) - megatron_patch/model/starcoder/language_model.py (365:376) duplicated block id: 5777 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (382:391) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (569:577) duplicated block id: 5778 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/mixtral_bak/transformer/attention.py (410:417) - megatron_patch/model/qwen1_5/transformer/attention.py (360:367) duplicated block id: 5779 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/llama3/model.py (112:128) - megatron_patch/model/llama3_1/model.py (164:180) duplicated block id: 5780 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (363:373) - megatron_patch/model/qwen/transformer.py (360:370) duplicated block id: 5781 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/galactica/checkpoint_reshaping_and_interoperability.py (531:539) - toolkits/model_checkpoints_convertor/starcoder/checkpoint_reshaping_and_interoperability.py (703:711) duplicated block id: 5782 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer_legacy.py (762:779) - megatron_patch/model/qwen2/transformer/attention.py (221:238) duplicated block id: 5783 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/falcon/language_model.py (287:295) - megatron_patch/model/qwen_vl/language_model.py (290:298) duplicated block id: 5784 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (301:320) - megatron_patch/model/starcoder/transformer.py (425:445) duplicated block id: 5785 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/language_model.py (299:307) - megatron_patch/model/falcon/language_model.py (287:295) duplicated block id: 5786 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/mixtral_bak/transformer/attention.py (333:340) - megatron_patch/model/qwen1_5/transformer/attention.py (526:533) duplicated block id: 5787 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/chatglm/checkpoint_reshaping_and_interoperability.py (343:350) - toolkits/model_checkpoints_convertor/yi/checkpoint_reshaping_and_interoperability.py (421:428) duplicated block id: 5788 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/glm/checkpoint_reshaping_and_interoperability.py (430:438) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (553:561) duplicated block id: 5789 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (255:273) - megatron_patch/model/llama3/transformer_legacy.py (390:406) duplicated block id: 5790 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/checkpoint_reshaping_and_interoperability.py (703:711) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (553:561) duplicated block id: 5791 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/llama2/transformer.py (1740:1747) - megatron_patch/model/llava/transformer.py (1200:1207) duplicated block id: 5792 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/mistral/transformer.py (1737:1744) - megatron_patch/model/qwen_vl/transformer.py (1200:1207) duplicated block id: 5793 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/language_model.py (345:353) - megatron_patch/model/llama/language_model.py (287:295) duplicated block id: 5794 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/hf2te.py (373:381) - toolkits/model_checkpoints_convertor/bloom/reward_model_to_megatron.py (702:710) duplicated block id: 5795 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/chatglm/checkpoint_reshaping_and_interoperability.py (343:350) - toolkits/model_checkpoints_convertor/galactica/checkpoint_reshaping_and_interoperability.py (445:452) duplicated block id: 5796 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/mistral/transformer.py (1198:1205) - megatron_patch/model/qwen1_5_megablocks/transformer.py (1602:1609) duplicated block id: 5797 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/llama2/transformer.py (1016:1025) - megatron_patch/model/qwen1_5_megablocks/transformer.py (1127:1134) duplicated block id: 5798 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/mixtral_bak/transformer/attention.py (342:350) - megatron_patch/model/qwen2_vl/attention_vision.py (504:512) duplicated block id: 5799 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/llama/language_model.py (287:295) - megatron_patch/model/llava/language_model.py (292:300) duplicated block id: 5800 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/llama/transformer.py (362:369) - megatron_patch/model/starcoder/transformer.py (524:531) duplicated block id: 5801 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llava/hf2mcore_llava.py (193:200) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_dense_and_moe_gqa.py (156:163) duplicated block id: 5802 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (1104:1117) - megatron_patch/model/falcon/transformer.py (1199:1212) duplicated block id: 5803 size: 8 cleaned lines of code in 2 files: - toolkits/pretrain_data_preprocessing/preprocess_data_megatron.py (195:204) - toolkits/sft_data_preprocessing/build_idxmap_sft_dataset.py (188:196) duplicated block id: 5804 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/llama/language_model.py (611:621) - megatron_patch/model/qwen1_5_megablocks/language_model.py (556:566) duplicated block id: 5805 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/galactica/transformer.py (598:605) - megatron_patch/model/glm130b/transformer.py (758:765) duplicated block id: 5806 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/mistral/transformer.py (1198:1205) - megatron_patch/model/qwen_vl/transformer.py (1739:1746) duplicated block id: 5807 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/reward_model_to_megatron.py (713:726) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (415:428) duplicated block id: 5808 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/llava/transformer.py (1200:1207) - megatron_patch/model/llava/transformer.py (1739:1746) duplicated block id: 5809 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer/attention.py (225:242) - megatron_patch/model/qwen/transformer.py (757:774) duplicated block id: 5810 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore.py (544:551) - toolkits/model_checkpoints_convertor/mistral/hf2mcore.py (245:252) duplicated block id: 5811 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (999:1006) - megatron_patch/model/qwen1_5_megablocks/transformer.py (1013:1020) duplicated block id: 5812 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/language_model.py (372:379) - megatron_patch/model/mistral/language_model.py (364:371) duplicated block id: 5813 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/language_model.py (634:644) - megatron_patch/model/chatglm/language_model.py (579:589) duplicated block id: 5814 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (534:546) - toolkits/model_checkpoints_convertor/baichuan2/hf2te.py (356:368) duplicated block id: 5815 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/language_model.py (303:311) - megatron_patch/model/llama3/language_model.py (284:292) duplicated block id: 5816 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/galactica/checkpoint_reshaping_and_interoperability.py (649:658) - toolkits/model_checkpoints_convertor/starcoder/checkpoint_reshaping_and_interoperability.py (823:832) duplicated block id: 5817 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/galactica/checkpoint_reshaping_and_interoperability.py (517:525) - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (395:404) duplicated block id: 5818 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/gpt_model.py (105:114) - megatron_patch/model/mistral/gpt_model.py (106:115) duplicated block id: 5819 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/transformer.py (1035:1043) - megatron_patch/model/llama2/transformer.py (1605:1613) duplicated block id: 5820 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (269:279) - megatron_patch/model/baichuan2/transformer.py (363:373) duplicated block id: 5821 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/qwen1_5_megablocks/language_model.py (333:340) - megatron_patch/model/starcoder/language_model.py (336:343) duplicated block id: 5822 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer_legacy.py (538:545) - megatron_patch/model/starcoder/transformer.py (593:600) duplicated block id: 5823 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llava/hf2mcore_llava.py (103:110) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (94:101) duplicated block id: 5824 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/qwen/transformer.py (333:341) - megatron_patch/model/starcoder/transformer.py (222:230) duplicated block id: 5825 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/llama2/transformer.py (1016:1025) - megatron_patch/model/llava/transformer.py (1242:1249) duplicated block id: 5826 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon40b/checkpoint_reshaping_and_interoperability.py (257:275) - toolkits/model_checkpoints_convertor/galactica/checkpoint_reshaping_and_interoperability.py (290:308) duplicated block id: 5827 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/falcon/language_model.py (600:610) - megatron_patch/model/llama2/language_model.py (557:567) duplicated block id: 5828 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/checkpoint_reshaping_and_interoperability.py (386:400) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (871:885) duplicated block id: 5829 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/model.py (194:219) - megatron_patch/model/llama3/model.py (173:194) duplicated block id: 5830 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/deepseek/hf2mcore_deepseek_v2_moe.py (374:382) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_mha_to_moe.py (160:167) duplicated block id: 5831 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/hf2te.py (424:433) - toolkits/model_checkpoints_convertor/galactica/checkpoint_reshaping_and_interoperability.py (603:612) duplicated block id: 5832 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer_legacy.py (966:975) - megatron_patch/model/qwen_vl/transformer.py (1242:1249) duplicated block id: 5833 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/hf2te.py (385:398) - toolkits/model_checkpoints_convertor/bloom/checkpoint_reshaping_and_interoperability.py (714:727) duplicated block id: 5834 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon/checkpoint_reshaping_and_interoperability.py (470:478) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (382:391) duplicated block id: 5835 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (255:273) - megatron_patch/model/qwen1_5_megablocks/transformer.py (282:298) duplicated block id: 5836 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore.py (139:146) - toolkits/model_checkpoints_convertor/llama/hf2mcore_llama3_1.py (80:87) duplicated block id: 5837 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/gpt_model.py (68:75) - megatron_patch/model/galactica/gpt_model.py (72:79) duplicated block id: 5838 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/hf2te.py (356:368) - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (587:599) duplicated block id: 5839 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (410:429) - megatron_patch/model/starcoder/transformer.py (425:445) duplicated block id: 5840 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/hf2te.py (111:118) - toolkits/model_checkpoints_convertor/llava/hf2mcore_llava.py (103:110) duplicated block id: 5841 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (815:822) - megatron_patch/model/llama3/transformer_legacy.py (903:910) duplicated block id: 5842 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (1605:1613) - megatron_patch/model/falcon/transformer.py (1151:1159) duplicated block id: 5843 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (154:163) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (191:200) duplicated block id: 5844 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/llama2/transformer.py (1201:1208) - megatron_patch/model/llava/transformer.py (1739:1746) duplicated block id: 5845 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/checkpoint_reshaping_and_interoperability.py (703:711) - toolkits/model_checkpoints_convertor/falcon/checkpoint_reshaping_and_interoperability.py (484:492) duplicated block id: 5846 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/llama2/transformer.py (1016:1025) - megatron_patch/model/qwen/transformer.py (1185:1192) duplicated block id: 5847 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon40b/checkpoint_reshaping_and_interoperability.py (473:481) - toolkits/model_checkpoints_convertor/glm/checkpoint_reshaping_and_interoperability.py (430:438) duplicated block id: 5848 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (716:723) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (666:673) duplicated block id: 5849 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/chatglm/checkpoint_reshaping_and_interoperability.py (330:341) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (463:474) duplicated block id: 5850 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/gpt_model.py (105:114) - megatron_patch/model/llama3/gpt_model.py (108:117) duplicated block id: 5851 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (76:87) - megatron_patch/model/qwen_vl/transformer.py (67:79) duplicated block id: 5852 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/falcon/language_model.py (287:295) - megatron_patch/model/qwen1_5_megablocks/language_model.py (288:296) duplicated block id: 5853 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/hf2te.py (112:119) - toolkits/model_checkpoints_convertor/llava/hf2mcore_llava.py (103:110) duplicated block id: 5854 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/falcon/transformer.py (744:754) - megatron_patch/model/qwen_vl/transformer.py (739:751) duplicated block id: 5855 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/mixtral_bak/layer_specs.py (67:74) - megatron_patch/model/qwen1_5/layer_specs.py (55:62) duplicated block id: 5856 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (1470:1478) - megatron_patch/model/galactica/transformer.py (779:787) duplicated block id: 5857 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/falcon/transformer.py (744:754) - megatron_patch/model/falcon/transformer.py (765:777) duplicated block id: 5858 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/falcon/language_model.py (287:295) - megatron_patch/model/llama2/language_model.py (289:297) duplicated block id: 5859 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/llama3_1/model.py (164:180) - megatron_patch/model/qwen2_vl/gpt_model.py (103:119) duplicated block id: 5860 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/transformer.py (374:381) - megatron_patch/model/starcoder/transformer.py (524:531) duplicated block id: 5861 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/reward_model_to_megatron.py (479:486) - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (668:675) duplicated block id: 5862 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (142:149) - toolkits/model_checkpoints_convertor/llama/hf2mcore_llama3_1.py (80:87) duplicated block id: 5863 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/reward_model_to_megatron.py (479:486) - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (716:723) duplicated block id: 5864 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (793:800) - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (829:836) duplicated block id: 5865 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (793:800) - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (781:788) duplicated block id: 5866 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/qwen2_moe/layer_specs.py (279:286) - megatron_patch/model/qwen2_vl/layer_specs.py (114:121) duplicated block id: 5867 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/llava/transformer.py (1200:1207) - megatron_patch/model/qwen1_5_megablocks/transformer.py (1602:1609) duplicated block id: 5868 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/transformer_config.py (130:165) - megatron_patch/model/mixtral_bak/transformer_config.py (98:109) duplicated block id: 5869 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_mha_to_moe.py (246:255) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_vl.py (702:713) duplicated block id: 5870 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/chatglm/checkpoint_reshaping_and_interoperability.py (330:341) - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (486:497) duplicated block id: 5871 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/llama2/language_model.py (374:386) - megatron_patch/model/qwen/language_model.py (375:387) duplicated block id: 5872 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/language_model.py (634:644) - megatron_patch/model/falcon/language_model.py (600:610) duplicated block id: 5873 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/language_model.py (600:610) - megatron_patch/model/qwen/language_model.py (539:549) duplicated block id: 5874 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/moe/router.py (144:160) - megatron_patch/model/qwen3_moe/moe/router.py (41:57) duplicated block id: 5875 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/hf2te.py (285:298) - toolkits/model_checkpoints_convertor/yi/checkpoint_reshaping_and_interoperability.py (404:417) duplicated block id: 5876 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/language_model.py (349:356) - megatron_patch/model/llava/language_model.py (337:344) duplicated block id: 5877 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (76:87) - megatron_patch/model/llava/transformer.py (67:79) duplicated block id: 5878 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/reward_model_to_megatron.py (385:399) - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (849:863) duplicated block id: 5879 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/llama2/transformer.py (1243:1250) - megatron_patch/model/qwen1_5_megablocks/transformer.py (900:909) duplicated block id: 5880 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llava/hf2mcore_llava.py (103:110) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (170:177) duplicated block id: 5881 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/layer_specs.py (89:96) - megatron_patch/model/qwen2_moe/layer_specs.py (171:178) duplicated block id: 5882 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llava/hf2mcore_llava.py (103:110) - toolkits/model_checkpoints_convertor/starcoder/checkpoint_reshaping_and_interoperability.py (121:128) duplicated block id: 5883 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/chatglm/checkpoint_reshaping_and_interoperability.py (343:350) - toolkits/model_checkpoints_convertor/falcon/checkpoint_reshaping_and_interoperability.py (408:415) duplicated block id: 5884 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/qwen2_moe/layer_specs.py (171:178) - megatron_patch/model/qwen3_moe/gpt_layer_specs.py (130:137) duplicated block id: 5885 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/llava/transformer.py (1739:1746) - megatron_patch/model/qwen_vl/transformer.py (1200:1207) duplicated block id: 5886 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/hf2te.py (373:381) - toolkits/model_checkpoints_convertor/starcoder/checkpoint_reshaping_and_interoperability.py (703:711) duplicated block id: 5887 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/falcon/transformer.py (429:449) - megatron_patch/model/llama2/transformer.py (398:417) duplicated block id: 5888 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/qwen2/transformer_block.py (440:448) - megatron_patch/model/qwen2_5_vl/transformer_block.py (617:638) duplicated block id: 5889 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/qwen2_vl/attention.py (153:160) - megatron_patch/model/qwen2_vl/attention_vision.py (447:454) duplicated block id: 5890 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/qwen2_vl/attention.py (153:160) - megatron_patch/model/qwen2_vl/attention_vision.py (457:464) duplicated block id: 5891 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/qwen_vl/transformer.py (540:547) - megatron_patch/model/starcoder/transformer.py (593:600) duplicated block id: 5892 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (85:92) - toolkits/model_checkpoints_convertor/llama/hf2mcore.py (119:127) duplicated block id: 5893 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (815:822) - megatron_patch/model/qwen/transformer.py (898:905) duplicated block id: 5894 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer_legacy.py (1294:1317) - megatron_patch/model/starcoder/transformer.py (1059:1082) duplicated block id: 5895 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer/mlp.py (109:116) - megatron_patch/model/qwen2/moe/experts.py (277:284) duplicated block id: 5896 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/llama2/transformer.py (1016:1025) - megatron_patch/model/llama2/transformer.py (1243:1250) duplicated block id: 5897 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon40b/checkpoint_reshaping_and_interoperability.py (761:768) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (666:673) duplicated block id: 5898 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon40b/checkpoint_reshaping_and_interoperability.py (761:768) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (713:720) duplicated block id: 5899 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/gpt_model.py (115:125) - megatron_patch/model/falcon40b/gpt_model.py (105:114) duplicated block id: 5900 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (363:373) - megatron_patch/model/llama3/transformer_legacy.py (365:375) duplicated block id: 5901 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/transformer_config.py (34:41) - megatron_patch/model/qwen2_moe/transformer_config.py (56:63) duplicated block id: 5902 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon/checkpoint_reshaping_and_interoperability.py (777:784) - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (781:788) duplicated block id: 5903 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon/checkpoint_reshaping_and_interoperability.py (777:784) - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (829:836) duplicated block id: 5904 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer/attention.py (360:367) - megatron_patch/model/qwen1_5/transformer/attention.py (526:533) duplicated block id: 5905 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/language_model.py (390:397) - megatron_patch/model/bloom/language_model.py (352:359) duplicated block id: 5906 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2megablocks_qwen1.5.py (299:306) - toolkits/model_checkpoints_convertor/qwen/hf2megablocks_qwen1.5.py (554:561) duplicated block id: 5907 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/language_model.py (334:341) - megatron_patch/model/qwen_vl/language_model.py (335:342) duplicated block id: 5908 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/llama3/layer_specs.py (55:62) - megatron_patch/model/mixtral/layer_specs.py (89:96) duplicated block id: 5909 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (1018:1027) - megatron_patch/model/llama3/transformer_legacy.py (1201:1208) duplicated block id: 5910 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/transformer.py (890:898) - megatron_patch/model/starcoder/transformer.py (1201:1209) duplicated block id: 5911 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (872:881) - megatron_patch/model/starcoder/transformer.py (974:982) duplicated block id: 5912 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/transformer.py (775:783) - megatron_patch/model/llama3/transformer_legacy.py (1565:1573) duplicated block id: 5913 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/transformer.py (553:566) - megatron_patch/model/llava/transformer.py (689:702) duplicated block id: 5914 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/model.py (210:220) - megatron_patch/model/qwen1_5/model.py (207:217) duplicated block id: 5915 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore.py (119:127) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (113:120) duplicated block id: 5916 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (496:508) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (594:606) duplicated block id: 5917 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/hf2te.py (257:270) - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (482:495) duplicated block id: 5918 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/galactica/transformer.py (779:787) - megatron_patch/model/llama3/transformer_legacy.py (1565:1573) duplicated block id: 5919 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (1243:1250) - megatron_patch/model/llama3/transformer_legacy.py (966:975) duplicated block id: 5920 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon40b/checkpoint_reshaping_and_interoperability.py (459:467) - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (395:404) duplicated block id: 5921 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/llava_mcore/transformer_config.py (80:87) - megatron_patch/model/llava_mcore/transformer_config.py (100:107) duplicated block id: 5922 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (587:599) - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (496:508) duplicated block id: 5923 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/language_model.py (287:295) - megatron_patch/model/qwen1_5_megablocks/language_model.py (288:296) duplicated block id: 5924 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/hf2te.py (360:367) - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (395:404) duplicated block id: 5925 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/transformer.py (953:966) - megatron_patch/model/llama/transformer.py (1007:1020) duplicated block id: 5926 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/galactica/checkpoint_reshaping_and_interoperability.py (67:74) - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (86:94) duplicated block id: 5927 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/llava_mcore/llava_spec.py (63:71) - megatron_patch/model/qwen1_5/layer_specs.py (67:74) duplicated block id: 5928 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/llava/transformer.py (1015:1024) - megatron_patch/model/qwen1_5_megablocks/transformer.py (1127:1134) duplicated block id: 5929 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/hf2te.py (385:398) - toolkits/model_checkpoints_convertor/bloom/reward_model_to_megatron.py (713:726) duplicated block id: 5930 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon40b/checkpoint_reshaping_and_interoperability.py (522:529) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (588:595) duplicated block id: 5931 size: 8 cleaned lines of code in 2 files: - megatron_patch/template/helper.py (69:77) - megatron_patch/template/helper.py (110:118) duplicated block id: 5932 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/deepseek/hf2mcore_deepseek_v2_moe.py (374:382) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_moe.py (309:316) duplicated block id: 5933 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/language_model.py (303:311) - megatron_patch/model/qwen_vl/language_model.py (290:298) duplicated block id: 5934 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/mistral/transformer.py (1013:1022) - megatron_patch/model/mistral/transformer.py (1240:1247) duplicated block id: 5935 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/hf2te.py (384:396) - toolkits/model_checkpoints_convertor/yi/checkpoint_reshaping_and_interoperability.py (513:525) duplicated block id: 5936 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (1116:1125) - megatron_patch/model/starcoder/transformer.py (962:971) duplicated block id: 5937 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_moe.py (367:374) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_dense_and_moe_gqa.py (741:748) duplicated block id: 5938 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/glm/checkpoint_reshaping_and_interoperability.py (430:438) - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (576:584) duplicated block id: 5939 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (999:1006) - megatron_patch/model/qwen_vl/transformer.py (1128:1135) duplicated block id: 5940 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (294:310) - megatron_patch/model/bloom/transformer.py (255:273) duplicated block id: 5941 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/reward_model_to_megatron.py (702:710) - toolkits/model_checkpoints_convertor/falcon/checkpoint_reshaping_and_interoperability.py (484:492) duplicated block id: 5942 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/llava_mcore/layer_specs.py (21:34) - megatron_patch/model/qwen2_moe/layer_specs.py (42:55) duplicated block id: 5943 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/language_model.py (287:295) - megatron_patch/model/qwen/language_model.py (289:297) duplicated block id: 5944 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/llama2/transformer.py (1016:1025) - megatron_patch/model/qwen_vl/transformer.py (1242:1249) duplicated block id: 5945 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_mha.py (314:323) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_vl.py (706:715) duplicated block id: 5946 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (97:108) - megatron_patch/model/qwen_vl/transformer.py (98:109) duplicated block id: 5947 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/model.py (154:189) - megatron_patch/model/qwen1_5/model.py (173:194) duplicated block id: 5948 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/llava/transformer.py (797:814) - megatron_patch/model/qwen2/transformer/attention.py (221:238) duplicated block id: 5949 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/qwen1_5_megablocks/transformer.py (1602:1609) - megatron_patch/model/qwen_vl/transformer.py (1200:1207) duplicated block id: 5950 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/qwen2_5_vl/visionmodel.py (217:235) - megatron_patch/model/qwen2_vl/visionmodel.py (171:190) duplicated block id: 5951 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/language_model.py (287:295) - megatron_patch/model/llava/language_model.py (292:300) duplicated block id: 5952 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/galactica/language_model.py (629:639) - megatron_patch/model/qwen/language_model.py (539:549) duplicated block id: 5953 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/transformer.py (775:783) - megatron_patch/model/qwen_vl/transformer.py (1604:1612) duplicated block id: 5954 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (998:1008) - megatron_patch/model/chatglm/transformer.py (729:739) duplicated block id: 5955 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/mistral/transformer.py (796:813) - megatron_patch/model/qwen2/transformer/attention.py (221:238) duplicated block id: 5956 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/qwen1_5/transformer/attention.py (360:367) - megatron_patch/model/qwen2/transformer/attention.py (531:538) duplicated block id: 5957 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer/attention.py (225:242) - megatron_patch/model/qwen_vl/transformer.py (797:814) duplicated block id: 5958 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/mixtral_bak/transformer/attention.py (199:216) - megatron_patch/model/qwen1_5_megablocks/transformer.py (689:706) duplicated block id: 5959 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/qwen_vl/transformer.py (1200:1207) - megatron_patch/model/qwen_vl/transformer.py (1739:1746) duplicated block id: 5960 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/mixtral_bak/layer_specs.py (67:74) - megatron_patch/model/qwen3_moe/gpt_layer_specs.py (157:164) duplicated block id: 5961 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/mixtral_bak/layer_specs.py (67:74) - megatron_patch/model/qwen3_moe/gpt_layer_specs.py (130:137) duplicated block id: 5962 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/falcon/language_model.py (334:341) - megatron_patch/model/llava/language_model.py (337:344) duplicated block id: 5963 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (772:782) - megatron_patch/model/starcoder/transformer.py (848:857) duplicated block id: 5964 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llava/hf2mcore_llava.py (208:215) - toolkits/model_checkpoints_convertor/llava/hf2mcore_llava.py (375:382) duplicated block id: 5965 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/transformer.py (453:460) - megatron_patch/model/glm130b/transformer.py (506:513) duplicated block id: 5966 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/llama/gpt_model.py (105:114) - megatron_patch/model/llama3/gpt_model.py (108:117) duplicated block id: 5967 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (1050:1058) - megatron_patch/model/starcoder/transformer.py (1201:1209) duplicated block id: 5968 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/mixtral_bak/transformer/attention.py (199:216) - megatron_patch/model/qwen_vl/transformer.py (797:814) duplicated block id: 5969 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/qwen2_vl/attention_vision.py (447:454) - megatron_patch/model/qwen2_vl/attention_vision.py (457:464) duplicated block id: 5970 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_dense_and_moe_gqa.py (619:626) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_dense_and_moe_gqa.py (741:748) duplicated block id: 5971 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/moe/router.py (20:38) - megatron_patch/model/mixtral_bak/moe/router.py (29:46) duplicated block id: 5972 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (815:822) - megatron_patch/model/qwen_vl/transformer.py (952:959) duplicated block id: 5973 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/language_model.py (385:397) - megatron_patch/model/qwen1_5_megablocks/language_model.py (373:385) duplicated block id: 5974 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/llama/language_model.py (606:614) - megatron_patch/model/starcoder/language_model.py (481:489) duplicated block id: 5975 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/galactica/transformer.py (779:787) - megatron_patch/model/mistral/transformer.py (1602:1610) duplicated block id: 5976 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/multi_latent_attention.py (202:209) - megatron_patch/model/qwen1_5/transformer/attention.py (362:369) duplicated block id: 5977 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/llama/transformer.py (945:953) - megatron_patch/model/starcoder/transformer.py (1201:1209) duplicated block id: 5978 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/language_model.py (345:353) - megatron_patch/model/bloom/language_model.py (305:313) duplicated block id: 5979 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/checkpoint_reshaping_and_interoperability.py (714:727) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (415:428) duplicated block id: 5980 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (531:538) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (621:628) duplicated block id: 5981 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/checkpoint_reshaping_and_interoperability.py (386:400) - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (849:863) duplicated block id: 5982 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/galactica/transformer.py (779:787) - megatron_patch/model/llama2/transformer.py (1605:1613) duplicated block id: 5983 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/llama3_1/model.py (278:289) - megatron_patch/model/mixtral_bak/model.py (205:216) duplicated block id: 5984 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (668:675) - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (829:836) duplicated block id: 5985 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/llama/language_model.py (334:341) - megatron_patch/model/qwen_vl/language_model.py (335:342) duplicated block id: 5986 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore_llama3_1.py (80:87) - toolkits/model_checkpoints_convertor/qwen/hf2megablocks_qwen1.5.py (133:140) duplicated block id: 5987 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_mha_to_moe.py (65:74) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_moe.py (62:71) duplicated block id: 5988 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/llama/language_model.py (611:621) - megatron_patch/model/qwen_vl/language_model.py (586:596) duplicated block id: 5989 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/model.py (118:129) - megatron_patch/model/mixtral_bak/model.py (96:107) duplicated block id: 5990 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/llama2/transformer.py (798:815) - megatron_patch/model/mixtral_bak/transformer/attention.py (199:216) duplicated block id: 5991 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/llama/transformer.py (945:953) - megatron_patch/model/qwen1_5_megablocks/transformer.py (1468:1476) duplicated block id: 5992 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/moe/moe_layer.py (111:118) - megatron_patch/model/qwen2/moe/moe_layer.py (114:121) duplicated block id: 5993 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/moe/experts.py (49:61) - megatron_patch/model/qwen2/moe/experts.py (52:64) duplicated block id: 5994 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/transformer.py (268:284) - megatron_patch/model/llama2/transformer.py (391:407) duplicated block id: 5995 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (86:94) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (96:103) duplicated block id: 5996 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/llama/language_model.py (611:621) - megatron_patch/model/llama2/language_model.py (557:567) duplicated block id: 5997 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/language_model.py (208:216) - megatron_patch/model/qwen2_vl/language_model_embedding.py (87:95) duplicated block id: 5998 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/qwen1_5_megablocks/transformer.py (1221:1244) - megatron_patch/model/starcoder/transformer.py (1059:1082) duplicated block id: 5999 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/mistral/transformer.py (1013:1022) - megatron_patch/model/qwen/transformer.py (1185:1192) duplicated block id: 6000 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/hf2te.py (285:298) - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (401:414) duplicated block id: 6001 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/chatglm/checkpoint_reshaping_and_interoperability.py (343:350) - toolkits/model_checkpoints_convertor/falcon40b/checkpoint_reshaping_and_interoperability.py (397:404) duplicated block id: 6002 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llava/hf2mcore_llava.py (357:364) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_dense_and_moe_gqa.py (650:657) duplicated block id: 6003 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon40b/checkpoint_reshaping_and_interoperability.py (143:150) - toolkits/model_checkpoints_convertor/llava/hf2mcore_llava.py (103:110) duplicated block id: 6004 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/falcon/gpt_model.py (105:114) - megatron_patch/model/qwen/gpt_model.py (106:115) duplicated block id: 6005 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/llava/transformer.py (399:418) - megatron_patch/model/starcoder/transformer.py (302:322) duplicated block id: 6006 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (1013:1020) - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (1029:1042) duplicated block id: 6007 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/bloom/language_model.py (305:313) - megatron_patch/model/qwen_vl/language_model.py (290:298) duplicated block id: 6008 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/language_model.py (344:351) - megatron_patch/model/starcoder/language_model.py (336:343) duplicated block id: 6009 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/reward_model_to_megatron.py (781:789) - toolkits/model_checkpoints_convertor/starcoder/checkpoint_reshaping_and_interoperability.py (797:805) duplicated block id: 6010 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (495:503) - toolkits/model_checkpoints_convertor/glm130b/checkpoint_reshaping_and_interoperability.py (411:419) duplicated block id: 6011 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/qwen1_5_megablocks/transformer.py (230:238) - megatron_patch/model/starcoder/transformer.py (222:230) duplicated block id: 6012 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/transformer.py (775:783) - megatron_patch/model/starcoder/transformer.py (1201:1209) duplicated block id: 6013 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (1605:1613) - megatron_patch/model/galactica/transformer.py (779:787) duplicated block id: 6014 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (731:743) - megatron_patch/model/falcon/transformer.py (744:754) duplicated block id: 6015 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/llama/transformer.py (945:953) - megatron_patch/model/mistral/transformer.py (1602:1610) duplicated block id: 6016 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/bloom/language_model.py (352:359) - megatron_patch/model/llama3/language_model.py (329:336) duplicated block id: 6017 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore_llama3_1.py (606:613) - toolkits/model_checkpoints_convertor/llava/hf2mcore_llava.py (208:215) duplicated block id: 6018 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/galactica/language_model.py (629:639) - megatron_patch/model/llama3/language_model.py (534:544) duplicated block id: 6019 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/qwen/transformer.py (831:842) - megatron_patch/model/starcoder/transformer.py (814:825) duplicated block id: 6020 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/glm130b/checkpoint_reshaping_and_interoperability.py (411:419) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (553:561) duplicated block id: 6021 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/llama3/model.py (104:115) - megatron_patch/model/mixtral/model.py (100:111) duplicated block id: 6022 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer/attention.py (360:367) - megatron_patch/model/mixtral_bak/transformer/attention.py (410:417) duplicated block id: 6023 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/llama3_1/layer_specs.py (56:63) - megatron_patch/model/mixtral/layer_specs.py (89:96) duplicated block id: 6024 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/hf2te.py (384:396) - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (587:599) duplicated block id: 6025 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/language_model.py (550:560) - megatron_patch/model/llama/language_model.py (611:621) duplicated block id: 6026 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/language_model.py (579:589) - megatron_patch/model/llava/language_model.py (617:627) duplicated block id: 6027 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (768:778) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (656:664) duplicated block id: 6028 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (495:503) - toolkits/model_checkpoints_convertor/bloom/checkpoint_reshaping_and_interoperability.py (703:711) duplicated block id: 6029 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer/mlp.py (149:157) - megatron_patch/model/mixtral_bak/transformer/mlp.py (122:130) duplicated block id: 6030 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/qwen1_5_megablocks/transformer.py (1085:1092) - megatron_patch/model/qwen_vl/transformer.py (1739:1746) duplicated block id: 6031 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/mixtral_bak/transformer/attention.py (333:340) - megatron_patch/model/qwen2/transformer/attention.py (531:538) duplicated block id: 6032 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/qwen/transformer.py (757:774) - megatron_patch/model/qwen1_5/transformer/attention.py (225:242) duplicated block id: 6033 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer/attention.py (225:242) - megatron_patch/model/llama3/transformer_legacy.py (762:779) duplicated block id: 6034 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/mistral/transformer.py (339:347) - megatron_patch/model/starcoder/transformer.py (222:230) duplicated block id: 6035 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/mistral/hf2mcore.py (315:322) - toolkits/model_checkpoints_convertor/qwen/hf2megablocks_qwen1.5.py (373:380) duplicated block id: 6036 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon40b/checkpoint_reshaping_and_interoperability.py (143:150) - toolkits/model_checkpoints_convertor/llama/hf2mcore_llama3_1.py (80:87) duplicated block id: 6037 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/transformer.py (953:966) - megatron_patch/model/galactica/transformer.py (843:856) duplicated block id: 6038 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/llama/language_model.py (287:295) - megatron_patch/model/qwen/language_model.py (289:297) duplicated block id: 6039 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (427:439) - toolkits/model_checkpoints_convertor/yi/checkpoint_reshaping_and_interoperability.py (513:525) duplicated block id: 6040 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/language_model.py (600:610) - megatron_patch/model/qwen_vl/language_model.py (586:596) duplicated block id: 6041 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/hf2te.py (357:370) - toolkits/model_checkpoints_convertor/starcoder/checkpoint_reshaping_and_interoperability.py (714:727) duplicated block id: 6042 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/qwen2/layer_specs.py (65:72) - megatron_patch/model/qwen3_moe/gpt_layer_specs.py (130:137) duplicated block id: 6043 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/llava/transformer.py (399:418) - megatron_patch/model/starcoder/transformer.py (425:445) duplicated block id: 6044 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/galactica/checkpoint_reshaping_and_interoperability.py (567:574) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (599:606) duplicated block id: 6045 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/transformer.py (437:445) - megatron_patch/model/llama/transformer.py (439:447) duplicated block id: 6046 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/qwen1_5/transformer/mlp.py (125:132) - megatron_patch/model/qwen2/moe/experts.py (277:284) duplicated block id: 6047 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon40b/checkpoint_reshaping_and_interoperability.py (761:768) - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (829:836) duplicated block id: 6048 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (403:419) - megatron_patch/model/galactica/transformer.py (201:218) duplicated block id: 6049 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/transformer.py (890:898) - megatron_patch/model/mistral/transformer.py (1602:1610) duplicated block id: 6050 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_mha_to_moe.py (160:167) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_dense_and_moe_gqa.py (526:534) duplicated block id: 6051 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/hf2te.py (360:367) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (382:391) duplicated block id: 6052 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/falcon/gpt_model.py (105:114) - megatron_patch/model/qwen1_5_megablocks/gpt_model.py (106:115) duplicated block id: 6053 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/galactica/gpt_model.py (105:114) - megatron_patch/model/qwen1_5_megablocks/gpt_model.py (106:115) duplicated block id: 6054 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (142:149) - toolkits/model_checkpoints_convertor/llava/hf2mcore_llava.py (103:110) duplicated block id: 6055 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (716:723) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (895:902) duplicated block id: 6056 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/transformer/attention.py (204:223) - megatron_patch/model/qwen2_vl/attention.py (194:213) duplicated block id: 6057 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/deepseek/hf2mcore_deepseek_v2_moe.py (117:124) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_dense_and_moe_gqa.py (109:116) duplicated block id: 6058 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (363:373) - megatron_patch/model/mistral/transformer.py (366:376) duplicated block id: 6059 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon/checkpoint_reshaping_and_interoperability.py (777:784) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (713:720) duplicated block id: 6060 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/transformer.py (712:719) - megatron_patch/model/glm130b/transformer.py (758:765) duplicated block id: 6061 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/transformer/attention.py (504:512) - megatron_patch/model/mixtral_bak/transformer/attention.py (342:350) duplicated block id: 6062 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon/checkpoint_reshaping_and_interoperability.py (777:784) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (666:673) duplicated block id: 6063 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/galactica/gpt_model.py (105:114) - megatron_patch/model/qwen/gpt_model.py (106:115) duplicated block id: 6064 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (797:804) - megatron_patch/model/llama2/transformer.py (940:947) duplicated block id: 6065 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/hf2te.py (394:402) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (586:594) duplicated block id: 6066 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/language_model.py (303:311) - megatron_patch/model/starcoder/language_model.py (290:298) duplicated block id: 6067 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/llama/gpt_model.py (105:114) - megatron_patch/model/qwen/gpt_model.py (106:115) duplicated block id: 6068 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (395:404) - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (562:570) duplicated block id: 6069 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/falcon/gpt_model.py (105:114) - megatron_patch/model/qwen_vl/gpt_model.py (107:116) duplicated block id: 6070 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/chatglm/checkpoint_reshaping_and_interoperability.py (330:341) - toolkits/model_checkpoints_convertor/yi/checkpoint_reshaping_and_interoperability.py (408:419) duplicated block id: 6071 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/deepseek/hf2mcore_deepseek_v2_moe.py (237:249) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_dense_and_moe_gqa.py (341:353) duplicated block id: 6072 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/language_model.py (303:311) - megatron_patch/model/qwen1_5_megablocks/language_model.py (288:296) duplicated block id: 6073 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/gpt_model.py (105:114) - megatron_patch/model/qwen_vl/gpt_model.py (107:116) duplicated block id: 6074 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (781:788) - toolkits/model_checkpoints_convertor/starcoder/checkpoint_reshaping_and_interoperability.py (494:501) duplicated block id: 6075 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llava/hf2mcore_llava.py (443:450) - toolkits/model_checkpoints_convertor/utils/__init__.py (73:80) duplicated block id: 6076 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (97:108) - megatron_patch/model/llama3/transformer_legacy.py (102:113) duplicated block id: 6077 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/llava_mcore/llava_spec.py (6:31) - megatron_patch/model/llava_mcore/vision/vit_layer_specs.py (6:31) duplicated block id: 6078 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (1114:1121) - megatron_patch/model/qwen/transformer.py (1189:1196) duplicated block id: 6079 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/qwen1_5/transformer/attention.py (360:367) - megatron_patch/model/qwen1_5/transformer/attention.py (526:533) duplicated block id: 6080 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (153:162) - megatron_patch/model/glm130b/transformer.py (187:196) duplicated block id: 6081 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/language_model.py (506:513) - megatron_patch/model/mistral/language_model.py (523:530) duplicated block id: 6082 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon/checkpoint_reshaping_and_interoperability.py (68:75) - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (86:94) duplicated block id: 6083 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/hf2te.py (214:224) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (389:399) duplicated block id: 6084 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/galactica/transformer.py (324:331) - megatron_patch/model/starcoder/transformer.py (524:531) duplicated block id: 6085 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (67:74) - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (86:94) duplicated block id: 6086 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/galactica/language_model.py (318:326) - megatron_patch/model/mistral/language_model.py (290:298) duplicated block id: 6087 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/language_model.py (371:382) - megatron_patch/model/starcoder/language_model.py (365:376) duplicated block id: 6088 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (713:720) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (962:969) duplicated block id: 6089 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (418:425) - toolkits/model_checkpoints_convertor/chatglm/checkpoint_reshaping_and_interoperability.py (343:350) duplicated block id: 6090 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/galactica/language_model.py (629:639) - megatron_patch/model/mistral/language_model.py (569:579) duplicated block id: 6091 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/hf2te.py (332:339) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (382:391) duplicated block id: 6092 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/galactica/transformer.py (201:218) - megatron_patch/model/starcoder/transformer.py (295:312) duplicated block id: 6093 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/language_model.py (180:193) - megatron_patch/model/qwen/language_model.py (167:180) duplicated block id: 6094 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/galactica/gpt_model.py (105:114) - megatron_patch/model/llama2/gpt_model.py (106:115) duplicated block id: 6095 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon40b/checkpoint_reshaping_and_interoperability.py (435:445) - toolkits/model_checkpoints_convertor/galactica/checkpoint_reshaping_and_interoperability.py (487:497) duplicated block id: 6096 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (414:426) - toolkits/model_checkpoints_convertor/yi/checkpoint_reshaping_and_interoperability.py (513:525) duplicated block id: 6097 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/transformer.py (775:783) - megatron_patch/model/llama2/transformer.py (1605:1613) duplicated block id: 6098 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/multi_latent_attention.py (202:209) - megatron_patch/model/qwen2/transformer/attention.py (358:365) duplicated block id: 6099 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/glm/checkpoint_reshaping_and_interoperability.py (441:454) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (415:428) duplicated block id: 6100 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/llava_mcore/vision/vit_layer_specs.py (76:83) - megatron_patch/model/qwen3_moe/gpt_layer_specs.py (227:234) duplicated block id: 6101 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon/checkpoint_reshaping_and_interoperability.py (484:492) - toolkits/model_checkpoints_convertor/starcoder/checkpoint_reshaping_and_interoperability.py (703:711) duplicated block id: 6102 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore.py (305:312) - toolkits/model_checkpoints_convertor/mistral/hf2mcore.py (451:458) duplicated block id: 6103 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (76:87) - megatron_patch/model/llama2/transformer.py (67:79) duplicated block id: 6104 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/reward_model_to_megatron.py (702:710) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (583:591) duplicated block id: 6105 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/reward_model_to_megatron.py (136:146) - toolkits/model_checkpoints_convertor/glm130b/checkpoint_reshaping_and_interoperability.py (108:117) duplicated block id: 6106 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/transformer.py (553:566) - megatron_patch/model/mistral/transformer.py (689:702) duplicated block id: 6107 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_mha.py (314:323) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_dense_and_moe_gqa.py (896:905) duplicated block id: 6108 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/galactica/language_model.py (238:246) - megatron_patch/model/qwen2_vl/language_model_embedding.py (87:95) duplicated block id: 6109 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (1114:1121) - megatron_patch/model/llama3/transformer_legacy.py (1205:1212) duplicated block id: 6110 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (938:945) - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (781:788) duplicated block id: 6111 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (938:945) - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (829:836) duplicated block id: 6112 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/mistral/hf2mcore.py (245:252) - toolkits/model_checkpoints_convertor/mistral/hf2mcore.py (451:458) duplicated block id: 6113 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/transformer.py (916:923) - megatron_patch/model/falcon/transformer.py (1301:1308) duplicated block id: 6114 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (145:154) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (123:132) duplicated block id: 6115 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/llama3_1/model.py (164:180) - megatron_patch/model/qwen2/model.py (113:129) duplicated block id: 6116 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/galactica/transformer.py (779:787) - megatron_patch/model/starcoder/transformer.py (1201:1209) duplicated block id: 6117 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/falcon/transformer.py (547:554) - megatron_patch/model/galactica/transformer.py (324:331) duplicated block id: 6118 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/qwen1_5_megablocks/transformer.py (1085:1092) - megatron_patch/model/qwen1_5_megablocks/transformer.py (1602:1609) duplicated block id: 6119 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/mistral/hf2mcore.py (478:485) - toolkits/model_checkpoints_convertor/qwen/hf2megablocks_qwen1.5.py (581:588) duplicated block id: 6120 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (429:437) - megatron_patch/model/glm130b/transformer.py (464:472) duplicated block id: 6121 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (76:87) - megatron_patch/model/glm130b/transformer.py (65:76) duplicated block id: 6122 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (921:934) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (699:711) duplicated block id: 6123 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/transformer.py (1035:1043) - megatron_patch/model/qwen1_5_megablocks/transformer.py (1468:1476) duplicated block id: 6124 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/qwen2_vl/attention.py (448:455) - megatron_patch/model/qwen2_vl/attention_vision.py (152:159) duplicated block id: 6125 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (294:310) - megatron_patch/model/galactica/transformer.py (201:218) duplicated block id: 6126 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon40b/checkpoint_reshaping_and_interoperability.py (761:768) - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (781:788) duplicated block id: 6127 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/galactica/transformer.py (201:218) - megatron_patch/model/qwen/transformer.py (385:401) duplicated block id: 6128 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (116:123) - toolkits/model_checkpoints_convertor/llama/hf2mcore_llama3_1.py (80:87) duplicated block id: 6129 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/galactica/transformer.py (779:787) - megatron_patch/model/llava/transformer.py (1604:1612) duplicated block id: 6130 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (780:791) - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (701:714) duplicated block id: 6131 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/llava/transformer.py (1015:1024) - megatron_patch/model/mistral/transformer.py (1240:1247) duplicated block id: 6132 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/transformer.py (402:409) - megatron_patch/model/glm130b/transformer.py (506:513) duplicated block id: 6133 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/reward_model_to_megatron.py (768:776) - toolkits/model_checkpoints_convertor/glm130b/checkpoint_reshaping_and_interoperability.py (476:484) duplicated block id: 6134 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/llama2/transformer.py (339:347) - megatron_patch/model/starcoder/transformer.py (222:230) duplicated block id: 6135 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/galactica/checkpoint_reshaping_and_interoperability.py (517:525) - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (464:473) duplicated block id: 6136 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/glm/checkpoint_reshaping_and_interoperability.py (532:540) - toolkits/model_checkpoints_convertor/glm130b/checkpoint_reshaping_and_interoperability.py (489:497) duplicated block id: 6137 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/qwen_vl/language_model.py (335:342) - megatron_patch/model/starcoder/language_model.py (336:343) duplicated block id: 6138 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llava/hf2mcore_llava.py (103:110) - toolkits/model_checkpoints_convertor/qwen/hf2megablocks_qwen1.5.py (133:140) duplicated block id: 6139 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (62:74) - megatron_patch/model/bloom/transformer.py (76:87) duplicated block id: 6140 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/llama3_1/layer_specs.py (67:74) - megatron_patch/model/llava_mcore/llava_spec.py (63:71) duplicated block id: 6141 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer_legacy.py (1201:1208) - megatron_patch/model/mistral/transformer.py (1013:1022) duplicated block id: 6142 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (781:788) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (962:969) duplicated block id: 6143 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (921:934) - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (814:827) duplicated block id: 6144 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (716:723) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (962:969) duplicated block id: 6145 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/glm/checkpoint_reshaping_and_interoperability.py (441:454) - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (428:441) duplicated block id: 6146 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore.py (305:312) - toolkits/model_checkpoints_convertor/llama/hf2mcore.py (544:551) duplicated block id: 6147 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/transformer.py (466:474) - megatron_patch/model/glm130b/transformer.py (523:531) duplicated block id: 6148 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (464:473) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (539:547) duplicated block id: 6149 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer_legacy.py (966:975) - megatron_patch/model/qwen1_5_megablocks/transformer.py (1127:1134) duplicated block id: 6150 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/qwen/language_model.py (375:387) - megatron_patch/model/qwen1_5_megablocks/language_model.py (373:385) duplicated block id: 6151 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/galactica/checkpoint_reshaping_and_interoperability.py (517:525) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (382:391) duplicated block id: 6152 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llava/hf2mcore_llava.py (403:410) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_moe.py (651:658) duplicated block id: 6153 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/checkpoint_reshaping_and_interoperability.py (136:146) - toolkits/model_checkpoints_convertor/glm/checkpoint_reshaping_and_interoperability.py (108:117) duplicated block id: 6154 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/transformer.py (890:898) - megatron_patch/model/llama2/transformer.py (1605:1613) duplicated block id: 6155 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (620:632) - megatron_patch/model/falcon/transformer.py (744:754) duplicated block id: 6156 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/checkpoint_reshaping_and_interoperability.py (703:711) - toolkits/model_checkpoints_convertor/chatglm/checkpoint_reshaping_and_interoperability.py (453:461) duplicated block id: 6157 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/galactica/language_model.py (318:326) - megatron_patch/model/llava/language_model.py (292:300) duplicated block id: 6158 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/language_model.py (180:193) - megatron_patch/model/qwen_vl/language_model.py (168:181) duplicated block id: 6159 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/language_model.py (345:353) - megatron_patch/model/falcon/language_model.py (287:295) duplicated block id: 6160 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/moe/experts.py (50:62) - megatron_patch/model/qwen1_5/moe/experts.py (32:44) duplicated block id: 6161 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/llama2/transformer.py (798:815) - megatron_patch/model/qwen2/transformer/attention.py (221:238) duplicated block id: 6162 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/llama/transformer.py (945:953) - megatron_patch/model/llama2/transformer.py (1605:1613) duplicated block id: 6163 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/moe/experts.py (671:681) - megatron_patch/model/qwen2/moe/experts.py (295:305) duplicated block id: 6164 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/language_model.py (180:193) - megatron_patch/model/qwen1_5_megablocks/language_model.py (166:179) duplicated block id: 6165 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/hf2te.py (332:339) - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (395:404) duplicated block id: 6166 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/hf2te.py (336:346) - toolkits/model_checkpoints_convertor/galactica/checkpoint_reshaping_and_interoperability.py (487:497) duplicated block id: 6167 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/reward_model_to_megatron.py (813:822) - toolkits/model_checkpoints_convertor/chatglm/checkpoint_reshaping_and_interoperability.py (579:588) duplicated block id: 6168 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (883:892) - megatron_patch/model/baichuan/transformer.py (1110:1117) duplicated block id: 6169 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/deepseek/hf2mcore_deepseek_v2_moe.py (287:295) - toolkits/model_checkpoints_convertor/deepseek/hf2mcore_deepseek_v3_moe.py (384:392) duplicated block id: 6170 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_dense_and_moe_gqa.py (380:388) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_moe.py (410:418) duplicated block id: 6171 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (716:723) - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (938:945) duplicated block id: 6172 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (559:566) - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (614:621) duplicated block id: 6173 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/moe/experts.py (50:62) - megatron_patch/model/mixtral_bak/moe/experts.py (32:45) duplicated block id: 6174 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/llama2/transformer.py (1243:1250) - megatron_patch/model/qwen/transformer.py (961:970) duplicated block id: 6175 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon40b/checkpoint_reshaping_and_interoperability.py (459:467) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (382:391) duplicated block id: 6176 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/qwen2/transformer_block.py (251:258) - megatron_patch/model/qwen2/transformer_block.py (300:307) duplicated block id: 6177 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (509:517) - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (464:473) duplicated block id: 6178 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (495:503) - toolkits/model_checkpoints_convertor/bloom/reward_model_to_megatron.py (702:710) duplicated block id: 6179 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/llava_mcore/layer_specs.py (93:100) - megatron_patch/model/llava_mcore/vision/vit_layer_specs.py (50:57) duplicated block id: 6180 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/galactica/language_model.py (624:632) - megatron_patch/model/starcoder/language_model.py (481:489) duplicated block id: 6181 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (395:404) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (569:577) duplicated block id: 6182 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/transformer.py (268:284) - megatron_patch/model/starcoder/transformer.py (295:312) duplicated block id: 6183 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/moe/moe_layer.py (111:118) - megatron_patch/model/qwen1_5/moe/moe_layer.py (89:96) duplicated block id: 6184 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_mha.py (304:311) - toolkits/model_checkpoints_convertor/utils/__init__.py (73:80) duplicated block id: 6185 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/falcon/transformer.py (865:876) - megatron_patch/model/starcoder/transformer.py (814:825) duplicated block id: 6186 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer/attention.py (225:242) - megatron_patch/model/llava/transformer.py (797:814) duplicated block id: 6187 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/model.py (194:219) - megatron_patch/model/qwen2/model.py (174:195) duplicated block id: 6188 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/llama/language_model.py (611:621) - megatron_patch/model/llama3/language_model.py (534:544) duplicated block id: 6189 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer_legacy.py (787:794) - megatron_patch/model/mistral/transformer.py (819:826) duplicated block id: 6190 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (1104:1117) - megatron_patch/model/chatglm/transformer.py (820:833) duplicated block id: 6191 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (587:599) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (414:426) duplicated block id: 6192 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (442:449) - megatron_patch/model/starcoder/transformer.py (593:600) duplicated block id: 6193 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/language_model.py (334:341) - megatron_patch/model/llava/language_model.py (337:344) duplicated block id: 6194 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (666:673) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (962:969) duplicated block id: 6195 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon40b/checkpoint_reshaping_and_interoperability.py (459:467) - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (464:473) duplicated block id: 6196 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (433:444) - toolkits/model_checkpoints_convertor/chatglm/checkpoint_reshaping_and_interoperability.py (330:341) duplicated block id: 6197 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (998:1008) - megatron_patch/model/falcon/transformer.py (1105:1115) duplicated block id: 6198 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/falcon/transformer.py (663:670) - megatron_patch/model/glm130b/transformer.py (506:513) duplicated block id: 6199 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/language_model.py (345:353) - megatron_patch/model/falcon40b/language_model.py (287:295) duplicated block id: 6200 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/deepseek/hf2mcore_deepseek_v2_moe.py (517:526) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_dense_and_moe_gqa.py (896:905) duplicated block id: 6201 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llava/hf2mcore_llava.py (103:110) - toolkits/model_checkpoints_convertor/mistral/hf2mcore.py (107:114) duplicated block id: 6202 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/galactica/gpt_model.py (105:114) - megatron_patch/model/llava/gpt_model.py (107:116) duplicated block id: 6203 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon40b/checkpoint_reshaping_and_interoperability.py (257:275) - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (339:357) duplicated block id: 6204 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/bloom/language_model.py (352:359) - megatron_patch/model/llava/language_model.py (337:344) duplicated block id: 6205 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/llava/language_model.py (571:578) - megatron_patch/model/qwen/language_model.py (495:502) duplicated block id: 6206 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (815:822) - megatron_patch/model/baichuan2/transformer.py (916:923) duplicated block id: 6207 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/glm130b/checkpoint_reshaping_and_interoperability.py (422:435) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (415:428) duplicated block id: 6208 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/mixtral_bak/transformer/mlp.py (122:130) - megatron_patch/model/qwen2/transformer/mlp.py (168:176) duplicated block id: 6209 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (688:700) - megatron_patch/model/galactica/transformer.py (607:616) duplicated block id: 6210 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/mixtral_bak/moe/moe_layer.py (38:45) - megatron_patch/model/qwen1_5/moe/moe_layer.py (43:50) duplicated block id: 6211 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/checkpoint_reshaping_and_interoperability.py (121:128) - toolkits/model_checkpoints_convertor/llama/hf2mcore_llama3_1.py (80:87) duplicated block id: 6212 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/checkpoint_reshaping_and_interoperability.py (782:790) - toolkits/model_checkpoints_convertor/starcoder/checkpoint_reshaping_and_interoperability.py (797:805) duplicated block id: 6213 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/language_model.py (600:610) - megatron_patch/model/qwen1_5_megablocks/language_model.py (556:566) duplicated block id: 6214 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer_legacy.py (1148:1157) - megatron_patch/model/qwen1_5_megablocks/transformer.py (1081:1090) duplicated block id: 6215 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore_llama3_1.py (80:87) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (171:178) duplicated block id: 6216 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_mha.py (214:221) - toolkits/model_checkpoints_convertor/qwen/hf2megablocks_qwen1.5.py (480:487) duplicated block id: 6217 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (999:1006) - megatron_patch/model/qwen/transformer.py (1074:1081) duplicated block id: 6218 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (255:273) - megatron_patch/model/starcoder/transformer.py (295:312) duplicated block id: 6219 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/qwen1_5/layer_specs.py (55:62) - megatron_patch/model/qwen3_moe/gpt_layer_specs.py (130:137) duplicated block id: 6220 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/llama/language_model.py (207:215) - megatron_patch/model/qwen2_vl/language_model_embedding.py (87:95) duplicated block id: 6221 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (815:822) - megatron_patch/model/llava/transformer.py (952:959) duplicated block id: 6222 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore.py (139:146) - toolkits/model_checkpoints_convertor/llava/hf2mcore_llava.py (103:110) duplicated block id: 6223 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/mistral/transformer.py (1333:1356) - megatron_patch/model/starcoder/transformer.py (1059:1082) duplicated block id: 6224 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/qwen1_5_megablocks/transformer.py (689:706) - megatron_patch/model/qwen2/transformer/attention.py (221:238) duplicated block id: 6225 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/hf2te.py (198:208) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (389:399) duplicated block id: 6226 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (1050:1058) - megatron_patch/model/qwen_vl/transformer.py (1604:1612) duplicated block id: 6227 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/gpt_model.py (115:125) - megatron_patch/model/falcon/gpt_model.py (105:114) duplicated block id: 6228 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (780:791) - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (814:827) duplicated block id: 6229 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/model.py (154:189) - megatron_patch/model/qwen2_vl/gpt_model.py (164:185) duplicated block id: 6230 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/qwen/transformer.py (1278:1301) - megatron_patch/model/starcoder/transformer.py (1059:1082) duplicated block id: 6231 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (1114:1121) - megatron_patch/model/mistral/transformer.py (1244:1251) duplicated block id: 6232 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/galactica/checkpoint_reshaping_and_interoperability.py (487:497) - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (538:548) duplicated block id: 6233 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/falcon/transformer.py (428:439) - megatron_patch/model/llama/transformer.py (244:256) duplicated block id: 6234 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon/checkpoint_reshaping_and_interoperability.py (259:277) - toolkits/model_checkpoints_convertor/yi/checkpoint_reshaping_and_interoperability.py (288:306) duplicated block id: 6235 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/llava/transformer.py (1200:1207) - megatron_patch/model/mistral/transformer.py (1737:1744) duplicated block id: 6236 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (97:108) - megatron_patch/model/llava/transformer.py (98:109) duplicated block id: 6237 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/moe/experts.py (50:62) - megatron_patch/model/qwen2/moe/experts.py (52:64) duplicated block id: 6238 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (797:804) - megatron_patch/model/qwen_vl/transformer.py (939:946) duplicated block id: 6239 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/qwen/transformer.py (1185:1192) - megatron_patch/model/qwen1_5_megablocks/transformer.py (900:909) duplicated block id: 6240 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore.py (471:479) - toolkits/model_checkpoints_convertor/mistral/hf2mcore.py (373:381) duplicated block id: 6241 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/falcon/gpt_model.py (105:114) - megatron_patch/model/llava/gpt_model.py (107:116) duplicated block id: 6242 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer_legacy.py (787:794) - megatron_patch/model/qwen1_5_megablocks/transformer.py (713:720) duplicated block id: 6243 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/transformer.py (1035:1043) - megatron_patch/model/llava/transformer.py (1604:1612) duplicated block id: 6244 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (481:489) - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (395:404) duplicated block id: 6245 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (1104:1117) - megatron_patch/model/falcon40b/transformer.py (938:951) duplicated block id: 6246 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (477:485) - megatron_patch/model/falcon40b/transformer.py (466:474) duplicated block id: 6247 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/llava/transformer.py (1242:1249) - megatron_patch/model/qwen_vl/transformer.py (1015:1024) duplicated block id: 6248 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/model.py (210:220) - megatron_patch/model/qwen2/model.py (208:218) duplicated block id: 6249 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/moe/router.py (155:162) - megatron_patch/model/mixtral/moe/router.py (283:290) duplicated block id: 6250 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/llama/language_model.py (287:295) - megatron_patch/model/qwen_vl/language_model.py (290:298) duplicated block id: 6251 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/mistral/transformer.py (1198:1205) - megatron_patch/model/mistral/transformer.py (1737:1744) duplicated block id: 6252 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/transformer.py (553:566) - megatron_patch/model/llama2/transformer.py (690:703) duplicated block id: 6253 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (116:123) - toolkits/model_checkpoints_convertor/llava/hf2mcore_llava.py (103:110) duplicated block id: 6254 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/bloom/language_model.py (156:165) - megatron_patch/model/starcoder/language_model.py (143:152) duplicated block id: 6255 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_moe.py (69:81) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_dense_and_moe_gqa.py (66:77) duplicated block id: 6256 size: 8 cleaned lines of code in 2 files: - toolkits/distributed_checkpoints_convertor/impl/general/h2m_synchronizer.py (234:246) - toolkits/distributed_checkpoints_convertor/impl/general/m2h_synchronizer.py (238:250) duplicated block id: 6257 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (457:467) - toolkits/model_checkpoints_convertor/galactica/checkpoint_reshaping_and_interoperability.py (487:497) duplicated block id: 6258 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/falcon/transformer.py (1214:1227) - megatron_patch/model/llama/transformer.py (1007:1020) duplicated block id: 6259 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/llava/transformer.py (1015:1024) - megatron_patch/model/qwen/transformer.py (1185:1192) duplicated block id: 6260 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/qwen2_vl/attention.py (448:455) - megatron_patch/model/qwen2_vl/attention_vision.py (457:464) duplicated block id: 6261 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/hf2te.py (422:430) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (586:594) duplicated block id: 6262 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer_legacy.py (338:346) - megatron_patch/model/starcoder/transformer.py (222:230) duplicated block id: 6263 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (1139:1156) - megatron_patch/model/glm130b/transformer.py (1170:1187) duplicated block id: 6264 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/gpt_model.py (115:125) - megatron_patch/model/llama/gpt_model.py (105:114) duplicated block id: 6265 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/llama3_1/layer_specs.py (56:63) - megatron_patch/model/qwen3_moe/gpt_layer_specs.py (130:137) duplicated block id: 6266 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/transformer.py (412:419) - megatron_patch/model/starcoder/transformer.py (524:531) duplicated block id: 6267 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (797:804) - megatron_patch/model/qwen/transformer.py (885:892) duplicated block id: 6268 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (1605:1613) - megatron_patch/model/llama/transformer.py (945:953) duplicated block id: 6269 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (626:633) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (632:639) duplicated block id: 6270 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/model.py (194:219) - megatron_patch/model/mixtral_bak/model.py (149:184) duplicated block id: 6271 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (1243:1250) - megatron_patch/model/qwen_vl/transformer.py (1015:1024) duplicated block id: 6272 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/llama3/language_model.py (490:497) - megatron_patch/model/qwen1_5_megablocks/language_model.py (510:517) duplicated block id: 6273 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/chatglm/checkpoint_reshaping_and_interoperability.py (330:341) - toolkits/model_checkpoints_convertor/falcon/checkpoint_reshaping_and_interoperability.py (395:406) duplicated block id: 6274 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (97:108) - megatron_patch/model/baichuan2/transformer.py (93:104) duplicated block id: 6275 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (1114:1121) - megatron_patch/model/llama2/transformer.py (1247:1254) duplicated block id: 6276 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/falcon/language_model.py (600:610) - megatron_patch/model/llava/language_model.py (617:627) duplicated block id: 6277 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/deepseek/hf2mcore_deepseek_v2_moe.py (107:115) - toolkits/model_checkpoints_convertor/deepseek/hf2mcore_deepseek_v3_moe.py (170:178) duplicated block id: 6278 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/hf2te.py (345:353) - toolkits/model_checkpoints_convertor/glm/checkpoint_reshaping_and_interoperability.py (430:438) duplicated block id: 6279 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (1470:1478) - megatron_patch/model/glm130b/transformer.py (1035:1043) duplicated block id: 6280 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/falcon/transformer.py (1105:1115) - megatron_patch/model/starcoder/transformer.py (1159:1169) duplicated block id: 6281 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/llama/transformer.py (945:953) - megatron_patch/model/llama3/transformer_legacy.py (1565:1573) duplicated block id: 6282 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer/attention.py (526:533) - megatron_patch/model/qwen1_5/transformer/attention.py (360:367) duplicated block id: 6283 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (668:675) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (895:902) duplicated block id: 6284 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/llama2/transformer.py (1201:1208) - megatron_patch/model/qwen_vl/transformer.py (1739:1746) duplicated block id: 6285 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/llama/language_model.py (334:341) - megatron_patch/model/llava/language_model.py (337:344) duplicated block id: 6286 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/hf2te.py (345:353) - toolkits/model_checkpoints_convertor/glm130b/checkpoint_reshaping_and_interoperability.py (411:419) duplicated block id: 6287 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore.py (459:466) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_mha_to_moe.py (128:135) duplicated block id: 6288 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/language_model.py (177:190) - megatron_patch/model/glm130b/language_model.py (180:193) duplicated block id: 6289 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/language_model.py (600:610) - megatron_patch/model/llama3/language_model.py (534:544) duplicated block id: 6290 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/mistral/transformer.py (1240:1247) - megatron_patch/model/qwen1_5_megablocks/transformer.py (900:909) duplicated block id: 6291 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/mixtral_bak/transformer/attention.py (396:411) - megatron_patch/model/qwen2_vl/attention.py (661:676) duplicated block id: 6292 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (998:1008) - megatron_patch/model/falcon40b/transformer.py (844:854) duplicated block id: 6293 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/transformer.py (775:783) - megatron_patch/model/qwen1_5_megablocks/transformer.py (1468:1476) duplicated block id: 6294 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (1203:1226) - megatron_patch/model/starcoder/transformer.py (1059:1082) duplicated block id: 6295 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/falcon/transformer.py (1188:1195) - megatron_patch/model/starcoder/transformer.py (1243:1250) duplicated block id: 6296 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/llama2/transformer.py (1201:1208) - megatron_patch/model/qwen1_5_megablocks/transformer.py (1602:1609) duplicated block id: 6297 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (551:558) - megatron_patch/model/starcoder/transformer.py (593:600) duplicated block id: 6298 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/llama/transformer.py (945:953) - megatron_patch/model/llava/transformer.py (1604:1612) duplicated block id: 6299 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/qwen2_vl/attention_vision.py (152:159) - megatron_patch/model/qwen2_vl/attention_vision.py (447:454) duplicated block id: 6300 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/qwen2_vl/attention_vision.py (152:159) - megatron_patch/model/qwen2_vl/attention_vision.py (457:464) duplicated block id: 6301 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/transformer.py (553:566) - megatron_patch/model/llama3/transformer_legacy.py (668:681) duplicated block id: 6302 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore_llama3_1.py (186:193) - toolkits/model_checkpoints_convertor/llama/hf2mcore_llama3_1.py (606:613) duplicated block id: 6303 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (1018:1027) - megatron_patch/model/qwen_vl/transformer.py (1242:1249) duplicated block id: 6304 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/bloom/language_model.py (305:313) - megatron_patch/model/qwen1_5_megablocks/language_model.py (288:296) duplicated block id: 6305 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/galactica/checkpoint_reshaping_and_interoperability.py (531:539) - toolkits/model_checkpoints_convertor/glm/checkpoint_reshaping_and_interoperability.py (430:438) duplicated block id: 6306 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/qwen2_vl/attention.py (458:465) - megatron_patch/model/qwen2_vl/attention_vision.py (447:454) duplicated block id: 6307 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (1018:1027) - megatron_patch/model/qwen1_5_megablocks/transformer.py (1127:1134) duplicated block id: 6308 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/gpt_model.py (115:126) - megatron_patch/model/starcoder/gpt_model.py (116:127) duplicated block id: 6309 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/bloom/language_model.py (305:313) - megatron_patch/model/qwen/language_model.py (289:297) duplicated block id: 6310 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/transformer.py (506:513) - megatron_patch/model/llama/transformer.py (463:470) duplicated block id: 6311 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/language_model.py (579:589) - megatron_patch/model/llama3/language_model.py (534:544) duplicated block id: 6312 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (497:510) - toolkits/model_checkpoints_convertor/starcoder/checkpoint_reshaping_and_interoperability.py (714:727) duplicated block id: 6313 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/llama/gpt_model.py (105:114) - megatron_patch/model/llava/gpt_model.py (107:116) duplicated block id: 6314 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/chatglm/checkpoint_reshaping_and_interoperability.py (330:341) - toolkits/model_checkpoints_convertor/galactica/checkpoint_reshaping_and_interoperability.py (432:443) duplicated block id: 6315 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (414:426) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (594:606) duplicated block id: 6316 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/qwen/transformer.py (392:411) - megatron_patch/model/starcoder/transformer.py (425:445) duplicated block id: 6317 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/checkpoint_reshaping_and_interoperability.py (814:823) - toolkits/model_checkpoints_convertor/galactica/checkpoint_reshaping_and_interoperability.py (649:658) duplicated block id: 6318 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/language_model.py (287:295) - megatron_patch/model/qwen_vl/language_model.py (290:298) duplicated block id: 6319 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (87:94) - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (86:94) duplicated block id: 6320 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/hf2te.py (373:381) - toolkits/model_checkpoints_convertor/glm/checkpoint_reshaping_and_interoperability.py (430:438) duplicated block id: 6321 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/llava/transformer.py (540:547) - megatron_patch/model/starcoder/transformer.py (593:600) duplicated block id: 6322 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (97:108) - megatron_patch/model/mistral/transformer.py (98:109) duplicated block id: 6323 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/language_model.py (506:513) - megatron_patch/model/llama2/language_model.py (511:518) duplicated block id: 6324 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer_legacy.py (1148:1157) - megatron_patch/model/mistral/transformer.py (1194:1203) duplicated block id: 6325 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/mixtral_bak/transformer/mlp.py (122:130) - megatron_patch/model/qwen1_5/transformer/mlp.py (165:173) duplicated block id: 6326 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (76:87) - megatron_patch/model/qwen/transformer.py (62:74) duplicated block id: 6327 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon/checkpoint_reshaping_and_interoperability.py (777:784) - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (668:675) duplicated block id: 6328 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/llama3/layer_specs.py (55:62) - megatron_patch/model/mixtral_bak/layer_specs.py (67:74) duplicated block id: 6329 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon/checkpoint_reshaping_and_interoperability.py (777:784) - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (716:723) duplicated block id: 6330 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/qwen/transformer.py (961:970) - megatron_patch/model/qwen_vl/transformer.py (1242:1249) duplicated block id: 6331 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (781:788) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (713:720) duplicated block id: 6332 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/hf2te.py (357:370) - toolkits/model_checkpoints_convertor/bloom/checkpoint_reshaping_and_interoperability.py (714:727) duplicated block id: 6333 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (1018:1027) - megatron_patch/model/qwen/transformer.py (1185:1192) duplicated block id: 6334 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/falcon/transformer.py (556:563) - megatron_patch/model/falcon40b/transformer.py (382:389) duplicated block id: 6335 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/galactica/language_model.py (402:413) - megatron_patch/model/starcoder/language_model.py (365:376) duplicated block id: 6336 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/checkpoint_reshaping_and_interoperability.py (703:711) - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (576:584) duplicated block id: 6337 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/language_model.py (287:295) - megatron_patch/model/llama3/language_model.py (284:292) duplicated block id: 6338 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (775:792) - megatron_patch/model/mixtral_bak/transformer/attention.py (199:216) duplicated block id: 6339 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (664:681) - megatron_patch/model/llama3/transformer/attention.py (225:242) duplicated block id: 6340 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/chatglm/checkpoint_reshaping_and_interoperability.py (343:350) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (506:513) duplicated block id: 6341 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/transformer.py (553:566) - megatron_patch/model/qwen_vl/transformer.py (689:702) duplicated block id: 6342 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/llama3/language_model.py (357:364) - megatron_patch/model/mistral/language_model.py (364:371) duplicated block id: 6343 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/hf2te.py (373:381) - toolkits/model_checkpoints_convertor/bloom/checkpoint_reshaping_and_interoperability.py (703:711) duplicated block id: 6344 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/falcon/transformer.py (1151:1159) - megatron_patch/model/qwen/transformer.py (1547:1555) duplicated block id: 6345 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/bloom/language_model.py (305:313) - megatron_patch/model/mistral/language_model.py (290:298) duplicated block id: 6346 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/moe/experts.py (641:648) - megatron_patch/model/qwen1_5/transformer/mlp.py (125:132) duplicated block id: 6347 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/gpt_model.py (113:124) - megatron_patch/model/starcoder/gpt_model.py (116:127) duplicated block id: 6348 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/bloom/language_model.py (381:392) - megatron_patch/model/starcoder/language_model.py (365:376) duplicated block id: 6349 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (255:273) - megatron_patch/model/qwen/transformer.py (385:401) duplicated block id: 6350 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (652:660) - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (768:778) duplicated block id: 6351 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/mistral/transformer.py (883:894) - megatron_patch/model/starcoder/transformer.py (814:825) duplicated block id: 6352 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (481:489) - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (464:473) duplicated block id: 6353 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/gpt_model.py (115:125) - megatron_patch/model/galactica/gpt_model.py (105:114) duplicated block id: 6354 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (523:531) - toolkits/model_checkpoints_convertor/bloom/checkpoint_reshaping_and_interoperability.py (703:711) duplicated block id: 6355 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/language_model.py (383:394) - megatron_patch/model/starcoder/language_model.py (365:376) duplicated block id: 6356 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/falcon/transformer.py (647:655) - megatron_patch/model/llama/transformer.py (439:447) duplicated block id: 6357 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/llama/transformer.py (945:953) - megatron_patch/model/qwen/transformer.py (1547:1555) duplicated block id: 6358 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/transformer.py (268:284) - megatron_patch/model/llama3/transformer_legacy.py (390:406) duplicated block id: 6359 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (86:94) - toolkits/model_checkpoints_convertor/qwen/hf2megablocks_qwen1.5.py (99:107) duplicated block id: 6360 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/llama/language_model.py (287:295) - megatron_patch/model/llama2/language_model.py (289:297) duplicated block id: 6361 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/qwen_vl/transformer.py (339:347) - megatron_patch/model/starcoder/transformer.py (222:230) duplicated block id: 6362 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (294:310) - megatron_patch/model/glm130b/transformer.py (268:284) duplicated block id: 6363 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/model.py (194:219) - megatron_patch/model/qwen2_vl/gpt_model.py (164:185) duplicated block id: 6364 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/language_model.py (532:539) - megatron_patch/model/llava/language_model.py (472:479) duplicated block id: 6365 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/chatglm/checkpoint_reshaping_and_interoperability.py (330:341) - toolkits/model_checkpoints_convertor/falcon40b/checkpoint_reshaping_and_interoperability.py (384:395) duplicated block id: 6366 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/language_model.py (590:597) - megatron_patch/model/llama2/language_model.py (511:518) duplicated block id: 6367 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/transformer.py (775:783) - megatron_patch/model/mistral/transformer.py (1602:1610) duplicated block id: 6368 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/mistral/hf2mcore.py (73:81) - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (86:94) duplicated block id: 6369 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (1243:1250) - megatron_patch/model/qwen/transformer.py (961:970) duplicated block id: 6370 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/galactica/language_model.py (318:326) - megatron_patch/model/llama2/language_model.py (289:297) duplicated block id: 6371 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llava/hf2mcore_llava.py (103:110) - toolkits/model_checkpoints_convertor/yi/checkpoint_reshaping_and_interoperability.py (143:150) duplicated block id: 6372 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (534:546) - toolkits/model_checkpoints_convertor/baichuan/hf2te.py (384:396) duplicated block id: 6373 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (363:373) - megatron_patch/model/qwen_vl/transformer.py (366:376) duplicated block id: 6374 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/glm130b/checkpoint_reshaping_and_interoperability.py (411:419) - toolkits/model_checkpoints_convertor/yi/checkpoint_reshaping_and_interoperability.py (502:510) duplicated block id: 6375 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_mha_to_moe.py (218:225) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_dense_and_moe_gqa.py (741:748) duplicated block id: 6376 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/language_model.py (579:589) - megatron_patch/model/qwen/language_model.py (539:549) duplicated block id: 6377 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/hf2te.py (394:402) - toolkits/model_checkpoints_convertor/galactica/checkpoint_reshaping_and_interoperability.py (565:573) duplicated block id: 6378 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/falcon/gpt_model.py (105:114) - megatron_patch/model/llama2/gpt_model.py (106:115) duplicated block id: 6379 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/mixtral_bak/layer_specs.py (67:74) - megatron_patch/model/qwen2/layer_specs.py (65:72) duplicated block id: 6380 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (363:373) - megatron_patch/model/qwen1_5_megablocks/transformer.py (257:267) duplicated block id: 6381 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/language_model.py (180:193) - megatron_patch/model/llama2/language_model.py (167:180) duplicated block id: 6382 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/language_model.py (299:307) - megatron_patch/model/bloom/language_model.py (305:313) duplicated block id: 6383 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/hf2te.py (452:461) - toolkits/model_checkpoints_convertor/galactica/checkpoint_reshaping_and_interoperability.py (603:612) duplicated block id: 6384 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_moe.py (413:420) - toolkits/model_checkpoints_convertor/utils/__init__.py (73:80) duplicated block id: 6385 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/bloom/language_model.py (352:359) - megatron_patch/model/qwen_vl/language_model.py (335:342) duplicated block id: 6386 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_mha.py (85:95) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_dense_and_moe_gqa.py (97:106) duplicated block id: 6387 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/multi_latent_attention.py (216:223) - megatron_patch/model/deepseek_v2/multi_latent_attention.py (242:249) duplicated block id: 6388 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (85:92) - toolkits/model_checkpoints_convertor/llama/hf2mcore.py (119:127) duplicated block id: 6389 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/falcon/transformer.py (1301:1308) - megatron_patch/model/galactica/transformer.py (924:931) duplicated block id: 6390 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/falcon/language_model.py (600:610) - megatron_patch/model/mistral/language_model.py (569:579) duplicated block id: 6391 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (1336:1359) - megatron_patch/model/starcoder/transformer.py (1059:1082) duplicated block id: 6392 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (716:723) - toolkits/model_checkpoints_convertor/starcoder/checkpoint_reshaping_and_interoperability.py (494:501) duplicated block id: 6393 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (523:531) - toolkits/model_checkpoints_convertor/starcoder/checkpoint_reshaping_and_interoperability.py (703:711) duplicated block id: 6394 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/llama3/model.py (207:217) - megatron_patch/model/mixtral/model.py (210:220) duplicated block id: 6395 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/mistral/language_model.py (523:530) - megatron_patch/model/qwen/language_model.py (495:502) duplicated block id: 6396 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/galactica/checkpoint_reshaping_and_interoperability.py (603:612) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (624:633) duplicated block id: 6397 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon40b/checkpoint_reshaping_and_interoperability.py (86:93) - toolkits/model_checkpoints_convertor/llama/hf2mcore.py (119:127) duplicated block id: 6398 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/language_model.py (590:597) - megatron_patch/model/mistral/language_model.py (523:530) duplicated block id: 6399 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/llava_mcore/vision/vit_layer_specs.py (50:57) - megatron_patch/model/qwen2_vl/layer_specs.py (56:63) duplicated block id: 6400 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/chatglm/checkpoint_reshaping_and_interoperability.py (453:461) - toolkits/model_checkpoints_convertor/glm/checkpoint_reshaping_and_interoperability.py (430:438) duplicated block id: 6401 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/llava_mcore/vision/vit_layer_specs.py (50:57) - megatron_patch/model/qwen2_vl/layer_specs.py (89:96) duplicated block id: 6402 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/transformer.py (554:561) - megatron_patch/model/glm130b/transformer.py (758:765) duplicated block id: 6403 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (86:94) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (95:102) duplicated block id: 6404 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/galactica/transformer.py (201:218) - megatron_patch/model/qwen1_5_megablocks/transformer.py (282:298) duplicated block id: 6405 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/bloom/gpt_model.py (48:58) - megatron_patch/model/falcon40b/gpt_model.py (50:60) duplicated block id: 6406 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/galactica/checkpoint_reshaping_and_interoperability.py (487:497) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (515:525) duplicated block id: 6407 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/language_model.py (180:193) - megatron_patch/model/llava/language_model.py (170:183) duplicated block id: 6408 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/llama3/layer_specs.py (67:74) - megatron_patch/model/llava_mcore/llava_spec.py (63:71) duplicated block id: 6409 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/gpt_model.py (105:114) - megatron_patch/model/llava/gpt_model.py (107:116) duplicated block id: 6410 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore_llama3_1.py (80:87) - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (166:173) duplicated block id: 6411 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (793:800) - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (716:723) duplicated block id: 6412 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (793:800) - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (668:675) duplicated block id: 6413 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/llava/transformer.py (1200:1207) - megatron_patch/model/qwen_vl/transformer.py (1739:1746) duplicated block id: 6414 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (775:792) - megatron_patch/model/qwen2/transformer/attention.py (221:238) duplicated block id: 6415 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (86:94) - toolkits/model_checkpoints_convertor/yi/checkpoint_reshaping_and_interoperability.py (68:75) duplicated block id: 6416 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/language_model.py (345:353) - megatron_patch/model/galactica/language_model.py (318:326) duplicated block id: 6417 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/hf2te.py (357:370) - toolkits/model_checkpoints_convertor/bloom/reward_model_to_megatron.py (713:726) duplicated block id: 6418 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/reward_model_to_megatron.py (713:726) - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (497:510) duplicated block id: 6419 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/galactica/language_model.py (365:372) - megatron_patch/model/llava/language_model.py (337:344) duplicated block id: 6420 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (688:700) - megatron_patch/model/llama/transformer.py (712:721) duplicated block id: 6421 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/checkpoint_reshaping_and_interoperability.py (814:823) - toolkits/model_checkpoints_convertor/yi/checkpoint_reshaping_and_interoperability.py (666:674) duplicated block id: 6422 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/model.py (100:111) - megatron_patch/model/qwen2/model.py (105:116) duplicated block id: 6423 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (378:385) - megatron_patch/model/starcoder/transformer.py (524:531) duplicated block id: 6424 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (999:1006) - megatron_patch/model/llava/transformer.py (1128:1135) duplicated block id: 6425 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/llama/gpt_model.py (105:114) - megatron_patch/model/mistral/gpt_model.py (106:115) duplicated block id: 6426 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/transformer.py (890:898) - megatron_patch/model/qwen/transformer.py (1547:1555) duplicated block id: 6427 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (668:675) - toolkits/model_checkpoints_convertor/starcoder/checkpoint_reshaping_and_interoperability.py (494:501) duplicated block id: 6428 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/multi_latent_attention.py (202:209) - megatron_patch/model/mixtral_bak/transformer/attention.py (335:342) duplicated block id: 6429 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/transformer.py (758:765) - megatron_patch/model/starcoder/transformer.py (871:878) duplicated block id: 6430 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (781:788) - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (829:836) duplicated block id: 6431 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/transformer.py (835:848) - megatron_patch/model/falcon/transformer.py (1214:1227) duplicated block id: 6432 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (76:87) - megatron_patch/model/llama3/transformer_legacy.py (71:83) duplicated block id: 6433 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/transformer.py (916:923) - megatron_patch/model/falcon40b/transformer.py (1040:1047) duplicated block id: 6434 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (829:836) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (666:673) duplicated block id: 6435 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/reward_model_to_megatron.py (121:128) - toolkits/model_checkpoints_convertor/llava/hf2mcore_llava.py (103:110) duplicated block id: 6436 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/checkpoint_reshaping_and_interoperability.py (703:711) - toolkits/model_checkpoints_convertor/yi/checkpoint_reshaping_and_interoperability.py (502:510) duplicated block id: 6437 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (506:518) - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (496:508) duplicated block id: 6438 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/falcon/transformer.py (547:554) - megatron_patch/model/falcon40b/transformer.py (374:381) duplicated block id: 6439 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/chatglm/checkpoint_reshaping_and_interoperability.py (75:82) - toolkits/model_checkpoints_convertor/llama/hf2mcore.py (119:127) duplicated block id: 6440 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (167:176) - toolkits/model_checkpoints_convertor/falcon40b/checkpoint_reshaping_and_interoperability.py (168:177) duplicated block id: 6441 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/galactica/language_model.py (629:639) - megatron_patch/model/qwen_vl/language_model.py (586:596) duplicated block id: 6442 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (395:404) - toolkits/model_checkpoints_convertor/yi/checkpoint_reshaping_and_interoperability.py (484:492) duplicated block id: 6443 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/bloom/gpt_model.py (48:58) - megatron_patch/model/galactica/gpt_model.py (50:60) duplicated block id: 6444 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (713:720) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (895:902) duplicated block id: 6445 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/qwen1_5_megablocks/transformer.py (777:788) - megatron_patch/model/starcoder/transformer.py (814:825) duplicated block id: 6446 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/falcon/language_model.py (600:610) - megatron_patch/model/qwen/language_model.py (539:549) duplicated block id: 6447 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/chatglm/checkpoint_reshaping_and_interoperability.py (330:341) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (493:504) duplicated block id: 6448 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer_legacy.py (966:975) - megatron_patch/model/llava/transformer.py (1242:1249) duplicated block id: 6449 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/multi_latent_attention.py (202:209) - megatron_patch/model/llama3/transformer/attention.py (362:369) duplicated block id: 6450 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/llava/transformer.py (797:814) - megatron_patch/model/qwen1_5/transformer/attention.py (225:242) duplicated block id: 6451 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer_legacy.py (762:779) - megatron_patch/model/qwen1_5/transformer/attention.py (225:242) duplicated block id: 6452 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/hf2te.py (360:367) - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (464:473) duplicated block id: 6453 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon40b/checkpoint_reshaping_and_interoperability.py (257:275) - toolkits/model_checkpoints_convertor/yi/checkpoint_reshaping_and_interoperability.py (288:306) duplicated block id: 6454 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/checkpoint_reshaping_and_interoperability.py (703:711) - toolkits/model_checkpoints_convertor/falcon40b/checkpoint_reshaping_and_interoperability.py (473:481) duplicated block id: 6455 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (534:546) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (414:426) duplicated block id: 6456 size: 8 cleaned lines of code in 2 files: - megatron_patch/tokenizer/tokenization_baichuan.py (32:40) - megatron_patch/tokenizer/tokenization_yi.py (29:37) duplicated block id: 6457 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/bloom/language_model.py (209:217) - megatron_patch/model/qwen2_vl/language_model_embedding.py (87:95) duplicated block id: 6458 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (872:881) - megatron_patch/model/bloom/transformer.py (828:836) duplicated block id: 6459 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/checkpoint_reshaping_and_interoperability.py (136:146) - toolkits/model_checkpoints_convertor/glm130b/checkpoint_reshaping_and_interoperability.py (108:117) duplicated block id: 6460 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (336:344) - megatron_patch/model/starcoder/transformer.py (222:230) duplicated block id: 6461 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (1114:1121) - megatron_patch/model/qwen1_5_megablocks/transformer.py (1131:1138) duplicated block id: 6462 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/falcon/transformer.py (556:563) - megatron_patch/model/llama/transformer.py (370:377) duplicated block id: 6463 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/transformer.py (927:934) - megatron_patch/model/starcoder/transformer.py (1243:1250) duplicated block id: 6464 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (142:149) - toolkits/model_checkpoints_convertor/llava/hf2mcore_llava.py (103:110) duplicated block id: 6465 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/mistral/transformer.py (399:418) - megatron_patch/model/starcoder/transformer.py (425:445) duplicated block id: 6466 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/chatglm/checkpoint_reshaping_and_interoperability.py (57:64) - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (86:94) duplicated block id: 6467 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/bloom/language_model.py (305:313) - megatron_patch/model/llama2/language_model.py (289:297) duplicated block id: 6468 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/hf2te.py (345:353) - toolkits/model_checkpoints_convertor/starcoder/checkpoint_reshaping_and_interoperability.py (703:711) duplicated block id: 6469 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (1050:1058) - megatron_patch/model/qwen1_5_megablocks/transformer.py (1468:1476) duplicated block id: 6470 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/language_model.py (132:153) - megatron_patch/model/starcoder/language_model.py (122:142) duplicated block id: 6471 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/checkpoint_reshaping_and_interoperability.py (703:711) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (583:591) duplicated block id: 6472 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (142:149) - toolkits/model_checkpoints_convertor/llama/hf2mcore_llama3_1.py (80:87) duplicated block id: 6473 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/hf2te.py (356:368) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (594:606) duplicated block id: 6474 size: 8 cleaned lines of code in 2 files: - megatron_patch/data/utils.py (164:172) - megatron_patch/data/utils.py (348:356) duplicated block id: 6475 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/hf2te.py (385:398) - toolkits/model_checkpoints_convertor/starcoder/checkpoint_reshaping_and_interoperability.py (714:727) duplicated block id: 6476 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer_legacy.py (1201:1208) - megatron_patch/model/qwen/transformer.py (961:970) duplicated block id: 6477 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (154:163) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (190:199) duplicated block id: 6478 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (822:829) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (666:673) duplicated block id: 6479 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (822:829) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (713:720) duplicated block id: 6480 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/glm/checkpoint_reshaping_and_interoperability.py (430:438) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (583:591) duplicated block id: 6481 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (614:621) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (621:628) duplicated block id: 6482 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/llama2/transformer.py (1243:1250) - megatron_patch/model/llava/transformer.py (1015:1024) duplicated block id: 6483 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/mistral/transformer.py (399:418) - megatron_patch/model/starcoder/transformer.py (302:322) duplicated block id: 6484 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/gpt_model.py (68:75) - megatron_patch/model/llama/gpt_model.py (71:78) duplicated block id: 6485 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/language_model.py (299:307) - megatron_patch/model/falcon40b/language_model.py (287:295) duplicated block id: 6486 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/glm130b/checkpoint_reshaping_and_interoperability.py (411:419) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (583:591) duplicated block id: 6487 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/hf2te.py (345:353) - toolkits/model_checkpoints_convertor/bloom/checkpoint_reshaping_and_interoperability.py (703:711) duplicated block id: 6488 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/qwen2_moe/layer_specs.py (121:142) - megatron_patch/model/qwen2_moe/layer_specs.py (182:203) duplicated block id: 6489 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon40b/checkpoint_reshaping_and_interoperability.py (473:481) - toolkits/model_checkpoints_convertor/starcoder/checkpoint_reshaping_and_interoperability.py (703:711) duplicated block id: 6490 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/llama/gpt_model.py (105:114) - megatron_patch/model/qwen_vl/gpt_model.py (107:116) duplicated block id: 6491 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/language_model.py (299:307) - megatron_patch/model/galactica/language_model.py (318:326) duplicated block id: 6492 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/falcon/language_model.py (287:295) - megatron_patch/model/qwen/language_model.py (289:297) duplicated block id: 6493 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/falcon/transformer.py (744:754) - megatron_patch/model/llama3/transformer_legacy.py (718:730) duplicated block id: 6494 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/llama/gpt_model.py (105:114) - megatron_patch/model/llama2/gpt_model.py (106:115) duplicated block id: 6495 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore_llama3_1.py (80:87) - toolkits/model_checkpoints_convertor/mistral/hf2mcore.py (107:114) duplicated block id: 6496 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/model.py (154:189) - megatron_patch/model/mixtral_bak/model.py (149:184) duplicated block id: 6497 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/falcon/transformer.py (547:554) - megatron_patch/model/llama/transformer.py (362:369) duplicated block id: 6498 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (495:503) - toolkits/model_checkpoints_convertor/starcoder/checkpoint_reshaping_and_interoperability.py (703:711) duplicated block id: 6499 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/falcon/language_model.py (600:610) - megatron_patch/model/qwen1_5_megablocks/language_model.py (556:566) duplicated block id: 6500 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/bloom/language_model.py (352:359) - megatron_patch/model/llama2/language_model.py (334:341) duplicated block id: 6501 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/reward_model_to_megatron.py (813:822) - toolkits/model_checkpoints_convertor/galactica/checkpoint_reshaping_and_interoperability.py (649:658) duplicated block id: 6502 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/mixtral_bak/model.py (82:90) - megatron_patch/model/qwen2_vl/gpt_model.py (78:86) duplicated block id: 6503 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/galactica/gpt_model.py (105:114) - megatron_patch/model/llama3/gpt_model.py (108:117) duplicated block id: 6504 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (97:108) - megatron_patch/model/qwen/transformer.py (93:103) duplicated block id: 6505 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/glm/checkpoint_reshaping_and_interoperability.py (441:454) - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (497:510) duplicated block id: 6506 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (363:373) - megatron_patch/model/llava/transformer.py (366:376) duplicated block id: 6507 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (829:836) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (895:902) duplicated block id: 6508 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (668:675) - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (938:945) duplicated block id: 6509 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (668:675) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (713:720) duplicated block id: 6510 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer/attention.py (360:367) - megatron_patch/model/qwen2/transformer/attention.py (531:538) duplicated block id: 6511 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (382:391) - toolkits/model_checkpoints_convertor/yi/checkpoint_reshaping_and_interoperability.py (484:492) duplicated block id: 6512 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (576:584) - toolkits/model_checkpoints_convertor/starcoder/checkpoint_reshaping_and_interoperability.py (703:711) duplicated block id: 6513 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/galactica/gpt_model.py (50:60) - megatron_patch/model/glm130b/gpt_model.py (48:58) duplicated block id: 6514 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/falcon/transformer.py (263:275) - megatron_patch/model/falcon/transformer.py (428:439) duplicated block id: 6515 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/llama2/transformer.py (1201:1208) - megatron_patch/model/mistral/transformer.py (1737:1744) duplicated block id: 6516 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/chatglm/checkpoint_reshaping_and_interoperability.py (125:132) - toolkits/model_checkpoints_convertor/llama/hf2mcore_llama3_1.py (80:87) duplicated block id: 6517 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/reward_model_to_megatron.py (785:794) - toolkits/model_checkpoints_convertor/glm/checkpoint_reshaping_and_interoperability.py (537:546) duplicated block id: 6518 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/mistral/transformer.py (1240:1247) - megatron_patch/model/qwen/transformer.py (961:970) duplicated block id: 6519 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/llama2/transformer.py (398:417) - megatron_patch/model/starcoder/transformer.py (425:445) duplicated block id: 6520 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/language_model.py (595:603) - megatron_patch/model/starcoder/language_model.py (481:489) duplicated block id: 6521 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/reward_model_to_megatron.py (702:710) - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (576:584) duplicated block id: 6522 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/qwen3_moe/gpt_layer_specs.py (130:137) - megatron_patch/model/qwen3_moe/gpt_layer_specs.py (157:164) duplicated block id: 6523 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (1094:1101) - megatron_patch/model/falcon40b/transformer.py (927:934) duplicated block id: 6524 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (427:439) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (594:606) duplicated block id: 6525 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (1018:1027) - megatron_patch/model/llava/transformer.py (1242:1249) duplicated block id: 6526 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/glm/checkpoint_reshaping_and_interoperability.py (430:438) - toolkits/model_checkpoints_convertor/yi/checkpoint_reshaping_and_interoperability.py (502:510) duplicated block id: 6527 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/llama3/language_model.py (490:497) - megatron_patch/model/mistral/language_model.py (523:530) duplicated block id: 6528 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/mistral/language_model.py (335:342) - megatron_patch/model/starcoder/language_model.py (336:343) duplicated block id: 6529 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/qwen1_5_megablocks/transformer.py (289:308) - megatron_patch/model/starcoder/transformer.py (425:445) duplicated block id: 6530 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/qwen2_vl/attention.py (448:455) - megatron_patch/model/qwen2_vl/attention.py (458:465) duplicated block id: 6531 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/galactica/transformer.py (733:743) - megatron_patch/model/starcoder/transformer.py (1159:1169) duplicated block id: 6532 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (429:442) - toolkits/model_checkpoints_convertor/baichuan/hf2te.py (285:298) duplicated block id: 6533 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/layer_specs.py (117:124) - megatron_patch/model/qwen2_vl/layer_specs.py (114:121) duplicated block id: 6534 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/checkpoint_reshaping_and_interoperability.py (121:128) - toolkits/model_checkpoints_convertor/llava/hf2mcore_llava.py (103:110) duplicated block id: 6535 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (1243:1250) - megatron_patch/model/llama2/transformer.py (1016:1025) duplicated block id: 6536 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/reward_model_to_megatron.py (713:726) - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (428:441) duplicated block id: 6537 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/checkpoint_reshaping_and_interoperability.py (714:727) - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (497:510) duplicated block id: 6538 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore.py (571:578) - toolkits/model_checkpoints_convertor/qwen/hf2megablocks_qwen1.5.py (581:588) duplicated block id: 6539 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/qwen2/transformer/attention.py (356:363) - megatron_patch/model/qwen2/transformer/attention.py (531:538) duplicated block id: 6540 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/language_model.py (600:610) - megatron_patch/model/mistral/language_model.py (569:579) duplicated block id: 6541 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/falcon/language_model.py (287:295) - megatron_patch/model/llama3/language_model.py (284:292) duplicated block id: 6542 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon/checkpoint_reshaping_and_interoperability.py (470:478) - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (464:473) duplicated block id: 6543 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer/attention.py (360:367) - megatron_patch/model/llama3/transformer/attention.py (526:533) duplicated block id: 6544 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/llama3_1/model.py (164:180) - megatron_patch/model/qwen1_5/model.py (112:128) duplicated block id: 6545 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon/checkpoint_reshaping_and_interoperability.py (143:150) - toolkits/model_checkpoints_convertor/llama/hf2mcore_llama3_1.py (80:87) duplicated block id: 6546 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/reward_model_to_megatron.py (702:710) - toolkits/model_checkpoints_convertor/falcon40b/checkpoint_reshaping_and_interoperability.py (473:481) duplicated block id: 6547 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/language_model.py (180:193) - megatron_patch/model/mistral/language_model.py (168:181) duplicated block id: 6548 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (523:531) - toolkits/model_checkpoints_convertor/glm/checkpoint_reshaping_and_interoperability.py (430:438) duplicated block id: 6549 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (999:1006) - megatron_patch/model/baichuan2/transformer.py (1131:1138) duplicated block id: 6550 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/llama2/language_model.py (334:341) - megatron_patch/model/starcoder/language_model.py (336:343) duplicated block id: 6551 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/llama2/transformer.py (1740:1747) - megatron_patch/model/mistral/transformer.py (1198:1205) duplicated block id: 6552 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/transformer.py (1035:1043) - megatron_patch/model/qwen/transformer.py (1547:1555) duplicated block id: 6553 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/mistral/transformer.py (796:813) - megatron_patch/model/qwen1_5/transformer/attention.py (225:242) duplicated block id: 6554 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/language_model.py (436:448) - megatron_patch/model/qwen1_5_megablocks/language_model.py (373:385) duplicated block id: 6555 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/chatglm/checkpoint_reshaping_and_interoperability.py (453:461) - toolkits/model_checkpoints_convertor/glm130b/checkpoint_reshaping_and_interoperability.py (411:419) duplicated block id: 6556 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (427:439) - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (587:599) duplicated block id: 6557 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/llama2/transformer.py (1243:1250) - megatron_patch/model/mistral/transformer.py (1013:1022) duplicated block id: 6558 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (1470:1478) - megatron_patch/model/falcon40b/transformer.py (890:898) duplicated block id: 6559 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/llama3/language_model.py (329:336) - megatron_patch/model/starcoder/language_model.py (336:343) duplicated block id: 6560 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (797:804) - megatron_patch/model/llama3/transformer_legacy.py (890:897) duplicated block id: 6561 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon40b/checkpoint_reshaping_and_interoperability.py (68:75) - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (86:94) duplicated block id: 6562 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (1605:1613) - megatron_patch/model/glm130b/transformer.py (1035:1043) duplicated block id: 6563 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llava/hf2mcore_llava.py (103:110) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (171:178) duplicated block id: 6564 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/transformer.py (1035:1043) - megatron_patch/model/starcoder/transformer.py (1201:1209) duplicated block id: 6565 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/llama/language_model.py (611:621) - megatron_patch/model/mistral/language_model.py (569:579) duplicated block id: 6566 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (1018:1027) - megatron_patch/model/llama2/transformer.py (1243:1250) duplicated block id: 6567 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llava/hf2mcore_llava.py (103:110) - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (106:113) duplicated block id: 6568 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/language_model.py (600:610) - megatron_patch/model/llava/language_model.py (617:627) duplicated block id: 6569 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/hf2te.py (111:118) - toolkits/model_checkpoints_convertor/llama/hf2mcore_llama3_1.py (80:87) duplicated block id: 6570 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore.py (670:678) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_moe.py (477:486) duplicated block id: 6571 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (534:546) - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (427:439) duplicated block id: 6572 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (562:570) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (382:391) duplicated block id: 6573 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/mistral/transformer.py (796:813) - megatron_patch/model/mixtral_bak/transformer/attention.py (199:216) duplicated block id: 6574 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (378:385) - megatron_patch/model/falcon/transformer.py (547:554) duplicated block id: 6575 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/model.py (246:256) - megatron_patch/model/qwen2/model.py (208:218) duplicated block id: 6576 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/gpt_model.py (68:75) - megatron_patch/model/glm130b/gpt_model.py (67:74) duplicated block id: 6577 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/galactica/language_model.py (318:326) - megatron_patch/model/qwen_vl/language_model.py (290:298) duplicated block id: 6578 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/transformer.py (1035:1043) - megatron_patch/model/llama3/transformer_legacy.py (1565:1573) duplicated block id: 6579 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/mistral/transformer.py (1013:1022) - megatron_patch/model/qwen1_5_megablocks/transformer.py (1127:1134) duplicated block id: 6580 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (485:495) - toolkits/model_checkpoints_convertor/galactica/checkpoint_reshaping_and_interoperability.py (487:497) duplicated block id: 6581 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/galactica/checkpoint_reshaping_and_interoperability.py (567:574) - toolkits/model_checkpoints_convertor/galactica/checkpoint_reshaping_and_interoperability.py (578:585) duplicated block id: 6582 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (1104:1117) - megatron_patch/model/llama/transformer.py (992:1005) duplicated block id: 6583 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon/checkpoint_reshaping_and_interoperability.py (143:150) - toolkits/model_checkpoints_convertor/llava/hf2mcore_llava.py (103:110) duplicated block id: 6584 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/model.py (210:220) - megatron_patch/model/qwen2_vl/gpt_model.py (195:205) duplicated block id: 6585 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon40b/checkpoint_reshaping_and_interoperability.py (257:275) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (358:376) duplicated block id: 6586 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (807:818) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (699:711) duplicated block id: 6587 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llava/hf2mcore_llava.py (403:410) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_dense_and_moe_gqa.py (689:696) duplicated block id: 6588 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llava/hf2mcore_llava.py (403:410) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_dense_and_moe_gqa.py (766:773) duplicated block id: 6589 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/falcon/language_model.py (600:610) - megatron_patch/model/llama3/language_model.py (534:544) duplicated block id: 6590 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (405:416) - toolkits/model_checkpoints_convertor/chatglm/checkpoint_reshaping_and_interoperability.py (330:341) duplicated block id: 6591 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer/attention.py (526:533) - megatron_patch/model/qwen2/transformer/attention.py (356:363) duplicated block id: 6592 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/llama2/transformer.py (822:829) - megatron_patch/model/llama3/transformer_legacy.py (787:794) duplicated block id: 6593 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/qwen/transformer.py (961:970) - megatron_patch/model/qwen1_5_megablocks/transformer.py (1127:1134) duplicated block id: 6594 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/checkpoint_reshaping_and_interoperability.py (703:711) - toolkits/model_checkpoints_convertor/galactica/checkpoint_reshaping_and_interoperability.py (531:539) duplicated block id: 6595 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer_legacy.py (787:794) - megatron_patch/model/qwen_vl/transformer.py (821:828) duplicated block id: 6596 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore.py (119:127) - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (109:116) duplicated block id: 6597 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/llama3_1/layer_specs.py (56:63) - megatron_patch/model/mixtral_bak/layer_specs.py (67:74) duplicated block id: 6598 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (410:429) - megatron_patch/model/falcon/transformer.py (429:449) duplicated block id: 6599 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/reward_model_to_megatron.py (702:710) - toolkits/model_checkpoints_convertor/galactica/checkpoint_reshaping_and_interoperability.py (531:539) duplicated block id: 6600 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/qwen2/transformer_block.py (83:96) - megatron_patch/model/qwen2_5_vl/transformer_block.py (176:201) duplicated block id: 6601 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/qwen1_5/transformer/attention.py (225:242) - megatron_patch/model/qwen_vl/transformer.py (797:814) duplicated block id: 6602 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/language_model.py (299:307) - megatron_patch/model/llama/language_model.py (287:295) duplicated block id: 6603 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/llama2/transformer.py (1016:1025) - megatron_patch/model/mistral/transformer.py (1240:1247) duplicated block id: 6604 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/bloom/language_model.py (305:313) - megatron_patch/model/llava/language_model.py (292:300) duplicated block id: 6605 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/qwen2/transformer/attention.py (221:238) - megatron_patch/model/qwen_vl/transformer.py (797:814) duplicated block id: 6606 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (97:108) - megatron_patch/model/qwen1_5_megablocks/transformer.py (96:107) duplicated block id: 6607 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/transformer.py (319:326) - megatron_patch/model/falcon/transformer.py (547:554) duplicated block id: 6608 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (1470:1478) - megatron_patch/model/falcon/transformer.py (1151:1159) duplicated block id: 6609 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/qwen2_vl/attention.py (458:465) - megatron_patch/model/qwen2_vl/attention_vision.py (152:159) duplicated block id: 6610 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/moe/experts.py (49:61) - megatron_patch/model/qwen1_5/moe/experts.py (32:44) duplicated block id: 6611 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/mixtral_bak/transformer/attention.py (396:411) - megatron_patch/model/qwen2_vl/attention_vision.py (660:675) duplicated block id: 6612 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/qwen1_5_megablocks/transformer.py (900:909) - megatron_patch/model/qwen1_5_megablocks/transformer.py (1127:1134) duplicated block id: 6613 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/galactica/checkpoint_reshaping_and_interoperability.py (85:92) - toolkits/model_checkpoints_convertor/llama/hf2mcore.py (119:127) duplicated block id: 6614 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/transformer/attention.py (204:223) - megatron_patch/model/qwen2_vl/attention_vision.py (193:212) duplicated block id: 6615 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/hf2te.py (257:270) - toolkits/model_checkpoints_convertor/yi/checkpoint_reshaping_and_interoperability.py (404:417) duplicated block id: 6616 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/moe/moe_layer.py (110:117) - megatron_patch/model/qwen2/moe/moe_layer.py (114:121) duplicated block id: 6617 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (401:414) - toolkits/model_checkpoints_convertor/baichuan2/hf2te.py (257:270) duplicated block id: 6618 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/falcon/transformer.py (1151:1159) - megatron_patch/model/llama3/transformer_legacy.py (1565:1573) duplicated block id: 6619 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/falcon/transformer.py (1151:1159) - megatron_patch/model/qwen_vl/transformer.py (1604:1612) duplicated block id: 6620 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/transformer.py (268:284) - megatron_patch/model/qwen1_5_megablocks/transformer.py (282:298) duplicated block id: 6621 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/llama3/language_model.py (490:497) - megatron_patch/model/llava/language_model.py (571:578) duplicated block id: 6622 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/bloom/language_model.py (352:359) - megatron_patch/model/qwen/language_model.py (334:341) duplicated block id: 6623 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/llama2/language_model.py (511:518) - megatron_patch/model/llama3/language_model.py (490:497) duplicated block id: 6624 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (464:473) - toolkits/model_checkpoints_convertor/yi/checkpoint_reshaping_and_interoperability.py (484:492) duplicated block id: 6625 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/gpt_model.py (48:58) - megatron_patch/model/falcon40b/gpt_model.py (50:60) duplicated block id: 6626 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_moe.py (309:316) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_dense_and_moe_gqa.py (526:534) duplicated block id: 6627 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/model.py (246:256) - megatron_patch/model/qwen1_5/model.py (207:217) duplicated block id: 6628 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/falcon/language_model.py (371:382) - megatron_patch/model/starcoder/language_model.py (365:376) duplicated block id: 6629 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/falcon/transformer.py (1151:1159) - megatron_patch/model/llama2/transformer.py (1605:1613) duplicated block id: 6630 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/hf2te.py (373:381) - toolkits/model_checkpoints_convertor/glm130b/checkpoint_reshaping_and_interoperability.py (411:419) duplicated block id: 6631 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer/mlp.py (109:116) - megatron_patch/model/mixtral/moe/experts.py (642:649) duplicated block id: 6632 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/deepseek/hf2mcore_deepseek_v2_moe.py (522:530) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_moe.py (670:678) duplicated block id: 6633 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/language_model.py (574:582) - megatron_patch/model/starcoder/language_model.py (481:489) duplicated block id: 6634 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (626:633) - toolkits/model_checkpoints_convertor/yi/checkpoint_reshaping_and_interoperability.py (541:548) duplicated block id: 6635 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/mistral/language_model.py (364:371) - megatron_patch/model/qwen1_5_megablocks/language_model.py (361:368) duplicated block id: 6636 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (363:373) - megatron_patch/model/llama2/transformer.py (366:376) duplicated block id: 6637 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/glm130b/checkpoint_reshaping_and_interoperability.py (422:435) - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (497:510) duplicated block id: 6638 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (429:442) - toolkits/model_checkpoints_convertor/baichuan2/hf2te.py (257:270) duplicated block id: 6639 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/model.py (154:189) - megatron_patch/model/qwen2/model.py (174:195) duplicated block id: 6640 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/transformer.py (319:326) - megatron_patch/model/starcoder/transformer.py (524:531) duplicated block id: 6641 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/qwen2_vl/layer_specs.py (106:113) - megatron_patch/model/qwen3_moe/gpt_layer_specs.py (300:307) duplicated block id: 6642 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (76:87) - megatron_patch/model/starcoder/transformer.py (68:80) duplicated block id: 6643 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (588:595) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (599:606) duplicated block id: 6644 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/hf2te.py (55:62) - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (86:94) duplicated block id: 6645 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/mistral/transformer.py (1013:1022) - megatron_patch/model/qwen_vl/transformer.py (1242:1249) duplicated block id: 6646 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (664:681) - megatron_patch/model/qwen2/transformer/attention.py (221:238) duplicated block id: 6647 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (822:829) - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (829:836) duplicated block id: 6648 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (822:829) - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (781:788) duplicated block id: 6649 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/galactica/checkpoint_reshaping_and_interoperability.py (142:149) - toolkits/model_checkpoints_convertor/llama/hf2mcore_llama3_1.py (80:87) duplicated block id: 6650 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore_llama3_1.py (80:87) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (94:101) duplicated block id: 6651 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/language_model.py (303:311) - megatron_patch/model/llama2/language_model.py (289:297) duplicated block id: 6652 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/qwen_vl/transformer.py (1335:1358) - megatron_patch/model/starcoder/transformer.py (1059:1082) duplicated block id: 6653 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (1305:1313) - megatron_patch/model/bloom/transformer.py (979:987) duplicated block id: 6654 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/qwen/transformer.py (533:540) - megatron_patch/model/starcoder/transformer.py (593:600) duplicated block id: 6655 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon40b/checkpoint_reshaping_and_interoperability.py (761:768) - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (668:675) duplicated block id: 6656 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon40b/checkpoint_reshaping_and_interoperability.py (761:768) - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (716:723) duplicated block id: 6657 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (506:518) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (414:426) duplicated block id: 6658 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (708:722) - toolkits/model_checkpoints_convertor/bloom/checkpoint_reshaping_and_interoperability.py (386:400) duplicated block id: 6659 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/layer_specs.py (89:96) - megatron_patch/model/qwen3_moe/gpt_layer_specs.py (130:137) duplicated block id: 6660 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon40b/checkpoint_reshaping_and_interoperability.py (522:529) - toolkits/model_checkpoints_convertor/galactica/checkpoint_reshaping_and_interoperability.py (567:574) duplicated block id: 6661 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (643:653) - megatron_patch/model/glm130b/transformer.py (743:753) duplicated block id: 6662 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/layer_specs.py (89:96) - megatron_patch/model/qwen3_moe/gpt_layer_specs.py (157:164) duplicated block id: 6663 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/falcon/transformer.py (1214:1227) - megatron_patch/model/galactica/transformer.py (843:856) duplicated block id: 6664 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/transformer.py (1035:1043) - megatron_patch/model/qwen_vl/transformer.py (1604:1612) duplicated block id: 6665 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon/checkpoint_reshaping_and_interoperability.py (484:492) - toolkits/model_checkpoints_convertor/glm/checkpoint_reshaping_and_interoperability.py (430:438) duplicated block id: 6666 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (1470:1478) - megatron_patch/model/chatglm/transformer.py (775:783) duplicated block id: 6667 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/language_model.py (506:513) - megatron_patch/model/qwen1_5_megablocks/language_model.py (510:517) duplicated block id: 6668 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (1605:1613) - megatron_patch/model/bloom/transformer.py (1050:1058) duplicated block id: 6669 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (793:800) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (713:720) duplicated block id: 6670 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/llava/transformer.py (1015:1024) - megatron_patch/model/llava/transformer.py (1242:1249) duplicated block id: 6671 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (793:800) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (666:673) duplicated block id: 6672 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/llava/transformer.py (1242:1249) - megatron_patch/model/qwen/transformer.py (961:970) duplicated block id: 6673 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (797:804) - megatron_patch/model/qwen1_5_megablocks/transformer.py (830:837) duplicated block id: 6674 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/gpt_model.py (105:114) - megatron_patch/model/llama2/gpt_model.py (106:115) duplicated block id: 6675 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/falcon/transformer.py (944:951) - megatron_patch/model/glm130b/transformer.py (758:765) duplicated block id: 6676 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/checkpoint_reshaping_and_interoperability.py (814:823) - toolkits/model_checkpoints_convertor/chatglm/checkpoint_reshaping_and_interoperability.py (579:588) duplicated block id: 6677 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (1114:1121) - megatron_patch/model/baichuan2/transformer.py (1247:1254) duplicated block id: 6678 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/galactica/checkpoint_reshaping_and_interoperability.py (487:497) - toolkits/model_checkpoints_convertor/yi/checkpoint_reshaping_and_interoperability.py (460:470) duplicated block id: 6679 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer_legacy.py (787:794) - megatron_patch/model/llava/transformer.py (821:828) duplicated block id: 6680 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/model.py (126:143) - megatron_patch/model/llama3_1/model.py (164:180) duplicated block id: 6681 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/deepseek/hf2mcore_deepseek_v2_moe.py (130:137) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_dense_and_moe_gqa.py (201:208) duplicated block id: 6682 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/deepseek/hf2mcore_deepseek_v2_moe.py (487:494) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_moe.py (636:643) duplicated block id: 6683 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/galactica/language_model.py (318:326) - megatron_patch/model/qwen1_5_megablocks/language_model.py (288:296) duplicated block id: 6684 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/qwen1_5_megablocks/transformer.py (1127:1134) - megatron_patch/model/qwen_vl/transformer.py (1015:1024) duplicated block id: 6685 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_mha.py (60:69) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_mha_to_moe.py (65:74) duplicated block id: 6686 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/hf2te.py (56:63) - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (86:94) duplicated block id: 6687 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/language_model.py (303:311) - megatron_patch/model/mistral/language_model.py (290:298) duplicated block id: 6688 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/llama2/transformer.py (1336:1359) - megatron_patch/model/starcoder/transformer.py (1059:1082) duplicated block id: 6689 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/llava/transformer.py (339:347) - megatron_patch/model/starcoder/transformer.py (222:230) duplicated block id: 6690 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/language_model.py (345:353) - megatron_patch/model/chatglm/language_model.py (303:311) duplicated block id: 6691 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/llava/language_model.py (337:344) - megatron_patch/model/starcoder/language_model.py (336:343) duplicated block id: 6692 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer_legacy.py (762:779) - megatron_patch/model/mixtral_bak/transformer/attention.py (199:216) duplicated block id: 6693 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/falcon/transformer.py (744:754) - megatron_patch/model/qwen/transformer.py (713:725) duplicated block id: 6694 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/llama/language_model.py (611:621) - megatron_patch/model/llava/language_model.py (617:627) duplicated block id: 6695 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/mistral/hf2mcore.py (373:381) - toolkits/model_checkpoints_convertor/qwen/hf2megablocks_qwen1.5.py (492:500) duplicated block id: 6696 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/language_model.py (590:597) - megatron_patch/model/llava/language_model.py (571:578) duplicated block id: 6697 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/llama/transformer.py (899:909) - megatron_patch/model/starcoder/transformer.py (1159:1169) duplicated block id: 6698 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_mha_to_moe.py (128:135) - toolkits/model_checkpoints_convertor/qwen/hf2megablocks_qwen1.5.py (480:487) duplicated block id: 6699 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (428:441) - toolkits/model_checkpoints_convertor/starcoder/checkpoint_reshaping_and_interoperability.py (714:727) duplicated block id: 6700 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/mistral/transformer.py (1737:1744) - megatron_patch/model/qwen1_5_megablocks/transformer.py (1085:1092) duplicated block id: 6701 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (735:749) - toolkits/model_checkpoints_convertor/bloom/reward_model_to_megatron.py (385:399) duplicated block id: 6702 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (632:639) - toolkits/model_checkpoints_convertor/yi/checkpoint_reshaping_and_interoperability.py (541:548) duplicated block id: 6703 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (815:822) - megatron_patch/model/mistral/transformer.py (950:957) duplicated block id: 6704 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/transformer.py (729:739) - megatron_patch/model/starcoder/transformer.py (1159:1169) duplicated block id: 6705 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (446:453) - toolkits/model_checkpoints_convertor/chatglm/checkpoint_reshaping_and_interoperability.py (343:350) duplicated block id: 6706 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/qwen/transformer.py (757:774) - megatron_patch/model/qwen2/transformer/attention.py (221:238) duplicated block id: 6707 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore_llama3_1.py (588:595) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_dense_and_moe_gqa.py (650:657) duplicated block id: 6708 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore.py (99:107) - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (86:94) duplicated block id: 6709 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/qwen_vl/transformer.py (885:896) - megatron_patch/model/starcoder/transformer.py (814:825) duplicated block id: 6710 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/transformer.py (835:848) - megatron_patch/model/falcon40b/transformer.py (953:966) duplicated block id: 6711 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/falcon/transformer.py (429:449) - megatron_patch/model/llama3/transformer_legacy.py (397:416) duplicated block id: 6712 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (1050:1058) - megatron_patch/model/mistral/transformer.py (1602:1610) duplicated block id: 6713 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer/attention.py (225:242) - megatron_patch/model/mistral/transformer.py (796:813) duplicated block id: 6714 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/falcon/language_model.py (287:295) - megatron_patch/model/mistral/language_model.py (290:298) duplicated block id: 6715 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer_legacy.py (1201:1208) - megatron_patch/model/llava/transformer.py (1015:1024) duplicated block id: 6716 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/language_model.py (579:589) - megatron_patch/model/qwen_vl/language_model.py (586:596) duplicated block id: 6717 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/hf2te.py (214:224) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (411:421) duplicated block id: 6718 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/layer_specs.py (31:44) - megatron_patch/model/qwen2_moe/layer_specs.py (42:55) duplicated block id: 6719 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/llama2/transformer.py (1016:1025) - megatron_patch/model/llama3/transformer_legacy.py (1201:1208) duplicated block id: 6720 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon/checkpoint_reshaping_and_interoperability.py (446:456) - toolkits/model_checkpoints_convertor/galactica/checkpoint_reshaping_and_interoperability.py (487:497) duplicated block id: 6721 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/gpt_model.py (132:141) - megatron_patch/model/falcon40b/gpt_model.py (105:114) duplicated block id: 6722 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/falcon/transformer.py (744:754) - megatron_patch/model/mistral/transformer.py (738:750) duplicated block id: 6723 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/hf2te.py (345:353) - toolkits/model_checkpoints_convertor/bloom/reward_model_to_megatron.py (702:710) duplicated block id: 6724 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/qwen2_vl/layer_specs.py (56:63) - megatron_patch/model/qwen2_vl/layer_specs.py (89:96) duplicated block id: 6725 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (1605:1613) - megatron_patch/model/falcon40b/transformer.py (890:898) duplicated block id: 6726 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/hf2te.py (332:339) - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (464:473) duplicated block id: 6727 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/chatglm/checkpoint_reshaping_and_interoperability.py (579:588) - toolkits/model_checkpoints_convertor/starcoder/checkpoint_reshaping_and_interoperability.py (823:832) duplicated block id: 6728 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/reward_model_to_megatron.py (479:486) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (713:720) duplicated block id: 6729 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/qwen1_5/moe/token_dispatcher.py (178:201) - megatron_patch/model/qwen2/moe/token_dispatcher.py (184:207) duplicated block id: 6730 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/qwen1_5/moe/moe_layer.py (89:96) - megatron_patch/model/qwen3_moe/moe/moe_layer.py (36:43) duplicated block id: 6731 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/hf2te.py (357:370) - toolkits/model_checkpoints_convertor/glm130b/checkpoint_reshaping_and_interoperability.py (422:435) duplicated block id: 6732 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/transformer.py (890:898) - megatron_patch/model/qwen1_5_megablocks/transformer.py (1468:1476) duplicated block id: 6733 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (815:822) - megatron_patch/model/qwen1_5_megablocks/transformer.py (843:850) duplicated block id: 6734 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/qwen1_5/moe/token_dispatcher.py (421:432) - megatron_patch/model/qwen2/moe/token_dispatcher.py (448:459) duplicated block id: 6735 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore_llama3_1.py (171:178) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_dense_and_moe_gqa.py (156:163) duplicated block id: 6736 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (382:391) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (539:547) duplicated block id: 6737 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon/checkpoint_reshaping_and_interoperability.py (259:277) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (358:376) duplicated block id: 6738 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/reward_model_to_megatron.py (479:486) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (666:673) duplicated block id: 6739 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/language_model.py (390:397) - megatron_patch/model/starcoder/language_model.py (336:343) duplicated block id: 6740 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/mixtral_bak/transformer/attention.py (333:340) - megatron_patch/model/mixtral_bak/transformer/attention.py (410:417) duplicated block id: 6741 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/qwen_vl/transformer.py (399:418) - megatron_patch/model/starcoder/transformer.py (425:445) duplicated block id: 6742 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/falcon/language_model.py (287:295) - megatron_patch/model/llava/language_model.py (292:300) duplicated block id: 6743 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore_llama3_1.py (80:87) - toolkits/model_checkpoints_convertor/yi/checkpoint_reshaping_and_interoperability.py (143:150) duplicated block id: 6744 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/checkpoint_reshaping_and_interoperability.py (480:487) - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (781:788) duplicated block id: 6745 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/checkpoint_reshaping_and_interoperability.py (480:487) - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (829:836) duplicated block id: 6746 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/language_model.py (349:356) - megatron_patch/model/qwen_vl/language_model.py (335:342) duplicated block id: 6747 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/transformer/attention.py (660:675) - megatron_patch/model/mixtral_bak/transformer/attention.py (396:411) duplicated block id: 6748 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/chatglm/checkpoint_reshaping_and_interoperability.py (453:461) - toolkits/model_checkpoints_convertor/starcoder/checkpoint_reshaping_and_interoperability.py (703:711) duplicated block id: 6749 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/galactica/transformer.py (201:218) - megatron_patch/model/llama2/transformer.py (391:407) duplicated block id: 6750 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/transformer.py (890:898) - megatron_patch/model/qwen_vl/transformer.py (1604:1612) duplicated block id: 6751 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/reward_model_to_megatron.py (702:710) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (553:561) duplicated block id: 6752 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/gpt_model.py (132:141) - megatron_patch/model/falcon/gpt_model.py (105:114) duplicated block id: 6753 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/llava/transformer.py (885:896) - megatron_patch/model/starcoder/transformer.py (814:825) duplicated block id: 6754 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (1018:1027) - megatron_patch/model/mistral/transformer.py (1240:1247) duplicated block id: 6755 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer_legacy.py (1148:1157) - megatron_patch/model/llava/transformer.py (1196:1205) duplicated block id: 6756 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore.py (459:466) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_moe.py (277:284) duplicated block id: 6757 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/language_model.py (287:295) - megatron_patch/model/mistral/language_model.py (290:298) duplicated block id: 6758 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore.py (119:127) - toolkits/model_checkpoints_convertor/yi/checkpoint_reshaping_and_interoperability.py (86:93) duplicated block id: 6759 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/llava_mcore/layer_specs.py (93:100) - megatron_patch/model/qwen2_vl/layer_specs.py (56:63) duplicated block id: 6760 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/transformer_block.py (137:162) - megatron_patch/model/qwen2/transformer_block.py (83:96) duplicated block id: 6761 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/galactica/language_model.py (365:372) - megatron_patch/model/qwen_vl/language_model.py (335:342) duplicated block id: 6762 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/llama/language_model.py (287:295) - megatron_patch/model/qwen1_5_megablocks/language_model.py (288:296) duplicated block id: 6763 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/hf2te.py (285:298) - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (482:495) duplicated block id: 6764 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/galactica/checkpoint_reshaping_and_interoperability.py (142:149) - toolkits/model_checkpoints_convertor/llava/hf2mcore_llava.py (103:110) duplicated block id: 6765 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/llava/transformer.py (1739:1746) - megatron_patch/model/qwen1_5_megablocks/transformer.py (1085:1092) duplicated block id: 6766 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/transformer.py (775:783) - megatron_patch/model/llava/transformer.py (1604:1612) duplicated block id: 6767 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon/checkpoint_reshaping_and_interoperability.py (484:492) - toolkits/model_checkpoints_convertor/glm130b/checkpoint_reshaping_and_interoperability.py (411:419) duplicated block id: 6768 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer/attention.py (526:533) - megatron_patch/model/mixtral_bak/transformer/attention.py (333:340) duplicated block id: 6769 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/gpt_model.py (68:75) - megatron_patch/model/falcon/gpt_model.py (72:79) duplicated block id: 6770 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/hf2te.py (112:119) - toolkits/model_checkpoints_convertor/llama/hf2mcore_llama3_1.py (80:87) duplicated block id: 6771 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (509:517) - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (395:404) duplicated block id: 6772 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/qwen1_5_megablocks/transformer.py (900:909) - megatron_patch/model/qwen_vl/transformer.py (1242:1249) duplicated block id: 6773 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/llava/transformer.py (1242:1249) - megatron_patch/model/mistral/transformer.py (1013:1022) duplicated block id: 6774 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/qwen1_5_megablocks/transformer.py (430:437) - megatron_patch/model/starcoder/transformer.py (593:600) duplicated block id: 6775 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (1050:1058) - megatron_patch/model/qwen/transformer.py (1547:1555) duplicated block id: 6776 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/transformer.py (1035:1043) - megatron_patch/model/mistral/transformer.py (1602:1610) duplicated block id: 6777 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/llava/transformer.py (1335:1358) - megatron_patch/model/starcoder/transformer.py (1059:1082) duplicated block id: 6778 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (395:404) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (539:547) duplicated block id: 6779 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/starcoder/checkpoint_reshaping_and_interoperability.py (703:711) - toolkits/model_checkpoints_convertor/yi/checkpoint_reshaping_and_interoperability.py (502:510) duplicated block id: 6780 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/language_model.py (287:295) - megatron_patch/model/llama2/language_model.py (289:297) duplicated block id: 6781 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/qwen/language_model.py (495:502) - megatron_patch/model/qwen1_5_megablocks/language_model.py (510:517) duplicated block id: 6782 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (301:320) - megatron_patch/model/falcon/transformer.py (429:449) duplicated block id: 6783 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/galactica/checkpoint_reshaping_and_interoperability.py (487:497) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (545:555) duplicated block id: 6784 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (496:508) - toolkits/model_checkpoints_convertor/yi/checkpoint_reshaping_and_interoperability.py (513:525) duplicated block id: 6785 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (666:673) - toolkits/model_checkpoints_convertor/starcoder/checkpoint_reshaping_and_interoperability.py (494:501) duplicated block id: 6786 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/language_model.py (180:193) - megatron_patch/model/llama3/language_model.py (156:169) duplicated block id: 6787 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/qwen/transformer.py (961:970) - megatron_patch/model/qwen/transformer.py (1185:1192) duplicated block id: 6788 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/qwen_vl/transformer.py (399:418) - megatron_patch/model/starcoder/transformer.py (302:322) duplicated block id: 6789 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/moe/experts.py (641:648) - megatron_patch/model/llama3/transformer/mlp.py (109:116) duplicated block id: 6790 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (509:517) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (382:391) duplicated block id: 6791 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/llama/gpt_model.py (105:114) - megatron_patch/model/qwen1_5_megablocks/gpt_model.py (106:115) duplicated block id: 6792 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/language_model.py (550:560) - megatron_patch/model/galactica/language_model.py (629:639) duplicated block id: 6793 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/falcon/language_model.py (207:215) - megatron_patch/model/qwen2_vl/language_model_embedding.py (87:95) duplicated block id: 6794 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/gpt_model.py (68:75) - megatron_patch/model/falcon40b/gpt_model.py (72:79) duplicated block id: 6795 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/galactica/checkpoint_reshaping_and_interoperability.py (578:585) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (588:595) duplicated block id: 6796 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/falcon/transformer.py (744:754) - megatron_patch/model/llama2/transformer.py (740:752) duplicated block id: 6797 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (716:723) - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (781:788) duplicated block id: 6798 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/transformer.py (844:854) - megatron_patch/model/starcoder/transformer.py (1159:1169) duplicated block id: 6799 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (583:591) - toolkits/model_checkpoints_convertor/starcoder/checkpoint_reshaping_and_interoperability.py (703:711) duplicated block id: 6800 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/hf2te.py (356:368) - toolkits/model_checkpoints_convertor/yi/checkpoint_reshaping_and_interoperability.py (513:525) duplicated block id: 6801 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (713:720) - toolkits/model_checkpoints_convertor/starcoder/checkpoint_reshaping_and_interoperability.py (494:501) duplicated block id: 6802 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/galactica/language_model.py (318:326) - megatron_patch/model/llama3/language_model.py (284:292) duplicated block id: 6803 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/language_model.py (550:560) - megatron_patch/model/falcon40b/language_model.py (600:610) duplicated block id: 6804 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/mixtral_bak/transformer/attention.py (410:417) - megatron_patch/model/qwen2/transformer/attention.py (356:363) duplicated block id: 6805 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/deepseek/hf2mcore_deepseek_v2_moe.py (364:373) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_moe.py (558:566) duplicated block id: 6806 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (797:804) - megatron_patch/model/llava/transformer.py (939:946) duplicated block id: 6807 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (1605:1613) - megatron_patch/model/chatglm/transformer.py (775:783) duplicated block id: 6808 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/mixtral_bak/layer_specs.py (67:74) - megatron_patch/model/qwen2_moe/layer_specs.py (171:178) duplicated block id: 6809 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/language_model.py (550:560) - megatron_patch/model/chatglm/language_model.py (579:589) duplicated block id: 6810 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore_llama3_1.py (186:193) - toolkits/model_checkpoints_convertor/llava/hf2mcore_llava.py (375:382) duplicated block id: 6811 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (781:788) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (895:902) duplicated block id: 6812 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/galactica/language_model.py (629:639) - megatron_patch/model/qwen1_5_megablocks/language_model.py (556:566) duplicated block id: 6813 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/hf2te.py (422:430) - toolkits/model_checkpoints_convertor/galactica/checkpoint_reshaping_and_interoperability.py (565:573) duplicated block id: 6814 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/mixtral_bak/moe/moe_layer.py (38:45) - megatron_patch/model/qwen2/moe/moe_layer.py (58:66) duplicated block id: 6815 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (998:1008) - megatron_patch/model/galactica/transformer.py (733:743) duplicated block id: 6816 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (780:791) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (699:711) duplicated block id: 6817 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/bloom/language_model.py (352:359) - megatron_patch/model/qwen1_5_megablocks/language_model.py (333:340) duplicated block id: 6818 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/falcon/gpt_model.py (105:114) - megatron_patch/model/llama3/gpt_model.py (108:117) duplicated block id: 6819 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/language_model.py (579:589) - megatron_patch/model/llama2/language_model.py (557:567) duplicated block id: 6820 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/gpt_model.py (105:114) - megatron_patch/model/qwen1_5_megablocks/gpt_model.py (106:115) duplicated block id: 6821 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/qwen/transformer.py (1185:1192) - megatron_patch/model/qwen_vl/transformer.py (1015:1024) duplicated block id: 6822 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/reward_model_to_megatron.py (702:710) - toolkits/model_checkpoints_convertor/chatglm/checkpoint_reshaping_and_interoperability.py (453:461) duplicated block id: 6823 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/model.py (246:256) - megatron_patch/model/qwen2_vl/gpt_model.py (195:205) duplicated block id: 6824 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/bloom/language_model.py (352:359) - megatron_patch/model/mistral/language_model.py (335:342) duplicated block id: 6825 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/transformer.py (890:898) - megatron_patch/model/llama3/transformer_legacy.py (1565:1573) duplicated block id: 6826 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/deepseek/hf2mcore_deepseek_v2_moe.py (515:524) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_mha_to_moe.py (246:255) duplicated block id: 6827 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (1050:1058) - megatron_patch/model/llava/transformer.py (1604:1612) duplicated block id: 6828 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/moe/router.py (280:288) - megatron_patch/model/qwen3_moe/moe/router.py (153:161) duplicated block id: 6829 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/falcon/transformer.py (428:439) - megatron_patch/model/falcon40b/transformer.py (256:268) duplicated block id: 6830 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/language_model.py (299:307) - megatron_patch/model/chatglm/language_model.py (303:311) duplicated block id: 6831 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (664:681) - megatron_patch/model/qwen1_5/transformer/attention.py (225:242) duplicated block id: 6832 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/layer_specs.py (38:51) - megatron_patch/model/qwen2_moe/layer_specs.py (42:55) duplicated block id: 6833 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (572:584) - megatron_patch/model/glm130b/transformer.py (553:566) duplicated block id: 6834 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (664:681) - megatron_patch/model/mixtral_bak/transformer/attention.py (199:216) duplicated block id: 6835 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (68:77) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_dense_and_moe_gqa.py (54:63) duplicated block id: 6836 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer/attention.py (225:242) - megatron_patch/model/qwen1_5_megablocks/transformer.py (689:706) duplicated block id: 6837 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/layer_specs.py (89:96) - megatron_patch/model/qwen1_5/layer_specs.py (55:62) duplicated block id: 6838 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/falcon/transformer.py (429:449) - megatron_patch/model/qwen/transformer.py (392:411) duplicated block id: 6839 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/falcon/transformer.py (1151:1159) - megatron_patch/model/starcoder/transformer.py (1201:1209) duplicated block id: 6840 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (495:503) - toolkits/model_checkpoints_convertor/glm/checkpoint_reshaping_and_interoperability.py (430:438) duplicated block id: 6841 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore.py (459:466) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_mha.py (214:221) duplicated block id: 6842 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/galactica/checkpoint_reshaping_and_interoperability.py (531:539) - toolkits/model_checkpoints_convertor/glm130b/checkpoint_reshaping_and_interoperability.py (411:419) duplicated block id: 6843 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer_legacy.py (966:975) - megatron_patch/model/llama3/transformer_legacy.py (1201:1208) duplicated block id: 6844 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (797:804) - megatron_patch/model/mistral/transformer.py (937:944) duplicated block id: 6845 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/language_model.py (600:610) - megatron_patch/model/llama2/language_model.py (557:567) duplicated block id: 6846 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/galactica/transformer.py (779:787) - megatron_patch/model/qwen1_5_megablocks/transformer.py (1468:1476) duplicated block id: 6847 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (559:566) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (621:628) duplicated block id: 6848 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/qwen1_5/transformer/attention.py (526:533) - megatron_patch/model/qwen2/transformer/attention.py (356:363) duplicated block id: 6849 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/gpt_model.py (48:58) - megatron_patch/model/falcon/gpt_model.py (50:60) duplicated block id: 6850 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/language_model.py (436:448) - megatron_patch/model/llama2/language_model.py (374:386) duplicated block id: 6851 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/qwen2_vl/attention.py (153:160) - megatron_patch/model/qwen2_vl/attention.py (458:465) duplicated block id: 6852 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/qwen2_vl/attention.py (153:160) - megatron_patch/model/qwen2_vl/attention.py (448:455) duplicated block id: 6853 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (822:829) - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (716:723) duplicated block id: 6854 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer_legacy.py (966:975) - megatron_patch/model/qwen/transformer.py (1185:1192) duplicated block id: 6855 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (807:818) - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (814:827) duplicated block id: 6856 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (822:829) - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (668:675) duplicated block id: 6857 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (523:531) - toolkits/model_checkpoints_convertor/bloom/reward_model_to_megatron.py (702:710) duplicated block id: 6858 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/language_model.py (385:397) - megatron_patch/model/llama2/language_model.py (374:386) duplicated block id: 6859 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/transformer.py (758:765) - megatron_patch/model/llama/transformer.py (703:710) duplicated block id: 6860 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/falcon/language_model.py (600:610) - megatron_patch/model/qwen_vl/language_model.py (586:596) duplicated block id: 6861 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer_legacy.py (1148:1157) - megatron_patch/model/qwen_vl/transformer.py (1196:1205) duplicated block id: 6862 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon/checkpoint_reshaping_and_interoperability.py (470:478) - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (395:404) duplicated block id: 6863 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/deepseek/hf2mcore_deepseek_v2_moe.py (502:509) - toolkits/model_checkpoints_convertor/llava/hf2mcore_llava.py (403:410) duplicated block id: 6864 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/glm130b/checkpoint_reshaping_and_interoperability.py (411:419) - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (576:584) duplicated block id: 6865 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/llama2/transformer.py (886:897) - megatron_patch/model/starcoder/transformer.py (814:825) duplicated block id: 6866 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/galactica/transformer.py (418:425) - megatron_patch/model/glm130b/transformer.py (506:513) duplicated block id: 6867 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/language_model.py (634:644) - megatron_patch/model/llama/language_model.py (611:621) duplicated block id: 6868 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/checkpoint_reshaping_and_interoperability.py (714:727) - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (428:441) duplicated block id: 6869 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (1243:1250) - megatron_patch/model/mistral/transformer.py (1013:1022) duplicated block id: 6870 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon/checkpoint_reshaping_and_interoperability.py (86:93) - toolkits/model_checkpoints_convertor/llama/hf2mcore.py (119:127) duplicated block id: 6871 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/starcoder/checkpoint_reshaping_and_interoperability.py (823:832) - toolkits/model_checkpoints_convertor/yi/checkpoint_reshaping_and_interoperability.py (666:674) duplicated block id: 6872 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/falcon/gpt_model.py (50:60) - megatron_patch/model/glm130b/gpt_model.py (48:58) duplicated block id: 6873 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (666:673) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (895:902) duplicated block id: 6874 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/language_model.py (634:644) - megatron_patch/model/falcon40b/language_model.py (600:610) duplicated block id: 6875 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/falcon/transformer.py (556:563) - megatron_patch/model/galactica/transformer.py (332:339) duplicated block id: 6876 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/llava_mcore/llava_spec.py (80:87) - megatron_patch/model/qwen3_moe/gpt_layer_specs.py (227:234) duplicated block id: 6877 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/galactica/gpt_model.py (105:114) - megatron_patch/model/mistral/gpt_model.py (106:115) duplicated block id: 6878 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (708:722) - toolkits/model_checkpoints_convertor/bloom/reward_model_to_megatron.py (385:399) duplicated block id: 6879 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore.py (119:127) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (114:121) duplicated block id: 6880 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_moe.py (277:284) - toolkits/model_checkpoints_convertor/qwen/hf2megablocks_qwen1.5.py (480:487) duplicated block id: 6881 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/gpt_model.py (132:141) - megatron_patch/model/galactica/gpt_model.py (105:114) duplicated block id: 6882 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (67:74) - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (86:94) duplicated block id: 6883 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/bloom/language_model.py (305:313) - megatron_patch/model/llama3/language_model.py (284:292) duplicated block id: 6884 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (999:1006) - megatron_patch/model/llama2/transformer.py (1129:1136) duplicated block id: 6885 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore_llama3_1.py (80:87) - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (106:113) duplicated block id: 6886 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/language_model.py (207:215) - megatron_patch/model/qwen2_vl/language_model_embedding.py (87:95) duplicated block id: 6887 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/chatglm/checkpoint_reshaping_and_interoperability.py (125:132) - toolkits/model_checkpoints_convertor/llava/hf2mcore_llava.py (103:110) duplicated block id: 6888 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/hf2te.py (384:396) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (594:606) duplicated block id: 6889 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/hf2te.py (198:208) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (411:421) duplicated block id: 6890 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (775:792) - megatron_patch/model/qwen1_5/transformer/attention.py (225:242) duplicated block id: 6891 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/qwen2/moe/moe_layer.py (114:121) - megatron_patch/model/qwen3_moe/moe/moe_layer.py (36:43) duplicated block id: 6892 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (562:570) - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (464:473) duplicated block id: 6893 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/bloom/gpt_model.py (67:74) - megatron_patch/model/chatglm/gpt_model.py (68:75) duplicated block id: 6894 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/reward_model_to_megatron.py (813:822) - toolkits/model_checkpoints_convertor/yi/checkpoint_reshaping_and_interoperability.py (666:674) duplicated block id: 6895 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (67:74) - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (86:94) duplicated block id: 6896 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/layer_specs.py (89:96) - megatron_patch/model/qwen2/layer_specs.py (65:72) duplicated block id: 6897 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore_llama3_1.py (80:87) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (170:177) duplicated block id: 6898 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (456:463) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (443:450) duplicated block id: 6899 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (681:694) - megatron_patch/model/glm130b/transformer.py (553:566) duplicated block id: 6900 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_mha.py (254:261) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_dense_and_moe_gqa.py (560:567) duplicated block id: 6901 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_dense_and_moe_gqa.py (603:610) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_dense_and_moe_gqa.py (667:674) duplicated block id: 6902 size: 8 cleaned lines of code in 2 files: - megatron_patch/generation/api.py (206:213) - megatron_patch/generation/api.py (256:263) duplicated block id: 6903 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (938:945) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (713:720) duplicated block id: 6904 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/moe/experts.py (642:649) - megatron_patch/model/qwen1_5/transformer/mlp.py (125:132) duplicated block id: 6905 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (938:945) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (666:673) duplicated block id: 6906 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/llama2/language_model.py (511:518) - megatron_patch/model/qwen/language_model.py (495:502) duplicated block id: 6907 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/moe/experts.py (670:680) - megatron_patch/model/qwen2/moe/experts.py (295:305) duplicated block id: 6908 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/model.py (194:219) - megatron_patch/model/qwen1_5/model.py (173:194) duplicated block id: 6909 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/mistral/transformer.py (1240:1247) - megatron_patch/model/qwen_vl/transformer.py (1015:1024) duplicated block id: 6910 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (688:700) - megatron_patch/model/falcon40b/transformer.py (721:730) duplicated block id: 6911 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/falcon/gpt_model.py (105:114) - megatron_patch/model/mistral/gpt_model.py (106:115) duplicated block id: 6912 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (66:78) - megatron_patch/model/bloom/transformer.py (76:87) duplicated block id: 6913 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/checkpoint_reshaping_and_interoperability.py (480:487) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (713:720) duplicated block id: 6914 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (255:273) - megatron_patch/model/llama2/transformer.py (391:407) duplicated block id: 6915 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/checkpoint_reshaping_and_interoperability.py (480:487) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (666:673) duplicated block id: 6916 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/language_model.py (532:539) - megatron_patch/model/qwen_vl/language_model.py (464:471) duplicated block id: 6917 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/model.py (246:256) - megatron_patch/model/llama3/model.py (207:217) duplicated block id: 6918 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/moe/router.py (25:42) - megatron_patch/model/qwen2/moe/router.py (117:134) duplicated block id: 6919 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (97:108) - megatron_patch/model/llama2/transformer.py (98:109) duplicated block id: 6920 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (1114:1121) - megatron_patch/model/qwen_vl/transformer.py (1246:1253) duplicated block id: 6921 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (1243:1250) - megatron_patch/model/qwen1_5_megablocks/transformer.py (900:909) duplicated block id: 6922 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/llava/transformer.py (797:814) - megatron_patch/model/mixtral_bak/transformer/attention.py (199:216) duplicated block id: 6923 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (403:419) - megatron_patch/model/bloom/transformer.py (255:273) duplicated block id: 6924 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/bloom/gpt_model.py (114:125) - megatron_patch/model/starcoder/gpt_model.py (116:127) duplicated block id: 6925 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (1050:1058) - megatron_patch/model/llama2/transformer.py (1605:1613) duplicated block id: 6926 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/qwen1_5/transformer/attention.py (225:242) - megatron_patch/model/qwen1_5_megablocks/transformer.py (689:706) duplicated block id: 6927 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (666:673) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (713:720) duplicated block id: 6928 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/llama2/transformer.py (1740:1747) - megatron_patch/model/qwen_vl/transformer.py (1200:1207) duplicated block id: 6929 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/galactica/transformer.py (201:218) - megatron_patch/model/llama3/transformer_legacy.py (390:406) duplicated block id: 6930 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/qwen2/layer_specs.py (115:122) - megatron_patch/model/qwen2_vl/layer_specs.py (106:113) duplicated block id: 6931 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/hf2te.py (385:398) - toolkits/model_checkpoints_convertor/glm130b/checkpoint_reshaping_and_interoperability.py (422:435) duplicated block id: 6932 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/mistral/language_model.py (364:371) - megatron_patch/model/qwen/language_model.py (362:369) duplicated block id: 6933 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (1104:1117) - megatron_patch/model/galactica/transformer.py (828:841) duplicated block id: 6934 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (1050:1058) - megatron_patch/model/llama3/transformer_legacy.py (1565:1573) duplicated block id: 6935 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/falcon/transformer.py (744:754) - megatron_patch/model/qwen1_5_megablocks/transformer.py (631:643) duplicated block id: 6936 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer_legacy.py (966:975) - megatron_patch/model/mistral/transformer.py (1240:1247) duplicated block id: 6937 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore_llama3_1.py (80:87) - toolkits/model_checkpoints_convertor/starcoder/checkpoint_reshaping_and_interoperability.py (121:128) duplicated block id: 6938 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/reward_model_to_megatron.py (136:146) - toolkits/model_checkpoints_convertor/glm/checkpoint_reshaping_and_interoperability.py (108:117) duplicated block id: 6939 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/llava/transformer.py (1015:1024) - megatron_patch/model/qwen_vl/transformer.py (1242:1249) duplicated block id: 6940 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/falcon/transformer.py (1151:1159) - megatron_patch/model/qwen1_5_megablocks/transformer.py (1468:1476) duplicated block id: 6941 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/chatglm/checkpoint_reshaping_and_interoperability.py (343:350) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (476:483) duplicated block id: 6942 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon40b/checkpoint_reshaping_and_interoperability.py (473:481) - toolkits/model_checkpoints_convertor/glm130b/checkpoint_reshaping_and_interoperability.py (411:419) duplicated block id: 6943 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/glm130b/checkpoint_reshaping_and_interoperability.py (422:435) - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (428:441) duplicated block id: 6944 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/chatglm/checkpoint_reshaping_and_interoperability.py (343:350) - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (499:506) duplicated block id: 6945 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/mistral/transformer.py (540:547) - megatron_patch/model/starcoder/transformer.py (593:600) duplicated block id: 6946 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/hf2te.py (385:398) - toolkits/model_checkpoints_convertor/glm/checkpoint_reshaping_and_interoperability.py (441:454) duplicated block id: 6947 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/llava/transformer.py (1739:1746) - megatron_patch/model/mistral/transformer.py (1198:1205) duplicated block id: 6948 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (506:518) - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (427:439) duplicated block id: 6949 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (175:184) - toolkits/model_checkpoints_convertor/falcon40b/checkpoint_reshaping_and_interoperability.py (168:177) duplicated block id: 6950 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (534:546) - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (496:508) duplicated block id: 6951 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer_legacy.py (1201:1208) - megatron_patch/model/qwen_vl/transformer.py (1015:1024) duplicated block id: 6952 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/falcon/transformer.py (1151:1159) - megatron_patch/model/mistral/transformer.py (1602:1610) duplicated block id: 6953 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/language_model.py (303:311) - megatron_patch/model/llava/language_model.py (292:300) duplicated block id: 6954 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/llama/transformer.py (945:953) - megatron_patch/model/qwen_vl/transformer.py (1604:1612) duplicated block id: 6955 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/falcon/transformer.py (744:754) - megatron_patch/model/llava/transformer.py (739:751) duplicated block id: 6956 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/model.py (210:220) - megatron_patch/model/mixtral_bak/model.py (197:207) duplicated block id: 6957 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (688:700) - megatron_patch/model/falcon/transformer.py (953:962) duplicated block id: 6958 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/llama3/model.py (173:194) - megatron_patch/model/mixtral/model.py (154:189) duplicated block id: 6959 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (1094:1101) - megatron_patch/model/falcon/transformer.py (1188:1195) duplicated block id: 6960 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/transformer.py (890:898) - megatron_patch/model/llava/transformer.py (1604:1612) duplicated block id: 6961 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/model.py (246:256) - megatron_patch/model/mixtral_bak/model.py (197:207) duplicated block id: 6962 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (1305:1313) - megatron_patch/model/glm130b/transformer.py (904:912) duplicated block id: 6963 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (829:836) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (962:969) duplicated block id: 6964 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (166:173) - toolkits/model_checkpoints_convertor/llava/hf2mcore_llava.py (103:110) duplicated block id: 6965 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/bloom/gpt_model.py (48:58) - megatron_patch/model/falcon/gpt_model.py (50:60) duplicated block id: 6966 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (849:860) - megatron_patch/model/starcoder/transformer.py (814:825) duplicated block id: 6967 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/llama3/layer_specs.py (55:62) - megatron_patch/model/qwen3_moe/gpt_layer_specs.py (130:137) duplicated block id: 6968 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (807:818) - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (701:714) duplicated block id: 6969 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (526:533) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (443:450) duplicated block id: 6970 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/hf2te.py (384:396) - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (506:518) duplicated block id: 6971 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (1470:1478) - megatron_patch/model/llama/transformer.py (945:953) duplicated block id: 6972 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (242:250) - megatron_patch/model/starcoder/transformer.py (222:230) duplicated block id: 6973 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/deepseek/hf2mcore_deepseek_v2_moe.py (517:526) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_mha.py (314:323) duplicated block id: 6974 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (1114:1121) - megatron_patch/model/llava/transformer.py (1246:1253) duplicated block id: 6975 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (553:561) - toolkits/model_checkpoints_convertor/starcoder/checkpoint_reshaping_and_interoperability.py (703:711) duplicated block id: 6976 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/galactica/transformer.py (779:787) - megatron_patch/model/qwen/transformer.py (1547:1555) duplicated block id: 6977 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (815:822) - megatron_patch/model/llama2/transformer.py (953:960) duplicated block id: 6978 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/llama2/transformer.py (1243:1250) - megatron_patch/model/qwen_vl/transformer.py (1015:1024) duplicated block id: 6979 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/gpt_model.py (48:58) - megatron_patch/model/galactica/gpt_model.py (50:60) duplicated block id: 6980 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/checkpoint_reshaping_and_interoperability.py (480:487) - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (716:723) duplicated block id: 6981 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (999:1006) - megatron_patch/model/mistral/transformer.py (1126:1133) duplicated block id: 6982 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/reward_model_to_megatron.py (702:710) - toolkits/model_checkpoints_convertor/yi/checkpoint_reshaping_and_interoperability.py (502:510) duplicated block id: 6983 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (1470:1478) - megatron_patch/model/bloom/transformer.py (1050:1058) duplicated block id: 6984 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/checkpoint_reshaping_and_interoperability.py (480:487) - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (668:675) duplicated block id: 6985 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/model.py (100:111) - megatron_patch/model/qwen1_5/model.py (104:115) duplicated block id: 6986 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/galactica/language_model.py (629:639) - megatron_patch/model/llama2/language_model.py (557:567) duplicated block id: 6987 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/llama2/transformer.py (1243:1250) - megatron_patch/model/llama3/transformer_legacy.py (966:975) duplicated block id: 6988 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (668:675) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (962:969) duplicated block id: 6989 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/falcon/language_model.py (595:603) - megatron_patch/model/starcoder/language_model.py (481:489) duplicated block id: 6990 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/transformer.py (775:783) - megatron_patch/model/qwen/transformer.py (1547:1555) duplicated block id: 6991 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/hf2te.py (308:318) - toolkits/model_checkpoints_convertor/galactica/checkpoint_reshaping_and_interoperability.py (487:497) duplicated block id: 6992 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/galactica/gpt_model.py (105:114) - megatron_patch/model/qwen_vl/gpt_model.py (107:116) duplicated block id: 6993 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/language_model.py (344:351) - megatron_patch/model/bloom/language_model.py (352:359) duplicated block id: 6994 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (829:836) - toolkits/model_checkpoints_convertor/starcoder/checkpoint_reshaping_and_interoperability.py (494:501) duplicated block id: 6995 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/llama2/transformer.py (1201:1208) - megatron_patch/model/llama2/transformer.py (1740:1747) duplicated block id: 6996 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/hf2te.py (357:370) - toolkits/model_checkpoints_convertor/glm/checkpoint_reshaping_and_interoperability.py (441:454) duplicated block id: 6997 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/llama2/transformer.py (1740:1747) - megatron_patch/model/qwen1_5_megablocks/transformer.py (1085:1092) duplicated block id: 6998 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer_legacy.py (397:416) - megatron_patch/model/starcoder/transformer.py (425:445) duplicated block id: 6999 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/language_model.py (218:231) - megatron_patch/model/glm130b/language_model.py (180:193) duplicated block id: 7000 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/llama2/transformer.py (798:815) - megatron_patch/model/llama3/transformer/attention.py (225:242) duplicated block id: 7001 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon/checkpoint_reshaping_and_interoperability.py (259:277) - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (339:357) duplicated block id: 7002 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/llama/language_model.py (287:295) - megatron_patch/model/mistral/language_model.py (290:298) duplicated block id: 7003 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (201:210) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (214:223) duplicated block id: 7004 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/mixtral_bak/transformer/attention.py (199:216) - megatron_patch/model/qwen/transformer.py (757:774) duplicated block id: 7005 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/llama2/transformer.py (1197:1206) - megatron_patch/model/llama3/transformer_legacy.py (1148:1157) duplicated block id: 7006 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (999:1006) - megatron_patch/model/llama3/transformer_legacy.py (1078:1085) duplicated block id: 7007 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (76:87) - megatron_patch/model/qwen1_5_megablocks/transformer.py (65:77) duplicated block id: 7008 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (91:98) - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (86:94) duplicated block id: 7009 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/reward_model_to_megatron.py (385:399) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (871:885) duplicated block id: 7010 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/language_model.py (634:644) - megatron_patch/model/galactica/language_model.py (629:639) duplicated block id: 7011 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/transformer.py (1040:1047) - megatron_patch/model/galactica/transformer.py (924:931) duplicated block id: 7012 size: 8 cleaned lines of code in 2 files: - rlhf/trlx/train_reward_model_bloom.py (16:23) - rlhf/trlx/trlx_bloom_rlhf.py (86:93) duplicated block id: 7013 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (531:538) - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (614:621) duplicated block id: 7014 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (775:792) - megatron_patch/model/llama3/transformer/attention.py (225:242) duplicated block id: 7015 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/language_model.py (506:513) - megatron_patch/model/llava/language_model.py (571:578) duplicated block id: 7016 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/falcon/transformer.py (744:754) - megatron_patch/model/starcoder/transformer.py (739:751) duplicated block id: 7017 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/qwen3_moe/gpt_layer_specs.py (99:107) - megatron_patch/model/qwen3_moe/gpt_layer_specs.py (204:212) duplicated block id: 7018 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon/checkpoint_reshaping_and_interoperability.py (259:277) - toolkits/model_checkpoints_convertor/galactica/checkpoint_reshaping_and_interoperability.py (290:308) duplicated block id: 7019 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (738:749) - megatron_patch/model/starcoder/transformer.py (814:825) duplicated block id: 7020 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (998:1008) - megatron_patch/model/llama/transformer.py (899:909) duplicated block id: 7021 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (701:714) - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (921:934) duplicated block id: 7022 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/llama2/language_model.py (362:369) - megatron_patch/model/mistral/language_model.py (364:371) duplicated block id: 7023 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (735:749) - toolkits/model_checkpoints_convertor/bloom/checkpoint_reshaping_and_interoperability.py (386:400) duplicated block id: 7024 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/transformer.py (743:753) - megatron_patch/model/starcoder/transformer.py (857:867) duplicated block id: 7025 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/gpt_model.py (50:60) - megatron_patch/model/glm130b/gpt_model.py (48:58) duplicated block id: 7026 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/moe/experts.py (49:61) - megatron_patch/model/mixtral_bak/moe/experts.py (32:45) duplicated block id: 7027 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/llama2/transformer.py (798:815) - megatron_patch/model/qwen1_5/transformer/attention.py (225:242) duplicated block id: 7028 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/layer_specs.py (151:158) - megatron_patch/model/qwen2_vl/layer_specs.py (106:113) duplicated block id: 7029 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/galactica/transformer.py (779:787) - megatron_patch/model/qwen_vl/transformer.py (1604:1612) duplicated block id: 7030 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/qwen_vl/transformer.py (1015:1024) - megatron_patch/model/qwen_vl/transformer.py (1242:1249) duplicated block id: 7031 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (403:419) - megatron_patch/model/glm130b/transformer.py (268:284) duplicated block id: 7032 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (415:428) - toolkits/model_checkpoints_convertor/starcoder/checkpoint_reshaping_and_interoperability.py (714:727) duplicated block id: 7033 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/checkpoint_reshaping_and_interoperability.py (769:777) - toolkits/model_checkpoints_convertor/glm130b/checkpoint_reshaping_and_interoperability.py (476:484) duplicated block id: 7034 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/reward_model_to_megatron.py (479:486) - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (781:788) duplicated block id: 7035 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/reward_model_to_megatron.py (479:486) - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (829:836) duplicated block id: 7036 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/language_model.py (303:311) - megatron_patch/model/qwen/language_model.py (289:297) duplicated block id: 7037 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore_llama3_1.py (453:460) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_dense_and_moe_gqa.py (856:863) duplicated block id: 7038 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/language_model.py (590:597) - megatron_patch/model/qwen1_5_megablocks/language_model.py (510:517) duplicated block id: 7039 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/reward_model_to_megatron.py (121:128) - toolkits/model_checkpoints_convertor/llama/hf2mcore_llama3_1.py (80:87) duplicated block id: 7040 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (1018:1027) - megatron_patch/model/baichuan2/transformer.py (1243:1250) duplicated block id: 7041 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/transformer.py (553:566) - megatron_patch/model/qwen1_5_megablocks/transformer.py (581:594) duplicated block id: 7042 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/moe/router.py (267:279) - megatron_patch/model/qwen3_moe/moe/router.py (138:150) duplicated block id: 7043 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/qwen/language_model.py (334:341) - megatron_patch/model/starcoder/language_model.py (336:343) duplicated block id: 7044 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/llava/transformer.py (1242:1249) - megatron_patch/model/qwen1_5_megablocks/transformer.py (900:909) duplicated block id: 7045 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/falcon/transformer.py (1151:1159) - megatron_patch/model/llava/transformer.py (1604:1612) duplicated block id: 7046 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer_legacy.py (1201:1208) - megatron_patch/model/qwen1_5_megablocks/transformer.py (900:909) duplicated block id: 7047 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/galactica/language_model.py (629:639) - megatron_patch/model/llava/language_model.py (617:627) duplicated block id: 7048 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/llava_mcore/llava_spec.py (63:71) - megatron_patch/model/qwen2/layer_specs.py (77:84) duplicated block id: 7049 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/llama/language_model.py (287:295) - megatron_patch/model/llama3/language_model.py (284:292) duplicated block id: 7050 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (464:473) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (569:577) duplicated block id: 7051 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/language_model.py (579:589) - megatron_patch/model/qwen1_5_megablocks/language_model.py (556:566) duplicated block id: 7052 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/transformer.py (553:566) - megatron_patch/model/qwen/transformer.py (663:676) duplicated block id: 7053 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/gpt_model.py (105:114) - megatron_patch/model/qwen/gpt_model.py (106:115) duplicated block id: 7054 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/llama2/transformer.py (539:546) - megatron_patch/model/starcoder/transformer.py (593:600) duplicated block id: 7055 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/qwen1_5/moe/router.py (254:263) - megatron_patch/model/qwen2/moe/router.py (341:350) duplicated block id: 7056 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (668:675) - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (716:723) duplicated block id: 7057 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/falcon/transformer.py (429:449) - megatron_patch/model/qwen1_5_megablocks/transformer.py (289:308) duplicated block id: 7058 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (797:804) - megatron_patch/model/baichuan2/transformer.py (903:910) duplicated block id: 7059 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/galactica/language_model.py (318:326) - megatron_patch/model/qwen/language_model.py (289:297) duplicated block id: 7060 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (523:531) - toolkits/model_checkpoints_convertor/glm130b/checkpoint_reshaping_and_interoperability.py (411:419) duplicated block id: 7061 size: 8 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (481:489) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (382:391) duplicated block id: 7062 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/falcon/language_model.py (334:341) - megatron_patch/model/qwen_vl/language_model.py (335:342) duplicated block id: 7063 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (1243:1250) - megatron_patch/model/llava/transformer.py (1015:1024) duplicated block id: 7064 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/llama/language_model.py (611:621) - megatron_patch/model/qwen/language_model.py (539:549) duplicated block id: 7065 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/falcon/transformer.py (547:554) - megatron_patch/model/glm130b/transformer.py (412:419) duplicated block id: 7066 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/moe/moe_layer.py (110:117) - megatron_patch/model/qwen1_5/moe/moe_layer.py (89:96) duplicated block id: 7067 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (76:87) - megatron_patch/model/mistral/transformer.py (67:79) duplicated block id: 7068 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/mixtral_bak/transformer/attention.py (342:350) - megatron_patch/model/qwen2_vl/attention.py (505:513) duplicated block id: 7069 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/language_model.py (550:560) - megatron_patch/model/falcon/language_model.py (600:610) duplicated block id: 7070 size: 8 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/gpt_model.py (132:141) - megatron_patch/model/llama/gpt_model.py (105:114) duplicated block id: 7071 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/transformer/attention.py (344:350) - megatron_patch/model/mixtral_bak/transformer/attention.py (227:233) duplicated block id: 7072 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/galactica/checkpoint_reshaping_and_interoperability.py (477:485) - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (529:537) duplicated block id: 7073 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/checkpoint_reshaping_and_interoperability.py (170:178) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (251:259) duplicated block id: 7074 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (947:954) - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (829:837) duplicated block id: 7075 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/llava/transformer.py (1702:1708) - megatron_patch/model/qwen_vl/transformer.py (1678:1684) duplicated block id: 7076 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/transformer.py (613:623) - megatron_patch/model/glm130b/transformer.py (821:831) duplicated block id: 7077 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (853:861) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (1140:1148) duplicated block id: 7078 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/language_model.py (424:432) - megatron_patch/model/chatglm/language_model.py (412:420) duplicated block id: 7079 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/mixtral_bak/moe/experts.py (81:87) - megatron_patch/model/mixtral_bak/moe/experts.py (116:122) duplicated block id: 7080 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/transformer/attention.py (707:713) - megatron_patch/model/qwen2_vl/attention_vision.py (695:701) duplicated block id: 7081 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/transformer.py (578:589) - megatron_patch/model/qwen/transformer.py (705:716) duplicated block id: 7082 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer/attention.py (26:32) - megatron_patch/model/qwen2_vl/attention_vision.py (15:21) duplicated block id: 7083 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/deepseek/hf2mcore_deepseek_v2_moe.py (504:510) - toolkits/model_checkpoints_convertor/deepseek/hf2mcore_deepseek_v3_moe.py (663:669) duplicated block id: 7084 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/llava_mcore/vision/vit_layer_specs.py (89:95) - megatron_patch/model/mixtral_bak/layer_specs.py (138:144) duplicated block id: 7085 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_mha_to_moe.py (49:57) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_moe.py (112:120) duplicated block id: 7086 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/qwen2/transformer/attention.py (249:255) - megatron_patch/model/qwen2_vl/attention.py (334:340) duplicated block id: 7087 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/transformer/attention.py (496:502) - megatron_patch/model/qwen2/transformer/attention.py (358:364) duplicated block id: 7088 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (1549:1555) - megatron_patch/model/mistral/transformer.py (1700:1706) duplicated block id: 7089 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/language_model.py (651:659) - megatron_patch/model/llama2/language_model.py (632:640) duplicated block id: 7090 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (409:416) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (581:587) duplicated block id: 7091 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/transformer.py (319:325) - megatron_patch/model/llama3/transformer_legacy.py (516:522) duplicated block id: 7092 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/hf2te.py (343:349) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (396:403) duplicated block id: 7093 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer_legacy.py (1639:1645) - megatron_patch/model/qwen/transformer.py (1645:1651) duplicated block id: 7094 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/language_model.py (89:95) - megatron_patch/model/qwen_vl/language_model.py (80:86) duplicated block id: 7095 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/transformer.py (282:296) - megatron_patch/model/mistral/transformer.py (424:438) duplicated block id: 7096 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/llama2/transformer.py (920:928) - megatron_patch/model/starcoder/transformer.py (848:855) duplicated block id: 7097 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (809:820) - megatron_patch/model/llama2/transformer.py (846:857) duplicated block id: 7098 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/galactica/checkpoint_reshaping_and_interoperability.py (186:194) - toolkits/model_checkpoints_convertor/glm130b/checkpoint_reshaping_and_interoperability.py (141:149) duplicated block id: 7099 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/hf2te.py (122:129) - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (193:200) duplicated block id: 7100 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/transformer/attention.py (496:502) - megatron_patch/model/mixtral_bak/transformer/attention.py (335:341) duplicated block id: 7101 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/glm/checkpoint_reshaping_and_interoperability.py (32:38) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (53:60) duplicated block id: 7102 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/gpt_model.py (51:60) - megatron_patch/model/starcoder/gpt_model.py (52:62) duplicated block id: 7103 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/qwen1_5/transformer/attention.py (210:217) - megatron_patch/model/qwen2_vl/attention_vision.py (231:238) duplicated block id: 7104 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (809:820) - megatron_patch/model/qwen1_5_megablocks/transformer.py (737:748) duplicated block id: 7105 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/galactica/language_model.py (89:95) - megatron_patch/model/qwen_vl/language_model.py (80:86) duplicated block id: 7106 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (834:842) - megatron_patch/model/qwen1_5_megablocks/transformer.py (857:865) duplicated block id: 7107 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/language_model.py (615:623) - megatron_patch/model/qwen_vl/language_model.py (661:669) duplicated block id: 7108 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/llava_mcore/layer_specs.py (105:111) - megatron_patch/model/qwen2_vl/layer_specs.py (114:120) duplicated block id: 7109 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/mixtral_bak/transformer/attention.py (443:449) - megatron_patch/model/qwen2_vl/attention_vision.py (695:701) duplicated block id: 7110 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/qwen/language_model.py (614:622) - megatron_patch/model/starcoder/language_model.py (555:563) duplicated block id: 7111 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/mixtral_bak/transformer/attention.py (431:437) - megatron_patch/model/qwen1_5/transformer/attention.py (559:565) duplicated block id: 7112 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/llama3/layer_specs.py (76:82) - megatron_patch/model/mixtral/layer_specs.py (121:127) duplicated block id: 7113 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (812:818) - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (957:963) duplicated block id: 7114 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/hf2te.py (123:130) - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (193:200) duplicated block id: 7115 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/moe/token_dispatcher.py (110:138) - megatron_patch/model/qwen2/moe/token_dispatcher.py (99:125) duplicated block id: 7116 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/language_model.py (89:95) - megatron_patch/model/falcon/language_model.py (89:95) duplicated block id: 7117 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/qwen2_moe/layer_specs.py (320:326) - megatron_patch/model/qwen3_moe/gpt_layer_specs.py (343:349) duplicated block id: 7118 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_moe.py (52:60) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_vl.py (54:62) duplicated block id: 7119 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/llama/transformer.py (362:368) - megatron_patch/model/mistral/transformer.py (518:524) duplicated block id: 7120 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/mixtral_bak/transformer/attention.py (431:437) - megatron_patch/model/qwen2_vl/attention.py (708:714) duplicated block id: 7121 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/falcon/transformer.py (1252:1258) - megatron_patch/model/llama/transformer.py (723:729) duplicated block id: 7122 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/llama/transformer.py (246:256) - megatron_patch/model/qwen1_5_megablocks/transformer.py (289:298) duplicated block id: 7123 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/qwen1_5/moe/experts.py (83:89) - megatron_patch/model/qwen2/moe/experts.py (142:148) duplicated block id: 7124 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_moe.py (199:206) - toolkits/model_checkpoints_convertor/qwen/hf2megablocks_qwen1.5.py (363:370) duplicated block id: 7125 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/qwen1_5_megablocks/transformer.py (1565:1571) - megatron_patch/model/qwen_vl/transformer.py (1678:1684) duplicated block id: 7126 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/moe/experts.py (671:679) - megatron_patch/model/qwen2/transformer/mlp.py (142:150) duplicated block id: 7127 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/reward_model_to_megatron.py (170:178) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (254:262) duplicated block id: 7128 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/transformer.py (412:418) - megatron_patch/model/mistral/transformer.py (518:524) duplicated block id: 7129 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/model.py (101:107) - megatron_patch/model/qwen1_5/model.py (89:95) duplicated block id: 7130 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/hf2te.py (178:190) - toolkits/model_checkpoints_convertor/chatglm/checkpoint_reshaping_and_interoperability.py (284:296) duplicated block id: 7131 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/galactica/checkpoint_reshaping_and_interoperability.py (529:535) - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (409:416) duplicated block id: 7132 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (588:594) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (621:627) duplicated block id: 7133 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/galactica/language_model.py (397:403) - megatron_patch/model/glm130b/language_model.py (380:386) duplicated block id: 7134 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/transformer.py (258:268) - megatron_patch/model/starcoder/transformer.py (425:435) duplicated block id: 7135 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/chatglm/checkpoint_reshaping_and_interoperability.py (75:81) - toolkits/model_checkpoints_convertor/mistral/hf2mcore.py (93:99) duplicated block id: 7136 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (929:937) - megatron_patch/model/qwen1_5_megablocks/transformer.py (946:954) duplicated block id: 7137 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/transformer.py (258:268) - megatron_patch/model/starcoder/transformer.py (302:312) duplicated block id: 7138 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/llava/transformer.py (1702:1708) - megatron_patch/model/mistral/transformer.py (1676:1682) duplicated block id: 7139 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/llama3_1/model.py (113:119) - megatron_patch/model/qwen1_5/model.py (98:104) duplicated block id: 7140 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/language_model.py (89:95) - megatron_patch/model/llama/language_model.py (89:95) duplicated block id: 7141 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/hf2te.py (396:402) - toolkits/model_checkpoints_convertor/galactica/checkpoint_reshaping_and_interoperability.py (578:584) duplicated block id: 7142 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/deepseek/hf2mcore_deepseek_v2_moe.py (325:331) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_dense_and_moe_gqa.py (466:472) duplicated block id: 7143 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/language_model.py (89:95) - megatron_patch/model/qwen/language_model.py (79:85) duplicated block id: 7144 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/deepseek/hf2mcore_deepseek_v2_moe.py (50:58) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_mha_to_moe.py (49:57) duplicated block id: 7145 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/chatglm/checkpoint_reshaping_and_interoperability.py (169:177) - toolkits/model_checkpoints_convertor/glm/checkpoint_reshaping_and_interoperability.py (141:149) duplicated block id: 7146 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/language_model.py (350:356) - megatron_patch/model/llama3/language_model.py (329:335) duplicated block id: 7147 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/llama3_1/layer_specs.py (76:82) - megatron_patch/model/mixtral_bak/layer_specs.py (99:105) duplicated block id: 7148 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/checkpoint_reshaping_and_interoperability.py (170:178) - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (272:280) duplicated block id: 7149 size: 7 cleaned lines of code in 2 files: - megatron_patch/training.py (324:330) - megatron_patch/training.py (763:769) duplicated block id: 7150 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/language_model.py (390:396) - megatron_patch/model/chatglm/language_model.py (350:356) duplicated block id: 7151 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/llama/transformer.py (362:368) - megatron_patch/model/llava/transformer.py (518:524) duplicated block id: 7152 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/falcon/transformer.py (547:553) - megatron_patch/model/qwen/transformer.py (511:517) duplicated block id: 7153 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer/attention.py (362:368) - megatron_patch/model/qwen2_vl/attention.py (497:503) duplicated block id: 7154 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (535:546) - toolkits/model_checkpoints_convertor/bloom/checkpoint_reshaping_and_interoperability.py (714:725) duplicated block id: 7155 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer/attention.py (137:143) - megatron_patch/model/qwen1_5/transformer/attention.py (316:322) duplicated block id: 7156 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/llama/language_model.py (89:95) - megatron_patch/model/qwen_vl/language_model.py (80:86) duplicated block id: 7157 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore_llama3_1.py (722:729) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_vl.py (327:333) duplicated block id: 7158 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/moe/shared_experts.py (172:180) - megatron_patch/model/mixtral/moe/experts.py (671:679) duplicated block id: 7159 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/language_model.py (581:587) - megatron_patch/model/qwen_vl/language_model.py (519:525) duplicated block id: 7160 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer_legacy.py (1663:1669) - megatron_patch/model/qwen/transformer.py (1621:1627) duplicated block id: 7161 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer/attention.py (137:143) - megatron_patch/model/qwen1_5/transformer/attention.py (325:331) duplicated block id: 7162 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/qwen1_5/moe/experts.py (118:124) - megatron_patch/model/qwen2/moe/experts.py (107:113) duplicated block id: 7163 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/layer_specs.py (105:112) - megatron_patch/model/qwen2_moe/layer_specs.py (238:246) duplicated block id: 7164 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/reward_model_to_megatron.py (170:178) - toolkits/model_checkpoints_convertor/galactica/checkpoint_reshaping_and_interoperability.py (186:194) duplicated block id: 7165 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/llama3_1/layer_specs.py (76:82) - megatron_patch/model/llava_mcore/llava_spec.py (73:79) duplicated block id: 7166 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_mha_to_moe.py (169:176) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_moe.py (116:122) duplicated block id: 7167 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (1203:1209) - megatron_patch/model/llama3/transformer_legacy.py (1700:1706) duplicated block id: 7168 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/llama/transformer.py (246:256) - megatron_patch/model/starcoder/transformer.py (425:435) duplicated block id: 7169 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/checkpoint_reshaping_and_interoperability.py (170:178) - toolkits/model_checkpoints_convertor/falcon40b/checkpoint_reshaping_and_interoperability.py (202:210) duplicated block id: 7170 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer/attention.py (137:143) - megatron_patch/model/mixtral_bak/transformer/attention.py (298:304) duplicated block id: 7171 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer/attention.py (137:143) - megatron_patch/model/mixtral_bak/transformer/attention.py (289:295) duplicated block id: 7172 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (196:204) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (201:209) duplicated block id: 7173 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/llama/transformer.py (246:256) - megatron_patch/model/starcoder/transformer.py (302:312) duplicated block id: 7174 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/qwen1_5_megablocks/transformer.py (811:818) - megatron_patch/model/starcoder/transformer.py (848:855) duplicated block id: 7175 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/llava_mcore/llava_spec.py (73:79) - megatron_patch/model/qwen1_5/layer_specs.py (76:82) duplicated block id: 7176 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/language_model.py (526:532) - megatron_patch/model/glm130b/language_model.py (468:474) duplicated block id: 7177 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/deepseek/hf2mcore_deepseek_v3_moe.py (127:135) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_vl.py (72:80) duplicated block id: 7178 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/transformer/attention.py (242:249) - megatron_patch/model/mixtral_bak/transformer/attention.py (184:191) duplicated block id: 7179 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (435:449) - megatron_patch/model/llama/transformer.py (270:284) duplicated block id: 7180 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/chatglm/checkpoint_reshaping_and_interoperability.py (451:457) - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (409:416) duplicated block id: 7181 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/qwen2/layer_specs.py (86:92) - megatron_patch/model/qwen2_moe/layer_specs.py (219:225) duplicated block id: 7182 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/language_model.py (90:96) - megatron_patch/model/llama2/language_model.py (79:85) duplicated block id: 7183 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/mixtral_bak/transformer/attention.py (111:117) - megatron_patch/model/mixtral_bak/transformer/attention.py (289:295) duplicated block id: 7184 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/mixtral_bak/transformer/attention.py (111:117) - megatron_patch/model/mixtral_bak/transformer/attention.py (298:304) duplicated block id: 7185 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/galactica/gpt_model.py (28:38) - megatron_patch/model/llama/gpt_model.py (28:39) duplicated block id: 7186 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/mixtral_bak/layer_specs.py (99:105) - megatron_patch/model/qwen2/layer_specs.py (86:92) duplicated block id: 7187 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/mistral/transformer.py (1676:1682) - megatron_patch/model/qwen_vl/transformer.py (1702:1708) duplicated block id: 7188 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (614:620) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (599:605) duplicated block id: 7189 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/qwen1_5/transformer/attention.py (210:217) - megatron_patch/model/qwen2_vl/attention.py (232:239) duplicated block id: 7190 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (614:620) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (588:594) duplicated block id: 7191 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (1549:1555) - megatron_patch/model/llama3/transformer_legacy.py (1663:1669) duplicated block id: 7192 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/transformer/attention.py (707:713) - megatron_patch/model/qwen2/transformer/attention.py (552:558) duplicated block id: 7193 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/qwen2_5_vl/transformer_config.py (62:68) - megatron_patch/model/qwen2_vl/transformer_config.py (54:60) duplicated block id: 7194 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/mixtral_bak/transformer/attention.py (227:233) - megatron_patch/model/qwen2_vl/attention_vision.py (333:339) duplicated block id: 7195 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/deepseek/hf2mcore_deepseek_v3_moe.py (42:50) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2.5_vl.py (45:52) duplicated block id: 7196 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (612:623) - megatron_patch/model/glm130b/transformer.py (578:589) duplicated block id: 7197 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/mixtral_bak/moe/token_dispatcher.py (275:282) - megatron_patch/model/qwen1_5/moe/token_dispatcher.py (238:245) duplicated block id: 7198 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/transformer/mlp.py (186:192) - megatron_patch/model/mixtral_bak/transformer/mlp.py (162:168) duplicated block id: 7199 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/transformer.py (319:325) - megatron_patch/model/qwen/transformer.py (511:517) duplicated block id: 7200 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/mistral/hf2mcore.py (188:194) - toolkits/model_checkpoints_convertor/qwen/hf2megablocks_qwen1.5.py (244:250) duplicated block id: 7201 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_mha_to_moe.py (248:255) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_dense_and_moe_gqa.py (896:903) duplicated block id: 7202 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon40b/checkpoint_reshaping_and_interoperability.py (202:210) - toolkits/model_checkpoints_convertor/glm130b/checkpoint_reshaping_and_interoperability.py (141:149) duplicated block id: 7203 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (507:518) - toolkits/model_checkpoints_convertor/starcoder/checkpoint_reshaping_and_interoperability.py (714:725) duplicated block id: 7204 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/hf2te.py (371:377) - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (409:416) duplicated block id: 7205 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (122:129) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (213:220) duplicated block id: 7206 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/mistral/hf2mcore.py (274:280) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_moe.py (262:268) duplicated block id: 7207 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/reward_model_to_megatron.py (755:763) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (617:625) duplicated block id: 7208 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/moe/experts.py (670:678) - megatron_patch/model/mixtral_bak/transformer/mlp.py (98:106) duplicated block id: 7209 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (1703:1709) - megatron_patch/model/mistral/transformer.py (1676:1682) duplicated block id: 7210 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (409:416) - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (574:580) duplicated block id: 7211 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/mixtral_bak/transformer/attention.py (443:449) - megatron_patch/model/qwen2/transformer/attention.py (552:558) duplicated block id: 7212 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (420:426) - megatron_patch/model/galactica/transformer.py (324:330) duplicated block id: 7213 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/qwen2/transformer/attention.py (533:539) - megatron_patch/model/qwen2_vl/attention_vision.py (677:683) duplicated block id: 7214 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/llama/transformer.py (246:256) - megatron_patch/model/llama3/transformer_legacy.py (397:406) duplicated block id: 7215 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/mixtral_bak/transformer/attention.py (227:233) - megatron_patch/model/qwen2_vl/attention.py (334:340) duplicated block id: 7216 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/mixtral_bak/moe/token_dispatcher.py (275:282) - megatron_patch/model/qwen2/moe/token_dispatcher.py (244:251) duplicated block id: 7217 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (493:499) - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (478:485) duplicated block id: 7218 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/moe/experts.py (671:679) - megatron_patch/model/qwen1_5/transformer/mlp.py (139:147) duplicated block id: 7219 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (1071:1077) - megatron_patch/model/qwen/transformer.py (1146:1152) duplicated block id: 7220 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/checkpoint_reshaping_and_interoperability.py (714:725) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (595:606) duplicated block id: 7221 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/mixtral_bak/moe/experts.py (81:87) - megatron_patch/model/qwen2/moe/experts.py (142:148) duplicated block id: 7222 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (199:207) - toolkits/model_checkpoints_convertor/bloom/reward_model_to_megatron.py (170:178) duplicated block id: 7223 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/glm/checkpoint_reshaping_and_interoperability.py (441:452) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (595:606) duplicated block id: 7224 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/qwen2/transformer/attention.py (310:316) - megatron_patch/model/qwen2_vl/attention.py (446:452) duplicated block id: 7225 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/mixtral_bak/transformer/attention.py (289:295) - megatron_patch/model/qwen1_5/transformer/attention.py (325:331) duplicated block id: 7226 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/moe/experts.py (218:224) - megatron_patch/model/qwen1_5/moe/experts.py (219:225) duplicated block id: 7227 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/transformer.py (374:380) - megatron_patch/model/llava/transformer.py (518:524) duplicated block id: 7228 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/llama3/layer_specs.py (104:110) - megatron_patch/model/qwen2/layer_specs.py (115:121) duplicated block id: 7229 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/transformer/attention.py (344:350) - megatron_patch/model/qwen2/transformer/attention.py (249:255) duplicated block id: 7230 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer/attention.py (559:565) - megatron_patch/model/qwen2_vl/attention.py (696:702) duplicated block id: 7231 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llava/hf2mcore_llava.py (528:535) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_vl.py (327:333) duplicated block id: 7232 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/glm130b/checkpoint_reshaping_and_interoperability.py (141:149) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (251:259) duplicated block id: 7233 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/bloom/gpt_model.py (28:39) - megatron_patch/model/galactica/gpt_model.py (28:38) duplicated block id: 7234 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/qwen2_vl/attention_vision.py (695:701) - megatron_patch/model/qwen2_vl/attention_vision.py (707:713) duplicated block id: 7235 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (161:168) - toolkits/model_checkpoints_convertor/baichuan/hf2te.py (123:130) duplicated block id: 7236 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/llava/transformer.py (966:974) - megatron_patch/model/qwen1_5_megablocks/transformer.py (857:865) duplicated block id: 7237 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/checkpoint_reshaping_and_interoperability.py (170:178) - toolkits/model_checkpoints_convertor/galactica/checkpoint_reshaping_and_interoperability.py (186:194) duplicated block id: 7238 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (172:180) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (201:209) duplicated block id: 7239 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (536:542) - toolkits/model_checkpoints_convertor/yi/checkpoint_reshaping_and_interoperability.py (552:558) duplicated block id: 7240 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/transformer.py (228:242) - megatron_patch/model/mistral/transformer.py (424:438) duplicated block id: 7241 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (829:837) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (853:861) duplicated block id: 7242 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_moe.py (116:122) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_moe.py (318:325) duplicated block id: 7243 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/transformer.py (282:296) - megatron_patch/model/qwen1_5_megablocks/transformer.py (314:328) duplicated block id: 7244 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/llama2/transformer.py (1679:1685) - megatron_patch/model/qwen/transformer.py (1645:1651) duplicated block id: 7245 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer/attention.py (325:331) - megatron_patch/model/qwen1_5/transformer/attention.py (137:143) duplicated block id: 7246 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer/attention.py (547:553) - megatron_patch/model/qwen2_vl/attention_vision.py (707:713) duplicated block id: 7247 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/transformer.py (275:284) - megatron_patch/model/qwen_vl/transformer.py (399:408) duplicated block id: 7248 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (197:205) - toolkits/model_checkpoints_convertor/bloom/reward_model_to_megatron.py (170:178) duplicated block id: 7249 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (164:172) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (214:222) duplicated block id: 7250 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/llava_mcore/vision/vit_layer_specs.py (89:95) - megatron_patch/model/qwen2_vl/layer_specs.py (106:112) duplicated block id: 7251 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/qwen1_5_megablocks/transformer.py (1587:1603) - megatron_patch/model/starcoder/transformer.py (1280:1296) duplicated block id: 7252 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer/attention.py (559:565) - megatron_patch/model/mixtral/transformer/attention.py (695:701) duplicated block id: 7253 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/language_model.py (412:420) - megatron_patch/model/mistral/language_model.py (419:427) duplicated block id: 7254 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (438:445) - megatron_patch/model/glm130b/transformer.py (474:481) duplicated block id: 7255 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/language_model.py (89:95) - megatron_patch/model/falcon40b/language_model.py (89:95) duplicated block id: 7256 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/mixtral_bak/moe/experts.py (81:87) - megatron_patch/model/qwen1_5/moe/experts.py (118:124) duplicated block id: 7257 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore.py (259:265) - toolkits/model_checkpoints_convertor/mistral/hf2mcore.py (398:404) duplicated block id: 7258 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (326:340) - megatron_patch/model/galactica/transformer.py (232:246) duplicated block id: 7259 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/language_model.py (468:474) - megatron_patch/model/llama/language_model.py (526:532) duplicated block id: 7260 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (262:273) - megatron_patch/model/mistral/transformer.py (399:408) duplicated block id: 7261 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/layer_specs.py (121:127) - megatron_patch/model/qwen2/layer_specs.py (86:92) duplicated block id: 7262 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/glm/checkpoint_reshaping_and_interoperability.py (506:514) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (617:625) duplicated block id: 7263 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/llama2/transformer.py (1703:1709) - megatron_patch/model/qwen/transformer.py (1621:1627) duplicated block id: 7264 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/llama2/transformer.py (619:630) - megatron_patch/model/mistral/transformer.py (620:631) duplicated block id: 7265 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/gpt_model.py (28:39) - megatron_patch/model/galactica/gpt_model.py (28:38) duplicated block id: 7266 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/moe/experts.py (370:376) - megatron_patch/model/deepseek_v2/moe/experts.py (462:468) duplicated block id: 7267 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/hf2te.py (194:206) - toolkits/model_checkpoints_convertor/chatglm/checkpoint_reshaping_and_interoperability.py (284:296) duplicated block id: 7268 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/qwen_vl/language_model.py (661:669) - megatron_patch/model/starcoder/language_model.py (555:563) duplicated block id: 7269 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/hf2te.py (123:130) - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (169:176) duplicated block id: 7270 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/qwen_vl/transformer.py (1724:1740) - megatron_patch/model/starcoder/transformer.py (1280:1296) duplicated block id: 7271 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (1003:1009) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (1033:1039) duplicated block id: 7272 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/mixtral_bak/moe/moe_utils.py (61:68) - megatron_patch/model/qwen/transformer.py (165:172) duplicated block id: 7273 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (529:535) - megatron_patch/model/starcoder/transformer.py (524:530) duplicated block id: 7274 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/transformer/attention.py (449:455) - megatron_patch/model/qwen2/transformer/attention.py (133:139) duplicated block id: 7275 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_moe.py (52:60) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2.5_vl.py (54:62) duplicated block id: 7276 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/qwen2_moe/layer_specs.py (312:318) - megatron_patch/model/qwen3_moe/gpt_layer_specs.py (333:339) duplicated block id: 7277 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/language_model.py (502:510) - megatron_patch/model/glm130b/language_model.py (499:507) duplicated block id: 7278 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/moe/experts.py (670:678) - megatron_patch/model/deepseek_v2/moe/shared_experts.py (172:180) duplicated block id: 7279 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/deepseek/hf2mcore_deepseek_v3_moe.py (103:111) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_moe.py (52:60) duplicated block id: 7280 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_mha.py (98:104) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_vl.py (113:119) duplicated block id: 7281 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/llava_mcore/llava_spec.py (73:79) - megatron_patch/model/qwen2/layer_specs.py (86:92) duplicated block id: 7282 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2megablocks_qwen1.5.py (119:125) - toolkits/model_checkpoints_convertor/yi/checkpoint_reshaping_and_interoperability.py (86:92) duplicated block id: 7283 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/language_model.py (651:659) - megatron_patch/model/mistral/language_model.py (644:652) duplicated block id: 7284 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/transformer/attention.py (707:713) - megatron_patch/model/mixtral_bak/transformer/attention.py (431:437) duplicated block id: 7285 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/qwen1_5/transformer/attention.py (137:143) - megatron_patch/model/qwen2/transformer/attention.py (312:318) duplicated block id: 7286 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/galactica/language_model.py (89:95) - megatron_patch/model/llama2/language_model.py (79:85) duplicated block id: 7287 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/llama2/transformer.py (1679:1685) - megatron_patch/model/mistral/transformer.py (1700:1706) duplicated block id: 7288 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/galactica/transformer.py (232:246) - megatron_patch/model/mistral/transformer.py (424:438) duplicated block id: 7289 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/bloom/language_model.py (91:97) - megatron_patch/model/llava/language_model.py (82:88) duplicated block id: 7290 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/bloom/language_model.py (585:593) - megatron_patch/model/qwen/language_model.py (614:622) duplicated block id: 7291 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/multi_latent_attention.py (162:168) - megatron_patch/model/qwen2_vl/attention.py (455:461) duplicated block id: 7292 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer_legacy.py (871:878) - megatron_patch/model/starcoder/transformer.py (848:855) duplicated block id: 7293 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/llama3_1/layer_specs.py (76:82) - megatron_patch/model/mixtral/layer_specs.py (121:127) duplicated block id: 7294 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (143:149) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (233:239) duplicated block id: 7295 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/checkpoint_reshaping_and_interoperability.py (334:343) - toolkits/model_checkpoints_convertor/starcoder/checkpoint_reshaping_and_interoperability.py (359:367) duplicated block id: 7296 size: 7 cleaned lines of code in 2 files: - megatron_patch/data/utils.py (193:200) - megatron_patch/data/utils.py (376:383) duplicated block id: 7297 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/checkpoint_reshaping_and_interoperability.py (602:608) - toolkits/model_checkpoints_convertor/starcoder/checkpoint_reshaping_and_interoperability.py (615:621) duplicated block id: 7298 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/qwen1_5/transformer/attention.py (137:143) - megatron_patch/model/qwen2/transformer/attention.py (321:327) duplicated block id: 7299 size: 7 cleaned lines of code in 2 files: - megatron_patch/data/utils.py (218:226) - megatron_patch/data/utils.py (403:411) duplicated block id: 7300 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/transformer/attention.py (449:455) - megatron_patch/model/mixtral/transformer/attention.py (458:464) duplicated block id: 7301 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/chatglm/checkpoint_reshaping_and_interoperability.py (75:81) - toolkits/model_checkpoints_convertor/qwen/hf2megablocks_qwen1.5.py (119:125) duplicated block id: 7302 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (883:891) - megatron_patch/model/starcoder/transformer.py (848:855) duplicated block id: 7303 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/transformer.py (594:600) - megatron_patch/model/falcon/transformer.py (1252:1258) duplicated block id: 7304 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/galactica/checkpoint_reshaping_and_interoperability.py (477:485) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (536:544) duplicated block id: 7305 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (378:384) - megatron_patch/model/qwen1_5_megablocks/transformer.py (408:414) duplicated block id: 7306 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (535:546) - toolkits/model_checkpoints_convertor/glm/checkpoint_reshaping_and_interoperability.py (441:452) duplicated block id: 7307 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/llama/transformer.py (270:284) - megatron_patch/model/qwen/transformer.py (417:431) duplicated block id: 7308 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/qwen2_vl/attention.py (708:714) - megatron_patch/model/qwen2_vl/attention_vision.py (695:701) duplicated block id: 7309 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/transformer/attention.py (27:33) - megatron_patch/model/qwen1_5/transformer/attention.py (26:32) duplicated block id: 7310 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/transformer/attention.py (695:701) - megatron_patch/model/qwen2_vl/attention.py (708:714) duplicated block id: 7311 size: 7 cleaned lines of code in 2 files: - toolkits/distributed_checkpoints_convertor/impl/deepseek_v3/h2m_synchronizer.py (18:25) - toolkits/distributed_checkpoints_convertor/impl/general/h2m_synchronizer.py (38:45) duplicated block id: 7312 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/gpt_model.py (27:38) - megatron_patch/model/falcon40b/gpt_model.py (28:38) duplicated block id: 7313 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/llama/transformer.py (270:284) - megatron_patch/model/mistral/transformer.py (424:438) duplicated block id: 7314 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/moe/moe_layer.py (120:127) - megatron_patch/model/qwen1_5/moe/moe_layer.py (95:102) duplicated block id: 7315 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/checkpoint_reshaping_and_interoperability.py (714:725) - toolkits/model_checkpoints_convertor/yi/checkpoint_reshaping_and_interoperability.py (514:525) duplicated block id: 7316 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/falcon/transformer.py (1237:1253) - megatron_patch/model/glm130b/transformer.py (1170:1186) duplicated block id: 7317 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/galactica/checkpoint_reshaping_and_interoperability.py (578:584) - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (614:620) duplicated block id: 7318 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/llama/transformer.py (246:256) - megatron_patch/model/llama2/transformer.py (398:407) duplicated block id: 7319 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (85:91) - toolkits/model_checkpoints_convertor/qwen/hf2megablocks_qwen1.5.py (119:125) duplicated block id: 7320 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/mixtral_bak/moe/experts.py (211:218) - megatron_patch/model/qwen1_5/moe/experts.py (203:210) duplicated block id: 7321 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (507:518) - toolkits/model_checkpoints_convertor/glm/checkpoint_reshaping_and_interoperability.py (441:452) duplicated block id: 7322 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/llava/transformer.py (171:178) - megatron_patch/model/mixtral_bak/moe/moe_utils.py (61:68) duplicated block id: 7323 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/glm130b/checkpoint_reshaping_and_interoperability.py (141:149) - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (235:243) duplicated block id: 7324 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (1213:1221) - megatron_patch/model/mistral/transformer.py (1209:1217) duplicated block id: 7325 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/llama2/language_model.py (632:640) - megatron_patch/model/starcoder/language_model.py (555:563) duplicated block id: 7326 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2.5_vl.py (78:86) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_dense_and_moe_gqa.py (56:64) duplicated block id: 7327 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/moe/experts.py (462:468) - megatron_patch/model/mixtral/moe/experts.py (371:377) duplicated block id: 7328 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/language_model.py (124:130) - megatron_patch/model/galactica/language_model.py (89:95) duplicated block id: 7329 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (143:149) - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (214:220) duplicated block id: 7330 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/transformer.py (282:296) - megatron_patch/model/llama3/transformer_legacy.py (422:436) duplicated block id: 7331 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/deepseek/hf2mcore_deepseek_v2_moe.py (201:207) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_dense_and_moe_gqa.py (314:320) duplicated block id: 7332 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/reward_model_to_megatron.py (170:178) - toolkits/model_checkpoints_convertor/falcon40b/checkpoint_reshaping_and_interoperability.py (202:210) duplicated block id: 7333 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/glm130b/checkpoint_reshaping_and_interoperability.py (141:149) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (231:239) duplicated block id: 7334 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/language_model.py (90:96) - megatron_patch/model/qwen_vl/language_model.py (80:86) duplicated block id: 7335 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/mistral/hf2mcore.py (197:203) - toolkits/model_checkpoints_convertor/mistral/hf2mcore.py (398:404) duplicated block id: 7336 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/llama/transformer.py (362:368) - megatron_patch/model/llama2/transformer.py (517:523) duplicated block id: 7337 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (287:301) - megatron_patch/model/starcoder/transformer.py (328:342) duplicated block id: 7338 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/qwen1_5/transformer/attention.py (316:322) - megatron_patch/model/qwen1_5/transformer/attention.py (325:331) duplicated block id: 7339 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/mixtral_bak/transformer/attention.py (298:304) - megatron_patch/model/qwen2/transformer/attention.py (133:139) duplicated block id: 7340 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/llava/transformer.py (518:524) - megatron_patch/model/starcoder/transformer.py (524:530) duplicated block id: 7341 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_mha_to_moe.py (49:57) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2.5_vl.py (54:62) duplicated block id: 7342 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/galactica/language_model.py (89:95) - megatron_patch/model/qwen/language_model.py (79:85) duplicated block id: 7343 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (287:301) - megatron_patch/model/qwen/transformer.py (417:431) duplicated block id: 7344 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/mistral/transformer.py (1676:1682) - megatron_patch/model/qwen/transformer.py (1645:1651) duplicated block id: 7345 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/language_model.py (615:623) - megatron_patch/model/qwen/language_model.py (614:622) duplicated block id: 7346 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/galactica/checkpoint_reshaping_and_interoperability.py (477:485) - toolkits/model_checkpoints_convertor/yi/checkpoint_reshaping_and_interoperability.py (451:459) duplicated block id: 7347 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/language_model.py (182:194) - megatron_patch/model/starcoder/language_model.py (167:179) duplicated block id: 7348 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (604:611) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (533:540) duplicated block id: 7349 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (772:778) - megatron_patch/model/bloom/transformer.py (1154:1160) duplicated block id: 7350 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/moe/token_dispatcher.py (540:549) - megatron_patch/model/qwen1_5/moe/token_dispatcher.py (475:483) duplicated block id: 7351 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/galactica/transformer.py (867:883) - megatron_patch/model/glm130b/transformer.py (1170:1186) duplicated block id: 7352 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/galactica/checkpoint_reshaping_and_interoperability.py (529:535) - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (478:485) duplicated block id: 7353 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/moe/experts.py (139:145) - megatron_patch/model/qwen2/moe/experts.py (107:113) duplicated block id: 7354 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (420:426) - megatron_patch/model/starcoder/transformer.py (524:530) duplicated block id: 7355 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/falcon/transformer.py (964:970) - megatron_patch/model/falcon40b/transformer.py (991:997) duplicated block id: 7356 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_mha.py (98:104) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_dense_and_moe_gqa.py (110:116) duplicated block id: 7357 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer_legacy.py (1663:1669) - megatron_patch/model/mistral/transformer.py (1676:1682) duplicated block id: 7358 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/llama/transformer.py (270:284) - megatron_patch/model/llama2/transformer.py (423:437) duplicated block id: 7359 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer/mlp.py (123:131) - megatron_patch/model/mixtral/moe/experts.py (671:679) duplicated block id: 7360 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/qwen/transformer.py (1645:1651) - megatron_patch/model/qwen_vl/transformer.py (1678:1684) duplicated block id: 7361 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/transformer/attention.py (458:464) - megatron_patch/model/qwen2/transformer/attention.py (133:139) duplicated block id: 7362 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/mixtral_bak/transformer/mlp.py (162:168) - megatron_patch/model/qwen2/transformer/mlp.py (208:214) duplicated block id: 7363 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (109:115) - toolkits/model_checkpoints_convertor/mistral/hf2mcore.py (93:99) duplicated block id: 7364 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/mixtral_bak/transformer/mlp.py (162:168) - megatron_patch/model/qwen2/transformer/mlp.py (259:265) duplicated block id: 7365 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon40b/checkpoint_reshaping_and_interoperability.py (471:477) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (396:403) duplicated block id: 7366 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/qwen1_5/transformer/attention.py (362:368) - megatron_patch/model/qwen2_vl/attention_vision.py (496:502) duplicated block id: 7367 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (535:546) - toolkits/model_checkpoints_convertor/glm130b/checkpoint_reshaping_and_interoperability.py (422:433) duplicated block id: 7368 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (1679:1685) - megatron_patch/model/llava/transformer.py (1702:1708) duplicated block id: 7369 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (396:403) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (581:587) duplicated block id: 7370 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/falcon/transformer.py (757:768) - megatron_patch/model/glm130b/transformer.py (578:589) duplicated block id: 7371 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/mistral/hf2mcore.py (93:99) - toolkits/model_checkpoints_convertor/yi/checkpoint_reshaping_and_interoperability.py (86:92) duplicated block id: 7372 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/mixtral_bak/transformer/attention.py (289:295) - megatron_patch/model/qwen1_5/transformer/attention.py (137:143) duplicated block id: 7373 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (1549:1555) - megatron_patch/model/qwen1_5_megablocks/transformer.py (1565:1571) duplicated block id: 7374 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/qwen2/transformer/attention.py (552:558) - megatron_patch/model/qwen2_vl/attention.py (708:714) duplicated block id: 7375 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/transformer.py (412:418) - megatron_patch/model/qwen/transformer.py (511:517) duplicated block id: 7376 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (478:485) - toolkits/model_checkpoints_convertor/yi/checkpoint_reshaping_and_interoperability.py (500:506) duplicated block id: 7377 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (809:820) - megatron_patch/model/glm130b/transformer.py (654:665) duplicated block id: 7378 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/llava/transformer.py (919:927) - megatron_patch/model/starcoder/transformer.py (848:855) duplicated block id: 7379 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/transformer_block.py (278:284) - megatron_patch/model/qwen2_5_vl/transformer_block.py (324:330) duplicated block id: 7380 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer/attention.py (547:553) - megatron_patch/model/qwen2/transformer/attention.py (564:570) duplicated block id: 7381 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/deepseek/hf2mcore_deepseek_v3_moe.py (103:111) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_mha_to_moe.py (49:57) duplicated block id: 7382 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/transformer.py (228:242) - megatron_patch/model/qwen1_5_megablocks/transformer.py (314:328) duplicated block id: 7383 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon40b/checkpoint_reshaping_and_interoperability.py (427:434) - toolkits/model_checkpoints_convertor/galactica/checkpoint_reshaping_and_interoperability.py (477:485) duplicated block id: 7384 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/moe/token_dispatcher.py (540:549) - megatron_patch/model/qwen2/moe/token_dispatcher.py (521:530) duplicated block id: 7385 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer/attention.py (325:331) - megatron_patch/model/mixtral_bak/transformer/attention.py (111:117) duplicated block id: 7386 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/mixtral_bak/transformer/attention.py (443:449) - megatron_patch/model/qwen2_vl/attention.py (696:702) duplicated block id: 7387 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/checkpoint_reshaping_and_interoperability.py (170:178) - toolkits/model_checkpoints_convertor/yi/checkpoint_reshaping_and_interoperability.py (184:192) duplicated block id: 7388 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/llama3/layer_specs.py (76:82) - megatron_patch/model/llava_mcore/vision/vit_layer_specs.py (69:75) duplicated block id: 7389 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/llava_mcore/vision/vit_layer_specs.py (89:95) - megatron_patch/model/qwen2/layer_specs.py (115:121) duplicated block id: 7390 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/language_model.py (709:717) - megatron_patch/model/starcoder/language_model.py (555:563) duplicated block id: 7391 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/llava_mcore/layer_specs.py (105:111) - megatron_patch/model/qwen2_moe/layer_specs.py (279:285) duplicated block id: 7392 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/model.py (101:107) - megatron_patch/model/llama3_1/model.py (104:110) duplicated block id: 7393 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (212:221) - megatron_patch/model/baichuan2/transformer.py (304:313) duplicated block id: 7394 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/starcoder/checkpoint_reshaping_and_interoperability.py (192:200) - toolkits/model_checkpoints_convertor/yi/checkpoint_reshaping_and_interoperability.py (184:192) duplicated block id: 7395 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2megablocks_qwen1.5.py (119:125) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (114:120) duplicated block id: 7396 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/hf2te.py (396:402) - toolkits/model_checkpoints_convertor/falcon40b/checkpoint_reshaping_and_interoperability.py (522:528) duplicated block id: 7397 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon40b/checkpoint_reshaping_and_interoperability.py (202:210) - toolkits/model_checkpoints_convertor/starcoder/checkpoint_reshaping_and_interoperability.py (192:200) duplicated block id: 7398 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/language_model.py (497:503) - megatron_patch/model/qwen_vl/language_model.py (519:525) duplicated block id: 7399 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_moe.py (52:60) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_moe.py (112:120) duplicated block id: 7400 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (812:818) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (981:987) duplicated block id: 7401 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/transformer.py (594:600) - megatron_patch/model/falcon40b/transformer.py (991:997) duplicated block id: 7402 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (161:168) - toolkits/model_checkpoints_convertor/baichuan2/hf2te.py (122:129) duplicated block id: 7403 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon/configuration_RW.py (71:79) - toolkits/model_checkpoints_convertor/falcon/configuration_RW.py (126:134) duplicated block id: 7404 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/transformer.py (282:296) - megatron_patch/model/llama2/transformer.py (423:437) duplicated block id: 7405 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (531:537) - toolkits/model_checkpoints_convertor/baichuan2/hf2te.py (396:402) duplicated block id: 7406 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/model.py (101:107) - megatron_patch/model/qwen2_vl/gpt_model.py (78:84) duplicated block id: 7407 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/transformer/attention.py (27:33) - megatron_patch/model/qwen2/transformer/attention.py (22:28) duplicated block id: 7408 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/bloom/language_model.py (585:593) - megatron_patch/model/qwen_vl/language_model.py (661:669) duplicated block id: 7409 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/llama3/layer_specs.py (76:82) - megatron_patch/model/qwen3_moe/gpt_layer_specs.py (240:246) duplicated block id: 7410 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/checkpoint_reshaping_and_interoperability.py (170:178) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (254:262) duplicated block id: 7411 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (440:447) - megatron_patch/model/falcon40b/transformer.py (422:429) duplicated block id: 7412 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/qwen1_5/transformer/attention.py (559:565) - megatron_patch/model/qwen2_vl/attention_vision.py (695:701) duplicated block id: 7413 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/transformer/attention.py (695:701) - megatron_patch/model/qwen2/transformer/attention.py (564:570) duplicated block id: 7414 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/falcon/transformer.py (512:523) - megatron_patch/model/glm130b/transformer.py (367:380) duplicated block id: 7415 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/deepseek/hf2mcore_deepseek_v3_moe.py (663:669) - toolkits/model_checkpoints_convertor/llama/hf2mcore_llama3_1.py (636:642) duplicated block id: 7416 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/moe/experts.py (138:144) - megatron_patch/model/qwen1_5/moe/experts.py (83:89) duplicated block id: 7417 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/llava/transformer.py (1678:1684) - megatron_patch/model/qwen1_5_megablocks/transformer.py (1565:1571) duplicated block id: 7418 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/llama3_1/layer_specs.py (104:110) - megatron_patch/model/qwen2/layer_specs.py (115:121) duplicated block id: 7419 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/language_model.py (124:130) - megatron_patch/model/bloom/language_model.py (91:97) duplicated block id: 7420 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (531:537) - toolkits/model_checkpoints_convertor/falcon40b/checkpoint_reshaping_and_interoperability.py (522:528) duplicated block id: 7421 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/qwen1_5_megablocks/language_model.py (631:639) - megatron_patch/model/starcoder/language_model.py (555:563) duplicated block id: 7422 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/glm/checkpoint_reshaping_and_interoperability.py (141:149) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (254:262) duplicated block id: 7423 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/gpt_model.py (27:38) - megatron_patch/model/galactica/gpt_model.py (28:38) duplicated block id: 7424 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/falcon/transformer.py (265:275) - megatron_patch/model/qwen1_5_megablocks/transformer.py (289:298) duplicated block id: 7425 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/language_model.py (499:507) - megatron_patch/model/llama3/language_model.py (486:494) duplicated block id: 7426 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (378:384) - megatron_patch/model/mistral/transformer.py (518:524) duplicated block id: 7427 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/llama/transformer.py (362:368) - megatron_patch/model/llama3/transformer_legacy.py (516:522) duplicated block id: 7428 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/llama/transformer.py (270:284) - megatron_patch/model/llava/transformer.py (424:438) duplicated block id: 7429 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/qwen1_5/transformer/attention.py (528:534) - megatron_patch/model/qwen2_vl/attention_vision.py (677:683) duplicated block id: 7430 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/llama/language_model.py (89:95) - megatron_patch/model/llama3/language_model.py (68:74) duplicated block id: 7431 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/galactica/checkpoint_reshaping_and_interoperability.py (578:584) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (621:627) duplicated block id: 7432 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/qwen1_5/transformer/attention.py (559:565) - megatron_patch/model/qwen2/transformer/attention.py (552:558) duplicated block id: 7433 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/mistral/hf2mcore.py (40:47) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (53:60) duplicated block id: 7434 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/qwen/transformer.py (1621:1627) - megatron_patch/model/qwen_vl/transformer.py (1702:1708) duplicated block id: 7435 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/checkpoint_reshaping_and_interoperability.py (756:764) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (617:625) duplicated block id: 7436 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/galactica/language_model.py (89:95) - megatron_patch/model/llava/language_model.py (82:88) duplicated block id: 7437 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/mixtral_bak/transformer/attention.py (289:295) - megatron_patch/model/mixtral_bak/transformer/attention.py (298:304) duplicated block id: 7438 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_dense_and_moe_gqa.py (207:213) - toolkits/model_checkpoints_convertor/qwen/hf2megablocks_qwen1.5.py (263:269) duplicated block id: 7439 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (809:820) - megatron_patch/model/llava/transformer.py (845:856) duplicated block id: 7440 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/mistral/transformer.py (1676:1682) - megatron_patch/model/mistral/transformer.py (1700:1706) duplicated block id: 7441 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore.py (250:256) - toolkits/model_checkpoints_convertor/qwen/hf2megablocks_qwen1.5.py (244:250) duplicated block id: 7442 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/hf2te.py (343:349) - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (478:485) duplicated block id: 7443 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_dense_and_moe_gqa.py (244:250) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_moe.py (262:268) duplicated block id: 7444 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/llama/transformer.py (751:757) - megatron_patch/model/starcoder/transformer.py (950:956) duplicated block id: 7445 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (559:565) - toolkits/model_checkpoints_convertor/baichuan2/hf2te.py (396:402) duplicated block id: 7446 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/transformer.py (374:380) - megatron_patch/model/llama2/transformer.py (517:523) duplicated block id: 7447 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer/attention.py (142:149) - megatron_patch/model/qwen2_vl/attention_vision.py (158:165) duplicated block id: 7448 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer_legacy.py (1639:1645) - megatron_patch/model/qwen_vl/transformer.py (1702:1708) duplicated block id: 7449 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/llama3/language_model.py (481:487) - megatron_patch/model/qwen_vl/language_model.py (519:525) duplicated block id: 7450 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (1549:1555) - megatron_patch/model/llava/transformer.py (1702:1708) duplicated block id: 7451 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/language_model.py (412:420) - megatron_patch/model/llama/language_model.py (398:406) duplicated block id: 7452 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/mixtral_bak/transformer/attention.py (431:437) - megatron_patch/model/qwen2_vl/attention_vision.py (707:713) duplicated block id: 7453 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/falcon/language_model.py (89:95) - megatron_patch/model/llama2/language_model.py (79:85) duplicated block id: 7454 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (493:499) - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (409:416) duplicated block id: 7455 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/moe/token_dispatcher.py (254:260) - megatron_patch/model/qwen1_5/moe/token_dispatcher.py (305:311) duplicated block id: 7456 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/chatglm/checkpoint_reshaping_and_interoperability.py (292:306) - toolkits/model_checkpoints_convertor/glm130b/checkpoint_reshaping_and_interoperability.py (265:279) duplicated block id: 7457 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/transformer/attention.py (677:683) - megatron_patch/model/qwen2/transformer/attention.py (533:539) duplicated block id: 7458 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (287:301) - megatron_patch/model/llama3/transformer_legacy.py (422:436) duplicated block id: 7459 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/llama2/transformer.py (171:178) - megatron_patch/model/mixtral_bak/moe/moe_utils.py (61:68) duplicated block id: 7460 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (378:384) - megatron_patch/model/qwen/transformer.py (511:517) duplicated block id: 7461 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer/attention.py (547:553) - megatron_patch/model/mixtral_bak/transformer/attention.py (443:449) duplicated block id: 7462 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/layer_specs.py (117:123) - megatron_patch/model/llava_mcore/layer_specs.py (105:111) duplicated block id: 7463 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (199:207) - toolkits/model_checkpoints_convertor/glm130b/checkpoint_reshaping_and_interoperability.py (141:149) duplicated block id: 7464 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/deepseek/hf2mcore_deepseek_v3_moe.py (42:50) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_vl.py (45:52) duplicated block id: 7465 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer/attention.py (547:553) - megatron_patch/model/qwen2_vl/attention.py (708:714) duplicated block id: 7466 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/mixtral_bak/transformer/mlp.py (154:160) - megatron_patch/model/qwen2/transformer/mlp.py (251:257) duplicated block id: 7467 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_mha_to_moe.py (169:176) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_dense_and_moe_gqa.py (197:203) duplicated block id: 7468 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer_legacy.py (1639:1645) - megatron_patch/model/mistral/transformer.py (1700:1706) duplicated block id: 7469 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/qwen/transformer.py (865:873) - megatron_patch/model/starcoder/transformer.py (848:855) duplicated block id: 7470 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/language_model.py (412:420) - megatron_patch/model/qwen1_5_megablocks/language_model.py (416:424) duplicated block id: 7471 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/reward_model_to_megatron.py (713:724) - toolkits/model_checkpoints_convertor/yi/checkpoint_reshaping_and_interoperability.py (514:525) duplicated block id: 7472 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/qwen2_vl/attention.py (696:702) - megatron_patch/model/qwen2_vl/attention_vision.py (707:713) duplicated block id: 7473 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/qwen1_5/layer_specs.py (76:82) - megatron_patch/model/qwen3_moe/gpt_layer_specs.py (240:246) duplicated block id: 7474 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (478:485) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (551:557) duplicated block id: 7475 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/transformer.py (258:268) - megatron_patch/model/llama2/transformer.py (398:407) duplicated block id: 7476 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2megablocks_qwen1.5.py (119:125) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (113:119) duplicated block id: 7477 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/transformer.py (578:589) - megatron_patch/model/llama3/transformer_legacy.py (710:721) duplicated block id: 7478 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/qwen2/transformer_block.py (251:257) - megatron_patch/model/qwen2/transformer_block.py (262:268) duplicated block id: 7479 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_mha.py (123:129) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_moe.py (262:268) duplicated block id: 7480 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/deepseek/hf2mcore_deepseek_v3_moe.py (42:50) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_dense_and_moe_gqa.py (29:36) duplicated block id: 7481 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/qwen1_5/transformer/mlp.py (99:105) - megatron_patch/model/qwen2/transformer/mlp.py (72:78) duplicated block id: 7482 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/falcon/transformer.py (547:553) - megatron_patch/model/mistral/transformer.py (518:524) duplicated block id: 7483 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/glm130b/checkpoint_reshaping_and_interoperability.py (32:38) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (53:60) duplicated block id: 7484 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/transformer.py (991:997) - megatron_patch/model/galactica/transformer.py (618:624) duplicated block id: 7485 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/hf2te.py (396:402) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (621:627) duplicated block id: 7486 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/falcon/transformer.py (964:970) - megatron_patch/model/falcon/transformer.py (1252:1258) duplicated block id: 7487 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/reward_model_to_megatron.py (601:607) - toolkits/model_checkpoints_convertor/starcoder/checkpoint_reshaping_and_interoperability.py (615:621) duplicated block id: 7488 size: 7 cleaned lines of code in 2 files: - toolkits/distributed_checkpoints_convertor/impl/deepseek_v3/h2m_synchronizer.py (36:42) - toolkits/distributed_checkpoints_convertor/impl/deepseek_v3/m2h_synchronizer.py (31:37) duplicated block id: 7489 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/llava_mcore/llava_model.py (206:218) - megatron_patch/model/qwen2_5_vl/model.py (125:136) duplicated block id: 7490 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/language_model.py (625:633) - megatron_patch/model/starcoder/language_model.py (555:563) duplicated block id: 7491 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_mha.py (162:169) - toolkits/model_checkpoints_convertor/qwen/hf2megablocks_qwen1.5.py (363:370) duplicated block id: 7492 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer/attention.py (137:143) - megatron_patch/model/mixtral/transformer/attention.py (458:464) duplicated block id: 7493 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/galactica/transformer.py (208:218) - megatron_patch/model/starcoder/transformer.py (425:435) duplicated block id: 7494 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer/attention.py (137:143) - megatron_patch/model/mixtral/transformer/attention.py (449:455) duplicated block id: 7495 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/transformer.py (374:380) - megatron_patch/model/qwen_vl/transformer.py (518:524) duplicated block id: 7496 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/language_model.py (90:96) - megatron_patch/model/llava/language_model.py (82:88) duplicated block id: 7497 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (109:115) - toolkits/model_checkpoints_convertor/qwen/hf2megablocks_qwen1.5.py (119:125) duplicated block id: 7498 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/moe/router.py (10:18) - megatron_patch/model/qwen1_5/moe/router.py (27:35) duplicated block id: 7499 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/galactica/transformer.py (324:330) - megatron_patch/model/mistral/transformer.py (518:524) duplicated block id: 7500 size: 7 cleaned lines of code in 2 files: - megatron_patch/tokenizer/tokenization_baichuan.py (12:20) - megatron_patch/tokenizer/tokenization_yi.py (9:17) duplicated block id: 7501 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/mistral/transformer.py (1722:1738) - megatron_patch/model/starcoder/transformer.py (1280:1296) duplicated block id: 7502 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/transformer.py (319:325) - megatron_patch/model/llava/transformer.py (518:524) duplicated block id: 7503 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_moe.py (132:138) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_moe.py (338:344) duplicated block id: 7504 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/language_model.py (89:95) - megatron_patch/model/llama3/language_model.py (68:74) duplicated block id: 7505 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (723:734) - megatron_patch/model/glm130b/transformer.py (578:589) duplicated block id: 7506 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/llama3/layer_specs.py (104:110) - megatron_patch/model/qwen2_vl/layer_specs.py (106:112) duplicated block id: 7507 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/transformer.py (282:296) - megatron_patch/model/qwen/transformer.py (417:431) duplicated block id: 7508 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/galactica/transformer.py (324:330) - megatron_patch/model/qwen1_5_megablocks/transformer.py (408:414) duplicated block id: 7509 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/deepseek/hf2mcore_deepseek_v3_moe.py (42:50) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_moe.py (43:53) duplicated block id: 7510 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (529:535) - megatron_patch/model/chatglm/transformer.py (319:325) duplicated block id: 7511 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer/attention.py (316:322) - megatron_patch/model/mixtral_bak/transformer/attention.py (111:117) duplicated block id: 7512 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/llama3_1/layer_specs.py (104:110) - megatron_patch/model/mixtral_bak/layer_specs.py (138:144) duplicated block id: 7513 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/transformer.py (228:242) - megatron_patch/model/starcoder/transformer.py (328:342) duplicated block id: 7514 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/falcon/language_model.py (89:95) - megatron_patch/model/llava/language_model.py (82:88) duplicated block id: 7515 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/qwen2/transformer_block.py (389:395) - megatron_patch/model/qwen2_5_vl/transformer_block.py (542:548) duplicated block id: 7516 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/gpt_model.py (26:37) - megatron_patch/model/chatglm/gpt_model.py (28:39) duplicated block id: 7517 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (420:426) - megatron_patch/model/glm130b/transformer.py (412:418) duplicated block id: 7518 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (1610:1616) - megatron_patch/model/baichuan2/transformer.py (1203:1209) duplicated block id: 7519 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/qwen1_5/transformer/attention.py (253:259) - megatron_patch/model/qwen2_vl/attention_vision.py (333:339) duplicated block id: 7520 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/transformer.py (578:589) - megatron_patch/model/qwen_vl/transformer.py (731:742) duplicated block id: 7521 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/llama2/transformer.py (1679:1685) - megatron_patch/model/llava/transformer.py (1702:1708) duplicated block id: 7522 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (1679:1685) - megatron_patch/model/qwen1_5_megablocks/transformer.py (1565:1571) duplicated block id: 7523 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/moe/experts.py (671:679) - megatron_patch/model/mixtral/transformer/mlp.py (129:137) duplicated block id: 7524 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/transformer.py (483:492) - megatron_patch/model/llama/transformer.py (437:445) duplicated block id: 7525 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (828:836) - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (976:984) duplicated block id: 7526 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/language_model.py (350:356) - megatron_patch/model/qwen/language_model.py (334:340) duplicated block id: 7527 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/qwen2/layer_specs.py (86:92) - megatron_patch/model/qwen3_moe/gpt_layer_specs.py (240:246) duplicated block id: 7528 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (435:449) - megatron_patch/model/galactica/transformer.py (232:246) duplicated block id: 7529 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (529:535) - megatron_patch/model/falcon/transformer.py (547:553) duplicated block id: 7530 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/language_model.py (89:95) - megatron_patch/model/bloom/language_model.py (91:97) duplicated block id: 7531 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/language_model.py (90:96) - megatron_patch/model/llama3/language_model.py (68:74) duplicated block id: 7532 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/language_model.py (90:96) - megatron_patch/model/qwen_vl/language_model.py (80:86) duplicated block id: 7533 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (1549:1555) - megatron_patch/model/baichuan/transformer.py (1573:1579) duplicated block id: 7534 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/transformer/attention.py (458:464) - megatron_patch/model/qwen2/transformer/attention.py (312:318) duplicated block id: 7535 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (378:384) - megatron_patch/model/llama2/transformer.py (517:523) duplicated block id: 7536 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/llama2/transformer.py (1703:1709) - megatron_patch/model/qwen_vl/transformer.py (1678:1684) duplicated block id: 7537 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (199:207) - toolkits/model_checkpoints_convertor/bloom/checkpoint_reshaping_and_interoperability.py (170:178) duplicated block id: 7538 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/transformer.py (412:418) - megatron_patch/model/llava/transformer.py (518:524) duplicated block id: 7539 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_dense_and_moe_gqa.py (664:670) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_moe.py (604:611) duplicated block id: 7540 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/galactica/language_model.py (211:224) - megatron_patch/model/starcoder/language_model.py (167:179) duplicated block id: 7541 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer/attention.py (362:368) - megatron_patch/model/mixtral/transformer/attention.py (496:502) duplicated block id: 7542 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (1703:1709) - megatron_patch/model/llama3/transformer_legacy.py (1639:1645) duplicated block id: 7543 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/mixtral_bak/model.py (82:88) - megatron_patch/model/qwen2/model.py (90:96) duplicated block id: 7544 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/falcon/transformer.py (1252:1258) - megatron_patch/model/falcon40b/transformer.py (734:740) duplicated block id: 7545 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/transformer/attention.py (165:171) - megatron_patch/model/qwen1_5/transformer/attention.py (316:322) duplicated block id: 7546 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/transformer/attention.py (165:171) - megatron_patch/model/qwen1_5/transformer/attention.py (325:331) duplicated block id: 7547 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/mistral/transformer.py (1700:1706) - megatron_patch/model/qwen/transformer.py (1621:1627) duplicated block id: 7548 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/language_model.py (503:509) - megatron_patch/model/glm130b/language_model.py (468:474) duplicated block id: 7549 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/multi_latent_attention.py (162:168) - megatron_patch/model/mixtral_bak/transformer/attention.py (295:301) duplicated block id: 7550 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (841:847) - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (957:963) duplicated block id: 7551 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (723:734) - megatron_patch/model/llama/transformer.py (529:540) duplicated block id: 7552 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/llava_mcore/vision/vit_layer_specs.py (46:52) - megatron_patch/model/mixtral/layer_specs.py (85:91) duplicated block id: 7553 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer/attention.py (559:565) - megatron_patch/model/mixtral_bak/transformer/attention.py (431:437) duplicated block id: 7554 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/language_model.py (124:130) - megatron_patch/model/llama/language_model.py (89:95) duplicated block id: 7555 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/llama3_1/model.py (104:110) - megatron_patch/model/mixtral_bak/model.py (82:88) duplicated block id: 7556 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/qwen/transformer.py (912:920) - megatron_patch/model/qwen1_5_megablocks/transformer.py (857:865) duplicated block id: 7557 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/qwen2_moe/layer_specs.py (327:333) - megatron_patch/model/qwen3_moe/gpt_layer_specs.py (351:357) duplicated block id: 7558 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/language_model.py (124:130) - megatron_patch/model/falcon/language_model.py (89:95) duplicated block id: 7559 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer_legacy.py (1639:1645) - megatron_patch/model/llama3/transformer_legacy.py (1663:1669) duplicated block id: 7560 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (559:565) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (599:605) duplicated block id: 7561 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (559:565) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (588:594) duplicated block id: 7562 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/language_model.py (615:623) - megatron_patch/model/mistral/language_model.py (644:652) duplicated block id: 7563 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/llava_mcore/vision/vit_layer_specs.py (89:95) - megatron_patch/model/mixtral/layer_specs.py (151:157) duplicated block id: 7564 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/bloom/language_model.py (183:195) - megatron_patch/model/starcoder/language_model.py (167:179) duplicated block id: 7565 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2.5_vl.py (661:667) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_dense_and_moe_gqa.py (604:610) duplicated block id: 7566 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer/attention.py (316:322) - megatron_patch/model/mixtral/transformer/attention.py (458:464) duplicated block id: 7567 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/qwen/transformer.py (1645:1651) - megatron_patch/model/qwen1_5_megablocks/transformer.py (1541:1547) duplicated block id: 7568 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/qwen/transformer.py (511:517) - megatron_patch/model/starcoder/transformer.py (524:530) duplicated block id: 7569 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (287:301) - megatron_patch/model/llama2/transformer.py (423:437) duplicated block id: 7570 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer/attention.py (325:331) - megatron_patch/model/mixtral/transformer/attention.py (449:455) duplicated block id: 7571 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore.py (332:338) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_moe.py (262:268) duplicated block id: 7572 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/llava_mcore/vision/vit_layer_specs.py (69:75) - megatron_patch/model/qwen2/layer_specs.py (86:92) duplicated block id: 7573 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/checkpoint_reshaping_and_interoperability.py (714:725) - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (588:599) duplicated block id: 7574 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon/checkpoint_reshaping_and_interoperability.py (482:488) - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (478:485) duplicated block id: 7575 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_dense_and_moe_gqa.py (604:610) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_vl.py (668:674) duplicated block id: 7576 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/qwen2/transformer_block.py (401:407) - megatron_patch/model/qwen2_5_vl/transformer_block.py (562:568) duplicated block id: 7577 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/reward_model_to_megatron.py (35:41) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (53:60) duplicated block id: 7578 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/qwen1_5/transformer/attention.py (547:553) - megatron_patch/model/qwen1_5/transformer/attention.py (559:565) duplicated block id: 7579 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/qwen1_5_megablocks/transformer.py (1541:1547) - megatron_patch/model/qwen1_5_megablocks/transformer.py (1565:1571) duplicated block id: 7580 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer/attention.py (142:149) - megatron_patch/model/qwen2_vl/attention.py (159:166) duplicated block id: 7581 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/multi_latent_attention.py (162:168) - megatron_patch/model/qwen2_vl/attention_vision.py (454:460) duplicated block id: 7582 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (599:605) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (621:627) duplicated block id: 7583 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/language_model.py (412:420) - megatron_patch/model/qwen/language_model.py (414:422) duplicated block id: 7584 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/hf2te.py (371:377) - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (478:485) duplicated block id: 7585 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/language_model.py (89:95) - megatron_patch/model/mistral/language_model.py (80:86) duplicated block id: 7586 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer/attention.py (316:322) - megatron_patch/model/mixtral/transformer/attention.py (165:171) duplicated block id: 7587 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer/attention.py (547:553) - megatron_patch/model/qwen1_5/transformer/attention.py (559:565) duplicated block id: 7588 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/mistral/transformer.py (171:178) - megatron_patch/model/mixtral_bak/moe/moe_utils.py (61:68) duplicated block id: 7589 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/hf2te.py (424:430) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (599:605) duplicated block id: 7590 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (231:239) - toolkits/model_checkpoints_convertor/starcoder/checkpoint_reshaping_and_interoperability.py (192:200) duplicated block id: 7591 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/reward_model_to_megatron.py (170:178) - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (272:280) duplicated block id: 7592 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (1071:1077) - megatron_patch/model/baichuan/transformer.py (1610:1616) duplicated block id: 7593 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/qwen1_5/transformer/attention.py (142:149) - megatron_patch/model/qwen2_vl/attention.py (159:166) duplicated block id: 7594 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon/checkpoint_reshaping_and_interoperability.py (554:562) - toolkits/model_checkpoints_convertor/galactica/checkpoint_reshaping_and_interoperability.py (614:622) duplicated block id: 7595 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/falcon/transformer.py (964:970) - megatron_patch/model/llama/transformer.py (1046:1052) duplicated block id: 7596 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/transformer.py (258:268) - megatron_patch/model/qwen/transformer.py (392:401) duplicated block id: 7597 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/transformer.py (319:325) - megatron_patch/model/llama2/transformer.py (517:523) duplicated block id: 7598 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/multi_latent_attention.py (255:261) - megatron_patch/model/deepseek_v2/multi_latent_attention.py (267:273) duplicated block id: 7599 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer_legacy.py (1639:1645) - megatron_patch/model/llava/transformer.py (1702:1708) duplicated block id: 7600 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/qwen2/transformer_block.py (262:268) - megatron_patch/model/qwen2/transformer_block.py (300:306) duplicated block id: 7601 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/multi_latent_attention.py (217:223) - megatron_patch/model/deepseek_v2/multi_latent_attention.py (267:273) duplicated block id: 7602 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/multi_latent_attention.py (217:223) - megatron_patch/model/deepseek_v2/multi_latent_attention.py (255:261) duplicated block id: 7603 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/falcon/transformer.py (265:275) - megatron_patch/model/qwen/transformer.py (392:401) duplicated block id: 7604 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (529:535) - megatron_patch/model/galactica/transformer.py (324:330) duplicated block id: 7605 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (435:449) - megatron_patch/model/bloom/transformer.py (287:301) duplicated block id: 7606 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/language_model.py (350:356) - megatron_patch/model/qwen_vl/language_model.py (335:341) duplicated block id: 7607 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/llama3/layer_specs.py (104:110) - megatron_patch/model/mixtral_bak/layer_specs.py (138:144) duplicated block id: 7608 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/reward_model_to_megatron.py (170:178) - toolkits/model_checkpoints_convertor/falcon/checkpoint_reshaping_and_interoperability.py (204:212) duplicated block id: 7609 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/language_model.py (90:96) - megatron_patch/model/llava/language_model.py (82:88) duplicated block id: 7610 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (396:403) - toolkits/model_checkpoints_convertor/yi/checkpoint_reshaping_and_interoperability.py (500:506) duplicated block id: 7611 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/multi_latent_attention.py (217:223) - megatron_patch/model/deepseek_v2/multi_latent_attention.py (231:237) duplicated block id: 7612 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (507:518) - toolkits/model_checkpoints_convertor/bloom/reward_model_to_megatron.py (713:724) duplicated block id: 7613 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (454:460) - toolkits/model_checkpoints_convertor/yi/checkpoint_reshaping_and_interoperability.py (552:558) duplicated block id: 7614 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer/attention.py (210:217) - megatron_patch/model/qwen2_vl/attention_vision.py (231:238) duplicated block id: 7615 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/llama/transformer.py (362:368) - megatron_patch/model/qwen_vl/transformer.py (518:524) duplicated block id: 7616 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/language_model.py (124:130) - megatron_patch/model/falcon40b/language_model.py (89:95) duplicated block id: 7617 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (829:837) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (1023:1031) duplicated block id: 7618 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer_legacy.py (1152:1158) - megatron_patch/model/llama3/transformer_legacy.py (1700:1706) duplicated block id: 7619 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (1679:1685) - megatron_patch/model/baichuan2/transformer.py (1703:1709) duplicated block id: 7620 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (531:537) - toolkits/model_checkpoints_convertor/galactica/checkpoint_reshaping_and_interoperability.py (578:584) duplicated block id: 7621 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (531:537) - toolkits/model_checkpoints_convertor/galactica/checkpoint_reshaping_and_interoperability.py (567:573) duplicated block id: 7622 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/model.py (101:107) - megatron_patch/model/llama3/model.py (89:95) duplicated block id: 7623 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/deepseek/hf2mcore_deepseek_v2_moe.py (104:112) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_moe.py (176:184) duplicated block id: 7624 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (1573:1579) - megatron_patch/model/llama2/transformer.py (1679:1685) duplicated block id: 7625 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer/attention.py (547:553) - megatron_patch/model/llama3/transformer/attention.py (559:565) duplicated block id: 7626 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/bloom/language_model.py (91:97) - megatron_patch/model/llama2/language_model.py (79:85) duplicated block id: 7627 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/language_model.py (625:633) - megatron_patch/model/bloom/language_model.py (585:593) duplicated block id: 7628 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer_legacy.py (1663:1669) - megatron_patch/model/qwen1_5_megablocks/transformer.py (1541:1547) duplicated block id: 7629 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (1071:1077) - megatron_patch/model/baichuan2/transformer.py (1203:1209) duplicated block id: 7630 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/mixtral_bak/model.py (99:107) - megatron_patch/model/qwen2_vl/gpt_model.py (98:106) duplicated block id: 7631 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/qwen2_5_vl/transformer_config.py (70:77) - megatron_patch/model/qwen2_vl/transformer_config.py (62:69) duplicated block id: 7632 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/qwen2/transformer/attention.py (358:364) - megatron_patch/model/qwen2_vl/attention_vision.py (496:502) duplicated block id: 7633 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/llama/transformer.py (529:540) - megatron_patch/model/mistral/transformer.py (730:741) duplicated block id: 7634 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/falcon/transformer.py (992:998) - megatron_patch/model/starcoder/transformer.py (950:956) duplicated block id: 7635 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/language_model.py (89:95) - megatron_patch/model/qwen1_5_megablocks/language_model.py (78:84) duplicated block id: 7636 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/llama/transformer.py (362:368) - megatron_patch/model/qwen1_5_megablocks/transformer.py (408:414) duplicated block id: 7637 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/language_model.py (90:96) - megatron_patch/model/qwen/language_model.py (79:85) duplicated block id: 7638 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/qwen1_5/transformer/attention.py (362:368) - megatron_patch/model/qwen2_vl/attention.py (497:503) duplicated block id: 7639 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/qwen1_5/transformer/attention.py (26:32) - megatron_patch/model/qwen2_vl/attention.py (15:21) duplicated block id: 7640 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/qwen2/transformer/attention.py (552:558) - megatron_patch/model/qwen2_vl/attention_vision.py (707:713) duplicated block id: 7641 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/mixtral_bak/transformer/attention.py (443:449) - megatron_patch/model/qwen1_5/transformer/attention.py (547:553) duplicated block id: 7642 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_moe.py (262:268) - toolkits/model_checkpoints_convertor/qwen/hf2megablocks_qwen1.5.py (325:331) duplicated block id: 7643 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon/checkpoint_reshaping_and_interoperability.py (659:665) - toolkits/model_checkpoints_convertor/falcon/checkpoint_reshaping_and_interoperability.py (677:683) duplicated block id: 7644 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_moe.py (262:268) - toolkits/model_checkpoints_convertor/qwen/hf2megablocks_qwen1.5.py (286:292) duplicated block id: 7645 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/galactica/transformer.py (324:330) - megatron_patch/model/qwen/transformer.py (511:517) duplicated block id: 7646 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_mha.py (107:113) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2.5_vl.py (122:128) duplicated block id: 7647 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore.py (269:275) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_dense_and_moe_gqa.py (207:213) duplicated block id: 7648 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llava/hf2mcore_llava.py (528:535) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2.5_vl.py (330:336) duplicated block id: 7649 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/moe/experts.py (17:29) - megatron_patch/model/qwen2/moe/experts.py (28:37) duplicated block id: 7650 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/qwen_vl/transformer.py (1678:1684) - megatron_patch/model/qwen_vl/transformer.py (1702:1708) duplicated block id: 7651 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/language_model.py (90:96) - megatron_patch/model/llama2/language_model.py (79:85) duplicated block id: 7652 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/deepseek/hf2mcore_deepseek_v2_moe.py (118:124) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_vl.py (113:119) duplicated block id: 7653 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/bloom/language_model.py (169:175) - megatron_patch/model/glm130b/language_model.py (169:176) duplicated block id: 7654 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/transformer.py (976:992) - megatron_patch/model/glm130b/transformer.py (1170:1186) duplicated block id: 7655 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/language_model.py (651:659) - megatron_patch/model/qwen/language_model.py (614:622) duplicated block id: 7656 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/llama3/layer_specs.py (104:110) - megatron_patch/model/mixtral/layer_specs.py (151:157) duplicated block id: 7657 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon/checkpoint_reshaping_and_interoperability.py (335:342) - toolkits/model_checkpoints_convertor/falcon40b/checkpoint_reshaping_and_interoperability.py (333:340) duplicated block id: 7658 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_mha.py (107:113) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_vl.py (122:128) duplicated block id: 7659 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (410:419) - megatron_patch/model/llama/transformer.py (246:256) duplicated block id: 7660 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/qwen1_5/transformer/attention.py (325:331) - megatron_patch/model/qwen2/transformer/attention.py (133:139) duplicated block id: 7661 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/qwen1_5/transformer/mlp.py (73:79) - megatron_patch/model/qwen1_5/transformer/mlp.py (99:105) duplicated block id: 7662 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/falcon/language_model.py (89:95) - megatron_patch/model/mistral/language_model.py (80:86) duplicated block id: 7663 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/qwen1_5/transformer/attention.py (316:322) - megatron_patch/model/qwen2/transformer/attention.py (133:139) duplicated block id: 7664 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (134:143) - megatron_patch/model/galactica/transformer.py (112:121) duplicated block id: 7665 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/language_model.py (586:594) - megatron_patch/model/glm130b/language_model.py (499:507) duplicated block id: 7666 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/falcon/transformer.py (547:553) - megatron_patch/model/llama2/transformer.py (517:523) duplicated block id: 7667 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/transformer.py (734:740) - megatron_patch/model/llama/transformer.py (1046:1052) duplicated block id: 7668 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/transformer/attention.py (242:249) - megatron_patch/model/qwen2/transformer/attention.py (206:213) duplicated block id: 7669 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/galactica/transformer.py (232:246) - megatron_patch/model/llama2/transformer.py (423:437) duplicated block id: 7670 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon/checkpoint_reshaping_and_interoperability.py (204:212) - toolkits/model_checkpoints_convertor/glm/checkpoint_reshaping_and_interoperability.py (141:149) duplicated block id: 7671 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/checkpoint_reshaping_and_interoperability.py (170:178) - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (235:243) duplicated block id: 7672 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (1573:1579) - megatron_patch/model/qwen1_5_megablocks/transformer.py (1541:1547) duplicated block id: 7673 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/llama2/transformer.py (1725:1741) - megatron_patch/model/starcoder/transformer.py (1280:1296) duplicated block id: 7674 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (976:984) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (930:938) duplicated block id: 7675 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/moe/experts.py (138:144) - megatron_patch/model/mixtral_bak/moe/experts.py (81:87) duplicated block id: 7676 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2.5_vl.py (143:149) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_moe.py (262:268) duplicated block id: 7677 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/chatglm/checkpoint_reshaping_and_interoperability.py (544:552) - toolkits/model_checkpoints_convertor/falcon/checkpoint_reshaping_and_interoperability.py (554:562) duplicated block id: 7678 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/mixtral_bak/transformer/attention.py (289:295) - megatron_patch/model/qwen2/transformer/attention.py (321:327) duplicated block id: 7679 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/galactica/checkpoint_reshaping_and_interoperability.py (529:535) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (396:403) duplicated block id: 7680 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/llama/transformer.py (637:648) - megatron_patch/model/starcoder/transformer.py (814:824) duplicated block id: 7681 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (507:518) - toolkits/model_checkpoints_convertor/glm130b/checkpoint_reshaping_and_interoperability.py (422:433) duplicated block id: 7682 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/language_model.py (124:130) - megatron_patch/model/chatglm/language_model.py (90:96) duplicated block id: 7683 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/language_model.py (615:623) - megatron_patch/model/qwen1_5_megablocks/language_model.py (631:639) duplicated block id: 7684 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon/checkpoint_reshaping_and_interoperability.py (848:855) - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (829:837) duplicated block id: 7685 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon/checkpoint_reshaping_and_interoperability.py (848:855) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (853:861) duplicated block id: 7686 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore_llama3_1.py (722:729) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2.5_vl.py (330:336) duplicated block id: 7687 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/qwen1_5/transformer/attention.py (559:565) - megatron_patch/model/qwen2_vl/attention.py (696:702) duplicated block id: 7688 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/mixtral_bak/transformer/attention.py (412:418) - megatron_patch/model/qwen2_vl/attention_vision.py (677:683) duplicated block id: 7689 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer_legacy.py (1700:1706) - megatron_patch/model/qwen/transformer.py (1146:1152) duplicated block id: 7690 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (149:158) - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (191:200) duplicated block id: 7691 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_dense_and_moe_gqa.py (322:328) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_moe.py (262:268) duplicated block id: 7692 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/moe/experts.py (371:377) - megatron_patch/model/mixtral/moe/experts.py (463:469) duplicated block id: 7693 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer_legacy.py (1639:1645) - megatron_patch/model/qwen1_5_megablocks/transformer.py (1565:1571) duplicated block id: 7694 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_dense_and_moe_gqa.py (183:189) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_moe.py (262:268) duplicated block id: 7695 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (1679:1685) - megatron_patch/model/qwen_vl/transformer.py (1702:1708) duplicated block id: 7696 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/moe/experts.py (138:144) - megatron_patch/model/qwen2/moe/experts.py (107:113) duplicated block id: 7697 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_dense_and_moe_gqa.py (322:328) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_moe.py (352:358) duplicated block id: 7698 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon/checkpoint_reshaping_and_interoperability.py (204:212) - toolkits/model_checkpoints_convertor/glm130b/checkpoint_reshaping_and_interoperability.py (141:149) duplicated block id: 7699 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/glm/checkpoint_reshaping_and_interoperability.py (441:452) - toolkits/model_checkpoints_convertor/yi/checkpoint_reshaping_and_interoperability.py (514:525) duplicated block id: 7700 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (326:340) - megatron_patch/model/chatglm/transformer.py (228:242) duplicated block id: 7701 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon40b/checkpoint_reshaping_and_interoperability.py (202:210) - toolkits/model_checkpoints_convertor/glm/checkpoint_reshaping_and_interoperability.py (141:149) duplicated block id: 7702 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/chatglm/checkpoint_reshaping_and_interoperability.py (169:177) - toolkits/model_checkpoints_convertor/starcoder/checkpoint_reshaping_and_interoperability.py (192:200) duplicated block id: 7703 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/qwen2/transformer/attention.py (206:213) - megatron_patch/model/qwen2_vl/attention_vision.py (231:238) duplicated block id: 7704 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (1213:1221) - megatron_patch/model/qwen_vl/transformer.py (1211:1219) duplicated block id: 7705 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (447:455) - megatron_patch/model/glm130b/transformer.py (483:492) duplicated block id: 7706 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/llama2/transformer.py (967:975) - megatron_patch/model/qwen1_5_megablocks/transformer.py (857:865) duplicated block id: 7707 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/llama/language_model.py (89:95) - megatron_patch/model/llava/language_model.py (82:88) duplicated block id: 7708 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore.py (259:265) - toolkits/model_checkpoints_convertor/llama/hf2mcore.py (491:497) duplicated block id: 7709 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/llama/transformer.py (529:540) - megatron_patch/model/qwen/transformer.py (705:716) duplicated block id: 7710 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/qwen1_5/transformer/attention.py (316:322) - megatron_patch/model/qwen2/transformer/attention.py (321:327) duplicated block id: 7711 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/language_model.py (412:420) - megatron_patch/model/llama3/language_model.py (408:416) duplicated block id: 7712 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/hf2te.py (343:349) - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (409:416) duplicated block id: 7713 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/glm130b/checkpoint_reshaping_and_interoperability.py (422:433) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (595:606) duplicated block id: 7714 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/transformer.py (972:978) - megatron_patch/model/starcoder/transformer.py (1175:1181) duplicated block id: 7715 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/mixtral_bak/moe/experts.py (116:122) - megatron_patch/model/qwen1_5/moe/experts.py (83:89) duplicated block id: 7716 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/moe/experts.py (219:225) - megatron_patch/model/qwen1_5/moe/experts.py (219:225) duplicated block id: 7717 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/transformer/mlp.py (178:184) - megatron_patch/model/mixtral_bak/transformer/mlp.py (154:160) duplicated block id: 7718 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (186:193) - toolkits/model_checkpoints_convertor/yi/checkpoint_reshaping_and_interoperability.py (154:161) duplicated block id: 7719 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/galactica/transformer.py (324:330) - megatron_patch/model/llama3/transformer_legacy.py (516:522) duplicated block id: 7720 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/llama/transformer.py (270:284) - megatron_patch/model/qwen_vl/transformer.py (424:438) duplicated block id: 7721 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/glm/checkpoint_reshaping_and_interoperability.py (141:149) - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (272:280) duplicated block id: 7722 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/falcon/transformer.py (647:654) - megatron_patch/model/galactica/transformer.py (402:409) duplicated block id: 7723 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (1573:1579) - megatron_patch/model/qwen_vl/transformer.py (1678:1684) duplicated block id: 7724 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/model.py (83:89) - megatron_patch/model/qwen2_vl/gpt_model.py (78:84) duplicated block id: 7725 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/bloom/language_model.py (585:593) - megatron_patch/model/llava/language_model.py (692:700) duplicated block id: 7726 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/transformer.py (374:380) - megatron_patch/model/qwen1_5_megablocks/transformer.py (408:414) duplicated block id: 7727 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (199:207) - toolkits/model_checkpoints_convertor/glm/checkpoint_reshaping_and_interoperability.py (141:149) duplicated block id: 7728 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/language_model.py (475:483) - megatron_patch/model/chatglm/language_model.py (412:420) duplicated block id: 7729 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/galactica/transformer.py (232:246) - megatron_patch/model/qwen_vl/transformer.py (424:438) duplicated block id: 7730 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/mixtral_bak/transformer/mlp.py (162:168) - megatron_patch/model/qwen1_5/transformer/mlp.py (205:211) duplicated block id: 7731 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/language_model.py (124:130) - megatron_patch/model/glm130b/language_model.py (90:96) duplicated block id: 7732 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/glm130b/checkpoint_reshaping_and_interoperability.py (377:384) - toolkits/model_checkpoints_convertor/starcoder/checkpoint_reshaping_and_interoperability.py (665:673) duplicated block id: 7733 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/llava_mcore/vision/vit_layer_specs.py (89:95) - megatron_patch/model/qwen3_moe/gpt_layer_specs.py (300:306) duplicated block id: 7734 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_dense_and_moe_gqa.py (651:657) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_dense_and_moe_gqa.py (706:712) duplicated block id: 7735 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_moe.py (160:166) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_moe.py (262:268) duplicated block id: 7736 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/transformer/attention.py (458:464) - megatron_patch/model/qwen1_5/transformer/attention.py (137:143) duplicated block id: 7737 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (1213:1221) - megatron_patch/model/llama2/transformer.py (1212:1220) duplicated block id: 7738 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (448:456) - toolkits/model_checkpoints_convertor/galactica/checkpoint_reshaping_and_interoperability.py (477:485) duplicated block id: 7739 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (1549:1555) - megatron_patch/model/llama2/transformer.py (1703:1709) duplicated block id: 7740 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer/attention.py (253:259) - megatron_patch/model/mixtral/transformer/attention.py (344:350) duplicated block id: 7741 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/falcon/transformer.py (547:553) - megatron_patch/model/qwen1_5_megablocks/transformer.py (408:414) duplicated block id: 7742 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/transformer.py (437:444) - megatron_patch/model/galactica/transformer.py (402:409) duplicated block id: 7743 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_dense_and_moe_gqa.py (424:431) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_vl.py (327:333) duplicated block id: 7744 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/llama/language_model.py (89:95) - megatron_patch/model/mistral/language_model.py (80:86) duplicated block id: 7745 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/gpt_model.py (28:39) - megatron_patch/model/falcon/gpt_model.py (28:38) duplicated block id: 7746 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/transformer.py (412:418) - megatron_patch/model/llama3/transformer_legacy.py (516:522) duplicated block id: 7747 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/llama/transformer.py (270:284) - megatron_patch/model/qwen1_5_megablocks/transformer.py (314:328) duplicated block id: 7748 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/llama/transformer.py (723:729) - megatron_patch/model/llama/transformer.py (1046:1052) duplicated block id: 7749 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/mixtral_bak/transformer/attention.py (298:304) - megatron_patch/model/qwen2/transformer/attention.py (312:318) duplicated block id: 7750 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/bloom/gpt_model.py (28:39) - megatron_patch/model/falcon/gpt_model.py (28:38) duplicated block id: 7751 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/llama2/transformer.py (1703:1709) - megatron_patch/model/llama3/transformer_legacy.py (1639:1645) duplicated block id: 7752 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/checkpoint_reshaping_and_interoperability.py (170:178) - toolkits/model_checkpoints_convertor/chatglm/checkpoint_reshaping_and_interoperability.py (169:177) duplicated block id: 7753 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/transformer.py (228:242) - megatron_patch/model/qwen/transformer.py (417:431) duplicated block id: 7754 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon40b/checkpoint_reshaping_and_interoperability.py (522:528) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (621:627) duplicated block id: 7755 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (334:342) - toolkits/model_checkpoints_convertor/starcoder/checkpoint_reshaping_and_interoperability.py (192:200) duplicated block id: 7756 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (1071:1077) - megatron_patch/model/qwen/transformer.py (1682:1688) duplicated block id: 7757 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (957:963) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (914:920) duplicated block id: 7758 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (172:180) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (214:222) duplicated block id: 7759 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/mistral/transformer.py (917:925) - megatron_patch/model/starcoder/transformer.py (848:855) duplicated block id: 7760 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/moe/experts.py (670:678) - megatron_patch/model/qwen1_5/transformer/mlp.py (139:147) duplicated block id: 7761 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (929:937) - megatron_patch/model/mistral/transformer.py (1059:1067) duplicated block id: 7762 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/llava/transformer.py (1678:1684) - megatron_patch/model/qwen_vl/transformer.py (1702:1708) duplicated block id: 7763 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (1203:1209) - megatron_patch/model/qwen/transformer.py (1682:1688) duplicated block id: 7764 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore.py (491:497) - toolkits/model_checkpoints_convertor/mistral/hf2mcore.py (197:203) duplicated block id: 7765 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/glm130b/checkpoint_reshaping_and_interoperability.py (141:149) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (254:262) duplicated block id: 7766 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (524:533) - megatron_patch/model/llama/transformer.py (522:531) duplicated block id: 7767 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/reward_model_to_megatron.py (170:178) - toolkits/model_checkpoints_convertor/chatglm/checkpoint_reshaping_and_interoperability.py (169:177) duplicated block id: 7768 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer/attention.py (253:259) - megatron_patch/model/qwen2_vl/attention_vision.py (333:339) duplicated block id: 7769 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_mha.py (314:321) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_mha_to_moe.py (248:255) duplicated block id: 7770 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_mha.py (98:104) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2.5_vl.py (113:119) duplicated block id: 7771 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/moe/experts.py (17:29) - megatron_patch/model/qwen2/moe/experts.py (28:37) duplicated block id: 7772 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/transformer.py (228:242) - megatron_patch/model/qwen_vl/transformer.py (424:438) duplicated block id: 7773 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (287:301) - megatron_patch/model/llava/transformer.py (424:438) duplicated block id: 7774 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (191:198) - toolkits/model_checkpoints_convertor/yi/checkpoint_reshaping_and_interoperability.py (154:161) duplicated block id: 7775 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (197:205) - toolkits/model_checkpoints_convertor/glm/checkpoint_reshaping_and_interoperability.py (141:149) duplicated block id: 7776 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/llama3_1/model.py (104:110) - megatron_patch/model/mixtral/model.py (83:89) duplicated block id: 7777 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore_llama3_1.py (279:285) - toolkits/model_checkpoints_convertor/llama/hf2mcore_llama3_1.py (491:497) duplicated block id: 7778 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/language_model.py (90:96) - megatron_patch/model/qwen1_5_megablocks/language_model.py (78:84) duplicated block id: 7779 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/layer_specs.py (121:128) - megatron_patch/model/qwen2/layer_specs.py (119:126) duplicated block id: 7780 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (574:580) - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (478:485) duplicated block id: 7781 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (542:551) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (653:662) duplicated block id: 7782 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (1071:1077) - megatron_patch/model/baichuan2/transformer.py (1740:1746) duplicated block id: 7783 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/galactica/language_model.py (89:95) - megatron_patch/model/qwen1_5_megablocks/language_model.py (78:84) duplicated block id: 7784 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (196:204) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (214:222) duplicated block id: 7785 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/llama/transformer.py (529:540) - megatron_patch/model/llava/transformer.py (731:742) duplicated block id: 7786 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/mistral/transformer.py (964:972) - megatron_patch/model/qwen1_5_megablocks/transformer.py (857:865) duplicated block id: 7787 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/mixtral_bak/transformer/attention.py (111:117) - megatron_patch/model/qwen1_5/transformer/attention.py (325:331) duplicated block id: 7788 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (409:416) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (551:557) duplicated block id: 7789 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/transformer.py (275:284) - megatron_patch/model/mistral/transformer.py (399:408) duplicated block id: 7790 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_moe.py (132:138) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_dense_and_moe_gqa.py (590:596) duplicated block id: 7791 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/qwen2/transformer/attention.py (22:28) - megatron_patch/model/qwen2_vl/attention_vision.py (15:21) duplicated block id: 7792 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/transformer.py (1050:1058) - megatron_patch/model/starcoder/transformer.py (1226:1234) duplicated block id: 7793 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/transformer/attention.py (496:502) - megatron_patch/model/qwen1_5/transformer/attention.py (362:368) duplicated block id: 7794 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (85:91) - toolkits/model_checkpoints_convertor/mistral/hf2mcore.py (93:99) duplicated block id: 7795 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (262:273) - megatron_patch/model/qwen_vl/transformer.py (399:408) duplicated block id: 7796 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer/attention.py (137:143) - megatron_patch/model/llama3/transformer/attention.py (316:322) duplicated block id: 7797 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/llama3/model.py (98:104) - megatron_patch/model/llama3_1/model.py (113:119) duplicated block id: 7798 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/bloom/language_model.py (376:382) - megatron_patch/model/glm130b/language_model.py (380:386) duplicated block id: 7799 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/mixtral_bak/transformer/attention.py (111:117) - megatron_patch/model/qwen1_5/transformer/attention.py (316:322) duplicated block id: 7800 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer/attention.py (137:143) - megatron_patch/model/llama3/transformer/attention.py (325:331) duplicated block id: 7801 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/deepseek/hf2mcore_deepseek_v3_moe.py (127:135) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2.5_vl.py (72:80) duplicated block id: 7802 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (1573:1579) - megatron_patch/model/mistral/transformer.py (1676:1682) duplicated block id: 7803 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (1679:1685) - megatron_patch/model/mistral/transformer.py (1700:1706) duplicated block id: 7804 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/falcon/transformer.py (429:439) - megatron_patch/model/galactica/transformer.py (208:218) duplicated block id: 7805 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/mixtral_bak/moe/moe_utils.py (61:68) - megatron_patch/model/qwen_vl/transformer.py (171:178) duplicated block id: 7806 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/llava_mcore/vision/vit_layer_specs.py (46:52) - megatron_patch/model/mixtral_bak/layer_specs.py (63:69) duplicated block id: 7807 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_mha_to_moe.py (65:72) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_dense_and_moe_gqa.py (54:61) duplicated block id: 7808 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (530:537) - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (604:611) duplicated block id: 7809 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/deepseek/hf2mcore_deepseek_v2_moe.py (192:198) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_dense_and_moe_gqa.py (306:312) duplicated block id: 7810 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer_legacy.py (1152:1158) - megatron_patch/model/qwen/transformer.py (1682:1688) duplicated block id: 7811 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/hf2te.py (33:39) - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (63:70) duplicated block id: 7812 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/transformer.py (734:740) - megatron_patch/model/falcon40b/transformer.py (991:997) duplicated block id: 7813 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/deepseek/hf2mcore_deepseek_v3_moe.py (103:111) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_mha.py (50:58) duplicated block id: 7814 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (1573:1579) - megatron_patch/model/llava/transformer.py (1678:1684) duplicated block id: 7815 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/llama/language_model.py (89:95) - megatron_patch/model/qwen/language_model.py (79:85) duplicated block id: 7816 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer/attention.py (316:322) - megatron_patch/model/llama3/transformer/attention.py (325:331) duplicated block id: 7817 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer/attention.py (314:320) - megatron_patch/model/qwen2_vl/attention.py (446:452) duplicated block id: 7818 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (929:937) - megatron_patch/model/baichuan2/transformer.py (1064:1072) duplicated block id: 7819 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon40b/checkpoint_reshaping_and_interoperability.py (471:477) - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (478:485) duplicated block id: 7820 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon40b/checkpoint_reshaping_and_interoperability.py (522:528) - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (614:620) duplicated block id: 7821 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/layer_specs.py (105:112) - megatron_patch/model/qwen3_moe/gpt_layer_specs.py (259:267) duplicated block id: 7822 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/llama/transformer.py (270:284) - megatron_patch/model/llama3/transformer_legacy.py (422:436) duplicated block id: 7823 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/mixtral_bak/layer_specs.py (99:105) - megatron_patch/model/qwen2_moe/layer_specs.py (219:225) duplicated block id: 7824 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/mistral/hf2mcore.py (477:483) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_dense_and_moe_gqa.py (740:746) duplicated block id: 7825 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/transformer.py (319:325) - megatron_patch/model/qwen1_5_megablocks/transformer.py (408:414) duplicated block id: 7826 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/transformer/attention.py (695:701) - megatron_patch/model/qwen1_5/transformer/attention.py (559:565) duplicated block id: 7827 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (301:310) - megatron_patch/model/llama/transformer.py (246:256) duplicated block id: 7828 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (947:954) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (853:861) duplicated block id: 7829 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/language_model.py (651:659) - megatron_patch/model/qwen1_5_megablocks/language_model.py (631:639) duplicated block id: 7830 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/llama/transformer.py (529:540) - megatron_patch/model/starcoder/transformer.py (731:742) duplicated block id: 7831 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer/attention.py (325:331) - megatron_patch/model/qwen2/transformer/attention.py (312:318) duplicated block id: 7832 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/mixtral_bak/moe/token_dispatcher.py (191:197) - megatron_patch/model/qwen2/moe/token_dispatcher.py (169:175) duplicated block id: 7833 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (1077:1085) - megatron_patch/model/glm130b/transformer.py (1050:1058) duplicated block id: 7834 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/glm130b/checkpoint_reshaping_and_interoperability.py (141:149) - toolkits/model_checkpoints_convertor/yi/checkpoint_reshaping_and_interoperability.py (184:192) duplicated block id: 7835 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (326:340) - megatron_patch/model/llama/transformer.py (270:284) duplicated block id: 7836 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/language_model.py (709:717) - megatron_patch/model/chatglm/language_model.py (651:659) duplicated block id: 7837 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_moe.py (366:372) - toolkits/model_checkpoints_convertor/qwen/hf2megablocks_qwen1.5.py (580:586) duplicated block id: 7838 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (301:310) - megatron_patch/model/falcon40b/transformer.py (258:268) duplicated block id: 7839 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/falcon/language_model.py (89:95) - megatron_patch/model/qwen1_5_megablocks/language_model.py (78:84) duplicated block id: 7840 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/llama3_1/layer_specs.py (104:110) - megatron_patch/model/qwen2_vl/layer_specs.py (106:112) duplicated block id: 7841 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore.py (570:576) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_dense_and_moe_gqa.py (740:746) duplicated block id: 7842 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/language_model.py (412:420) - megatron_patch/model/falcon/language_model.py (398:406) duplicated block id: 7843 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/transformer.py (228:242) - megatron_patch/model/llava/transformer.py (424:438) duplicated block id: 7844 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (162:169) - toolkits/model_checkpoints_convertor/yi/checkpoint_reshaping_and_interoperability.py (154:161) duplicated block id: 7845 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon/checkpoint_reshaping_and_interoperability.py (86:92) - toolkits/model_checkpoints_convertor/mistral/hf2mcore.py (93:99) duplicated block id: 7846 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer/attention.py (547:553) - megatron_patch/model/mixtral/transformer/attention.py (707:713) duplicated block id: 7847 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_mha_to_moe.py (49:57) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_dense_and_moe_gqa.py (38:46) duplicated block id: 7848 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/reward_model_to_megatron.py (673:680) - toolkits/model_checkpoints_convertor/glm/checkpoint_reshaping_and_interoperability.py (398:405) duplicated block id: 7849 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore.py (373:379) - toolkits/model_checkpoints_convertor/qwen/hf2megablocks_qwen1.5.py (373:379) duplicated block id: 7850 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/reward_model_to_megatron.py (170:178) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (231:239) duplicated block id: 7851 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer/attention.py (559:565) - megatron_patch/model/qwen2_vl/attention_vision.py (695:701) duplicated block id: 7852 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/galactica/transformer.py (232:246) - megatron_patch/model/llava/transformer.py (424:438) duplicated block id: 7853 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer/attention.py (559:565) - megatron_patch/model/qwen1_5/transformer/attention.py (547:553) duplicated block id: 7854 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/llama/transformer.py (529:540) - megatron_patch/model/llama3/transformer_legacy.py (710:721) duplicated block id: 7855 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/transformer_layer.py (259:267) - megatron_patch/model/qwen2/transformer_layer.py (163:171) duplicated block id: 7856 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (449:456) - megatron_patch/model/falcon40b/transformer.py (437:444) duplicated block id: 7857 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (853:861) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (1023:1031) duplicated block id: 7858 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer/attention.py (210:217) - megatron_patch/model/mixtral/transformer/attention.py (242:249) duplicated block id: 7859 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/language_model.py (651:659) - megatron_patch/model/qwen_vl/language_model.py (661:669) duplicated block id: 7860 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer/attention.py (316:322) - megatron_patch/model/qwen1_5/transformer/attention.py (137:143) duplicated block id: 7861 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (467:473) - toolkits/model_checkpoints_convertor/yi/checkpoint_reshaping_and_interoperability.py (552:558) duplicated block id: 7862 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (262:273) - megatron_patch/model/starcoder/transformer.py (425:435) duplicated block id: 7863 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_mha_to_moe.py (49:57) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_vl.py (54:62) duplicated block id: 7864 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (1679:1685) - megatron_patch/model/llama2/transformer.py (1703:1709) duplicated block id: 7865 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/deepseek/hf2mcore_deepseek_v3_moe.py (663:669) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_dense_and_moe_gqa.py (768:774) duplicated block id: 7866 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_moe.py (116:122) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_dense_and_moe_gqa.py (536:543) duplicated block id: 7867 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/bloom/language_model.py (91:97) - megatron_patch/model/qwen/language_model.py (79:85) duplicated block id: 7868 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/glm/checkpoint_reshaping_and_interoperability.py (141:149) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (251:259) duplicated block id: 7869 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (559:565) - toolkits/model_checkpoints_convertor/baichuan/hf2te.py (424:430) duplicated block id: 7870 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/transformer.py (228:242) - megatron_patch/model/llama3/transformer_legacy.py (422:436) duplicated block id: 7871 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/deepseek/hf2mcore_deepseek_v3_moe.py (663:669) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_dense_and_moe_gqa.py (691:697) duplicated block id: 7872 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/mistral/language_model.py (644:652) - megatron_patch/model/starcoder/language_model.py (555:563) duplicated block id: 7873 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/transformer/attention.py (165:171) - megatron_patch/model/mixtral_bak/transformer/attention.py (298:304) duplicated block id: 7874 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/transformer/attention.py (165:171) - megatron_patch/model/mixtral_bak/transformer/attention.py (289:295) duplicated block id: 7875 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (529:535) - megatron_patch/model/llama/transformer.py (362:368) duplicated block id: 7876 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/qwen/transformer.py (1621:1627) - megatron_patch/model/qwen/transformer.py (1645:1651) duplicated block id: 7877 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/galactica/language_model.py (89:95) - megatron_patch/model/llama3/language_model.py (68:74) duplicated block id: 7878 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon40b/checkpoint_reshaping_and_interoperability.py (824:832) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (853:861) duplicated block id: 7879 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/glm/checkpoint_reshaping_and_interoperability.py (388:395) - toolkits/model_checkpoints_convertor/starcoder/checkpoint_reshaping_and_interoperability.py (665:673) duplicated block id: 7880 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/deepseek/hf2mcore_deepseek_v2_moe.py (50:58) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_mha.py (50:58) duplicated block id: 7881 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/transformer_block.py (529:536) - megatron_patch/model/qwen2/transformer_block.py (427:434) duplicated block id: 7882 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/mixtral_bak/moe/experts.py (211:218) - megatron_patch/model/qwen2/moe/experts.py (370:377) duplicated block id: 7883 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/qwen2/transformer/attention.py (533:539) - megatron_patch/model/qwen2_vl/attention.py (678:684) duplicated block id: 7884 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/qwen1_5_megablocks/transformer.py (408:414) - megatron_patch/model/starcoder/transformer.py (524:530) duplicated block id: 7885 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/qwen1_5_megablocks/transformer.py (857:865) - megatron_patch/model/qwen_vl/transformer.py (966:974) duplicated block id: 7886 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/galactica/gpt_model.py (28:38) - megatron_patch/model/glm130b/gpt_model.py (28:39) duplicated block id: 7887 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/llava/transformer.py (1678:1684) - megatron_patch/model/mistral/transformer.py (1700:1706) duplicated block id: 7888 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/qwen1_5/layer_specs.py (76:82) - megatron_patch/model/qwen2_moe/layer_specs.py (219:225) duplicated block id: 7889 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (85:91) - toolkits/model_checkpoints_convertor/qwen/hf2megablocks_qwen1.5.py (119:125) duplicated block id: 7890 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (1679:1685) - megatron_patch/model/qwen/transformer.py (1645:1651) duplicated block id: 7891 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (921:928) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (853:861) duplicated block id: 7892 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/mixtral_bak/moe/token_dispatcher.py (191:197) - megatron_patch/model/qwen1_5/moe/token_dispatcher.py (166:172) duplicated block id: 7893 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_mha.py (50:58) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2.5_vl.py (54:62) duplicated block id: 7894 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/reward_model_to_megatron.py (713:724) - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (588:599) duplicated block id: 7895 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (529:535) - megatron_patch/model/bloom/transformer.py (378:384) duplicated block id: 7896 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/moe/token_dispatcher.py (254:260) - megatron_patch/model/qwen2/moe/token_dispatcher.py (313:319) duplicated block id: 7897 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/falcon/transformer.py (429:439) - megatron_patch/model/glm130b/transformer.py (275:284) duplicated block id: 7898 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/transformer/attention.py (458:464) - megatron_patch/model/mixtral_bak/transformer/attention.py (111:117) duplicated block id: 7899 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/qwen2/transformer_block.py (427:434) - megatron_patch/model/qwen2_5_vl/transformer_block.py (598:605) duplicated block id: 7900 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (570:579) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (653:662) duplicated block id: 7901 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (1703:1709) - megatron_patch/model/llava/transformer.py (1678:1684) duplicated block id: 7902 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/transformer/attention.py (677:683) - megatron_patch/model/qwen1_5/transformer/attention.py (528:534) duplicated block id: 7903 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (493:499) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (396:403) duplicated block id: 7904 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (326:340) - megatron_patch/model/falcon40b/transformer.py (282:296) duplicated block id: 7905 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/llava/transformer.py (1702:1708) - megatron_patch/model/qwen1_5_megablocks/transformer.py (1541:1547) duplicated block id: 7906 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer_legacy.py (516:522) - megatron_patch/model/starcoder/transformer.py (524:530) duplicated block id: 7907 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (262:273) - megatron_patch/model/falcon/transformer.py (429:439) duplicated block id: 7908 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/falcon/gpt_model.py (28:38) - megatron_patch/model/glm130b/gpt_model.py (28:39) duplicated block id: 7909 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/language_model.py (615:623) - megatron_patch/model/llama2/language_model.py (632:640) duplicated block id: 7910 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/mistral/hf2mcore.py (207:213) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_moe.py (126:132) duplicated block id: 7911 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/transformer.py (380:388) - megatron_patch/model/glm130b/transformer.py (483:492) duplicated block id: 7912 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/falcon/language_model.py (181:193) - megatron_patch/model/starcoder/language_model.py (167:179) duplicated block id: 7913 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/mistral/hf2mcore.py (93:99) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (114:120) duplicated block id: 7914 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/mixtral_bak/transformer/attention.py (335:341) - megatron_patch/model/qwen2_vl/attention.py (497:503) duplicated block id: 7915 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/falcon/transformer.py (265:275) - megatron_patch/model/llama3/transformer_legacy.py (397:406) duplicated block id: 7916 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/transformer.py (578:589) - megatron_patch/model/qwen1_5_megablocks/transformer.py (623:634) duplicated block id: 7917 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/transformer.py (412:418) - megatron_patch/model/llama2/transformer.py (517:523) duplicated block id: 7918 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/mixtral_bak/transformer/attention.py (111:117) - megatron_patch/model/qwen2/transformer/attention.py (321:327) duplicated block id: 7919 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/qwen1_5/transformer/attention.py (547:553) - megatron_patch/model/qwen2_vl/attention.py (708:714) duplicated block id: 7920 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/transformer.py (382:389) - megatron_patch/model/falcon40b/transformer.py (437:444) duplicated block id: 7921 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/transformer.py (282:296) - megatron_patch/model/starcoder/transformer.py (328:342) duplicated block id: 7922 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/starcoder/checkpoint_reshaping_and_interoperability.py (714:725) - toolkits/model_checkpoints_convertor/yi/checkpoint_reshaping_and_interoperability.py (514:525) duplicated block id: 7923 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/galactica/checkpoint_reshaping_and_interoperability.py (477:485) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (506:514) duplicated block id: 7924 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/mistral/transformer.py (518:524) - megatron_patch/model/starcoder/transformer.py (524:530) duplicated block id: 7925 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer/attention.py (528:534) - megatron_patch/model/qwen2_vl/attention.py (678:684) duplicated block id: 7926 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/moe/experts.py (670:678) - megatron_patch/model/qwen2/transformer/mlp.py (142:150) duplicated block id: 7927 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (68:75) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_mha_to_moe.py (65:72) duplicated block id: 7928 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/mixtral_bak/transformer/attention.py (298:304) - megatron_patch/model/qwen1_5/transformer/attention.py (316:322) duplicated block id: 7929 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (1703:1709) - megatron_patch/model/qwen1_5_megablocks/transformer.py (1541:1547) duplicated block id: 7930 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/transformer.py (275:284) - megatron_patch/model/starcoder/transformer.py (425:435) duplicated block id: 7931 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/galactica/transformer.py (324:330) - megatron_patch/model/qwen_vl/transformer.py (518:524) duplicated block id: 7932 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/falcon/gpt_model.py (51:60) - megatron_patch/model/starcoder/gpt_model.py (52:62) duplicated block id: 7933 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (476:484) - toolkits/model_checkpoints_convertor/galactica/checkpoint_reshaping_and_interoperability.py (477:485) duplicated block id: 7934 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/model.py (83:89) - megatron_patch/model/qwen1_5/model.py (89:95) duplicated block id: 7935 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/gpt_model.py (28:38) - megatron_patch/model/llama/gpt_model.py (28:39) duplicated block id: 7936 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/reward_model_to_megatron.py (170:178) - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (235:243) duplicated block id: 7937 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer/attention.py (137:143) - megatron_patch/model/qwen2/transformer/attention.py (312:318) duplicated block id: 7938 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/falcon/transformer.py (547:553) - megatron_patch/model/llava/transformer.py (518:524) duplicated block id: 7939 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/moe/router.py (154:160) - megatron_patch/model/qwen3_moe/moe/router.py (94:100) duplicated block id: 7940 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/glm/checkpoint_reshaping_and_interoperability.py (141:149) - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (235:243) duplicated block id: 7941 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/glm130b/checkpoint_reshaping_and_interoperability.py (141:149) - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (272:280) duplicated block id: 7942 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/llama3_1/layer_specs.py (104:110) - megatron_patch/model/mixtral/layer_specs.py (151:157) duplicated block id: 7943 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/galactica/language_model.py (89:95) - megatron_patch/model/mistral/language_model.py (80:86) duplicated block id: 7944 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/mixtral_bak/transformer/attention.py (111:117) - megatron_patch/model/qwen2/transformer/attention.py (312:318) duplicated block id: 7945 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer/attention.py (137:143) - megatron_patch/model/qwen2/transformer/attention.py (321:327) duplicated block id: 7946 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/language_model.py (350:356) - megatron_patch/model/llama2/language_model.py (334:340) duplicated block id: 7947 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (378:384) - megatron_patch/model/qwen_vl/transformer.py (518:524) duplicated block id: 7948 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/deepseek/hf2mcore_deepseek_v3_moe.py (23:39) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_dense_and_moe_gqa.py (9:27) duplicated block id: 7949 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/falcon/language_model.py (366:372) - megatron_patch/model/glm130b/language_model.py (380:386) duplicated block id: 7950 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/deepseek/hf2mcore_deepseek_v2_moe.py (384:390) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_dense_and_moe_gqa.py (197:203) duplicated block id: 7951 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/qwen1_5/transformer/mlp.py (73:79) - megatron_patch/model/qwen2/transformer/mlp.py (98:104) duplicated block id: 7952 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/transformer.py (412:418) - megatron_patch/model/qwen1_5_megablocks/transformer.py (408:414) duplicated block id: 7953 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/qwen2/transformer/attention.py (206:213) - megatron_patch/model/qwen2_vl/attention.py (232:239) duplicated block id: 7954 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/model.py (103:111) - megatron_patch/model/qwen2_vl/gpt_model.py (98:106) duplicated block id: 7955 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/llama/transformer.py (529:540) - megatron_patch/model/qwen1_5_megablocks/transformer.py (623:634) duplicated block id: 7956 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (809:820) - megatron_patch/model/mistral/transformer.py (843:854) duplicated block id: 7957 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer_legacy.py (175:182) - megatron_patch/model/mixtral_bak/moe/moe_utils.py (61:68) duplicated block id: 7958 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/mixtral_bak/transformer/attention.py (116:123) - megatron_patch/model/qwen2_vl/attention.py (159:166) duplicated block id: 7959 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/hf2te.py (424:430) - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (614:620) duplicated block id: 7960 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (1213:1221) - megatron_patch/model/llava/transformer.py (1211:1219) duplicated block id: 7961 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer/attention.py (253:259) - megatron_patch/model/qwen2_vl/attention.py (334:340) duplicated block id: 7962 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (1573:1579) - megatron_patch/model/llama3/transformer_legacy.py (1639:1645) duplicated block id: 7963 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (420:426) - megatron_patch/model/falcon40b/transformer.py (374:380) duplicated block id: 7964 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (68:75) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_mha.py (60:67) duplicated block id: 7965 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer/attention.py (362:368) - megatron_patch/model/qwen2_vl/attention_vision.py (496:502) duplicated block id: 7966 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/hf2te.py (424:430) - toolkits/model_checkpoints_convertor/galactica/checkpoint_reshaping_and_interoperability.py (578:584) duplicated block id: 7967 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/qwen2/transformer/attention.py (358:364) - megatron_patch/model/qwen2_vl/attention.py (497:503) duplicated block id: 7968 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/qwen1_5/transformer/attention.py (547:553) - megatron_patch/model/qwen2/transformer/attention.py (564:570) duplicated block id: 7969 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/language_model.py (89:95) - megatron_patch/model/llava/language_model.py (82:88) duplicated block id: 7970 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/llama2/transformer.py (517:523) - megatron_patch/model/starcoder/transformer.py (524:530) duplicated block id: 7971 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (857:865) - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (976:984) duplicated block id: 7972 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/multi_latent_attention.py (111:117) - megatron_patch/model/mixtral_bak/transformer/attention.py (227:233) duplicated block id: 7973 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/mistral/hf2mcore.py (432:438) - toolkits/model_checkpoints_convertor/qwen/hf2megablocks_qwen1.5.py (535:541) duplicated block id: 7974 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/checkpoint_reshaping_and_interoperability.py (585:594) - toolkits/model_checkpoints_convertor/bloom/reward_model_to_megatron.py (584:593) duplicated block id: 7975 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/transformer.py (282:296) - megatron_patch/model/qwen_vl/transformer.py (424:438) duplicated block id: 7976 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/reward_model_to_megatron.py (170:178) - toolkits/model_checkpoints_convertor/yi/checkpoint_reshaping_and_interoperability.py (184:192) duplicated block id: 7977 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/bloom/language_model.py (91:97) - megatron_patch/model/qwen_vl/language_model.py (80:86) duplicated block id: 7978 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (420:426) - megatron_patch/model/falcon/transformer.py (547:553) duplicated block id: 7979 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/transformer/attention.py (165:171) - megatron_patch/model/mixtral/transformer/attention.py (458:464) duplicated block id: 7980 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon40b/checkpoint_reshaping_and_interoperability.py (86:92) - toolkits/model_checkpoints_convertor/mistral/hf2mcore.py (93:99) duplicated block id: 7981 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon/checkpoint_reshaping_and_interoperability.py (659:665) - toolkits/model_checkpoints_convertor/falcon40b/checkpoint_reshaping_and_interoperability.py (662:668) duplicated block id: 7982 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_moe.py (52:60) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_dense_and_moe_gqa.py (38:46) duplicated block id: 7983 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/language_model.py (90:96) - megatron_patch/model/mistral/language_model.py (80:86) duplicated block id: 7984 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (287:301) - megatron_patch/model/mistral/transformer.py (424:438) duplicated block id: 7985 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/transformer/attention.py (165:171) - megatron_patch/model/mixtral/transformer/attention.py (449:455) duplicated block id: 7986 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/galactica/transformer.py (324:330) - megatron_patch/model/llava/transformer.py (518:524) duplicated block id: 7987 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (435:449) - megatron_patch/model/chatglm/transformer.py (228:242) duplicated block id: 7988 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (805:811) - megatron_patch/model/llama/transformer.py (751:757) duplicated block id: 7989 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/language_model.py (90:96) - megatron_patch/model/mistral/language_model.py (80:86) duplicated block id: 7990 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore_llama3_1.py (589:595) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_dense_and_moe_gqa.py (706:712) duplicated block id: 7991 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/transformer/attention.py (695:701) - megatron_patch/model/qwen2_vl/attention_vision.py (707:713) duplicated block id: 7992 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/qwen1_5/moe/router.py (190:206) - megatron_patch/model/qwen2/moe/router.py (273:289) duplicated block id: 7993 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/gpt_model.py (26:37) - megatron_patch/model/bloom/gpt_model.py (28:39) duplicated block id: 7994 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (396:403) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (551:557) duplicated block id: 7995 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/qwen2/transformer/attention.py (249:255) - megatron_patch/model/qwen2_vl/attention_vision.py (333:339) duplicated block id: 7996 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/transformer/attention.py (458:464) - megatron_patch/model/mixtral_bak/transformer/attention.py (289:295) duplicated block id: 7997 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/chatglm/checkpoint_reshaping_and_interoperability.py (451:457) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (396:403) duplicated block id: 7998 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (507:518) - toolkits/model_checkpoints_convertor/bloom/checkpoint_reshaping_and_interoperability.py (714:725) duplicated block id: 7999 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer/mlp.py (189:195) - megatron_patch/model/mixtral_bak/transformer/mlp.py (162:168) duplicated block id: 8000 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_mha.py (50:58) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_dense_and_moe_gqa.py (38:46) duplicated block id: 8001 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (68:75) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_moe.py (62:69) duplicated block id: 8002 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/transformer.py (161:171) - megatron_patch/model/glm130b/transformer.py (242:253) duplicated block id: 8003 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/mixtral_bak/transformer/attention.py (184:191) - megatron_patch/model/qwen2_vl/attention.py (232:239) duplicated block id: 8004 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (930:938) - megatron_patch/model/qwen1_5_megablocks/transformer.py (857:865) duplicated block id: 8005 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/llava_mcore/llava_model.py (206:218) - megatron_patch/model/qwen2_vl/model.py (125:136) duplicated block id: 8006 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_mha_to_moe.py (189:195) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_moe.py (132:138) duplicated block id: 8007 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/language_model.py (350:356) - megatron_patch/model/llava/language_model.py (337:343) duplicated block id: 8008 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_mha.py (330:337) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_mha_to_moe.py (256:263) duplicated block id: 8009 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon/checkpoint_reshaping_and_interoperability.py (86:92) - toolkits/model_checkpoints_convertor/qwen/hf2megablocks_qwen1.5.py (119:125) duplicated block id: 8010 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/mistral/transformer.py (1676:1682) - megatron_patch/model/qwen1_5_megablocks/transformer.py (1565:1571) duplicated block id: 8011 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/transformer.py (802:808) - megatron_patch/model/glm130b/transformer.py (1185:1191) duplicated block id: 8012 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/hf2te.py (258:270) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (490:502) duplicated block id: 8013 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/llama3/model.py (89:95) - megatron_patch/model/mixtral/model.py (83:89) duplicated block id: 8014 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer/attention.py (528:534) - megatron_patch/model/mixtral/transformer/attention.py (677:683) duplicated block id: 8015 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/mistral/transformer.py (1700:1706) - megatron_patch/model/qwen_vl/transformer.py (1678:1684) duplicated block id: 8016 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_dense_and_moe_gqa.py (56:64) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_vl.py (78:86) duplicated block id: 8017 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (164:172) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (201:209) duplicated block id: 8018 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/transformer.py (412:418) - megatron_patch/model/qwen_vl/transformer.py (518:524) duplicated block id: 8019 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/deepseek/hf2mcore_deepseek_v2_moe.py (384:390) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_moe.py (116:122) duplicated block id: 8020 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon/configuration_RW.py (71:79) - toolkits/model_checkpoints_convertor/falcon40b/configuration_RW.py (67:75) duplicated block id: 8021 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/transformer/attention.py (677:683) - megatron_patch/model/mixtral_bak/transformer/attention.py (412:418) duplicated block id: 8022 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/reward_model_to_megatron.py (170:178) - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (334:342) duplicated block id: 8023 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/checkpoint_reshaping_and_interoperability.py (674:681) - toolkits/model_checkpoints_convertor/glm/checkpoint_reshaping_and_interoperability.py (398:405) duplicated block id: 8024 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/bloom/language_model.py (585:593) - megatron_patch/model/qwen1_5_megablocks/language_model.py (631:639) duplicated block id: 8025 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/layer_specs.py (121:127) - megatron_patch/model/qwen3_moe/gpt_layer_specs.py (240:246) duplicated block id: 8026 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/falcon/transformer.py (265:275) - megatron_patch/model/llama2/transformer.py (398:407) duplicated block id: 8027 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/multi_latent_attention.py (243:249) - megatron_patch/model/deepseek_v2/multi_latent_attention.py (267:273) duplicated block id: 8028 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (1595:1611) - megatron_patch/model/starcoder/transformer.py (1280:1296) duplicated block id: 8029 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/transformer_block.py (496:502) - megatron_patch/model/qwen2/transformer_block.py (401:407) duplicated block id: 8030 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/multi_latent_attention.py (243:249) - megatron_patch/model/deepseek_v2/multi_latent_attention.py (255:261) duplicated block id: 8031 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/gpt_model.py (28:38) - megatron_patch/model/glm130b/gpt_model.py (28:39) duplicated block id: 8032 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/falcon/gpt_model.py (28:38) - megatron_patch/model/llama/gpt_model.py (28:39) duplicated block id: 8033 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/qwen1_5/transformer/attention.py (314:320) - megatron_patch/model/qwen2_vl/attention_vision.py (445:451) duplicated block id: 8034 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (134:143) - megatron_patch/model/falcon40b/transformer.py (162:171) duplicated block id: 8035 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (1740:1746) - megatron_patch/model/llama3/transformer_legacy.py (1152:1158) duplicated block id: 8036 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (53:60) - toolkits/model_checkpoints_convertor/qwen/hf2megablocks_qwen1.5.py (64:71) duplicated block id: 8037 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/galactica/checkpoint_reshaping_and_interoperability.py (85:91) - toolkits/model_checkpoints_convertor/mistral/hf2mcore.py (93:99) duplicated block id: 8038 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/deepseek/hf2mcore_deepseek_v3_moe.py (42:50) - toolkits/model_checkpoints_convertor/llama/hf2mcore_llama3_1.py (30:37) duplicated block id: 8039 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (272:280) - toolkits/model_checkpoints_convertor/starcoder/checkpoint_reshaping_and_interoperability.py (192:200) duplicated block id: 8040 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/transformer_block.py (393:399) - megatron_patch/model/qwen2/transformer_block.py (322:328) duplicated block id: 8041 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/multi_latent_attention.py (162:168) - megatron_patch/model/llama3/transformer/attention.py (322:328) duplicated block id: 8042 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/qwen1_5_megablocks/transformer.py (1541:1547) - megatron_patch/model/qwen_vl/transformer.py (1702:1708) duplicated block id: 8043 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/qwen2/transformer/attention.py (552:558) - megatron_patch/model/qwen2/transformer/attention.py (564:570) duplicated block id: 8044 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/mixtral_bak/transformer/attention.py (298:304) - megatron_patch/model/qwen1_5/transformer/attention.py (137:143) duplicated block id: 8045 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/transformer.py (578:589) - megatron_patch/model/starcoder/transformer.py (731:742) duplicated block id: 8046 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_moe.py (262:268) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_vl.py (143:149) duplicated block id: 8047 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/mixtral_bak/transformer/attention.py (431:437) - megatron_patch/model/qwen2/transformer/attention.py (564:570) duplicated block id: 8048 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/llama2/transformer.py (1703:1709) - megatron_patch/model/mistral/transformer.py (1676:1682) duplicated block id: 8049 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/transformer_block.py (483:489) - megatron_patch/model/qwen2_5_vl/transformer_block.py (542:548) duplicated block id: 8050 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/qwen1_5/transformer/attention.py (142:149) - megatron_patch/model/qwen2_vl/attention_vision.py (158:165) duplicated block id: 8051 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon40b/checkpoint_reshaping_and_interoperability.py (824:832) - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (829:837) duplicated block id: 8052 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/galactica/checkpoint_reshaping_and_interoperability.py (567:573) - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (614:620) duplicated block id: 8053 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/multi_latent_attention.py (162:168) - megatron_patch/model/qwen1_5/transformer/attention.py (322:328) duplicated block id: 8054 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (410:419) - megatron_patch/model/falcon/transformer.py (265:275) duplicated block id: 8055 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/galactica/checkpoint_reshaping_and_interoperability.py (567:573) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (621:627) duplicated block id: 8056 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/hf2te.py (396:402) - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (614:620) duplicated block id: 8057 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/language_model.py (625:633) - megatron_patch/model/chatglm/language_model.py (651:659) duplicated block id: 8058 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/chatglm/checkpoint_reshaping_and_interoperability.py (451:457) - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (478:485) duplicated block id: 8059 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (929:937) - megatron_patch/model/llama3/transformer_legacy.py (1012:1020) duplicated block id: 8060 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/language_model.py (90:96) - megatron_patch/model/llama3/language_model.py (68:74) duplicated block id: 8061 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/qwen1_5/transformer/attention.py (26:32) - megatron_patch/model/qwen2_vl/attention_vision.py (15:21) duplicated block id: 8062 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/transformer/attention.py (165:171) - megatron_patch/model/qwen2/transformer/attention.py (312:318) duplicated block id: 8063 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/llama/transformer.py (270:284) - megatron_patch/model/starcoder/transformer.py (328:342) duplicated block id: 8064 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/transformer.py (422:429) - megatron_patch/model/galactica/transformer.py (393:400) duplicated block id: 8065 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/transformer/attention.py (165:171) - megatron_patch/model/qwen2/transformer/attention.py (321:327) duplicated block id: 8066 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/multi_latent_attention.py (111:117) - megatron_patch/model/qwen2/transformer/attention.py (249:255) duplicated block id: 8067 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/llama3_1/layer_specs.py (76:82) - megatron_patch/model/qwen3_moe/gpt_layer_specs.py (240:246) duplicated block id: 8068 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (531:537) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (588:594) duplicated block id: 8069 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (531:537) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (599:605) duplicated block id: 8070 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/qwen2/transformer_block.py (322:328) - megatron_patch/model/qwen2_5_vl/transformer_block.py (441:447) duplicated block id: 8071 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/llama3_1/layer_specs.py (104:110) - megatron_patch/model/llava_mcore/vision/vit_layer_specs.py (89:95) duplicated block id: 8072 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/transformer.py (319:325) - megatron_patch/model/mistral/transformer.py (518:524) duplicated block id: 8073 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/galactica/transformer.py (208:218) - megatron_patch/model/mistral/transformer.py (399:408) duplicated block id: 8074 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/galactica/checkpoint_reshaping_and_interoperability.py (186:194) - toolkits/model_checkpoints_convertor/starcoder/checkpoint_reshaping_and_interoperability.py (192:200) duplicated block id: 8075 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (574:580) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (396:403) duplicated block id: 8076 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (53:60) - toolkits/model_checkpoints_convertor/starcoder/checkpoint_reshaping_and_interoperability.py (35:41) duplicated block id: 8077 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/glm130b/checkpoint_reshaping_and_interoperability.py (463:471) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (617:625) duplicated block id: 8078 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/transformer/attention.py (707:713) - megatron_patch/model/qwen2_vl/attention.py (696:702) duplicated block id: 8079 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/multi_latent_attention.py (231:237) - megatron_patch/model/deepseek_v2/multi_latent_attention.py (243:249) duplicated block id: 8080 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/multi_latent_attention.py (231:237) - megatron_patch/model/deepseek_v2/multi_latent_attention.py (255:261) duplicated block id: 8081 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/multi_latent_attention.py (231:237) - megatron_patch/model/deepseek_v2/multi_latent_attention.py (267:273) duplicated block id: 8082 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/mixtral_bak/transformer/attention.py (335:341) - megatron_patch/model/qwen2_vl/attention_vision.py (496:502) duplicated block id: 8083 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/llama3/layer_specs.py (104:110) - megatron_patch/model/qwen1_5/layer_specs.py (105:111) duplicated block id: 8084 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/transformer.py (374:380) - megatron_patch/model/qwen/transformer.py (511:517) duplicated block id: 8085 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/mixtral_bak/moe/experts.py (116:122) - megatron_patch/model/qwen2/moe/experts.py (107:113) duplicated block id: 8086 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (478:485) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (581:587) duplicated block id: 8087 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/qwen2/transformer_block.py (230:236) - megatron_patch/model/qwen2_5_vl/transformer_block.py (324:330) duplicated block id: 8088 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/hf2te.py (424:430) - toolkits/model_checkpoints_convertor/falcon40b/checkpoint_reshaping_and_interoperability.py (522:528) duplicated block id: 8089 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore.py (292:298) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_moe.py (262:268) duplicated block id: 8090 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (612:623) - megatron_patch/model/llama/transformer.py (529:540) duplicated block id: 8091 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (287:301) - megatron_patch/model/qwen1_5_megablocks/transformer.py (314:328) duplicated block id: 8092 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer/attention.py (210:217) - megatron_patch/model/qwen2_vl/attention.py (232:239) duplicated block id: 8093 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (1610:1616) - megatron_patch/model/llama3/transformer_legacy.py (1152:1158) duplicated block id: 8094 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (1725:1741) - megatron_patch/model/starcoder/transformer.py (1280:1296) duplicated block id: 8095 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (559:565) - toolkits/model_checkpoints_convertor/galactica/checkpoint_reshaping_and_interoperability.py (567:573) duplicated block id: 8096 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (559:565) - toolkits/model_checkpoints_convertor/galactica/checkpoint_reshaping_and_interoperability.py (578:584) duplicated block id: 8097 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (535:546) - toolkits/model_checkpoints_convertor/bloom/reward_model_to_megatron.py (713:724) duplicated block id: 8098 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/gpt_model.py (26:37) - megatron_patch/model/llama/gpt_model.py (28:39) duplicated block id: 8099 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/galactica/transformer.py (232:246) - megatron_patch/model/qwen1_5_megablocks/transformer.py (314:328) duplicated block id: 8100 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/mixtral_bak/transformer/attention.py (116:123) - megatron_patch/model/qwen2_vl/attention_vision.py (158:165) duplicated block id: 8101 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/language_model.py (380:386) - megatron_patch/model/llama/language_model.py (366:372) duplicated block id: 8102 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/qwen1_5/transformer/attention.py (253:259) - megatron_patch/model/qwen2_vl/attention.py (334:340) duplicated block id: 8103 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/transformer/attention.py (695:701) - megatron_patch/model/mixtral_bak/transformer/attention.py (443:449) duplicated block id: 8104 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (1740:1746) - megatron_patch/model/qwen/transformer.py (1146:1152) duplicated block id: 8105 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_dense_and_moe_gqa.py (197:203) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_dense_and_moe_gqa.py (536:543) duplicated block id: 8106 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/gpt_model.py (26:37) - megatron_patch/model/baichuan2/gpt_model.py (27:38) duplicated block id: 8107 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/language_model.py (89:95) - megatron_patch/model/llama2/language_model.py (79:85) duplicated block id: 8108 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/llama3/layer_specs.py (104:110) - megatron_patch/model/qwen3_moe/gpt_layer_specs.py (300:306) duplicated block id: 8109 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (1203:1209) - megatron_patch/model/baichuan2/transformer.py (1740:1746) duplicated block id: 8110 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (301:310) - megatron_patch/model/falcon/transformer.py (265:275) duplicated block id: 8111 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/reward_model_to_megatron.py (713:724) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (595:606) duplicated block id: 8112 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_moe.py (318:325) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_dense_and_moe_gqa.py (197:203) duplicated block id: 8113 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/falcon/language_model.py (89:95) - megatron_patch/model/llama3/language_model.py (68:74) duplicated block id: 8114 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer/attention.py (316:322) - megatron_patch/model/qwen2/transformer/attention.py (133:139) duplicated block id: 8115 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/deepseek/hf2mcore_deepseek_v2_moe.py (22:39) - toolkits/model_checkpoints_convertor/deepseek/hf2mcore_deepseek_v3_moe.py (23:39) duplicated block id: 8116 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (435:449) - megatron_patch/model/falcon40b/transformer.py (282:296) duplicated block id: 8117 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/language_model.py (181:193) - megatron_patch/model/starcoder/language_model.py (167:179) duplicated block id: 8118 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/checkpoint_reshaping_and_interoperability.py (170:178) - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (334:342) duplicated block id: 8119 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (521:527) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (396:403) duplicated block id: 8120 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_dense_and_moe_gqa.py (257:263) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_dense_and_moe_gqa.py (335:341) duplicated block id: 8121 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (378:384) - megatron_patch/model/llava/transformer.py (518:524) duplicated block id: 8122 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/language_model.py (651:659) - megatron_patch/model/llava/language_model.py (692:700) duplicated block id: 8123 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (199:207) - toolkits/model_checkpoints_convertor/starcoder/checkpoint_reshaping_and_interoperability.py (192:200) duplicated block id: 8124 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/gpt_model.py (28:39) - megatron_patch/model/falcon40b/gpt_model.py (28:38) duplicated block id: 8125 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/moe/experts.py (670:678) - megatron_patch/model/mixtral/transformer/mlp.py (129:137) duplicated block id: 8126 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (1213:1221) - megatron_patch/model/qwen1_5_megablocks/transformer.py (1096:1104) duplicated block id: 8127 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/layer_specs.py (121:127) - megatron_patch/model/qwen2_moe/layer_specs.py (219:225) duplicated block id: 8128 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/glm130b/checkpoint_reshaping_and_interoperability.py (141:149) - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (334:342) duplicated block id: 8129 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/language_model.py (412:420) - megatron_patch/model/llama2/language_model.py (417:425) duplicated block id: 8130 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/qwen1_5/transformer/attention.py (528:534) - megatron_patch/model/qwen2_vl/attention.py (678:684) duplicated block id: 8131 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (378:384) - megatron_patch/model/llama3/transformer_legacy.py (516:522) duplicated block id: 8132 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore.py (65:72) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (53:60) duplicated block id: 8133 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/galactica/transformer.py (400:408) - megatron_patch/model/glm130b/transformer.py (483:492) duplicated block id: 8134 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/llama3_1/layer_specs.py (104:110) - megatron_patch/model/qwen3_moe/gpt_layer_specs.py (300:306) duplicated block id: 8135 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (326:340) - megatron_patch/model/bloom/transformer.py (287:301) duplicated block id: 8136 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/language_model.py (89:95) - megatron_patch/model/chatglm/language_model.py (90:96) duplicated block id: 8137 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/transformer.py (258:268) - megatron_patch/model/qwen1_5_megablocks/transformer.py (289:298) duplicated block id: 8138 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (1703:1709) - megatron_patch/model/qwen/transformer.py (1621:1627) duplicated block id: 8139 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/hf2te.py (371:377) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (396:403) duplicated block id: 8140 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (960:966) - megatron_patch/model/starcoder/transformer.py (916:922) duplicated block id: 8141 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/language_model.py (90:96) - megatron_patch/model/qwen/language_model.py (79:85) duplicated block id: 8142 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/transformer/attention.py (449:455) - megatron_patch/model/mixtral_bak/transformer/attention.py (298:304) duplicated block id: 8143 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_mha_to_moe.py (217:223) - toolkits/model_checkpoints_convertor/qwen/hf2megablocks_qwen1.5.py (580:586) duplicated block id: 8144 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/qwen2/transformer/attention.py (564:570) - megatron_patch/model/qwen2_vl/attention_vision.py (695:701) duplicated block id: 8145 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/reward_model_to_megatron.py (584:593) - toolkits/model_checkpoints_convertor/starcoder/checkpoint_reshaping_and_interoperability.py (599:608) duplicated block id: 8146 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/falcon/language_model.py (89:95) - megatron_patch/model/qwen_vl/language_model.py (80:86) duplicated block id: 8147 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (1549:1555) - megatron_patch/model/qwen_vl/transformer.py (1702:1708) duplicated block id: 8148 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/transformer/attention.py (449:455) - megatron_patch/model/mixtral_bak/transformer/attention.py (111:117) duplicated block id: 8149 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon40b/checkpoint_reshaping_and_interoperability.py (86:92) - toolkits/model_checkpoints_convertor/qwen/hf2megablocks_qwen1.5.py (119:125) duplicated block id: 8150 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (1549:1555) - megatron_patch/model/qwen/transformer.py (1645:1651) duplicated block id: 8151 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (860:869) - megatron_patch/model/starcoder/transformer.py (995:1004) duplicated block id: 8152 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/transformer.py (258:268) - megatron_patch/model/llama3/transformer_legacy.py (397:406) duplicated block id: 8153 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon/checkpoint_reshaping_and_interoperability.py (482:488) - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (409:416) duplicated block id: 8154 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/deepseek/hf2mcore_deepseek_v2_moe.py (41:48) - toolkits/model_checkpoints_convertor/deepseek/hf2mcore_deepseek_v3_moe.py (42:50) duplicated block id: 8155 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/glm/checkpoint_reshaping_and_interoperability.py (141:149) - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (334:342) duplicated block id: 8156 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/hf2te.py (396:402) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (599:605) duplicated block id: 8157 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_dense_and_moe_gqa.py (138:144) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_moe.py (262:268) duplicated block id: 8158 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (251:259) - toolkits/model_checkpoints_convertor/starcoder/checkpoint_reshaping_and_interoperability.py (192:200) duplicated block id: 8159 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer/attention.py (325:331) - megatron_patch/model/qwen2/transformer/attention.py (133:139) duplicated block id: 8160 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_dense_and_moe_gqa.py (618:624) - toolkits/model_checkpoints_convertor/qwen/hf2megablocks_qwen1.5.py (580:586) duplicated block id: 8161 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/language_model.py (615:623) - megatron_patch/model/llava/language_model.py (692:700) duplicated block id: 8162 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/mistral/transformer.py (1700:1706) - megatron_patch/model/qwen1_5_megablocks/transformer.py (1541:1547) duplicated block id: 8163 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/mistral/hf2mcore.py (232:238) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_moe.py (262:268) duplicated block id: 8164 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (617:625) - toolkits/model_checkpoints_convertor/starcoder/checkpoint_reshaping_and_interoperability.py (753:761) duplicated block id: 8165 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (929:937) - megatron_patch/model/llava/transformer.py (1061:1069) duplicated block id: 8166 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/qwen1_5/transformer/attention.py (137:143) - megatron_patch/model/qwen1_5/transformer/attention.py (325:331) duplicated block id: 8167 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/language_model.py (366:372) - megatron_patch/model/glm130b/language_model.py (380:386) duplicated block id: 8168 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/deepseek/hf2mcore_deepseek_v2_moe.py (50:58) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_moe.py (52:60) duplicated block id: 8169 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/multi_latent_attention.py (111:117) - megatron_patch/model/qwen1_5/transformer/attention.py (253:259) duplicated block id: 8170 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/qwen1_5/transformer/attention.py (137:143) - megatron_patch/model/qwen1_5/transformer/attention.py (316:322) duplicated block id: 8171 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/bloom/gpt_model.py (28:39) - megatron_patch/model/falcon40b/gpt_model.py (28:38) duplicated block id: 8172 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/qwen/transformer.py (1621:1627) - megatron_patch/model/qwen1_5_megablocks/transformer.py (1565:1571) duplicated block id: 8173 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (461:468) - megatron_patch/model/starcoder/transformer.py (621:628) duplicated block id: 8174 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/language_model.py (89:95) - megatron_patch/model/glm130b/language_model.py (90:96) duplicated block id: 8175 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/llama/language_model.py (89:95) - megatron_patch/model/llama2/language_model.py (79:85) duplicated block id: 8176 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/mixtral_bak/transformer/attention.py (184:191) - megatron_patch/model/qwen2_vl/attention_vision.py (231:238) duplicated block id: 8177 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (929:937) - megatron_patch/model/llama2/transformer.py (1062:1070) duplicated block id: 8178 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer/attention.py (325:331) - megatron_patch/model/mixtral/transformer/attention.py (165:171) duplicated block id: 8179 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/qwen2/transformer/attention.py (312:318) - megatron_patch/model/qwen2/transformer/attention.py (321:327) duplicated block id: 8180 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/transformer/attention.py (242:249) - megatron_patch/model/qwen1_5/transformer/attention.py (210:217) duplicated block id: 8181 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/transformer.py (578:589) - megatron_patch/model/mistral/transformer.py (730:741) duplicated block id: 8182 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (809:820) - megatron_patch/model/qwen_vl/transformer.py (845:856) duplicated block id: 8183 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/qwen2/transformer/mlp.py (72:78) - megatron_patch/model/qwen2/transformer/mlp.py (98:104) duplicated block id: 8184 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/llama/transformer.py (246:256) - megatron_patch/model/qwen/transformer.py (392:401) duplicated block id: 8185 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_mha.py (50:58) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_vl.py (54:62) duplicated block id: 8186 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (420:426) - megatron_patch/model/bloom/transformer.py (378:384) duplicated block id: 8187 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/qwen2/transformer/attention.py (138:145) - megatron_patch/model/qwen2_vl/attention.py (159:166) duplicated block id: 8188 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (262:273) - megatron_patch/model/llava/transformer.py (399:408) duplicated block id: 8189 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (287:301) - megatron_patch/model/qwen_vl/transformer.py (424:438) duplicated block id: 8190 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer/attention.py (316:322) - megatron_patch/model/qwen1_5/transformer/attention.py (325:331) duplicated block id: 8191 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer_legacy.py (917:925) - megatron_patch/model/qwen1_5_megablocks/transformer.py (857:865) duplicated block id: 8192 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer/attention.py (325:331) - megatron_patch/model/qwen1_5/transformer/attention.py (316:322) duplicated block id: 8193 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/galactica/transformer.py (618:624) - megatron_patch/model/llama/transformer.py (1046:1052) duplicated block id: 8194 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/transformer.py (1170:1186) - megatron_patch/model/llama/transformer.py (1031:1047) duplicated block id: 8195 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/transformer.py (275:284) - megatron_patch/model/llava/transformer.py (399:408) duplicated block id: 8196 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/language_model.py (499:507) - megatron_patch/model/qwen/language_model.py (491:499) duplicated block id: 8197 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/qwen1_5/transformer/attention.py (547:553) - megatron_patch/model/qwen2_vl/attention_vision.py (707:713) duplicated block id: 8198 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (197:205) - toolkits/model_checkpoints_convertor/glm130b/checkpoint_reshaping_and_interoperability.py (141:149) duplicated block id: 8199 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/moe/experts.py (370:376) - megatron_patch/model/mixtral/moe/experts.py (463:469) duplicated block id: 8200 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/checkpoint_reshaping_and_interoperability.py (170:178) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (231:239) duplicated block id: 8201 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/bloom/language_model.py (91:97) - megatron_patch/model/qwen1_5_megablocks/language_model.py (78:84) duplicated block id: 8202 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (169:176) - toolkits/model_checkpoints_convertor/baichuan2/hf2te.py (122:129) duplicated block id: 8203 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/transformer.py (859:875) - megatron_patch/model/glm130b/transformer.py (1170:1186) duplicated block id: 8204 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/hf2te.py (424:430) - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (531:537) duplicated block id: 8205 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/llama2/transformer.py (1679:1685) - megatron_patch/model/llama2/transformer.py (1703:1709) duplicated block id: 8206 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer/attention.py (314:320) - megatron_patch/model/qwen2_vl/attention_vision.py (445:451) duplicated block id: 8207 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/falcon/transformer.py (547:553) - megatron_patch/model/qwen_vl/transformer.py (518:524) duplicated block id: 8208 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/llama/language_model.py (181:193) - megatron_patch/model/starcoder/language_model.py (167:179) duplicated block id: 8209 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon/checkpoint_reshaping_and_interoperability.py (204:212) - toolkits/model_checkpoints_convertor/starcoder/checkpoint_reshaping_and_interoperability.py (192:200) duplicated block id: 8210 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (921:928) - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (829:837) duplicated block id: 8211 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/transformer/attention.py (344:350) - megatron_patch/model/qwen1_5/transformer/attention.py (253:259) duplicated block id: 8212 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer/attention.py (316:322) - megatron_patch/model/mixtral_bak/transformer/attention.py (298:304) duplicated block id: 8213 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/hf2te.py (34:40) - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (63:70) duplicated block id: 8214 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (1017:1023) - megatron_patch/model/glm130b/transformer.py (972:978) duplicated block id: 8215 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/bloom/language_model.py (585:593) - megatron_patch/model/llama2/language_model.py (632:640) duplicated block id: 8216 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/galactica/transformer.py (324:330) - megatron_patch/model/llama2/transformer.py (517:523) duplicated block id: 8217 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (420:426) - megatron_patch/model/chatglm/transformer.py (319:325) duplicated block id: 8218 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon40b/checkpoint_reshaping_and_interoperability.py (471:477) - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (409:416) duplicated block id: 8219 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/mistral/hf2mcore.py (93:99) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (113:119) duplicated block id: 8220 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/llama3/layer_specs.py (104:110) - megatron_patch/model/llava_mcore/vision/vit_layer_specs.py (89:95) duplicated block id: 8221 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/llama/transformer.py (362:368) - megatron_patch/model/qwen/transformer.py (511:517) duplicated block id: 8222 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/glm/checkpoint_reshaping_and_interoperability.py (141:149) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (231:239) duplicated block id: 8223 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/mixtral_bak/transformer/attention.py (289:295) - megatron_patch/model/qwen2/transformer/attention.py (133:139) duplicated block id: 8224 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (420:426) - megatron_patch/model/llama/transformer.py (362:368) duplicated block id: 8225 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/deepseek/hf2mcore_deepseek_v2_moe.py (118:124) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2.5_vl.py (113:119) duplicated block id: 8226 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/falcon/transformer.py (265:275) - megatron_patch/model/starcoder/transformer.py (425:435) duplicated block id: 8227 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer_legacy.py (1663:1669) - megatron_patch/model/qwen_vl/transformer.py (1678:1684) duplicated block id: 8228 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/qwen_vl/transformer.py (919:927) - megatron_patch/model/starcoder/transformer.py (848:855) duplicated block id: 8229 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/llama/language_model.py (89:95) - megatron_patch/model/qwen1_5_megablocks/language_model.py (78:84) duplicated block id: 8230 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/llava_mcore/vision/vit_layer_specs.py (69:75) - megatron_patch/model/qwen1_5/layer_specs.py (76:82) duplicated block id: 8231 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (254:262) - toolkits/model_checkpoints_convertor/starcoder/checkpoint_reshaping_and_interoperability.py (192:200) duplicated block id: 8232 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (197:205) - toolkits/model_checkpoints_convertor/starcoder/checkpoint_reshaping_and_interoperability.py (192:200) duplicated block id: 8233 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/glm130b/checkpoint_reshaping_and_interoperability.py (422:433) - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (588:599) duplicated block id: 8234 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/llama2/transformer.py (1703:1709) - megatron_patch/model/qwen1_5_megablocks/transformer.py (1541:1547) duplicated block id: 8235 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/llama2/transformer.py (1679:1685) - megatron_patch/model/qwen_vl/transformer.py (1702:1708) duplicated block id: 8236 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (1573:1579) - megatron_patch/model/qwen/transformer.py (1621:1627) duplicated block id: 8237 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (588:599) - toolkits/model_checkpoints_convertor/starcoder/checkpoint_reshaping_and_interoperability.py (714:725) duplicated block id: 8238 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/qwen1_5/moe/experts.py (83:89) - megatron_patch/model/qwen1_5/moe/experts.py (118:124) duplicated block id: 8239 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/falcon/transformer.py (265:275) - megatron_patch/model/starcoder/transformer.py (302:312) duplicated block id: 8240 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/language_model.py (625:633) - megatron_patch/model/glm130b/language_model.py (615:623) duplicated block id: 8241 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/bloom/language_model.py (585:593) - megatron_patch/model/mistral/language_model.py (644:652) duplicated block id: 8242 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/model.py (83:89) - megatron_patch/model/qwen2/model.py (90:96) duplicated block id: 8243 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/qwen2/moe/experts.py (107:113) - megatron_patch/model/qwen2/moe/experts.py (142:148) duplicated block id: 8244 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (914:920) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (981:987) duplicated block id: 8245 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/llama/transformer.py (529:540) - megatron_patch/model/llama2/transformer.py (732:743) duplicated block id: 8246 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer/attention.py (559:565) - megatron_patch/model/qwen2/transformer/attention.py (552:558) duplicated block id: 8247 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (529:535) - megatron_patch/model/falcon40b/transformer.py (374:380) duplicated block id: 8248 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/transformer.py (578:589) - megatron_patch/model/llava/transformer.py (731:742) duplicated block id: 8249 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/deepseek/hf2mcore_deepseek_v2_moe.py (408:414) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_moe.py (571:577) duplicated block id: 8250 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (1679:1685) - megatron_patch/model/llama3/transformer_legacy.py (1663:1669) duplicated block id: 8251 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/llama3/layer_specs.py (76:82) - megatron_patch/model/mixtral_bak/layer_specs.py (99:105) duplicated block id: 8252 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/transformer/attention.py (458:464) - megatron_patch/model/qwen1_5/transformer/attention.py (316:322) duplicated block id: 8253 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/bloom/language_model.py (91:97) - megatron_patch/model/mistral/language_model.py (80:86) duplicated block id: 8254 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (449:456) - megatron_patch/model/falcon/transformer.py (647:654) duplicated block id: 8255 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/llama2/transformer.py (1679:1685) - megatron_patch/model/qwen1_5_megablocks/transformer.py (1565:1571) duplicated block id: 8256 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (535:546) - toolkits/model_checkpoints_convertor/starcoder/checkpoint_reshaping_and_interoperability.py (714:725) duplicated block id: 8257 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/galactica/language_model.py (552:558) - megatron_patch/model/glm130b/language_model.py (468:474) duplicated block id: 8258 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/moe/experts.py (139:145) - megatron_patch/model/mixtral_bak/moe/experts.py (81:87) duplicated block id: 8259 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/transformer/attention.py (449:455) - megatron_patch/model/qwen1_5/transformer/attention.py (325:331) duplicated block id: 8260 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (1120:1128) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (1132:1140) duplicated block id: 8261 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (1029:1041) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (1043:1049) duplicated block id: 8262 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/mlp.py (121:129) - megatron_patch/model/mixtral/moe/experts.py (671:679) duplicated block id: 8263 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon/checkpoint_reshaping_and_interoperability.py (438:445) - toolkits/model_checkpoints_convertor/galactica/checkpoint_reshaping_and_interoperability.py (477:485) duplicated block id: 8264 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (1071:1077) - megatron_patch/model/llama3/transformer_legacy.py (1152:1158) duplicated block id: 8265 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/hf2te.py (180:192) - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (279:291) duplicated block id: 8266 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (521:527) - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (478:485) duplicated block id: 8267 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/transformer.py (594:600) - megatron_patch/model/llama/transformer.py (1046:1052) duplicated block id: 8268 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/transformer/attention.py (449:455) - megatron_patch/model/qwen2/transformer/attention.py (321:327) duplicated block id: 8269 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/falcon/transformer.py (547:553) - megatron_patch/model/llama3/transformer_legacy.py (516:522) duplicated block id: 8270 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/gpt_model.py (27:38) - megatron_patch/model/falcon/gpt_model.py (28:38) duplicated block id: 8271 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/chatglm/checkpoint_reshaping_and_interoperability.py (292:306) - toolkits/model_checkpoints_convertor/glm/checkpoint_reshaping_and_interoperability.py (265:279) duplicated block id: 8272 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore.py (540:546) - toolkits/model_checkpoints_convertor/qwen/hf2megablocks_qwen1.5.py (550:556) duplicated block id: 8273 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (235:243) - toolkits/model_checkpoints_convertor/starcoder/checkpoint_reshaping_and_interoperability.py (192:200) duplicated block id: 8274 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/llama3/layer_specs.py (76:82) - megatron_patch/model/llava_mcore/llava_spec.py (73:79) duplicated block id: 8275 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/qwen1_5/transformer/attention.py (325:331) - megatron_patch/model/qwen2/transformer/attention.py (312:318) duplicated block id: 8276 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/galactica/transformer.py (232:246) - megatron_patch/model/starcoder/transformer.py (328:342) duplicated block id: 8277 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/reward_model_to_megatron.py (170:178) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (251:259) duplicated block id: 8278 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/llama3_1/layer_specs.py (76:82) - megatron_patch/model/llava_mcore/vision/vit_layer_specs.py (69:75) duplicated block id: 8279 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/language_model.py (412:420) - megatron_patch/model/falcon40b/language_model.py (398:406) duplicated block id: 8280 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/qwen/language_model.py (486:492) - megatron_patch/model/qwen_vl/language_model.py (519:525) duplicated block id: 8281 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/mixtral_bak/layer_specs.py (99:105) - megatron_patch/model/qwen3_moe/gpt_layer_specs.py (240:246) duplicated block id: 8282 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/chatglm/checkpoint_reshaping_and_interoperability.py (169:177) - toolkits/model_checkpoints_convertor/glm130b/checkpoint_reshaping_and_interoperability.py (141:149) duplicated block id: 8283 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_mha.py (280:286) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_mha_to_moe.py (239:246) duplicated block id: 8284 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/multi_latent_attention.py (162:168) - megatron_patch/model/mixtral/transformer/attention.py (455:461) duplicated block id: 8285 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/mlp.py (178:184) - megatron_patch/model/mixtral_bak/transformer/mlp.py (162:168) duplicated block id: 8286 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/checkpoint_reshaping_and_interoperability.py (35:41) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (53:60) duplicated block id: 8287 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/transformer.py (698:709) - megatron_patch/model/starcoder/transformer.py (814:824) duplicated block id: 8288 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/language_model.py (412:420) - megatron_patch/model/galactica/language_model.py (429:437) duplicated block id: 8289 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/transformer.py (319:325) - megatron_patch/model/qwen_vl/transformer.py (518:524) duplicated block id: 8290 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/galactica/transformer.py (208:218) - megatron_patch/model/llava/transformer.py (399:408) duplicated block id: 8291 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/bloom/language_model.py (91:97) - megatron_patch/model/llama3/language_model.py (68:74) duplicated block id: 8292 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/glm130b/checkpoint_reshaping_and_interoperability.py (422:433) - toolkits/model_checkpoints_convertor/yi/checkpoint_reshaping_and_interoperability.py (514:525) duplicated block id: 8293 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/model.py (101:107) - megatron_patch/model/qwen2/model.py (90:96) duplicated block id: 8294 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/checkpoint_reshaping_and_interoperability.py (170:178) - toolkits/model_checkpoints_convertor/falcon/checkpoint_reshaping_and_interoperability.py (204:212) duplicated block id: 8295 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2.5_vl.py (330:336) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_dense_and_moe_gqa.py (424:431) duplicated block id: 8296 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/qwen/transformer.py (1667:1683) - megatron_patch/model/starcoder/transformer.py (1280:1296) duplicated block id: 8297 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/multi_latent_attention.py (111:117) - megatron_patch/model/llama3/transformer/attention.py (253:259) duplicated block id: 8298 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/qwen/transformer.py (1146:1152) - megatron_patch/model/qwen/transformer.py (1682:1688) duplicated block id: 8299 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/transformer.py (991:997) - megatron_patch/model/llama/transformer.py (723:729) duplicated block id: 8300 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/model.py (101:107) - megatron_patch/model/mixtral_bak/model.py (82:88) duplicated block id: 8301 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/falcon/language_model.py (526:532) - megatron_patch/model/glm130b/language_model.py (468:474) duplicated block id: 8302 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/moe/experts.py (671:679) - megatron_patch/model/mixtral_bak/transformer/mlp.py (98:106) duplicated block id: 8303 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (134:143) - megatron_patch/model/falcon/transformer.py (169:178) duplicated block id: 8304 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer/attention.py (316:322) - megatron_patch/model/qwen2/transformer/attention.py (321:327) duplicated block id: 8305 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (529:535) - megatron_patch/model/glm130b/transformer.py (412:418) duplicated block id: 8306 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/galactica/transformer.py (232:246) - megatron_patch/model/qwen/transformer.py (417:431) duplicated block id: 8307 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/llava/transformer.py (1702:1708) - megatron_patch/model/qwen/transformer.py (1621:1627) duplicated block id: 8308 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/galactica/checkpoint_reshaping_and_interoperability.py (85:91) - toolkits/model_checkpoints_convertor/qwen/hf2megablocks_qwen1.5.py (119:125) duplicated block id: 8309 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/glm/checkpoint_reshaping_and_interoperability.py (398:405) - toolkits/model_checkpoints_convertor/starcoder/checkpoint_reshaping_and_interoperability.py (678:685) duplicated block id: 8310 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/moe/experts.py (139:145) - megatron_patch/model/qwen1_5/moe/experts.py (83:89) duplicated block id: 8311 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/qwen2/transformer/attention.py (310:316) - megatron_patch/model/qwen2_vl/attention_vision.py (445:451) duplicated block id: 8312 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/transformer.py (382:389) - megatron_patch/model/falcon/transformer.py (647:654) duplicated block id: 8313 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/llava/transformer.py (1678:1684) - megatron_patch/model/llava/transformer.py (1702:1708) duplicated block id: 8314 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (410:419) - megatron_patch/model/falcon40b/transformer.py (258:268) duplicated block id: 8315 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (1703:1709) - megatron_patch/model/llama2/transformer.py (1679:1685) duplicated block id: 8316 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer/attention.py (528:534) - megatron_patch/model/qwen2_vl/attention_vision.py (677:683) duplicated block id: 8317 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (829:837) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (1140:1148) duplicated block id: 8318 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (559:565) - toolkits/model_checkpoints_convertor/falcon40b/checkpoint_reshaping_and_interoperability.py (522:528) duplicated block id: 8319 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/language_model.py (378:384) - megatron_patch/model/glm130b/language_model.py (380:386) duplicated block id: 8320 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (929:937) - megatron_patch/model/qwen_vl/transformer.py (1061:1069) duplicated block id: 8321 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (521:527) - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (409:416) duplicated block id: 8322 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (423:431) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (476:484) duplicated block id: 8323 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (1071:1077) - megatron_patch/model/llama3/transformer_legacy.py (1700:1706) duplicated block id: 8324 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (595:606) - toolkits/model_checkpoints_convertor/starcoder/checkpoint_reshaping_and_interoperability.py (714:725) duplicated block id: 8325 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/qwen_vl/transformer.py (518:524) - megatron_patch/model/starcoder/transformer.py (524:530) duplicated block id: 8326 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/language_model.py (90:96) - megatron_patch/model/qwen1_5_megablocks/language_model.py (78:84) duplicated block id: 8327 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/glm/checkpoint_reshaping_and_interoperability.py (441:452) - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (588:599) duplicated block id: 8328 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/galactica/gpt_model.py (51:60) - megatron_patch/model/starcoder/gpt_model.py (52:62) duplicated block id: 8329 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/llava_mcore/vision/vit_layer_specs.py (89:95) - megatron_patch/model/qwen1_5/layer_specs.py (105:111) duplicated block id: 8330 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/chatglm/checkpoint_reshaping_and_interoperability.py (544:552) - toolkits/model_checkpoints_convertor/falcon40b/checkpoint_reshaping_and_interoperability.py (558:566) duplicated block id: 8331 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/llama3_1/layer_specs.py (76:82) - megatron_patch/model/qwen2_moe/layer_specs.py (219:225) duplicated block id: 8332 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_mha.py (50:58) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_moe.py (112:120) duplicated block id: 8333 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/deepseek/hf2mcore_deepseek_v2_moe.py (333:340) - toolkits/model_checkpoints_convertor/deepseek/hf2mcore_deepseek_v3_moe.py (504:511) duplicated block id: 8334 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/falcon/transformer.py (1252:1258) - megatron_patch/model/galactica/transformer.py (618:624) duplicated block id: 8335 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/falcon/language_model.py (89:95) - megatron_patch/model/qwen/language_model.py (79:85) duplicated block id: 8336 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (542:551) - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (647:656) duplicated block id: 8337 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (570:579) - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (647:656) duplicated block id: 8338 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (197:205) - toolkits/model_checkpoints_convertor/bloom/checkpoint_reshaping_and_interoperability.py (170:178) duplicated block id: 8339 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/language_model.py (709:717) - megatron_patch/model/glm130b/language_model.py (615:623) duplicated block id: 8340 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (85:91) - toolkits/model_checkpoints_convertor/mistral/hf2mcore.py (93:99) duplicated block id: 8341 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (929:937) - megatron_patch/model/qwen/transformer.py (1007:1015) duplicated block id: 8342 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/mlp.py (121:129) - megatron_patch/model/deepseek_v2/moe/experts.py (670:678) duplicated block id: 8343 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/qwen2/transformer/attention.py (564:570) - megatron_patch/model/qwen2_vl/attention.py (696:702) duplicated block id: 8344 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/transformer.py (374:380) - megatron_patch/model/llama3/transformer_legacy.py (516:522) duplicated block id: 8345 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/qwen2/transformer/attention.py (22:28) - megatron_patch/model/qwen2_vl/attention.py (15:21) duplicated block id: 8346 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/language_model.py (709:717) - megatron_patch/model/bloom/language_model.py (585:593) duplicated block id: 8347 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (409:416) - toolkits/model_checkpoints_convertor/yi/checkpoint_reshaping_and_interoperability.py (500:506) duplicated block id: 8348 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon40b/checkpoint_reshaping_and_interoperability.py (558:566) - toolkits/model_checkpoints_convertor/galactica/checkpoint_reshaping_and_interoperability.py (614:622) duplicated block id: 8349 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/galactica/checkpoint_reshaping_and_interoperability.py (186:194) - toolkits/model_checkpoints_convertor/glm/checkpoint_reshaping_and_interoperability.py (141:149) duplicated block id: 8350 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/transformer.py (282:296) - megatron_patch/model/llava/transformer.py (424:438) duplicated block id: 8351 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (1549:1555) - megatron_patch/model/baichuan2/transformer.py (1703:1709) duplicated block id: 8352 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/qwen1_5/transformer/attention.py (314:320) - megatron_patch/model/qwen2_vl/attention.py (446:452) duplicated block id: 8353 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/transformer/attention.py (707:713) - megatron_patch/model/qwen1_5/transformer/attention.py (547:553) duplicated block id: 8354 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/gpt_model.py (26:37) - megatron_patch/model/glm130b/gpt_model.py (28:39) duplicated block id: 8355 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon/checkpoint_reshaping_and_interoperability.py (482:488) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (396:403) duplicated block id: 8356 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2.5_vl.py (192:198) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_moe.py (262:268) duplicated block id: 8357 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (166:173) - megatron_patch/model/mixtral_bak/moe/moe_utils.py (61:68) duplicated block id: 8358 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/llama2/transformer.py (1679:1685) - megatron_patch/model/llama3/transformer_legacy.py (1663:1669) duplicated block id: 8359 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/llava/transformer.py (1678:1684) - megatron_patch/model/qwen/transformer.py (1645:1651) duplicated block id: 8360 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/llama3_1/layer_specs.py (104:110) - megatron_patch/model/qwen1_5/layer_specs.py (105:111) duplicated block id: 8361 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/transformer/attention.py (449:455) - megatron_patch/model/qwen1_5/transformer/attention.py (137:143) duplicated block id: 8362 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/galactica/transformer.py (208:218) - megatron_patch/model/qwen_vl/transformer.py (399:408) duplicated block id: 8363 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/qwen2/transformer/attention.py (133:139) - megatron_patch/model/qwen2/transformer/attention.py (321:327) duplicated block id: 8364 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/qwen2_vl/attention.py (696:702) - megatron_patch/model/qwen2_vl/attention.py (708:714) duplicated block id: 8365 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer/attention.py (325:331) - megatron_patch/model/mixtral_bak/transformer/attention.py (289:295) duplicated block id: 8366 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/language_model.py (344:350) - megatron_patch/model/chatglm/language_model.py (350:356) duplicated block id: 8367 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/language_model.py (89:95) - megatron_patch/model/galactica/language_model.py (89:95) duplicated block id: 8368 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/transformer.py (578:589) - megatron_patch/model/llama2/transformer.py (732:743) duplicated block id: 8369 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/llama/gpt_model.py (50:59) - megatron_patch/model/starcoder/gpt_model.py (52:62) duplicated block id: 8370 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/mlp.py (170:176) - megatron_patch/model/mixtral_bak/transformer/mlp.py (154:160) duplicated block id: 8371 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/qwen2/transformer/attention.py (133:139) - megatron_patch/model/qwen2/transformer/attention.py (312:318) duplicated block id: 8372 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/llava/transformer.py (1724:1740) - megatron_patch/model/starcoder/transformer.py (1280:1296) duplicated block id: 8373 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/falcon/transformer.py (757:768) - megatron_patch/model/llama/transformer.py (529:540) duplicated block id: 8374 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (805:811) - megatron_patch/model/falcon/transformer.py (992:998) duplicated block id: 8375 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/mistral/transformer.py (620:631) - megatron_patch/model/qwen1_5_megablocks/transformer.py (510:521) duplicated block id: 8376 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer_legacy.py (1685:1701) - megatron_patch/model/starcoder/transformer.py (1280:1296) duplicated block id: 8377 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (773:782) - megatron_patch/model/bloom/transformer.py (635:643) duplicated block id: 8378 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llava/hf2mcore_llava.py (358:364) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_dense_and_moe_gqa.py (706:712) duplicated block id: 8379 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/hf2te.py (286:298) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (490:502) duplicated block id: 8380 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/language_model.py (350:356) - megatron_patch/model/qwen1_5_megablocks/language_model.py (333:339) duplicated block id: 8381 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (190:197) - toolkits/model_checkpoints_convertor/yi/checkpoint_reshaping_and_interoperability.py (154:161) duplicated block id: 8382 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (524:533) - megatron_patch/model/glm130b/transformer.py (571:580) duplicated block id: 8383 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (1573:1579) - megatron_patch/model/baichuan2/transformer.py (1679:1685) duplicated block id: 8384 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer_legacy.py (1663:1669) - megatron_patch/model/llava/transformer.py (1678:1684) duplicated block id: 8385 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/moe/experts.py (670:678) - megatron_patch/model/llama3/transformer/mlp.py (123:131) duplicated block id: 8386 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_dense_and_moe_gqa.py (664:670) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_dense_and_moe_gqa.py (722:728) duplicated block id: 8387 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (841:847) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (981:987) duplicated block id: 8388 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_moe.py (262:268) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_vl.py (192:198) duplicated block id: 8389 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/transformer.py (228:242) - megatron_patch/model/llama2/transformer.py (423:437) duplicated block id: 8390 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/multi_latent_attention.py (162:168) - megatron_patch/model/qwen2/transformer/attention.py (318:324) duplicated block id: 8391 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/mixtral_bak/transformer/attention.py (431:437) - megatron_patch/model/mixtral_bak/transformer/attention.py (443:449) duplicated block id: 8392 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/transformer.py (374:380) - megatron_patch/model/mistral/transformer.py (518:524) duplicated block id: 8393 size: 7 cleaned lines of code in 2 files: - megatron_patch/tokenizer/tokenization_baichuan.py (44:50) - megatron_patch/tokenizer/tokenization_yi.py (41:47) duplicated block id: 8394 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (1610:1616) - megatron_patch/model/qwen/transformer.py (1146:1152) duplicated block id: 8395 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/glm/checkpoint_reshaping_and_interoperability.py (141:149) - toolkits/model_checkpoints_convertor/yi/checkpoint_reshaping_and_interoperability.py (184:192) duplicated block id: 8396 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/llama3/layer_specs.py (76:82) - megatron_patch/model/qwen2_moe/layer_specs.py (219:225) duplicated block id: 8397 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (1703:1709) - megatron_patch/model/qwen_vl/transformer.py (1678:1684) duplicated block id: 8398 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/galactica/transformer.py (232:246) - megatron_patch/model/llama3/transformer_legacy.py (422:436) duplicated block id: 8399 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/llama2/transformer.py (1703:1709) - megatron_patch/model/llava/transformer.py (1678:1684) duplicated block id: 8400 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/llava/language_model.py (692:700) - megatron_patch/model/starcoder/language_model.py (555:563) duplicated block id: 8401 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer/attention.py (26:32) - megatron_patch/model/qwen2_vl/attention.py (15:21) duplicated block id: 8402 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/transformer/attention.py (695:701) - megatron_patch/model/mixtral/transformer/attention.py (707:713) duplicated block id: 8403 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/mixtral_bak/transformer/attention.py (412:418) - megatron_patch/model/qwen2_vl/attention.py (678:684) duplicated block id: 8404 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer/attention.py (26:32) - megatron_patch/model/mixtral/transformer/attention.py (27:33) duplicated block id: 8405 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore.py (525:531) - toolkits/model_checkpoints_convertor/qwen/hf2megablocks_qwen1.5.py (535:541) duplicated block id: 8406 size: 7 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/hf2te.py (424:430) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (621:627) duplicated block id: 8407 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/language_model.py (350:356) - megatron_patch/model/mistral/language_model.py (335:341) duplicated block id: 8408 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/qwen2/transformer/attention.py (138:145) - megatron_patch/model/qwen2_vl/attention_vision.py (158:165) duplicated block id: 8409 size: 7 cleaned lines of code in 2 files: - megatron_patch/model/llama/transformer.py (529:540) - megatron_patch/model/qwen_vl/transformer.py (731:742) duplicated block id: 8410 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/mixtral_bak/transformer/attention.py (288:293) - megatron_patch/model/qwen2_vl/attention_vision.py (446:451) duplicated block id: 8411 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/mistral/transformer.py (706:711) - megatron_patch/model/mixtral_bak/transformer/attention.py (366:371) duplicated block id: 8412 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/gpt_model.py (100:107) - megatron_patch/model/qwen/gpt_model.py (108:115) duplicated block id: 8413 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/language_model.py (406:411) - megatron_patch/model/starcoder/language_model.py (371:376) duplicated block id: 8414 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/gpt_model.py (100:107) - megatron_patch/model/llava/gpt_model.py (109:116) duplicated block id: 8415 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer_legacy.py (1240:1245) - megatron_patch/model/mistral/transformer.py (1097:1102) duplicated block id: 8416 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_moe.py (388:393) - toolkits/model_checkpoints_convertor/qwen/hf2megablocks_qwen1.5.py (594:599) duplicated block id: 8417 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (899:908) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (978:987) duplicated block id: 8418 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/transformer.py (204:210) - megatron_patch/model/llava/transformer.py (339:345) duplicated block id: 8419 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore.py (381:386) - toolkits/model_checkpoints_convertor/mistral/hf2mcore.py (322:327) duplicated block id: 8420 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/llama3_1/layer_specs.py (69:74) - megatron_patch/model/mixtral_bak/layer_specs.py (92:97) duplicated block id: 8421 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/bloom/gpt_model.py (101:108) - megatron_patch/model/qwen_vl/gpt_model.py (109:116) duplicated block id: 8422 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/hf2te.py (172:177) - toolkits/model_checkpoints_convertor/baichuan2/hf2te.py (171:176) duplicated block id: 8423 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (599:604) - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (513:518) duplicated block id: 8424 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/language_model.py (518:523) - megatron_patch/model/falcon/language_model.py (542:547) duplicated block id: 8425 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (925:935) - megatron_patch/model/galactica/transformer.py (670:680) duplicated block id: 8426 size: 6 cleaned lines of code in 2 files: - toolkits/distributed_checkpoints_convertor/impl/general/h2m_synchronizer.py (146:152) - toolkits/distributed_checkpoints_convertor/impl/general/m2h_synchronizer.py (112:118) duplicated block id: 8427 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/language_model.py (408:413) - megatron_patch/model/falcon40b/language_model.py (356:361) duplicated block id: 8428 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/llama2/transformer.py (707:712) - megatron_patch/model/qwen1_5/transformer/attention.py (476:481) duplicated block id: 8429 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/hf2te.py (261:270) - toolkits/model_checkpoints_convertor/falcon40b/checkpoint_reshaping_and_interoperability.py (384:393) duplicated block id: 8430 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/chatglm/checkpoint_reshaping_and_interoperability.py (256:261) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (318:323) duplicated block id: 8431 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/hf2te.py (187:192) - toolkits/model_checkpoints_convertor/chatglm/checkpoint_reshaping_and_interoperability.py (256:261) duplicated block id: 8432 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (1071:1076) - megatron_patch/model/qwen_vl/transformer.py (1200:1205) duplicated block id: 8433 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/galactica/transformer.py (553:563) - megatron_patch/model/glm130b/transformer.py (724:734) duplicated block id: 8434 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/language_model.py (367:372) - megatron_patch/model/qwen_vl/language_model.py (353:358) duplicated block id: 8435 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon40b/checkpoint_reshaping_and_interoperability.py (542:550) - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (479:487) duplicated block id: 8436 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (613:621) - megatron_patch/model/llama3/transformer_legacy.py (837:844) duplicated block id: 8437 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (673:678) - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (513:518) duplicated block id: 8438 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/mistral/hf2mcore.py (407:412) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_moe.py (328:333) duplicated block id: 8439 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/galactica/language_model.py (166:183) - megatron_patch/model/qwen/language_model.py (130:148) duplicated block id: 8440 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/llama/transformer.py (868:873) - megatron_patch/model/mistral/transformer.py (1406:1411) duplicated block id: 8441 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon/checkpoint_reshaping_and_interoperability.py (422:427) - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (816:821) duplicated block id: 8442 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/transformer/attention.py (496:501) - megatron_patch/model/mixtral_bak/transformer/attention.py (412:417) duplicated block id: 8443 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/transformer.py (510:518) - megatron_patch/model/qwen_vl/transformer.py (885:892) duplicated block id: 8444 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/moe/router.py (202:207) - megatron_patch/model/qwen2/moe/router.py (269:274) duplicated block id: 8445 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (211:219) - megatron_patch/model/llama/transformer.py (215:223) duplicated block id: 8446 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/galactica/transformer.py (702:707) - megatron_patch/model/qwen_vl/transformer.py (1408:1413) duplicated block id: 8447 size: 6 cleaned lines of code in 2 files: - megatron_patch/arguments.py (418:423) - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (60:65) duplicated block id: 8448 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2megablocks_qwen1.5.py (123:131) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (151:158) duplicated block id: 8449 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/qwen/language_model.py (352:357) - megatron_patch/model/starcoder/language_model.py (352:357) duplicated block id: 8450 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/llama2/language_model.py (402:410) - megatron_patch/model/starcoder/language_model.py (378:386) duplicated block id: 8451 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/galactica/checkpoint_reshaping_and_interoperability.py (441:446) - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (311:316) duplicated block id: 8452 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/language_model.py (414:425) - megatron_patch/model/qwen1_5_megablocks/language_model.py (433:444) duplicated block id: 8453 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/glm/checkpoint_reshaping_and_interoperability.py (358:363) - toolkits/model_checkpoints_convertor/starcoder/checkpoint_reshaping_and_interoperability.py (401:406) duplicated block id: 8454 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_dense_and_moe_gqa.py (9:22) - toolkits/model_checkpoints_convertor/qwen/hf2megablocks_qwen1.5.py (8:22) duplicated block id: 8455 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/qwen1_5/transformer/attention.py (384:389) - megatron_patch/model/qwen2_vl/attention.py (521:526) duplicated block id: 8456 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (981:987) - megatron_patch/model/qwen/transformer.py (1382:1388) duplicated block id: 8457 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (402:411) - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (362:371) duplicated block id: 8458 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/language_model.py (370:375) - megatron_patch/model/llama2/language_model.py (352:357) duplicated block id: 8459 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/bloom/language_model.py (368:373) - megatron_patch/model/mistral/language_model.py (353:358) duplicated block id: 8460 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (588:593) - megatron_patch/model/mixtral/transformer/attention.py (619:624) duplicated block id: 8461 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/transformer.py (163:171) - megatron_patch/model/llama/transformer.py (215:223) duplicated block id: 8462 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/moe/experts.py (220:225) - megatron_patch/model/qwen2/moe/experts.py (324:329) duplicated block id: 8463 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/checkpoint_reshaping_and_interoperability.py (624:629) - toolkits/model_checkpoints_convertor/falcon40b/checkpoint_reshaping_and_interoperability.py (682:687) duplicated block id: 8464 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/galactica/transformer.py (154:160) - megatron_patch/model/mistral/transformer.py (339:345) duplicated block id: 8465 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/language_model.py (478:483) - megatron_patch/model/llama/language_model.py (533:538) duplicated block id: 8466 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_moe.py (380:385) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_vl.py (328:333) duplicated block id: 8467 size: 6 cleaned lines of code in 2 files: - toolkits/pretrain_data_preprocessing/clean_raw_text.py (29:34) - toolkits/pretrain_data_preprocessing/preprocess_wudao2.py (28:33) duplicated block id: 8468 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/language_model.py (423:428) - megatron_patch/model/llama2/language_model.py (362:367) duplicated block id: 8469 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/transformer/attention.py (135:140) - megatron_patch/model/qwen1_5/transformer/mlp.py (110:115) duplicated block id: 8470 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (1210:1218) - megatron_patch/model/llama/transformer.py (1102:1110) duplicated block id: 8471 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/moe/experts.py (71:77) - megatron_patch/model/mixtral_bak/moe/experts.py (51:57) duplicated block id: 8472 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (916:921) - toolkits/model_checkpoints_convertor/chatglm/checkpoint_reshaping_and_interoperability.py (583:588) duplicated block id: 8473 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/transformer_layer.py (171:187) - megatron_patch/model/qwen2/transformer_layer.py (129:145) duplicated block id: 8474 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/falcon/transformer.py (467:477) - megatron_patch/model/glm130b/transformer.py (312:322) duplicated block id: 8475 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer/mlp.py (94:99) - megatron_patch/model/qwen1_5/transformer/attention.py (107:112) duplicated block id: 8476 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (145:151) - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (196:202) duplicated block id: 8477 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/mistral/transformer.py (1198:1203) - megatron_patch/model/qwen/transformer.py (1682:1687) duplicated block id: 8478 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (1276:1281) - megatron_patch/model/falcon/transformer.py (1074:1079) duplicated block id: 8479 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/language_model.py (409:417) - megatron_patch/model/starcoder/language_model.py (378:386) duplicated block id: 8480 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/deepseek/hf2mcore_deepseek_v3_moe.py (570:575) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_moe.py (572:577) duplicated block id: 8481 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/chatglm/checkpoint_reshaping_and_interoperability.py (367:372) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (751:756) duplicated block id: 8482 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/transformer.py (504:510) - megatron_patch/model/qwen/transformer.py (613:619) duplicated block id: 8483 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (490:495) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (837:842) duplicated block id: 8484 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/llama2/transformer.py (1201:1206) - megatron_patch/model/qwen/transformer.py (1682:1687) duplicated block id: 8485 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/checkpoint_reshaping_and_interoperability.py (624:629) - toolkits/model_checkpoints_convertor/bloom/reward_model_to_megatron.py (378:383) duplicated block id: 8486 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (1440:1446) - megatron_patch/model/bloom/transformer.py (981:987) duplicated block id: 8487 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (183:189) - megatron_patch/model/qwen1_5_megablocks/transformer.py (230:236) duplicated block id: 8488 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/deepseek/hf2mcore_deepseek_v2_moe.py (393:398) - toolkits/model_checkpoints_convertor/mistral/hf2mcore.py (407:412) duplicated block id: 8489 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (588:593) - megatron_patch/model/qwen2/transformer/attention.py (478:483) duplicated block id: 8490 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/transformer/attention.py (496:501) - megatron_patch/model/qwen2/transformer/attention.py (533:538) duplicated block id: 8491 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/falcon/language_model.py (414:425) - megatron_patch/model/llama2/language_model.py (434:445) duplicated block id: 8492 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (599:604) - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (816:821) duplicated block id: 8493 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (1276:1281) - megatron_patch/model/glm130b/transformer.py (890:895) duplicated block id: 8494 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/transformer.py (890:895) - megatron_patch/model/llava/transformer.py (1408:1413) duplicated block id: 8495 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/mistral/transformer.py (1246:1251) - megatron_patch/model/starcoder/transformer.py (962:967) duplicated block id: 8496 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/bloom/gpt_model.py (89:96) - megatron_patch/model/falcon/gpt_model.py (94:100) duplicated block id: 8497 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/bloom/gpt_model.py (101:108) - megatron_patch/model/mistral/gpt_model.py (108:115) duplicated block id: 8498 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/falcon/transformer.py (211:217) - megatron_patch/model/qwen_vl/transformer.py (339:345) duplicated block id: 8499 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (464:469) - megatron_patch/model/falcon/transformer.py (663:668) duplicated block id: 8500 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/checkpoint_reshaping_and_interoperability.py (337:343) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (742:748) duplicated block id: 8501 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/language_model.py (680:685) - megatron_patch/model/starcoder/language_model.py (558:563) duplicated block id: 8502 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/transformer.py (217:223) - megatron_patch/model/llava/transformer.py (339:345) duplicated block id: 8503 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (981:987) - megatron_patch/model/mistral/transformer.py (1437:1443) duplicated block id: 8504 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/hf2te.py (314:319) - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (373:378) duplicated block id: 8505 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (284:289) - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (299:304) duplicated block id: 8506 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/transformer.py (735:741) - megatron_patch/model/mistral/transformer.py (918:925) duplicated block id: 8507 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/bloom/positional_embeddings.py (127:133) - megatron_patch/model/bloom/transformer.py (724:730) duplicated block id: 8508 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/language_model.py (434:440) - megatron_patch/model/glm130b/language_model.py (391:397) duplicated block id: 8509 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/falcon/transformer.py (965:970) - megatron_patch/model/llama/transformer.py (795:800) duplicated block id: 8510 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/transformer/attention.py (135:140) - megatron_patch/model/mixtral_bak/transformer/mlp.py (72:77) duplicated block id: 8511 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (336:342) - megatron_patch/model/chatglm/transformer.py (134:140) duplicated block id: 8512 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/layer_specs.py (29:35) - megatron_patch/model/qwen3_moe/moe_module_specs.py (18:24) duplicated block id: 8513 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/mlp.py (72:77) - megatron_patch/model/qwen2/transformer/attention.py (103:108) duplicated block id: 8514 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore.py (722:727) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_moe.py (531:536) duplicated block id: 8515 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/mistral/hf2mcore.py (478:483) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_moe.py (367:372) duplicated block id: 8516 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (673:678) - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (703:708) duplicated block id: 8517 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/mistral/transformer.py (1097:1102) - megatron_patch/model/qwen1_5_megablocks/transformer.py (1166:1171) duplicated block id: 8518 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/llama/language_model.py (414:425) - megatron_patch/model/llama2/language_model.py (434:445) duplicated block id: 8519 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer/attention.py (362:367) - megatron_patch/model/qwen2_vl/attention.py (678:683) duplicated block id: 8520 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon40b/checkpoint_reshaping_and_interoperability.py (259:275) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (338:354) duplicated block id: 8521 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/language_model.py (362:367) - megatron_patch/model/starcoder/language_model.py (352:357) duplicated block id: 8522 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/qwen2/transformer_block.py (222:227) - megatron_patch/model/qwen2/transformer_block.py (251:256) duplicated block id: 8523 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/checkpoint_reshaping_and_interoperability.py (337:343) - toolkits/model_checkpoints_convertor/falcon40b/checkpoint_reshaping_and_interoperability.py (640:646) duplicated block id: 8524 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/qwen2/transformer_block.py (222:227) - megatron_patch/model/qwen2/transformer_block.py (262:267) duplicated block id: 8525 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/mixtral_bak/transformer/mlp.py (135:145) - megatron_patch/model/qwen1_5/transformer/mlp.py (179:189) duplicated block id: 8526 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (673:678) - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (442:447) duplicated block id: 8527 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/qwen2/transformer_block.py (222:227) - megatron_patch/model/qwen2/transformer_block.py (300:305) duplicated block id: 8528 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (391:397) - megatron_patch/model/falcon/transformer.py (409:415) duplicated block id: 8529 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/checkpoint_reshaping_and_interoperability.py (337:343) - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (807:813) duplicated block id: 8530 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (970:975) - megatron_patch/model/baichuan/transformer.py (1149:1154) duplicated block id: 8531 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/language_model.py (408:413) - megatron_patch/model/bloom/language_model.py (368:373) duplicated block id: 8532 size: 6 cleaned lines of code in 2 files: - megatron_patch/arguments.py (418:423) - toolkits/model_checkpoints_convertor/galactica/checkpoint_reshaping_and_interoperability.py (36:41) duplicated block id: 8533 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer/mlp.py (35:58) - megatron_patch/model/mixtral_bak/transformer/mlp.py (19:42) duplicated block id: 8534 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (442:447) - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (311:316) duplicated block id: 8535 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/language_model.py (49:54) - megatron_patch/model/llava/language_model.py (50:55) duplicated block id: 8536 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/hf2te.py (286:291) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (837:842) duplicated block id: 8537 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (299:304) - toolkits/model_checkpoints_convertor/yi/checkpoint_reshaping_and_interoperability.py (271:276) duplicated block id: 8538 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/gpt_model.py (134:141) - megatron_patch/model/glm130b/gpt_model.py (100:107) duplicated block id: 8539 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (299:304) - toolkits/model_checkpoints_convertor/galactica/checkpoint_reshaping_and_interoperability.py (273:278) duplicated block id: 8540 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (814:825) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (943:954) duplicated block id: 8541 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/deepseek/hf2mcore_deepseek_v3_moe.py (23:34) - toolkits/model_checkpoints_convertor/qwen/hf2megablocks_qwen1.5.py (8:22) duplicated block id: 8542 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (402:411) - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (299:308) duplicated block id: 8543 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/reward_model_to_megatron.py (337:343) - toolkits/model_checkpoints_convertor/falcon40b/checkpoint_reshaping_and_interoperability.py (640:646) duplicated block id: 8544 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/transformer.py (857:862) - megatron_patch/model/starcoder/transformer.py (1174:1179) duplicated block id: 8545 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2megablocks_qwen1.5.py (253:258) - toolkits/model_checkpoints_convertor/qwen/hf2megablocks_qwen1.5.py (512:517) duplicated block id: 8546 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (1123:1128) - toolkits/model_checkpoints_convertor/yi/checkpoint_reshaping_and_interoperability.py (669:674) duplicated block id: 8547 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/mlp.py (72:77) - megatron_patch/model/qwen2_vl/attention.py (122:127) duplicated block id: 8548 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/transformer.py (406:413) - megatron_patch/model/glm130b/transformer.py (438:445) duplicated block id: 8549 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/model.py (110:115) - megatron_patch/model/llama3/model.py (98:103) duplicated block id: 8550 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/llama3/language_model.py (389:394) - megatron_patch/model/llava/language_model.py (414:419) duplicated block id: 8551 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/hf2te.py (187:192) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (341:346) duplicated block id: 8552 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/falcon/language_model.py (356:361) - megatron_patch/model/mistral/language_model.py (353:358) duplicated block id: 8553 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer_legacy.py (685:690) - megatron_patch/model/mixtral/transformer/attention.py (619:624) duplicated block id: 8554 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer_legacy.py (1240:1245) - megatron_patch/model/qwen/transformer.py (1045:1050) duplicated block id: 8555 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/qwen/transformer.py (1156:1163) - megatron_patch/model/qwen1_5_megablocks/transformer.py (1096:1103) duplicated block id: 8556 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (373:378) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (603:608) duplicated block id: 8557 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/transformer.py (312:322) - megatron_patch/model/starcoder/transformer.py (463:473) duplicated block id: 8558 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/falcon/transformer.py (1074:1079) - megatron_patch/model/llava/transformer.py (1408:1413) duplicated block id: 8559 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (1123:1128) - toolkits/model_checkpoints_convertor/starcoder/checkpoint_reshaping_and_interoperability.py (827:832) duplicated block id: 8560 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (38:46) - megatron_patch/model/chatglm/transformer.py (37:45) duplicated block id: 8561 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer/attention.py (384:389) - megatron_patch/model/qwen2/transformer/attention.py (381:386) duplicated block id: 8562 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (599:604) - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (442:447) duplicated block id: 8563 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/reward_model_to_megatron.py (378:383) - toolkits/model_checkpoints_convertor/glm/checkpoint_reshaping_and_interoperability.py (358:363) duplicated block id: 8564 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/deepseek/hf2mcore_deepseek_v2_moe.py (395:400) - toolkits/model_checkpoints_convertor/deepseek/hf2mcore_deepseek_v2_moe.py (443:448) duplicated block id: 8565 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/language_model.py (623:630) - megatron_patch/model/qwen_vl/language_model.py (674:680) duplicated block id: 8566 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/hf2te.py (261:270) - toolkits/model_checkpoints_convertor/falcon/checkpoint_reshaping_and_interoperability.py (395:404) duplicated block id: 8567 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/bloom/language_model.py (368:373) - megatron_patch/model/llama2/language_model.py (352:357) duplicated block id: 8568 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/transformer.py (906:912) - megatron_patch/model/llama2/transformer.py (1440:1446) duplicated block id: 8569 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (691:697) - toolkits/model_checkpoints_convertor/falcon/checkpoint_reshaping_and_interoperability.py (637:643) duplicated block id: 8570 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/galactica/transformer.py (542:550) - megatron_patch/model/llama2/transformer.py (886:893) duplicated block id: 8571 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/hf2te.py (261:270) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (463:472) duplicated block id: 8572 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/language_model.py (367:372) - megatron_patch/model/llama3/language_model.py (347:352) duplicated block id: 8573 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/mistral/hf2mcore.py (8:19) - toolkits/model_checkpoints_convertor/qwen/hf2megablocks_qwen1.5.py (8:22) duplicated block id: 8574 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/moe/experts.py (273:278) - megatron_patch/model/deepseek_v2/moe/experts.py (286:291) duplicated block id: 8575 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/language_model.py (659:665) - megatron_patch/model/qwen1_5_megablocks/language_model.py (644:650) duplicated block id: 8576 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/galactica/checkpoint_reshaping_and_interoperability.py (441:446) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (290:295) duplicated block id: 8577 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (1102:1107) - megatron_patch/model/baichuan2/transformer.py (1282:1287) duplicated block id: 8578 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (603:608) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (751:756) duplicated block id: 8579 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (60:66) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_mha.py (52:58) duplicated block id: 8580 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/deepseek/hf2mcore_deepseek_v2_moe.py (22:34) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_moe.py (7:20) duplicated block id: 8581 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/moe/experts.py (273:278) - megatron_patch/model/deepseek_v2/moe/experts.py (301:306) duplicated block id: 8582 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (1282:1287) - megatron_patch/model/qwen/transformer.py (1045:1050) duplicated block id: 8583 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (464:469) - megatron_patch/model/glm130b/transformer.py (506:511) duplicated block id: 8584 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/llava/transformer.py (706:711) - megatron_patch/model/qwen2_vl/attention_vision.py (619:624) duplicated block id: 8585 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (1357:1362) - megatron_patch/model/llama2/transformer.py (1490:1495) duplicated block id: 8586 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/qwen2/transformer/attention.py (533:538) - megatron_patch/model/qwen2_vl/attention_vision.py (496:501) duplicated block id: 8587 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/mistral/hf2mcore.py (8:19) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_moe.py (7:20) duplicated block id: 8588 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/transformer.py (813:818) - megatron_patch/model/llama3/transformer_legacy.py (1367:1372) duplicated block id: 8589 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/deepseek/hf2mcore_deepseek_v2_moe.py (243:251) - toolkits/model_checkpoints_convertor/deepseek/hf2mcore_deepseek_v3_moe.py (296:303) duplicated block id: 8590 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/transformer_block.py (239:244) - megatron_patch/model/qwen2/transformer_block.py (168:173) duplicated block id: 8591 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (284:289) - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (286:291) duplicated block id: 8592 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (603:608) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (490:495) duplicated block id: 8593 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (286:291) - toolkits/model_checkpoints_convertor/galactica/checkpoint_reshaping_and_interoperability.py (273:278) duplicated block id: 8594 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (374:379) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (502:507) duplicated block id: 8595 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/falcon/language_model.py (414:425) - megatron_patch/model/qwen1_5_megablocks/language_model.py (433:444) duplicated block id: 8596 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/llama2/transformer.py (1282:1287) - megatron_patch/model/qwen1_5_megablocks/transformer.py (984:989) duplicated block id: 8597 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/language_model.py (456:461) - megatron_patch/model/qwen_vl/language_model.py (406:411) duplicated block id: 8598 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/chatglm/checkpoint_reshaping_and_interoperability.py (367:372) - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (703:708) duplicated block id: 8599 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/deepseek/hf2mcore_deepseek_v2_moe.py (74:81) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (70:77) duplicated block id: 8600 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/llama/language_model.py (414:425) - megatron_patch/model/qwen/language_model.py (431:442) duplicated block id: 8601 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/llama2/transformer.py (767:772) - megatron_patch/model/mistral/transformer.py (765:770) duplicated block id: 8602 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (183:189) - megatron_patch/model/llava/transformer.py (339:345) duplicated block id: 8603 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/transformer.py (695:700) - megatron_patch/model/llama2/transformer.py (1409:1414) duplicated block id: 8604 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/qwen2_5_vl/transformer_block.py (325:330) - megatron_patch/model/qwen2_5_vl/transformer_block.py (563:568) duplicated block id: 8605 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (691:697) - toolkits/model_checkpoints_convertor/falcon40b/checkpoint_reshaping_and_interoperability.py (640:646) duplicated block id: 8606 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (633:638) - megatron_patch/model/llama2/transformer.py (642:647) duplicated block id: 8607 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/mixtral_bak/transformer/attention.py (366:371) - megatron_patch/model/qwen_vl/transformer.py (706:711) duplicated block id: 8608 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/reward_model_to_megatron.py (623:628) - toolkits/model_checkpoints_convertor/falcon40b/checkpoint_reshaping_and_interoperability.py (682:687) duplicated block id: 8609 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/falcon/language_model.py (356:361) - megatron_patch/model/llava/language_model.py (355:360) duplicated block id: 8610 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_moe.py (106:111) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_dense_and_moe_gqa.py (109:114) duplicated block id: 8611 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/chatglm/checkpoint_reshaping_and_interoperability.py (367:372) - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (442:447) duplicated block id: 8612 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore.py (500:505) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_moe.py (328:333) duplicated block id: 8613 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/transformer.py (781:791) - megatron_patch/model/starcoder/transformer.py (1088:1098) duplicated block id: 8614 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (360:365) - toolkits/model_checkpoints_convertor/yi/checkpoint_reshaping_and_interoperability.py (435:440) duplicated block id: 8615 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (1161:1169) - megatron_patch/model/bloom/transformer.py (860:868) duplicated block id: 8616 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (624:634) - megatron_patch/model/falcon40b/transformer.py (661:671) duplicated block id: 8617 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (614:620) - toolkits/model_checkpoints_convertor/falcon40b/checkpoint_reshaping_and_interoperability.py (560:566) duplicated block id: 8618 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/llama2/transformer.py (1100:1105) - megatron_patch/model/llava/transformer.py (1281:1286) duplicated block id: 8619 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon/checkpoint_reshaping_and_interoperability.py (422:427) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (751:756) duplicated block id: 8620 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (78:84) - megatron_patch/model/starcoder/transformer.py (83:90) duplicated block id: 8621 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (1282:1287) - megatron_patch/model/llama3/transformer_legacy.py (1050:1055) duplicated block id: 8622 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/checkpoint_reshaping_and_interoperability.py (624:629) - toolkits/model_checkpoints_convertor/starcoder/checkpoint_reshaping_and_interoperability.py (401:406) duplicated block id: 8623 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (916:921) - toolkits/model_checkpoints_convertor/galactica/checkpoint_reshaping_and_interoperability.py (653:658) duplicated block id: 8624 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore_llama3_1.py (462:468) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_dense_and_moe_gqa.py (865:871) duplicated block id: 8625 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore.py (500:505) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_mha_to_moe.py (179:184) duplicated block id: 8626 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (1282:1287) - megatron_patch/model/qwen1_5_megablocks/transformer.py (984:989) duplicated block id: 8627 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/falcon/language_model.py (356:361) - megatron_patch/model/llama2/language_model.py (352:357) duplicated block id: 8628 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (751:756) - toolkits/model_checkpoints_convertor/yi/checkpoint_reshaping_and_interoperability.py (435:440) duplicated block id: 8629 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (121:126) - megatron_patch/model/chatglm/transformer.py (81:86) duplicated block id: 8630 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/transformer/attention.py (619:624) - megatron_patch/model/qwen_vl/transformer.py (706:711) duplicated block id: 8631 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/mlp.py (26:49) - megatron_patch/model/mixtral_bak/transformer/mlp.py (19:42) duplicated block id: 8632 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2.5_vl.py (605:610) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_dense_and_moe_gqa.py (605:610) duplicated block id: 8633 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon/checkpoint_reshaping_and_interoperability.py (697:702) - toolkits/model_checkpoints_convertor/glm130b/checkpoint_reshaping_and_interoperability.py (353:358) duplicated block id: 8634 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/language_model.py (441:452) - megatron_patch/model/falcon40b/language_model.py (414:425) duplicated block id: 8635 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/gpt_model.py (102:109) - megatron_patch/model/galactica/gpt_model.py (107:114) duplicated block id: 8636 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/galactica/language_model.py (387:392) - megatron_patch/model/qwen_vl/language_model.py (353:358) duplicated block id: 8637 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (1071:1076) - megatron_patch/model/llava/transformer.py (1739:1744) duplicated block id: 8638 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/llama2/transformer.py (707:712) - megatron_patch/model/qwen2_vl/attention_vision.py (619:624) duplicated block id: 8639 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/multi_latent_attention.py (202:207) - megatron_patch/model/qwen1_5/transformer/attention.py (528:533) duplicated block id: 8640 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/moe/experts.py (220:225) - megatron_patch/model/mixtral/moe/experts.py (704:709) duplicated block id: 8641 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2.5_vl.py (221:226) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_dense_and_moe_gqa.py (426:431) duplicated block id: 8642 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/llama/transformer.py (192:198) - megatron_patch/model/llava/transformer.py (339:345) duplicated block id: 8643 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/deepseek/hf2mcore_deepseek_v3_moe.py (222:227) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_moe.py (593:598) duplicated block id: 8644 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer_legacy.py (1172:1179) - megatron_patch/model/qwen1_5_megablocks/transformer.py (1096:1103) duplicated block id: 8645 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (520:525) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (837:842) duplicated block id: 8646 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon40b/checkpoint_reshaping_and_interoperability.py (124:131) - toolkits/model_checkpoints_convertor/qwen/hf2megablocks_qwen1.5.py (123:131) duplicated block id: 8647 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/transformer.py (437:443) - megatron_patch/model/glm130b/transformer.py (485:492) duplicated block id: 8648 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/model.py (92:97) - megatron_patch/model/qwen2/model.py (99:104) duplicated block id: 8649 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/hf2te.py (317:322) - toolkits/model_checkpoints_convertor/chatglm/checkpoint_reshaping_and_interoperability.py (431:436) duplicated block id: 8650 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/hf2te.py (171:176) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (318:323) duplicated block id: 8651 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (988:996) - megatron_patch/model/bloom/transformer.py (737:746) duplicated block id: 8652 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/llama2/transformer.py (1100:1105) - megatron_patch/model/llama3/transformer_legacy.py (1240:1245) duplicated block id: 8653 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer_legacy.py (1050:1055) - megatron_patch/model/qwen/transformer.py (1224:1229) duplicated block id: 8654 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore.py (344:352) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_mha.py (135:143) duplicated block id: 8655 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/transformer.py (191:196) - megatron_patch/model/llama/transformer.py (166:171) duplicated block id: 8656 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (916:921) - toolkits/model_checkpoints_convertor/yi/checkpoint_reshaping_and_interoperability.py (669:674) duplicated block id: 8657 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/bloom/language_model.py (368:373) - megatron_patch/model/qwen/language_model.py (352:357) duplicated block id: 8658 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/bloom/positional_embeddings.py (71:78) - megatron_patch/model/chatglm/positional_embeddings.py (75:82) duplicated block id: 8659 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/hf2te.py (314:319) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (751:756) duplicated block id: 8660 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (123:130) - toolkits/model_checkpoints_convertor/mistral/hf2mcore.py (97:105) duplicated block id: 8661 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/checkpoint_reshaping_and_interoperability.py (379:384) - toolkits/model_checkpoints_convertor/bloom/checkpoint_reshaping_and_interoperability.py (624:629) duplicated block id: 8662 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/llava_mcore/vision/vit_layer_specs.py (76:81) - megatron_patch/model/qwen2_moe/layer_specs.py (228:233) duplicated block id: 8663 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/transformer.py (504:510) - megatron_patch/model/qwen_vl/transformer.py (639:645) duplicated block id: 8664 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/llava_mcore/llava_spec.py (40:45) - megatron_patch/model/qwen2_vl/layer_specs.py (76:81) duplicated block id: 8665 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (884:891) - megatron_patch/model/glm130b/transformer.py (735:741) duplicated block id: 8666 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (299:308) - toolkits/model_checkpoints_convertor/yi/checkpoint_reshaping_and_interoperability.py (405:414) duplicated block id: 8667 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (1282:1287) - megatron_patch/model/llava/transformer.py (1099:1104) duplicated block id: 8668 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon40b/checkpoint_reshaping_and_interoperability.py (542:550) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (466:474) duplicated block id: 8669 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_dense_and_moe_gqa.py (741:746) - toolkits/model_checkpoints_convertor/qwen/hf2megablocks_qwen1.5.py (581:586) duplicated block id: 8670 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (522:528) - megatron_patch/model/glm130b/transformer.py (504:510) duplicated block id: 8671 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (290:295) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (502:507) duplicated block id: 8672 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/llama/language_model.py (356:361) - megatron_patch/model/llava/language_model.py (355:360) duplicated block id: 8673 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer/attention.py (384:389) - megatron_patch/model/mixtral/transformer/attention.py (520:525) duplicated block id: 8674 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/moe/router.py (10:17) - megatron_patch/model/qwen2/moe/router.py (26:33) duplicated block id: 8675 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon40b/checkpoint_reshaping_and_interoperability.py (640:646) - toolkits/model_checkpoints_convertor/starcoder/checkpoint_reshaping_and_interoperability.py (361:367) duplicated block id: 8676 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/language_model.py (659:665) - megatron_patch/model/llama2/language_model.py (645:651) duplicated block id: 8677 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/deepseek/hf2mcore_deepseek_v2_moe.py (123:128) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_moe.py (113:118) duplicated block id: 8678 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (673:678) - toolkits/model_checkpoints_convertor/chatglm/checkpoint_reshaping_and_interoperability.py (367:372) duplicated block id: 8679 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/llama/transformer.py (868:873) - megatron_patch/model/qwen1_5_megablocks/transformer.py (1294:1299) duplicated block id: 8680 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llava/hf2mcore_llava.py (530:535) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_moe.py (380:385) duplicated block id: 8681 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/multi_latent_attention.py (101:106) - megatron_patch/model/mixtral/transformer/mlp.py (96:101) duplicated block id: 8682 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (849:856) - megatron_patch/model/bloom/transformer.py (613:621) duplicated block id: 8683 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/mixtral_bak/transformer/attention.py (81:86) - megatron_patch/model/qwen1_5/transformer/mlp.py (110:115) duplicated block id: 8684 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/language_model.py (414:425) - megatron_patch/model/llama3/language_model.py (425:436) duplicated block id: 8685 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/bloom/gpt_model.py (89:96) - megatron_patch/model/llama/gpt_model.py (93:100) duplicated block id: 8686 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/multi_latent_attention.py (101:106) - megatron_patch/model/qwen2/transformer/mlp.py (109:114) duplicated block id: 8687 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/transformer/attention.py (707:712) - megatron_patch/model/mixtral_bak/transformer/attention.py (348:353) duplicated block id: 8688 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/llama/transformer.py (912:917) - megatron_patch/model/starcoder/transformer.py (1174:1179) duplicated block id: 8689 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/language_model.py (414:425) - megatron_patch/model/qwen/language_model.py (431:442) duplicated block id: 8690 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/llava/transformer.py (1308:1321) - megatron_patch/model/starcoder/transformer.py (1008:1021) duplicated block id: 8691 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/hf2te.py (261:270) - toolkits/model_checkpoints_convertor/galactica/checkpoint_reshaping_and_interoperability.py (432:441) duplicated block id: 8692 size: 6 cleaned lines of code in 2 files: - megatron_patch/arguments.py (418:423) - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (57:62) duplicated block id: 8693 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_mha.py (243:249) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_vl.py (588:594) duplicated block id: 8694 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (613:621) - megatron_patch/model/qwen1_5_megablocks/transformer.py (777:784) duplicated block id: 8695 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/checkpoint_reshaping_and_interoperability.py (337:343) - toolkits/model_checkpoints_convertor/falcon/checkpoint_reshaping_and_interoperability.py (637:643) duplicated block id: 8696 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/mixtral_bak/moe/router.py (186:201) - megatron_patch/model/qwen1_5/moe/router.py (193:208) duplicated block id: 8697 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (613:621) - megatron_patch/model/llama/transformer.py (637:645) duplicated block id: 8698 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (981:987) - megatron_patch/model/llava/transformer.py (1439:1445) duplicated block id: 8699 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (691:697) - toolkits/model_checkpoints_convertor/starcoder/checkpoint_reshaping_and_interoperability.py (361:367) duplicated block id: 8700 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (539:548) - megatron_patch/model/llama/transformer.py (536:545) duplicated block id: 8701 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (442:447) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (603:608) duplicated block id: 8702 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/transformer.py (717:724) - megatron_patch/model/llama3/transformer_legacy.py (856:863) duplicated block id: 8703 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/gpt_model.py (102:109) - megatron_patch/model/falcon40b/gpt_model.py (107:114) duplicated block id: 8704 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (1139:1146) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (1141:1148) duplicated block id: 8705 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/qwen_vl/transformer.py (83:89) - megatron_patch/model/starcoder/transformer.py (83:90) duplicated block id: 8706 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/model.py (110:115) - megatron_patch/model/qwen2/model.py (99:104) duplicated block id: 8707 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/transformer.py (645:653) - megatron_patch/model/llava/transformer.py (885:892) duplicated block id: 8708 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/language_model.py (408:413) - megatron_patch/model/chatglm/language_model.py (367:372) duplicated block id: 8709 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (211:219) - megatron_patch/model/chatglm/transformer.py (163:171) duplicated block id: 8710 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (286:291) - toolkits/model_checkpoints_convertor/yi/checkpoint_reshaping_and_interoperability.py (271:276) duplicated block id: 8711 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/qwen1_5_megablocks/transformer.py (984:989) - megatron_patch/model/qwen_vl/transformer.py (1281:1286) duplicated block id: 8712 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/mistral/transformer.py (1279:1284) - megatron_patch/model/qwen/transformer.py (1045:1050) duplicated block id: 8713 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_dense_and_moe_gqa.py (123:128) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_vl.py (125:130) duplicated block id: 8714 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/galactica/language_model.py (387:392) - megatron_patch/model/llava/language_model.py (355:360) duplicated block id: 8715 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (633:638) - megatron_patch/model/qwen/transformer.py (615:620) duplicated block id: 8716 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/moe/moe_layer.py (157:164) - megatron_patch/model/qwen3_moe/moe/moe_layer.py (89:96) duplicated block id: 8717 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/galactica/transformer.py (702:707) - megatron_patch/model/llava/transformer.py (1408:1413) duplicated block id: 8718 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/qwen1_5/transformer/attention.py (384:389) - megatron_patch/model/qwen2_vl/attention_vision.py (520:525) duplicated block id: 8719 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/qwen2/transformer_block.py (168:173) - megatron_patch/model/qwen2_5_vl/transformer_block.py (278:283) duplicated block id: 8720 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/qwen2/transformer_block.py (306:320) - megatron_patch/model/qwen2_5_vl/transformer_block.py (395:409) duplicated block id: 8721 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/transformer.py (729:735) - megatron_patch/model/glm130b/transformer.py (924:930) duplicated block id: 8722 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/falcon/transformer.py (211:217) - megatron_patch/model/llama2/transformer.py (339:345) duplicated block id: 8723 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (786:795) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (1091:1100) duplicated block id: 8724 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (588:593) - megatron_patch/model/qwen2_vl/attention.py (620:625) duplicated block id: 8725 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/qwen2/transformer/attention.py (478:483) - megatron_patch/model/qwen_vl/transformer.py (706:711) duplicated block id: 8726 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/language_model.py (370:375) - megatron_patch/model/qwen/language_model.py (352:357) duplicated block id: 8727 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/mixtral_bak/transformer/attention.py (288:293) - megatron_patch/model/qwen2_vl/attention.py (447:452) duplicated block id: 8728 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/language_model.py (659:665) - megatron_patch/model/mistral/language_model.py (657:663) duplicated block id: 8729 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/qwen1_5_megablocks/language_model.py (644:650) - megatron_patch/model/starcoder/language_model.py (563:569) duplicated block id: 8730 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/transformer.py (510:518) - megatron_patch/model/qwen/transformer.py (831:838) duplicated block id: 8731 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer_legacy.py (222:229) - megatron_patch/model/mixtral_bak/moe/token_dispatcher.py (106:113) duplicated block id: 8732 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (738:745) - megatron_patch/model/chatglm/transformer.py (510:518) duplicated block id: 8733 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/moe/moe_layer.py (57:62) - megatron_patch/model/qwen1_5/moe/moe_layer.py (48:53) duplicated block id: 8734 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/transformer.py (406:413) - megatron_patch/model/starcoder/transformer.py (537:544) duplicated block id: 8735 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/mistral/language_model.py (657:663) - megatron_patch/model/starcoder/language_model.py (563:569) duplicated block id: 8736 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (464:469) - megatron_patch/model/galactica/transformer.py (418:423) duplicated block id: 8737 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/language_model.py (197:203) - megatron_patch/model/llama3/language_model.py (173:179) duplicated block id: 8738 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/reward_model_to_megatron.py (490:497) - toolkits/model_checkpoints_convertor/falcon/checkpoint_reshaping_and_interoperability.py (798:805) duplicated block id: 8739 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/llama/transformer.py (868:873) - megatron_patch/model/llama3/transformer_legacy.py (1367:1372) duplicated block id: 8740 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/transformer.py (108:113) - megatron_patch/model/glm130b/transformer.py (191:196) duplicated block id: 8741 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/hf2te.py (314:319) - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (599:604) duplicated block id: 8742 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore_llama3_1.py (724:729) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_dense_and_moe_gqa.py (352:357) duplicated block id: 8743 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer/attention.py (391:396) - megatron_patch/model/qwen2_vl/attention_vision.py (530:535) duplicated block id: 8744 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/llama2/transformer.py (1282:1287) - megatron_patch/model/llava/transformer.py (1099:1104) duplicated block id: 8745 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/falcon/transformer.py (583:589) - megatron_patch/model/starcoder/transformer.py (537:544) duplicated block id: 8746 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon/checkpoint_reshaping_and_interoperability.py (539:546) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (466:474) duplicated block id: 8747 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/transformer.py (134:140) - megatron_patch/model/qwen_vl/transformer.py (339:345) duplicated block id: 8748 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/transformer.py (924:930) - megatron_patch/model/llama/transformer.py (899:905) duplicated block id: 8749 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/qwen2/moe/moe_layer.py (120:125) - megatron_patch/model/qwen3_moe/moe/moe_layer.py (50:55) duplicated block id: 8750 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/chatglm/checkpoint_reshaping_and_interoperability.py (579:586) - toolkits/model_checkpoints_convertor/glm/checkpoint_reshaping_and_interoperability.py (565:572) duplicated block id: 8751 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/galactica/transformer.py (733:739) - megatron_patch/model/glm130b/transformer.py (924:930) duplicated block id: 8752 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer_legacy.py (1700:1705) - megatron_patch/model/llava/transformer.py (1200:1205) duplicated block id: 8753 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (700:705) - toolkits/model_checkpoints_convertor/chatglm/checkpoint_reshaping_and_interoperability.py (367:372) duplicated block id: 8754 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (1610:1615) - megatron_patch/model/llama2/transformer.py (1201:1206) duplicated block id: 8755 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/bloom/language_model.py (593:599) - megatron_patch/model/mistral/language_model.py (657:663) duplicated block id: 8756 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/transformer/mlp.py (96:101) - megatron_patch/model/qwen2_vl/attention.py (122:127) duplicated block id: 8757 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/bloom/language_model.py (464:469) - megatron_patch/model/llama/language_model.py (542:547) duplicated block id: 8758 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/llama/transformer.py (192:198) - megatron_patch/model/llama2/transformer.py (339:345) duplicated block id: 8759 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon40b/checkpoint_reshaping_and_interoperability.py (393:398) - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (311:316) duplicated block id: 8760 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (737:746) - megatron_patch/model/starcoder/transformer.py (1049:1057) duplicated block id: 8761 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_vl.py (221:226) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_vl.py (328:333) duplicated block id: 8762 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/hf2te.py (172:177) - toolkits/model_checkpoints_convertor/chatglm/checkpoint_reshaping_and_interoperability.py (256:261) duplicated block id: 8763 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (1740:1745) - megatron_patch/model/qwen_vl/transformer.py (1200:1205) duplicated block id: 8764 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/qwen1_5_megablocks/transformer.py (1166:1171) - megatron_patch/model/qwen_vl/transformer.py (1099:1104) duplicated block id: 8765 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (1249:1254) - megatron_patch/model/starcoder/transformer.py (962:967) duplicated block id: 8766 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (211:219) - megatron_patch/model/galactica/transformer.py (177:185) duplicated block id: 8767 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/mistral/transformer.py (1097:1102) - megatron_patch/model/mistral/transformer.py (1279:1284) duplicated block id: 8768 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer/attention.py (107:112) - megatron_patch/model/qwen2/transformer/mlp.py (109:114) duplicated block id: 8769 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/language_model.py (408:413) - megatron_patch/model/llama/language_model.py (356:361) duplicated block id: 8770 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/bloom/language_model.py (593:599) - megatron_patch/model/qwen/language_model.py (627:633) duplicated block id: 8771 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (816:821) - toolkits/model_checkpoints_convertor/yi/checkpoint_reshaping_and_interoperability.py (435:440) duplicated block id: 8772 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/multi_latent_attention.py (202:207) - megatron_patch/model/qwen2/transformer/attention.py (533:538) duplicated block id: 8773 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/transformer.py (181:188) - megatron_patch/model/llava/transformer.py (161:168) duplicated block id: 8774 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (703:708) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (490:495) duplicated block id: 8775 size: 6 cleaned lines of code in 2 files: - megatron_patch/tokenizer/tokenization_baichuan.py (65:70) - megatron_patch/tokenizer/tokenization_yi.py (67:72) duplicated block id: 8776 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/moe/experts.py (704:709) - megatron_patch/model/mixtral/moe/experts.py (798:803) duplicated block id: 8777 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/llama/transformer.py (192:198) - megatron_patch/model/llama3/transformer_legacy.py (338:344) duplicated block id: 8778 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/galactica/transformer.py (702:707) - megatron_patch/model/qwen1_5_megablocks/transformer.py (1294:1299) duplicated block id: 8779 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/galactica/transformer.py (355:362) - megatron_patch/model/starcoder/transformer.py (537:544) duplicated block id: 8780 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (336:342) - megatron_patch/model/galactica/transformer.py (154:160) duplicated block id: 8781 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/mixtral_bak/transformer/attention.py (81:86) - megatron_patch/model/qwen2/transformer/mlp.py (109:114) duplicated block id: 8782 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/gpt_model.py (102:109) - megatron_patch/model/qwen1_5_megablocks/gpt_model.py (108:115) duplicated block id: 8783 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/transformer.py (510:518) - megatron_patch/model/qwen1_5_megablocks/transformer.py (777:784) duplicated block id: 8784 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (229:234) - megatron_patch/model/galactica/transformer.py (193:198) duplicated block id: 8785 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/language_model.py (654:659) - megatron_patch/model/falcon/language_model.py (680:685) duplicated block id: 8786 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/transformer.py (217:223) - megatron_patch/model/qwen/transformer.py (333:339) duplicated block id: 8787 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer_legacy.py (1050:1055) - megatron_patch/model/llama3/transformer_legacy.py (1240:1245) duplicated block id: 8788 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/language_model.py (460:468) - megatron_patch/model/starcoder/language_model.py (378:386) duplicated block id: 8789 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/galactica/transformer.py (154:160) - megatron_patch/model/llama3/transformer_legacy.py (338:344) duplicated block id: 8790 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/transformer/attention.py (135:140) - megatron_patch/model/qwen2/transformer/mlp.py (109:114) duplicated block id: 8791 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (885:891) - megatron_patch/model/starcoder/transformer.py (1021:1027) duplicated block id: 8792 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/hf2te.py (187:192) - toolkits/model_checkpoints_convertor/yi/checkpoint_reshaping_and_interoperability.py (271:276) duplicated block id: 8793 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/mixtral_bak/layer_specs.py (56:61) - megatron_patch/model/qwen2_vl/layer_specs.py (76:81) duplicated block id: 8794 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/mistral/language_model.py (404:412) - megatron_patch/model/starcoder/language_model.py (378:386) duplicated block id: 8795 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/deepseek/hf2mcore_deepseek_v3_moe.py (45:50) - toolkits/model_checkpoints_convertor/llava/hf2mcore_llava.py (32:37) duplicated block id: 8796 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/reward_model_to_megatron.py (817:822) - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (1123:1128) duplicated block id: 8797 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/deepseek/hf2mcore_deepseek_v2_moe.py (443:448) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_moe.py (330:335) duplicated block id: 8798 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (1203:1208) - megatron_patch/model/llama2/transformer.py (1740:1745) duplicated block id: 8799 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/deepseek/hf2mcore_deepseek_v3_moe.py (23:34) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_mha.py (6:18) duplicated block id: 8800 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (183:189) - megatron_patch/model/llama3/transformer_legacy.py (338:344) duplicated block id: 8801 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/llava/transformer.py (706:711) - megatron_patch/model/qwen2_vl/attention.py (620:625) duplicated block id: 8802 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/glm130b/checkpoint_reshaping_and_interoperability.py (353:358) - toolkits/model_checkpoints_convertor/starcoder/checkpoint_reshaping_and_interoperability.py (401:406) duplicated block id: 8803 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/qwen/language_model.py (399:407) - megatron_patch/model/starcoder/language_model.py (378:386) duplicated block id: 8804 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/llama/transformer.py (868:873) - megatron_patch/model/qwen/transformer.py (1351:1356) duplicated block id: 8805 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/llama2/transformer.py (707:712) - megatron_patch/model/mixtral_bak/transformer/attention.py (366:371) duplicated block id: 8806 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer_legacy.py (1050:1055) - megatron_patch/model/llava/transformer.py (1281:1286) duplicated block id: 8807 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/mixtral_bak/transformer/mlp.py (72:77) - megatron_patch/model/qwen2_vl/attention.py (122:127) duplicated block id: 8808 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon40b/checkpoint_reshaping_and_interoperability.py (640:646) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (828:834) duplicated block id: 8809 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/language_model.py (638:644) - megatron_patch/model/chatglm/language_model.py (659:665) duplicated block id: 8810 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (673:678) - toolkits/model_checkpoints_convertor/baichuan2/hf2te.py (286:291) duplicated block id: 8811 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/language_model.py (227:235) - megatron_patch/model/glm130b/language_model.py (228:236) duplicated block id: 8812 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (613:621) - megatron_patch/model/llava/transformer.py (885:892) duplicated block id: 8813 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/mistral/transformer.py (706:711) - megatron_patch/model/qwen1_5/transformer/attention.py (476:481) duplicated block id: 8814 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/llama/language_model.py (414:425) - megatron_patch/model/qwen1_5_megablocks/language_model.py (433:444) duplicated block id: 8815 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/llama/transformer.py (836:846) - megatron_patch/model/starcoder/transformer.py (1088:1098) duplicated block id: 8816 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/chatglm/checkpoint_reshaping_and_interoperability.py (407:414) - toolkits/model_checkpoints_convertor/yi/checkpoint_reshaping_and_interoperability.py (444:451) duplicated block id: 8817 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (1102:1107) - megatron_patch/model/qwen_vl/transformer.py (1281:1286) duplicated block id: 8818 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/hf2te.py (258:267) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (278:287) duplicated block id: 8819 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/transformer.py (906:912) - megatron_patch/model/llama3/transformer_legacy.py (1398:1404) duplicated block id: 8820 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/checkpoint_reshaping_and_interoperability.py (379:384) - toolkits/model_checkpoints_convertor/bloom/reward_model_to_megatron.py (623:628) duplicated block id: 8821 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (698:703) - megatron_patch/model/qwen2_vl/attention.py (620:625) duplicated block id: 8822 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (703:708) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (751:756) duplicated block id: 8823 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/llava_mcore/vision/vit_layer_specs.py (76:81) - megatron_patch/model/qwen3_moe/gpt_layer_specs.py (249:254) duplicated block id: 8824 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/qwen1_5/transformer/attention.py (362:367) - megatron_patch/model/qwen2_vl/attention_vision.py (677:682) duplicated block id: 8825 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (1357:1362) - megatron_patch/model/mistral/transformer.py (1487:1492) duplicated block id: 8826 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (336:342) - megatron_patch/model/falcon40b/transformer.py (204:210) duplicated block id: 8827 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (442:447) - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (703:708) duplicated block id: 8828 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (430:439) - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (362:371) duplicated block id: 8829 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/mlp.py (72:77) - megatron_patch/model/qwen1_5/transformer/attention.py (107:112) duplicated block id: 8830 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/deepseek/hf2mcore_deepseek_v2_moe.py (117:122) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_moe.py (106:111) duplicated block id: 8831 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (588:593) - megatron_patch/model/mixtral_bak/transformer/attention.py (366:371) duplicated block id: 8832 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/qwen2_vl/attention_vision.py (619:624) - megatron_patch/model/qwen_vl/transformer.py (706:711) duplicated block id: 8833 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (432:437) - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (673:678) duplicated block id: 8834 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/moe/experts.py (798:803) - megatron_patch/model/qwen2/moe/experts.py (324:329) duplicated block id: 8835 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (360:365) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (837:842) duplicated block id: 8836 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/bloom/gpt_model.py (101:108) - megatron_patch/model/llava/gpt_model.py (109:116) duplicated block id: 8837 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (60:66) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_mha_to_moe.py (51:57) duplicated block id: 8838 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/reward_model_to_megatron.py (623:628) - toolkits/model_checkpoints_convertor/falcon/checkpoint_reshaping_and_interoperability.py (697:702) duplicated block id: 8839 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/qwen/language_model.py (395:400) - megatron_patch/model/qwen_vl/language_model.py (406:411) duplicated block id: 8840 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/language_model.py (492:503) - megatron_patch/model/falcon40b/language_model.py (414:425) duplicated block id: 8841 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (607:616) - toolkits/model_checkpoints_convertor/baichuan/hf2te.py (482:490) duplicated block id: 8842 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/galactica/transformer.py (542:550) - megatron_patch/model/mistral/transformer.py (883:890) duplicated block id: 8843 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/language_model.py (49:54) - megatron_patch/model/llama3/language_model.py (36:41) duplicated block id: 8844 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/mistral/transformer.py (1279:1284) - megatron_patch/model/qwen1_5_megablocks/transformer.py (984:989) duplicated block id: 8845 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer/attention.py (476:481) - megatron_patch/model/llama3/transformer_legacy.py (685:690) duplicated block id: 8846 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/transformer.py (33:41) - megatron_patch/model/starcoder/transformer.py (42:50) duplicated block id: 8847 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/qwen/transformer.py (1191:1196) - megatron_patch/model/starcoder/transformer.py (962:967) duplicated block id: 8848 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/transformer/mlp.py (37:60) - megatron_patch/model/mixtral_bak/transformer/mlp.py (19:42) duplicated block id: 8849 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore_llama3_1.py (648:653) - toolkits/model_checkpoints_convertor/llama/hf2mcore_llama3_1.py (724:729) duplicated block id: 8850 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/transformer.py (510:518) - megatron_patch/model/llama/transformer.py (637:645) duplicated block id: 8851 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/transformer/attention.py (135:140) - megatron_patch/model/mixtral/transformer/mlp.py (96:101) duplicated block id: 8852 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/llava/transformer.py (706:711) - megatron_patch/model/qwen1_5/transformer/attention.py (476:481) duplicated block id: 8853 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/model.py (110:115) - megatron_patch/model/llama3_1/model.py (113:118) duplicated block id: 8854 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (360:365) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (520:525) duplicated block id: 8855 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/qwen/transformer.py (1682:1687) - megatron_patch/model/qwen_vl/transformer.py (1200:1205) duplicated block id: 8856 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/language_model.py (441:452) - megatron_patch/model/llama/language_model.py (414:425) duplicated block id: 8857 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/language_model.py (659:665) - megatron_patch/model/llava/language_model.py (705:711) duplicated block id: 8858 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/transformer.py (813:818) - megatron_patch/model/llama2/transformer.py (1409:1414) duplicated block id: 8859 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/qwen/transformer.py (78:84) - megatron_patch/model/starcoder/transformer.py (83:90) duplicated block id: 8860 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/galactica/transformer.py (355:362) - megatron_patch/model/glm130b/transformer.py (438:445) duplicated block id: 8861 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon/checkpoint_reshaping_and_interoperability.py (422:427) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (360:365) duplicated block id: 8862 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/galactica/language_model.py (387:392) - megatron_patch/model/qwen1_5_megablocks/language_model.py (351:356) duplicated block id: 8863 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (38:46) - megatron_patch/model/starcoder/transformer.py (42:50) duplicated block id: 8864 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/galactica/checkpoint_reshaping_and_interoperability.py (598:606) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (466:474) duplicated block id: 8865 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (1139:1146) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (854:861) duplicated block id: 8866 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/mistral/hf2mcore.py (97:105) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (151:158) duplicated block id: 8867 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (635:641) - megatron_patch/model/qwen1_5_megablocks/transformer.py (812:818) duplicated block id: 8868 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (700:705) - toolkits/model_checkpoints_convertor/galactica/checkpoint_reshaping_and_interoperability.py (459:464) duplicated block id: 8869 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/bloom/gpt_model.py (105:111) - megatron_patch/model/starcoder/gpt_model.py (108:114) duplicated block id: 8870 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon/checkpoint_reshaping_and_interoperability.py (422:427) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (603:608) duplicated block id: 8871 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/transformer.py (81:86) - megatron_patch/model/galactica/transformer.py (98:103) duplicated block id: 8872 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/qwen2_vl/rope_utils.py (20:27) - megatron_patch/model/qwen2_vl/rope_utils.py (28:35) duplicated block id: 8873 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/qwen1_5/transformer/attention.py (528:533) - megatron_patch/model/qwen2_vl/attention_vision.py (496:501) duplicated block id: 8874 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (990:997) - megatron_patch/model/qwen/transformer.py (1065:1072) duplicated block id: 8875 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (1409:1414) - megatron_patch/model/falcon40b/transformer.py (813:818) duplicated block id: 8876 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/deepseek/hf2mcore_deepseek_v3_moe.py (663:668) - toolkits/model_checkpoints_convertor/llava/hf2mcore_llava.py (405:410) duplicated block id: 8877 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/llama2/transformer.py (1249:1254) - megatron_patch/model/starcoder/transformer.py (962:967) duplicated block id: 8878 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/qwen1_5/transformer/mlp.py (110:115) - megatron_patch/model/qwen2_vl/attention.py (122:127) duplicated block id: 8879 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore.py (357:362) - toolkits/model_checkpoints_convertor/qwen/hf2megablocks_qwen1.5.py (350:355) duplicated block id: 8880 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (990:997) - megatron_patch/model/mistral/transformer.py (1117:1124) duplicated block id: 8881 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/language_model.py (405:410) - megatron_patch/model/qwen_vl/language_model.py (406:411) duplicated block id: 8882 size: 6 cleaned lines of code in 2 files: - megatron_patch/arguments.py (418:423) - toolkits/model_checkpoints_convertor/falcon40b/checkpoint_reshaping_and_interoperability.py (37:42) duplicated block id: 8883 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/transformer.py (134:140) - megatron_patch/model/qwen/transformer.py (333:339) duplicated block id: 8884 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/mistral/hf2mcore.py (407:412) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_mha_to_moe.py (179:184) duplicated block id: 8885 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/qwen1_5_megablocks/language_model.py (351:356) - megatron_patch/model/starcoder/language_model.py (352:357) duplicated block id: 8886 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/mixtral_bak/transformer/mlp.py (170:177) - megatron_patch/model/qwen2/transformer/mlp.py (216:223) duplicated block id: 8887 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/mixtral_bak/transformer/mlp.py (72:77) - megatron_patch/model/qwen2_vl/attention_vision.py (121:126) duplicated block id: 8888 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/checkpoint_reshaping_and_interoperability.py (491:498) - toolkits/model_checkpoints_convertor/starcoder/checkpoint_reshaping_and_interoperability.py (510:517) duplicated block id: 8889 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/llava/language_model.py (414:419) - megatron_patch/model/qwen/language_model.py (395:400) duplicated block id: 8890 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (242:248) - megatron_patch/model/falcon40b/transformer.py (204:210) duplicated block id: 8891 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/llama3/language_model.py (619:625) - megatron_patch/model/starcoder/language_model.py (563:569) duplicated block id: 8892 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/galactica/transformer.py (542:550) - megatron_patch/model/qwen1_5_megablocks/transformer.py (777:784) duplicated block id: 8893 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/transformer/attention.py (677:682) - megatron_patch/model/qwen2/transformer/attention.py (358:363) duplicated block id: 8894 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/mistral/hf2mcore.py (478:483) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_dense_and_moe_gqa.py (619:624) duplicated block id: 8895 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (588:593) - megatron_patch/model/llama3/transformer/attention.py (476:481) duplicated block id: 8896 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/language_model.py (405:410) - megatron_patch/model/llava/language_model.py (414:419) duplicated block id: 8897 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/transformer.py (1149:1154) - megatron_patch/model/glm130b/transformer.py (1158:1163) duplicated block id: 8898 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/transformer.py (1149:1154) - megatron_patch/model/glm130b/transformer.py (1165:1170) duplicated block id: 8899 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (700:705) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (603:608) duplicated block id: 8900 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/language_model.py (356:361) - megatron_patch/model/llama3/language_model.py (347:352) duplicated block id: 8901 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/galactica/checkpoint_reshaping_and_interoperability.py (653:658) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (1018:1023) duplicated block id: 8902 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/mixtral_bak/transformer/attention.py (348:353) - megatron_patch/model/qwen2_vl/attention.py (708:713) duplicated block id: 8903 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/mixtral_bak/transformer/attention.py (348:353) - megatron_patch/model/qwen2_vl/attention.py (696:701) duplicated block id: 8904 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer/mlp.py (94:99) - megatron_patch/model/qwen2_vl/attention.py (122:127) duplicated block id: 8905 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (700:705) - toolkits/model_checkpoints_convertor/baichuan2/hf2te.py (286:291) duplicated block id: 8906 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (172:178) - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (145:151) duplicated block id: 8907 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (1071:1076) - megatron_patch/model/llava/transformer.py (1200:1205) duplicated block id: 8908 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (948:954) - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (1139:1146) duplicated block id: 8909 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/llava/transformer.py (1281:1286) - megatron_patch/model/qwen/transformer.py (1045:1050) duplicated block id: 8910 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/bloom/language_model.py (199:205) - megatron_patch/model/llama3/language_model.py (173:179) duplicated block id: 8911 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (700:705) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (360:365) duplicated block id: 8912 size: 6 cleaned lines of code in 2 files: - megatron_patch/arguments.py (418:423) - toolkits/model_checkpoints_convertor/chatglm/checkpoint_reshaping_and_interoperability.py (32:37) duplicated block id: 8913 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/mixtral_bak/transformer/attention.py (348:353) - megatron_patch/model/qwen1_5/transformer/attention.py (547:552) duplicated block id: 8914 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/mixtral_bak/transformer/attention.py (348:353) - megatron_patch/model/qwen1_5/transformer/attention.py (559:564) duplicated block id: 8915 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llava/hf2mcore_llava.py (454:459) - toolkits/model_checkpoints_convertor/llava/hf2mcore_llava.py (530:535) duplicated block id: 8916 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer_legacy.py (1152:1157) - megatron_patch/model/qwen1_5_megablocks/transformer.py (1602:1607) duplicated block id: 8917 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/hf2te.py (286:291) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (360:365) duplicated block id: 8918 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/falcon/transformer.py (1074:1079) - megatron_patch/model/qwen_vl/transformer.py (1408:1413) duplicated block id: 8919 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/language_model.py (356:361) - megatron_patch/model/mistral/language_model.py (353:358) duplicated block id: 8920 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer_legacy.py (685:690) - megatron_patch/model/qwen2_vl/attention.py (620:625) duplicated block id: 8921 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer/attention.py (547:552) - megatron_patch/model/mixtral_bak/transformer/attention.py (348:353) duplicated block id: 8922 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (579:588) - toolkits/model_checkpoints_convertor/baichuan2/hf2te.py (454:462) duplicated block id: 8923 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon40b/checkpoint_reshaping_and_interoperability.py (819:824) - toolkits/model_checkpoints_convertor/galactica/checkpoint_reshaping_and_interoperability.py (653:658) duplicated block id: 8924 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/deepseek/hf2mcore_deepseek_v2_moe.py (531:536) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_moe.py (681:686) duplicated block id: 8925 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (673:678) - toolkits/model_checkpoints_convertor/yi/checkpoint_reshaping_and_interoperability.py (435:440) duplicated block id: 8926 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_mha_to_moe.py (239:244) - toolkits/model_checkpoints_convertor/qwen/hf2megablocks_qwen1.5.py (594:599) duplicated block id: 8927 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/transformer.py (890:895) - megatron_patch/model/qwen1_5_megablocks/transformer.py (1294:1299) duplicated block id: 8928 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer_legacy.py (685:690) - megatron_patch/model/mixtral_bak/transformer/attention.py (366:371) duplicated block id: 8929 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/llama/language_model.py (356:361) - megatron_patch/model/qwen1_5_megablocks/language_model.py (351:356) duplicated block id: 8930 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/qwen1_5/transformer/attention.py (476:481) - megatron_patch/model/qwen_vl/transformer.py (706:711) duplicated block id: 8931 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/transformer.py (510:518) - megatron_patch/model/glm130b/transformer.py (698:706) duplicated block id: 8932 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/moe/experts.py (704:709) - megatron_patch/model/qwen2/moe/experts.py (392:397) duplicated block id: 8933 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (307:313) - megatron_patch/model/qwen1_5_megablocks/transformer.py (203:209) duplicated block id: 8934 size: 6 cleaned lines of code in 2 files: - megatron_patch/arguments.py (418:423) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (65:70) duplicated block id: 8935 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_dense_and_moe_gqa.py (119:124) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_dense_and_moe_gqa.py (556:561) duplicated block id: 8936 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/qwen3_moe/gpt_layer_specs.py (339:344) - megatron_patch/model/qwen3_moe/gpt_layer_specs.py (357:362) duplicated block id: 8937 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (284:289) - toolkits/model_checkpoints_convertor/baichuan/hf2te.py (187:192) duplicated block id: 8938 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/reward_model_to_megatron.py (337:343) - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (807:813) duplicated block id: 8939 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llava/hf2mcore_llava.py (454:459) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_vl.py (328:333) duplicated block id: 8940 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (215:224) - megatron_patch/model/starcoder/transformer.py (256:264) duplicated block id: 8941 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/galactica/transformer.py (542:550) - megatron_patch/model/qwen_vl/transformer.py (885:892) duplicated block id: 8942 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (70:77) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_vl.py (78:85) duplicated block id: 8943 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/chatglm/checkpoint_reshaping_and_interoperability.py (528:536) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (466:474) duplicated block id: 8944 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/language_model.py (356:361) - megatron_patch/model/llava/language_model.py (355:360) duplicated block id: 8945 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/moe/moe_layer.py (121:126) - megatron_patch/model/qwen2/moe/moe_layer.py (120:125) duplicated block id: 8946 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/transformer.py (695:700) - megatron_patch/model/qwen1_5_megablocks/transformer.py (1294:1299) duplicated block id: 8947 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/gpt_model.py (100:107) - megatron_patch/model/qwen_vl/gpt_model.py (109:116) duplicated block id: 8948 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/hf2te.py (187:192) - toolkits/model_checkpoints_convertor/falcon/checkpoint_reshaping_and_interoperability.py (257:262) duplicated block id: 8949 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/qwen2_moe/layer_specs.py (228:233) - megatron_patch/model/qwen3_moe/gpt_layer_specs.py (227:232) duplicated block id: 8950 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/galactica/checkpoint_reshaping_and_interoperability.py (459:464) - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (816:821) duplicated block id: 8951 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_dense_and_moe_gqa.py (352:357) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_dense_and_moe_gqa.py (426:431) duplicated block id: 8952 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/qwen3_moe/gpt_layer_specs.py (331:336) - megatron_patch/model/qwen3_moe/gpt_layer_specs.py (349:354) duplicated block id: 8953 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/transformer.py (521:531) - megatron_patch/model/glm130b/transformer.py (724:734) duplicated block id: 8954 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (673:678) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (490:495) duplicated block id: 8955 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/qwen/transformer.py (1224:1229) - megatron_patch/model/qwen_vl/transformer.py (1099:1104) duplicated block id: 8956 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/qwen/transformer.py (1682:1687) - megatron_patch/model/qwen1_5_megablocks/transformer.py (1085:1090) duplicated block id: 8957 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (240:247) - megatron_patch/model/falcon/transformer.py (407:414) duplicated block id: 8958 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/bloom/gpt_model.py (101:108) - megatron_patch/model/llama2/gpt_model.py (108:115) duplicated block id: 8959 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/transformer.py (890:895) - megatron_patch/model/llama3/transformer_legacy.py (1367:1372) duplicated block id: 8960 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llava/hf2mcore_llava.py (32:37) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_dense_and_moe_gqa.py (31:36) duplicated block id: 8961 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer/mlp.py (94:99) - megatron_patch/model/qwen2/transformer/attention.py (103:108) duplicated block id: 8962 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/chatglm/checkpoint_reshaping_and_interoperability.py (528:536) - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (479:487) duplicated block id: 8963 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/galactica/transformer.py (35:43) - megatron_patch/model/starcoder/transformer.py (42:50) duplicated block id: 8964 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (336:342) - megatron_patch/model/llama/transformer.py (192:198) duplicated block id: 8965 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/qwen/transformer.py (680:685) - megatron_patch/model/qwen1_5/transformer/attention.py (476:481) duplicated block id: 8966 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/llama/transformer.py (192:198) - megatron_patch/model/mistral/transformer.py (339:345) duplicated block id: 8967 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (242:248) - megatron_patch/model/llama/transformer.py (192:198) duplicated block id: 8968 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/qwen/transformer.py (680:685) - megatron_patch/model/qwen2/transformer/attention.py (478:483) duplicated block id: 8969 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (278:287) - toolkits/model_checkpoints_convertor/yi/checkpoint_reshaping_and_interoperability.py (405:414) duplicated block id: 8970 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/galactica/transformer.py (154:160) - megatron_patch/model/qwen/transformer.py (333:339) duplicated block id: 8971 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_moe.py (439:445) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_dense_and_moe_gqa.py (793:799) duplicated block id: 8972 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/chatglm/checkpoint_reshaping_and_interoperability.py (106:113) - toolkits/model_checkpoints_convertor/mistral/hf2mcore.py (97:105) duplicated block id: 8973 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/transformer.py (735:741) - megatron_patch/model/qwen_vl/transformer.py (920:927) duplicated block id: 8974 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/mixtral_bak/transformer/mlp.py (19:42) - megatron_patch/model/qwen1_5/transformer/mlp.py (35:58) duplicated block id: 8975 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/deepseek/hf2mcore_deepseek_v2_moe.py (22:34) - toolkits/model_checkpoints_convertor/qwen/hf2megablocks_qwen1.5.py (8:22) duplicated block id: 8976 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore.py (275:280) - toolkits/model_checkpoints_convertor/mistral/hf2mcore.py (417:422) duplicated block id: 8977 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (884:891) - megatron_patch/model/bloom/transformer.py (635:641) duplicated block id: 8978 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/mixtral_bak/transformer/attention.py (348:353) - megatron_patch/model/qwen2_vl/attention_vision.py (695:700) duplicated block id: 8979 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/mixtral_bak/transformer/attention.py (348:353) - megatron_patch/model/qwen2_vl/attention_vision.py (707:712) duplicated block id: 8980 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (698:703) - megatron_patch/model/llama3/transformer/attention.py (476:481) duplicated block id: 8981 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/llama/positional_embeddings.py (39:44) - megatron_patch/model/llama2/rotary_pos_embedding.py (65:70) duplicated block id: 8982 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/transformer.py (813:818) - megatron_patch/model/qwen/transformer.py (1351:1356) duplicated block id: 8983 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/llava/transformer.py (1200:1205) - megatron_patch/model/qwen/transformer.py (1682:1687) duplicated block id: 8984 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/qwen/transformer.py (1146:1151) - megatron_patch/model/qwen1_5_megablocks/transformer.py (1602:1607) duplicated block id: 8985 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/galactica/language_model.py (445:456) - megatron_patch/model/llama2/language_model.py (434:445) duplicated block id: 8986 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/reward_model_to_megatron.py (673:679) - toolkits/model_checkpoints_convertor/glm130b/checkpoint_reshaping_and_interoperability.py (387:393) duplicated block id: 8987 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/qwen1_5_megablocks/language_model.py (401:409) - megatron_patch/model/starcoder/language_model.py (378:386) duplicated block id: 8988 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (1357:1362) - megatron_patch/model/llama3/transformer_legacy.py (1448:1453) duplicated block id: 8989 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/moe/experts.py (165:175) - megatron_patch/model/qwen2/moe/experts.py (247:257) duplicated block id: 8990 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/language_model.py (722:728) - megatron_patch/model/chatglm/language_model.py (659:665) duplicated block id: 8991 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (828:834) - toolkits/model_checkpoints_convertor/starcoder/checkpoint_reshaping_and_interoperability.py (361:367) duplicated block id: 8992 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/mixtral_bak/transformer/attention.py (81:86) - megatron_patch/model/mixtral_bak/transformer/mlp.py (72:77) duplicated block id: 8993 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (664:670) - toolkits/model_checkpoints_convertor/bloom/reward_model_to_megatron.py (337:343) duplicated block id: 8994 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/llama2/transformer.py (707:712) - megatron_patch/model/mixtral/transformer/attention.py (619:624) duplicated block id: 8995 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (1409:1414) - megatron_patch/model/falcon/transformer.py (1074:1079) duplicated block id: 8996 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/llama/language_model.py (356:361) - megatron_patch/model/llama2/language_model.py (352:357) duplicated block id: 8997 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/qwen_vl/transformer.py (1248:1253) - megatron_patch/model/starcoder/transformer.py (962:967) duplicated block id: 8998 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/llava/transformer.py (83:89) - megatron_patch/model/starcoder/transformer.py (83:90) duplicated block id: 8999 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/moe/experts.py (287:292) - megatron_patch/model/mixtral/moe/experts.py (302:307) duplicated block id: 9000 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/mistral/hf2mcore.py (425:430) - toolkits/model_checkpoints_convertor/qwen/hf2megablocks_qwen1.5.py (528:533) duplicated block id: 9001 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llava/hf2mcore_llava.py (454:459) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_dense_and_moe_gqa.py (426:431) duplicated block id: 9002 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/deepseek/hf2mcore_deepseek_v3_moe.py (23:34) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_mha_to_moe.py (6:17) duplicated block id: 9003 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/llama2/transformer.py (1100:1105) - megatron_patch/model/qwen_vl/transformer.py (1281:1286) duplicated block id: 9004 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/qwen1_5_megablocks/transformer.py (598:603) - megatron_patch/model/qwen2_vl/attention.py (620:625) duplicated block id: 9005 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/transformer.py (347:354) - megatron_patch/model/starcoder/transformer.py (537:544) duplicated block id: 9006 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/language_model.py (198:204) - megatron_patch/model/llama3/language_model.py (173:179) duplicated block id: 9007 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/deepseek/hf2mcore_deepseek_v3_moe.py (222:227) - toolkits/model_checkpoints_convertor/deepseek/hf2mcore_deepseek_v3_moe.py (593:598) duplicated block id: 9008 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (1409:1414) - megatron_patch/model/chatglm/transformer.py (695:700) duplicated block id: 9009 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/mixtral_bak/moe/token_dispatcher.py (246:252) - megatron_patch/model/qwen1_5/moe/token_dispatcher.py (207:213) duplicated block id: 9010 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (700:705) - toolkits/model_checkpoints_convertor/falcon/checkpoint_reshaping_and_interoperability.py (422:427) duplicated block id: 9011 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/qwen1_5/moe/token_dispatcher.py (224:230) - megatron_patch/model/qwen2/moe/token_dispatcher.py (234:240) duplicated block id: 9012 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/hf2te.py (286:291) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (603:608) duplicated block id: 9013 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/transformer.py (37:45) - megatron_patch/model/starcoder/transformer.py (42:50) duplicated block id: 9014 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/falcon/transformer.py (1307:1314) - megatron_patch/model/starcoder/transformer.py (1351:1358) duplicated block id: 9015 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (981:987) - megatron_patch/model/qwen_vl/transformer.py (1439:1445) duplicated block id: 9016 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/reward_model_to_megatron.py (337:343) - toolkits/model_checkpoints_convertor/falcon/checkpoint_reshaping_and_interoperability.py (637:643) duplicated block id: 9017 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/mlp.py (72:77) - megatron_patch/model/llama3/transformer/attention.py (107:112) duplicated block id: 9018 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/mistral/hf2mcore.py (8:19) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_mha_to_moe.py (6:17) duplicated block id: 9019 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (432:437) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (751:756) duplicated block id: 9020 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (1276:1281) - megatron_patch/model/falcon40b/transformer.py (813:818) duplicated block id: 9021 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/bloom/language_model.py (593:599) - megatron_patch/model/qwen1_5_megablocks/language_model.py (644:650) duplicated block id: 9022 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_mha.py (6:18) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_dense_and_moe_gqa.py (9:22) duplicated block id: 9023 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/hf2te.py (289:298) - toolkits/model_checkpoints_convertor/falcon/checkpoint_reshaping_and_interoperability.py (395:404) duplicated block id: 9024 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/language_model.py (638:644) - megatron_patch/model/glm130b/language_model.py (623:630) duplicated block id: 9025 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore.py (510:515) - toolkits/model_checkpoints_convertor/mistral/hf2mcore.py (213:218) duplicated block id: 9026 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/layer_specs.py (36:42) - megatron_patch/model/qwen3_moe/moe_module_specs.py (18:24) duplicated block id: 9027 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/llama3/model.py (98:103) - megatron_patch/model/mixtral/model.py (92:97) duplicated block id: 9028 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/galactica/language_model.py (166:183) - megatron_patch/model/llama2/language_model.py (130:148) duplicated block id: 9029 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/reward_model_to_megatron.py (378:383) - toolkits/model_checkpoints_convertor/bloom/reward_model_to_megatron.py (623:628) duplicated block id: 9030 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon/checkpoint_reshaping_and_interoperability.py (404:409) - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (311:316) duplicated block id: 9031 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/mistral/hf2mcore.py (8:19) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_mha.py (6:18) duplicated block id: 9032 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/hf2te.py (286:291) - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (703:708) duplicated block id: 9033 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/moe/experts.py (219:224) - megatron_patch/model/deepseek_v2/moe/experts.py (703:708) duplicated block id: 9034 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/transformer.py (81:86) - megatron_patch/model/falcon/transformer.py (155:160) duplicated block id: 9035 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (157:164) - megatron_patch/model/glm130b/transformer.py (181:188) duplicated block id: 9036 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/transformer.py (735:741) - megatron_patch/model/qwen1_5_megablocks/transformer.py (812:818) duplicated block id: 9037 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/moe/experts.py (274:279) - megatron_patch/model/mixtral/moe/experts.py (287:292) duplicated block id: 9038 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/llama2/language_model.py (398:403) - megatron_patch/model/llava/language_model.py (414:419) duplicated block id: 9039 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon40b/checkpoint_reshaping_and_interoperability.py (825:832) - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (919:926) duplicated block id: 9040 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (588:593) - megatron_patch/model/qwen2_vl/attention_vision.py (619:624) duplicated block id: 9041 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/moe/experts.py (274:279) - megatron_patch/model/mixtral/moe/experts.py (302:307) duplicated block id: 9042 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer/attention.py (362:367) - megatron_patch/model/mixtral/transformer/attention.py (677:682) duplicated block id: 9043 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore_llama3_1.py (32:37) - toolkits/model_checkpoints_convertor/llava/hf2mcore_llava.py (32:37) duplicated block id: 9044 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/falcon/transformer.py (596:601) - megatron_patch/model/falcon40b/transformer.py (419:424) duplicated block id: 9045 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/llama/language_model.py (356:361) - megatron_patch/model/mistral/language_model.py (353:358) duplicated block id: 9046 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (1139:1146) - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (919:926) duplicated block id: 9047 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/language_model.py (659:665) - megatron_patch/model/llama3/language_model.py (619:625) duplicated block id: 9048 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/transformer.py (695:700) - megatron_patch/model/qwen/transformer.py (1351:1356) duplicated block id: 9049 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/hf2te.py (286:291) - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (442:447) duplicated block id: 9050 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/deepseek/hf2mcore_deepseek_v2_moe.py (43:48) - toolkits/model_checkpoints_convertor/llava/hf2mcore_llava.py (32:37) duplicated block id: 9051 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/language_model.py (362:367) - megatron_patch/model/falcon40b/language_model.py (356:361) duplicated block id: 9052 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (432:437) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (603:608) duplicated block id: 9053 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/moe/experts.py (70:76) - megatron_patch/model/mixtral_bak/moe/experts.py (51:57) duplicated block id: 9054 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/language_model.py (518:523) - megatron_patch/model/falcon40b/language_model.py (542:547) duplicated block id: 9055 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/checkpoint_reshaping_and_interoperability.py (585:592) - toolkits/model_checkpoints_convertor/glm130b/checkpoint_reshaping_and_interoperability.py (314:321) duplicated block id: 9056 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/transformer.py (217:223) - megatron_patch/model/llama3/transformer_legacy.py (338:344) duplicated block id: 9057 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/language_model.py (370:375) - megatron_patch/model/llama3/language_model.py (347:352) duplicated block id: 9058 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/llama2/rotary_pos_embedding.py (65:70) - megatron_patch/model/qwen1_5_megablocks/rotary_pos_embedding.py (44:49) duplicated block id: 9059 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (700:705) - toolkits/model_checkpoints_convertor/falcon40b/checkpoint_reshaping_and_interoperability.py (411:416) duplicated block id: 9060 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/falcon/transformer.py (1074:1079) - megatron_patch/model/qwen/transformer.py (1351:1356) duplicated block id: 9061 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/gpt_model.py (118:125) - megatron_patch/model/bloom/gpt_model.py (101:108) duplicated block id: 9062 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (432:437) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (360:365) duplicated block id: 9063 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (738:745) - megatron_patch/model/falcon40b/transformer.py (645:653) duplicated block id: 9064 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (859:866) - megatron_patch/model/falcon/transformer.py (1012:1020) duplicated block id: 9065 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/gpt_model.py (118:125) - megatron_patch/model/glm130b/gpt_model.py (100:107) duplicated block id: 9066 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore.py (473:479) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_vl.py (574:581) duplicated block id: 9067 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (414:419) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (290:295) duplicated block id: 9068 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer_legacy.py (1172:1179) - megatron_patch/model/mistral/transformer.py (1209:1216) duplicated block id: 9069 size: 6 cleaned lines of code in 2 files: - toolkits/pretrain_data_preprocessing/preprocess_data_megatron.py (207:213) - toolkits/sft_data_preprocessing/build_idxmap_sft_dataset.py (197:203) duplicated block id: 9070 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/transformer.py (163:171) - megatron_patch/model/galactica/transformer.py (177:185) duplicated block id: 9071 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/llava/transformer.py (1281:1286) - megatron_patch/model/mistral/transformer.py (1097:1102) duplicated block id: 9072 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/galactica/transformer.py (542:550) - megatron_patch/model/llama3/transformer_legacy.py (837:844) duplicated block id: 9073 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/hf2te.py (187:192) - toolkits/model_checkpoints_convertor/galactica/checkpoint_reshaping_and_interoperability.py (273:278) duplicated block id: 9074 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/qwen1_5_megablocks/transformer.py (1194:1207) - megatron_patch/model/starcoder/transformer.py (1008:1021) duplicated block id: 9075 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer/attention.py (559:564) - megatron_patch/model/mixtral_bak/transformer/attention.py (348:353) duplicated block id: 9076 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/chatglm/checkpoint_reshaping_and_interoperability.py (583:588) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (1018:1023) duplicated block id: 9077 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/reward_model_to_megatron.py (584:591) - toolkits/model_checkpoints_convertor/glm/checkpoint_reshaping_and_interoperability.py (319:326) duplicated block id: 9078 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/starcoder/checkpoint_reshaping_and_interoperability.py (401:406) - toolkits/model_checkpoints_convertor/starcoder/checkpoint_reshaping_and_interoperability.py (637:642) duplicated block id: 9079 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (942:947) - toolkits/model_checkpoints_convertor/galactica/checkpoint_reshaping_and_interoperability.py (653:658) duplicated block id: 9080 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/transformer.py (134:140) - megatron_patch/model/qwen1_5_megablocks/transformer.py (230:236) duplicated block id: 9081 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/mistral/transformer.py (1279:1284) - megatron_patch/model/qwen_vl/transformer.py (1099:1104) duplicated block id: 9082 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/language_model.py (370:375) - megatron_patch/model/qwen_vl/language_model.py (353:358) duplicated block id: 9083 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/transformer.py (890:895) - megatron_patch/model/qwen/transformer.py (1351:1356) duplicated block id: 9084 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/llama/transformer.py (393:400) - megatron_patch/model/starcoder/transformer.py (537:544) duplicated block id: 9085 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/galactica/transformer.py (702:707) - megatron_patch/model/llama3/transformer_legacy.py (1367:1372) duplicated block id: 9086 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/mixtral_bak/layer_specs.py (92:97) - megatron_patch/model/qwen2/layer_specs.py (79:84) duplicated block id: 9087 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/chatglm/checkpoint_reshaping_and_interoperability.py (431:436) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (524:529) duplicated block id: 9088 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/galactica/transformer.py (154:160) - megatron_patch/model/qwen_vl/transformer.py (339:345) duplicated block id: 9089 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/transformer.py (181:188) - megatron_patch/model/mistral/transformer.py (161:168) duplicated block id: 9090 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (613:621) - megatron_patch/model/qwen_vl/transformer.py (885:892) duplicated block id: 9091 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/llama2/transformer.py (1282:1287) - megatron_patch/model/qwen_vl/transformer.py (1099:1104) duplicated block id: 9092 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/multi_latent_attention.py (85:90) - megatron_patch/model/llama3/transformer/attention.py (93:98) duplicated block id: 9093 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (807:813) - toolkits/model_checkpoints_convertor/starcoder/checkpoint_reshaping_and_interoperability.py (361:367) duplicated block id: 9094 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/galactica/transformer.py (702:707) - megatron_patch/model/qwen/transformer.py (1351:1356) duplicated block id: 9095 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (816:821) - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (703:708) duplicated block id: 9096 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/model.py (241:246) - megatron_patch/model/deepseek_v2/transformer_layer.py (309:314) duplicated block id: 9097 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/deepseek/hf2mcore_deepseek_v3_moe.py (23:34) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_moe.py (7:20) duplicated block id: 9098 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (513:518) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (751:756) duplicated block id: 9099 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (599:604) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (751:756) duplicated block id: 9100 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/moe/moe_layer.py (121:126) - megatron_patch/model/qwen1_5/moe/moe_layer.py (95:100) duplicated block id: 9101 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (1102:1107) - megatron_patch/model/llava/transformer.py (1281:1286) duplicated block id: 9102 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llava/hf2mcore_llava.py (530:535) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2.5_vl.py (221:226) duplicated block id: 9103 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/moe/experts.py (703:708) - megatron_patch/model/qwen1_5/moe/experts.py (220:225) duplicated block id: 9104 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/transformer.py (504:510) - megatron_patch/model/qwen1_5_megablocks/transformer.py (531:537) duplicated block id: 9105 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/deepseek/hf2mcore_deepseek_v2_moe.py (409:414) - toolkits/model_checkpoints_convertor/deepseek/hf2mcore_deepseek_v3_moe.py (570:575) duplicated block id: 9106 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (816:821) - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (442:447) duplicated block id: 9107 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/transformer.py (661:671) - megatron_patch/model/glm130b/transformer.py (724:734) duplicated block id: 9108 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (599:604) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (490:495) duplicated block id: 9109 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/transformer.py (510:518) - megatron_patch/model/llava/transformer.py (885:892) duplicated block id: 9110 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/qwen/transformer.py (1224:1229) - megatron_patch/model/qwen1_5_megablocks/transformer.py (984:989) duplicated block id: 9111 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/falcon/transformer.py (37:45) - megatron_patch/model/glm130b/transformer.py (40:48) duplicated block id: 9112 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/mistral/hf2mcore.py (478:483) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_mha_to_moe.py (218:223) duplicated block id: 9113 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon/checkpoint_reshaping_and_interoperability.py (849:855) - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (919:926) duplicated block id: 9114 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/llama3_1/model.py (30:35) - megatron_patch/model/qwen3_moe/gpt_layer_specs.py (39:44) duplicated block id: 9115 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (613:621) - megatron_patch/model/glm130b/transformer.py (698:706) duplicated block id: 9116 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/galactica/checkpoint_reshaping_and_interoperability.py (459:464) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (837:842) duplicated block id: 9117 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (635:641) - megatron_patch/model/llama3/transformer_legacy.py (872:878) duplicated block id: 9118 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer_legacy.py (1050:1055) - megatron_patch/model/qwen1_5_megablocks/transformer.py (1166:1171) duplicated block id: 9119 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/deepseek/hf2mcore_deepseek_v2_moe.py (272:278) - toolkits/model_checkpoints_convertor/deepseek/hf2mcore_deepseek_v3_moe.py (367:372) duplicated block id: 9120 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon40b/checkpoint_reshaping_and_interoperability.py (640:646) - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (807:813) duplicated block id: 9121 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (948:954) - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (919:926) duplicated block id: 9122 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/transformer.py (438:445) - megatron_patch/model/starcoder/transformer.py (537:544) duplicated block id: 9123 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/reward_model_to_megatron.py (490:497) - toolkits/model_checkpoints_convertor/starcoder/checkpoint_reshaping_and_interoperability.py (510:517) duplicated block id: 9124 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/qwen3_moe/moe/router.py (74:79) - megatron_patch/model/qwen3_moe/moe/router.py (118:124) duplicated block id: 9125 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/transformer.py (695:700) - megatron_patch/model/llava/transformer.py (1408:1413) duplicated block id: 9126 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (1357:1362) - megatron_patch/model/llava/transformer.py (1489:1494) duplicated block id: 9127 size: 6 cleaned lines of code in 2 files: - toolkits/pretrain_data_preprocessing/preprocess_data_megatron.py (387:394) - toolkits/sft_data_preprocessing/build_idxmap_sft_dataset.py (377:384) duplicated block id: 9128 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/bloom/language_model.py (593:599) - megatron_patch/model/llama3/language_model.py (619:625) duplicated block id: 9129 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/moe/token_dispatcher.py (474:479) - megatron_patch/model/qwen1_5/moe/token_dispatcher.py (424:429) duplicated block id: 9130 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (1357:1362) - megatron_patch/model/qwen_vl/transformer.py (1489:1494) duplicated block id: 9131 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/moe/moe_layer.py (121:126) - megatron_patch/model/qwen3_moe/moe/moe_layer.py (50:55) duplicated block id: 9132 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (1282:1287) - megatron_patch/model/qwen_vl/transformer.py (1099:1104) duplicated block id: 9133 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (442:447) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (490:495) duplicated block id: 9134 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/qwen1_5/transformer/attention.py (362:367) - megatron_patch/model/qwen2_vl/attention.py (678:683) duplicated block id: 9135 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/language_model.py (356:361) - megatron_patch/model/qwen1_5_megablocks/language_model.py (351:356) duplicated block id: 9136 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/reward_model_to_megatron.py (337:343) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (828:834) duplicated block id: 9137 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/qwen2/transformer/attention.py (358:363) - megatron_patch/model/qwen2_vl/attention_vision.py (677:682) duplicated block id: 9138 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/hf2te.py (289:298) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (463:472) duplicated block id: 9139 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (849:856) - megatron_patch/model/falcon40b/transformer.py (645:653) duplicated block id: 9140 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon/checkpoint_reshaping_and_interoperability.py (404:409) - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (374:379) duplicated block id: 9141 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/gpt_model.py (102:109) - megatron_patch/model/llama2/gpt_model.py (108:115) duplicated block id: 9142 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore_llama3_1.py (724:729) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2.5_vl.py (221:226) duplicated block id: 9143 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (700:705) - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (703:708) duplicated block id: 9144 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (738:745) - megatron_patch/model/bloom/transformer.py (613:621) duplicated block id: 9145 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_vl.py (574:581) - toolkits/model_checkpoints_convertor/qwen/hf2megablocks_qwen1.5.py (494:500) duplicated block id: 9146 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (145:151) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (214:220) duplicated block id: 9147 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/language_model.py (654:659) - megatron_patch/model/falcon40b/language_model.py (680:685) duplicated block id: 9148 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/falcon/transformer.py (211:217) - megatron_patch/model/llava/transformer.py (339:345) duplicated block id: 9149 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/language_model.py (623:630) - megatron_patch/model/llama2/language_model.py (645:651) duplicated block id: 9150 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/moe/moe_layer.py (120:125) - megatron_patch/model/qwen3_moe/moe/moe_layer.py (50:55) duplicated block id: 9151 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/qwen1_5_megablocks/rotary_pos_embedding.py (44:49) - megatron_patch/model/qwen1_5_megablocks/rotary_pos_embedding.py (65:70) duplicated block id: 9152 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/layer_specs.py (132:139) - megatron_patch/model/qwen2/layer_specs.py (127:134) duplicated block id: 9153 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (925:935) - megatron_patch/model/glm130b/transformer.py (858:868) duplicated block id: 9154 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/model.py (110:115) - megatron_patch/model/qwen1_5/model.py (98:103) duplicated block id: 9155 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/language_model.py (423:428) - megatron_patch/model/llama3/language_model.py (357:362) duplicated block id: 9156 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (513:518) - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (703:708) duplicated block id: 9157 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/hf2te.py (345:350) - toolkits/model_checkpoints_convertor/chatglm/checkpoint_reshaping_and_interoperability.py (431:436) duplicated block id: 9158 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/moe/moe_layer.py (58:63) - megatron_patch/model/qwen2/moe/moe_layer.py (64:69) duplicated block id: 9159 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/llama3/layer_specs.py (42:47) - megatron_patch/model/qwen2_vl/layer_specs.py (76:81) duplicated block id: 9160 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (700:705) - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (442:447) duplicated block id: 9161 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (751:756) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (520:525) duplicated block id: 9162 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (981:987) - megatron_patch/model/llama3/transformer_legacy.py (1398:1404) duplicated block id: 9163 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/chatglm/checkpoint_reshaping_and_interoperability.py (256:261) - toolkits/model_checkpoints_convertor/galactica/checkpoint_reshaping_and_interoperability.py (273:278) duplicated block id: 9164 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/chatglm/checkpoint_reshaping_and_interoperability.py (583:588) - toolkits/model_checkpoints_convertor/falcon40b/checkpoint_reshaping_and_interoperability.py (819:824) duplicated block id: 9165 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/bloom/language_model.py (368:373) - megatron_patch/model/qwen1_5_megablocks/language_model.py (351:356) duplicated block id: 9166 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (700:705) - toolkits/model_checkpoints_convertor/yi/checkpoint_reshaping_and_interoperability.py (435:440) duplicated block id: 9167 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (469:476) - toolkits/model_checkpoints_convertor/chatglm/checkpoint_reshaping_and_interoperability.py (407:414) duplicated block id: 9168 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/bloom/language_model.py (593:599) - megatron_patch/model/llava/language_model.py (705:711) duplicated block id: 9169 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (1016:1021) - megatron_patch/model/falcon40b/transformer.py (857:862) duplicated block id: 9170 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/llama/transformer.py (868:873) - megatron_patch/model/qwen_vl/transformer.py (1408:1413) duplicated block id: 9171 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (460:465) - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (700:705) duplicated block id: 9172 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (442:447) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (751:756) duplicated block id: 9173 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (242:248) - megatron_patch/model/galactica/transformer.py (154:160) duplicated block id: 9174 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/galactica/transformer.py (702:707) - megatron_patch/model/llama2/transformer.py (1409:1414) duplicated block id: 9175 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (299:304) - toolkits/model_checkpoints_convertor/baichuan2/hf2te.py (171:176) duplicated block id: 9176 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/model.py (92:97) - megatron_patch/model/qwen1_5/model.py (98:103) duplicated block id: 9177 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/gpt_model.py (118:125) - megatron_patch/model/chatglm/gpt_model.py (102:109) duplicated block id: 9178 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (513:518) - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (442:447) duplicated block id: 9179 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/falcon/transformer.py (37:45) - megatron_patch/model/starcoder/transformer.py (42:50) duplicated block id: 9180 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore.py (584:589) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_mha.py (280:285) duplicated block id: 9181 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/deepseek/hf2mcore_deepseek_v3_moe.py (23:34) - toolkits/model_checkpoints_convertor/llama/hf2mcore.py (10:25) duplicated block id: 9182 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (299:304) - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (322:327) duplicated block id: 9183 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/transformer/attention.py (520:525) - megatron_patch/model/qwen1_5/transformer/attention.py (384:389) duplicated block id: 9184 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/llama/positional_embeddings.py (39:44) - megatron_patch/model/llava/rotary_pos_embedding.py (63:68) duplicated block id: 9185 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/llama2/transformer.py (1740:1745) - megatron_patch/model/qwen/transformer.py (1146:1151) duplicated block id: 9186 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/mixtral_bak/layer_specs.py (20:38) - megatron_patch/model/qwen1_5/layer_specs.py (20:37) duplicated block id: 9187 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/qwen2_vl/attention.py (620:625) - megatron_patch/model/qwen_vl/transformer.py (706:711) duplicated block id: 9188 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer_legacy.py (1700:1705) - megatron_patch/model/mistral/transformer.py (1198:1203) duplicated block id: 9189 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/mlp.py (72:77) - megatron_patch/model/deepseek_v2/multi_latent_attention.py (101:106) duplicated block id: 9190 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/falcon/language_model.py (414:425) - megatron_patch/model/llama3/language_model.py (425:436) duplicated block id: 9191 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon40b/checkpoint_reshaping_and_interoperability.py (411:416) - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (442:447) duplicated block id: 9192 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/transformer/attention.py (619:624) - megatron_patch/model/qwen1_5_megablocks/transformer.py (598:603) duplicated block id: 9193 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (1071:1076) - megatron_patch/model/llama2/transformer.py (1740:1745) duplicated block id: 9194 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/falcon/transformer.py (211:217) - megatron_patch/model/qwen1_5_megablocks/transformer.py (230:236) duplicated block id: 9195 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (373:378) - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (703:708) duplicated block id: 9196 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore_llama3_1.py (648:653) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2.5_vl.py (331:336) duplicated block id: 9197 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/chatglm/checkpoint_reshaping_and_interoperability.py (367:372) - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (816:821) duplicated block id: 9198 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (299:304) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (318:323) duplicated block id: 9199 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (284:289) - toolkits/model_checkpoints_convertor/baichuan2/hf2te.py (171:176) duplicated block id: 9200 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/qwen1_5/transformer/mlp.py (110:115) - megatron_patch/model/qwen2_vl/attention_vision.py (121:126) duplicated block id: 9201 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/falcon/transformer.py (865:872) - megatron_patch/model/galactica/transformer.py (542:550) duplicated block id: 9202 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (432:437) - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (442:447) duplicated block id: 9203 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (229:234) - megatron_patch/model/falcon40b/transformer.py (243:248) duplicated block id: 9204 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/transformer.py (645:653) - megatron_patch/model/qwen_vl/transformer.py (885:892) duplicated block id: 9205 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon40b/checkpoint_reshaping_and_interoperability.py (411:416) - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (703:708) duplicated block id: 9206 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/llava/transformer.py (706:711) - megatron_patch/model/mixtral/transformer/attention.py (619:624) duplicated block id: 9207 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore.py (473:479) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2.5_vl.py (567:574) duplicated block id: 9208 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_vl.py (237:242) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_vl.py (343:348) duplicated block id: 9209 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/transformer.py (906:912) - megatron_patch/model/qwen1_5_megablocks/transformer.py (1325:1331) duplicated block id: 9210 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/transformer.py (204:210) - megatron_patch/model/qwen/transformer.py (333:339) duplicated block id: 9211 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/language_model.py (370:375) - megatron_patch/model/qwen1_5_megablocks/language_model.py (351:356) duplicated block id: 9212 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (432:437) - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (703:708) duplicated block id: 9213 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/transformer.py (217:223) - megatron_patch/model/qwen_vl/transformer.py (339:345) duplicated block id: 9214 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/transformer_layer.py (154:159) - megatron_patch/model/qwen2/transformer_layer.py (118:123) duplicated block id: 9215 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/language_model.py (408:413) - megatron_patch/model/glm130b/language_model.py (370:375) duplicated block id: 9216 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (460:465) - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (703:708) duplicated block id: 9217 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/moe/moe_layer.py (32:43) - megatron_patch/model/qwen1_5/moe/moe_layer.py (31:42) duplicated block id: 9218 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/glm130b/checkpoint_reshaping_and_interoperability.py (387:393) - toolkits/model_checkpoints_convertor/starcoder/checkpoint_reshaping_and_interoperability.py (678:684) duplicated block id: 9219 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/mistral/hf2mcore.py (286:294) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_mha.py (135:143) duplicated block id: 9220 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (922:928) - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (919:926) duplicated block id: 9221 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (664:670) - toolkits/model_checkpoints_convertor/starcoder/checkpoint_reshaping_and_interoperability.py (361:367) duplicated block id: 9222 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_mha.py (135:143) - toolkits/model_checkpoints_convertor/qwen/hf2megablocks_qwen1.5.py (337:345) duplicated block id: 9223 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/mistral/transformer.py (706:711) - megatron_patch/model/mixtral/transformer/attention.py (619:624) duplicated block id: 9224 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/llama2/rotary_pos_embedding.py (44:49) - megatron_patch/model/llama2/rotary_pos_embedding.py (65:70) duplicated block id: 9225 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (664:670) - toolkits/model_checkpoints_convertor/bloom/checkpoint_reshaping_and_interoperability.py (337:343) duplicated block id: 9226 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (373:378) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (490:495) duplicated block id: 9227 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/mixtral_bak/transformer/mlp.py (135:145) - megatron_patch/model/qwen2/transformer/mlp.py (182:192) duplicated block id: 9228 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/mixtral_bak/transformer/attention.py (348:353) - megatron_patch/model/qwen2/transformer/attention.py (564:569) duplicated block id: 9229 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (373:378) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (751:756) duplicated block id: 9230 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/qwen/transformer.py (1045:1050) - megatron_patch/model/qwen1_5_megablocks/transformer.py (1166:1171) duplicated block id: 9231 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llava/hf2mcore_llava.py (32:37) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_moe.py (48:53) duplicated block id: 9232 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/falcon/language_model.py (414:425) - megatron_patch/model/mistral/language_model.py (436:447) duplicated block id: 9233 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2megablocks_qwen1.5.py (269:274) - toolkits/model_checkpoints_convertor/qwen/hf2megablocks_qwen1.5.py (520:525) duplicated block id: 9234 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (464:469) - megatron_patch/model/chatglm/transformer.py (402:407) duplicated block id: 9235 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/transformer.py (204:210) - megatron_patch/model/mistral/transformer.py (339:345) duplicated block id: 9236 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/qwen2/transformer_block.py (231:236) - megatron_patch/model/qwen2_5_vl/transformer_block.py (563:568) duplicated block id: 9237 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore.py (584:589) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_mha_to_moe.py (239:244) duplicated block id: 9238 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/moe/moe_layer.py (150:157) - megatron_patch/model/qwen3_moe/moe/moe_layer.py (89:96) duplicated block id: 9239 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/mixtral_bak/transformer/attention.py (412:417) - megatron_patch/model/qwen2_vl/attention_vision.py (496:501) duplicated block id: 9240 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/chatglm/checkpoint_reshaping_and_interoperability.py (407:414) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (529:536) duplicated block id: 9241 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/bloom/language_model.py (228:236) - megatron_patch/model/glm130b/language_model.py (228:236) duplicated block id: 9242 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon/checkpoint_reshaping_and_interoperability.py (637:643) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (742:748) duplicated block id: 9243 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2.5_vl.py (221:226) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2.5_vl.py (331:336) duplicated block id: 9244 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore.py (571:576) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_mha_to_moe.py (218:223) duplicated block id: 9245 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/transformer.py (81:86) - megatron_patch/model/falcon40b/transformer.py (148:153) duplicated block id: 9246 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/falcon/transformer.py (1074:1079) - megatron_patch/model/qwen1_5_megablocks/transformer.py (1294:1299) duplicated block id: 9247 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (1203:1208) - megatron_patch/model/qwen1_5_megablocks/transformer.py (1602:1607) duplicated block id: 9248 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/transformer_config.py (529:535) - megatron_patch/model/mixtral_bak/transformer_config.py (246:252) duplicated block id: 9249 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon40b/checkpoint_reshaping_and_interoperability.py (411:416) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (751:756) duplicated block id: 9250 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_dense_and_moe_gqa.py (352:357) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_vl.py (328:333) duplicated block id: 9251 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (460:465) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (360:365) duplicated block id: 9252 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/llama2/transformer.py (83:89) - megatron_patch/model/starcoder/transformer.py (83:90) duplicated block id: 9253 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/checkpoint_reshaping_and_interoperability.py (585:592) - toolkits/model_checkpoints_convertor/glm/checkpoint_reshaping_and_interoperability.py (319:326) duplicated block id: 9254 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/gpt_model.py (102:109) - megatron_patch/model/mistral/gpt_model.py (108:115) duplicated block id: 9255 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/qwen/transformer.py (1156:1163) - megatron_patch/model/qwen_vl/transformer.py (1211:1218) duplicated block id: 9256 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/mixtral_bak/transformer/attention.py (348:353) - megatron_patch/model/qwen2/transformer/attention.py (552:557) duplicated block id: 9257 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/language_model.py (49:54) - megatron_patch/model/baichuan2/language_model.py (57:62) duplicated block id: 9258 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (1018:1023) - toolkits/model_checkpoints_convertor/yi/checkpoint_reshaping_and_interoperability.py (669:674) duplicated block id: 9259 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (460:465) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (603:608) duplicated block id: 9260 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/llama2/rotary_pos_embedding.py (65:70) - megatron_patch/model/llava/rotary_pos_embedding.py (42:47) duplicated block id: 9261 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/chatglm/checkpoint_reshaping_and_interoperability.py (431:436) - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (547:552) duplicated block id: 9262 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (311:316) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (502:507) duplicated block id: 9263 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (466:471) - toolkits/model_checkpoints_convertor/chatglm/checkpoint_reshaping_and_interoperability.py (431:436) duplicated block id: 9264 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/transformer.py (813:818) - megatron_patch/model/qwen_vl/transformer.py (1408:1413) duplicated block id: 9265 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (460:465) - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (442:447) duplicated block id: 9266 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/language_model.py (362:367) - megatron_patch/model/falcon/language_model.py (356:361) duplicated block id: 9267 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon40b/checkpoint_reshaping_and_interoperability.py (411:416) - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (816:821) duplicated block id: 9268 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (242:248) - megatron_patch/model/chatglm/transformer.py (134:140) duplicated block id: 9269 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/language_model.py (408:413) - megatron_patch/model/falcon/language_model.py (356:361) duplicated block id: 9270 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/hf2te.py (286:295) - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (362:371) duplicated block id: 9271 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/falcon/language_model.py (680:685) - megatron_patch/model/glm130b/language_model.py (618:623) duplicated block id: 9272 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/gpt_model.py (100:107) - megatron_patch/model/llama3/gpt_model.py (110:117) duplicated block id: 9273 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/galactica/checkpoint_reshaping_and_interoperability.py (649:656) - toolkits/model_checkpoints_convertor/glm/checkpoint_reshaping_and_interoperability.py (565:572) duplicated block id: 9274 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/deepseek/hf2mcore_deepseek_v3_moe.py (551:560) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_dense_and_moe_gqa.py (516:525) duplicated block id: 9275 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (286:291) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (318:323) duplicated block id: 9276 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/hf2te.py (171:176) - toolkits/model_checkpoints_convertor/yi/checkpoint_reshaping_and_interoperability.py (271:276) duplicated block id: 9277 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/hf2te.py (171:176) - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (322:327) duplicated block id: 9278 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/qwen1_5_megablocks/transformer.py (984:989) - megatron_patch/model/qwen1_5_megablocks/transformer.py (1166:1171) duplicated block id: 9279 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon40b/checkpoint_reshaping_and_interoperability.py (819:824) - toolkits/model_checkpoints_convertor/yi/checkpoint_reshaping_and_interoperability.py (669:674) duplicated block id: 9280 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (700:705) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (490:495) duplicated block id: 9281 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/qwen1_5/transformer/mlp.py (110:115) - megatron_patch/model/qwen2/transformer/attention.py (103:108) duplicated block id: 9282 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (1210:1218) - megatron_patch/model/falcon40b/transformer.py (1047:1055) duplicated block id: 9283 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/qwen3_moe/gpt_layer_specs.py (227:232) - megatron_patch/model/qwen3_moe/gpt_layer_specs.py (249:254) duplicated block id: 9284 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/multi_latent_attention.py (101:106) - megatron_patch/model/mixtral_bak/transformer/mlp.py (72:77) duplicated block id: 9285 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llava/hf2mcore_llava.py (530:535) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_vl.py (221:226) duplicated block id: 9286 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (981:987) - megatron_patch/model/qwen1_5_megablocks/transformer.py (1325:1331) duplicated block id: 9287 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/transformer.py (890:895) - megatron_patch/model/llama2/transformer.py (1409:1414) duplicated block id: 9288 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (123:130) - toolkits/model_checkpoints_convertor/qwen/hf2megablocks_qwen1.5.py (123:131) duplicated block id: 9289 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/transformer.py (181:188) - megatron_patch/model/llama3/transformer_legacy.py (165:172) duplicated block id: 9290 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/llama2/transformer.py (1201:1206) - megatron_patch/model/llama3/transformer_legacy.py (1700:1705) duplicated block id: 9291 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2.5_vl.py (567:574) - toolkits/model_checkpoints_convertor/qwen/hf2megablocks_qwen1.5.py (494:500) duplicated block id: 9292 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (1282:1287) - megatron_patch/model/llama2/transformer.py (1100:1105) duplicated block id: 9293 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/language_model.py (362:367) - megatron_patch/model/galactica/language_model.py (387:392) duplicated block id: 9294 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/galactica/gpt_model.py (107:114) - megatron_patch/model/glm130b/gpt_model.py (100:107) duplicated block id: 9295 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/mistral/transformer.py (1097:1102) - megatron_patch/model/qwen_vl/transformer.py (1281:1286) duplicated block id: 9296 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore.py (629:635) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_dense_and_moe_gqa.py (793:799) duplicated block id: 9297 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/llava/transformer.py (1099:1104) - megatron_patch/model/mistral/transformer.py (1279:1284) duplicated block id: 9298 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/language_model.py (638:644) - megatron_patch/model/bloom/language_model.py (593:599) duplicated block id: 9299 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/llava/transformer.py (1099:1104) - megatron_patch/model/llava/transformer.py (1281:1286) duplicated block id: 9300 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/gpt_model.py (100:107) - megatron_patch/model/mistral/gpt_model.py (108:115) duplicated block id: 9301 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (1610:1615) - megatron_patch/model/qwen_vl/transformer.py (1200:1205) duplicated block id: 9302 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer_legacy.py (1267:1280) - megatron_patch/model/starcoder/transformer.py (1008:1021) duplicated block id: 9303 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/bloom/gpt_model.py (101:108) - megatron_patch/model/qwen1_5_megablocks/gpt_model.py (108:115) duplicated block id: 9304 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/language_model.py (367:372) - megatron_patch/model/llava/language_model.py (355:360) duplicated block id: 9305 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/llava/transformer.py (706:711) - megatron_patch/model/mixtral_bak/transformer/attention.py (366:371) duplicated block id: 9306 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/language_model.py (680:685) - megatron_patch/model/glm130b/language_model.py (618:623) duplicated block id: 9307 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/mixtral_bak/moe/router.py (33:46) - megatron_patch/model/qwen2/moe/router.py (117:130) duplicated block id: 9308 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/multi_latent_attention.py (202:207) - megatron_patch/model/mixtral_bak/transformer/attention.py (412:417) duplicated block id: 9309 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/language_model.py (49:54) - megatron_patch/model/llama2/language_model.py (47:52) duplicated block id: 9310 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/transformer_block.py (348:362) - megatron_patch/model/qwen2/transformer_block.py (306:320) duplicated block id: 9311 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore.py (205:212) - toolkits/model_checkpoints_convertor/mistral/hf2mcore.py (150:157) duplicated block id: 9312 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/galactica/transformer.py (542:550) - megatron_patch/model/glm130b/transformer.py (698:706) duplicated block id: 9313 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/galactica/transformer.py (542:550) - megatron_patch/model/qwen/transformer.py (831:838) duplicated block id: 9314 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2.5_vl.py (125:130) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_dense_and_moe_gqa.py (123:128) duplicated block id: 9315 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/gpt_model.py (102:109) - megatron_patch/model/qwen_vl/gpt_model.py (109:116) duplicated block id: 9316 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/llava/rotary_pos_embedding.py (63:68) - megatron_patch/model/qwen1_5_megablocks/rotary_pos_embedding.py (44:49) duplicated block id: 9317 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/language_model.py (362:367) - megatron_patch/model/bloom/language_model.py (368:373) duplicated block id: 9318 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/language_model.py (623:630) - megatron_patch/model/llava/language_model.py (705:711) duplicated block id: 9319 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/mixtral_bak/transformer/mlp.py (19:42) - megatron_patch/model/qwen2/transformer/mlp.py (34:57) duplicated block id: 9320 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (336:342) - megatron_patch/model/falcon/transformer.py (211:217) duplicated block id: 9321 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/bloom/positional_embeddings.py (80:85) - megatron_patch/model/bloom/positional_embeddings.py (93:98) duplicated block id: 9322 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/transformer.py (510:518) - megatron_patch/model/llama2/transformer.py (886:893) duplicated block id: 9323 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (391:397) - megatron_patch/model/starcoder/transformer.py (406:412) duplicated block id: 9324 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (773:780) - megatron_patch/model/glm130b/transformer.py (735:741) duplicated block id: 9325 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/falcon/language_model.py (680:685) - megatron_patch/model/starcoder/language_model.py (558:563) duplicated block id: 9326 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/llava/transformer.py (1099:1104) - megatron_patch/model/qwen_vl/transformer.py (1281:1286) duplicated block id: 9327 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/gpt_model.py (102:109) - megatron_patch/model/llama/gpt_model.py (107:114) duplicated block id: 9328 size: 6 cleaned lines of code in 2 files: - toolkits/distributed_checkpoints_convertor/impl/general/h2m_synchronizer.py (135:141) - toolkits/distributed_checkpoints_convertor/impl/general/h2m_synchronizer.py (169:174) duplicated block id: 9329 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/llama2/transformer.py (707:712) - megatron_patch/model/qwen2/transformer/attention.py (478:483) duplicated block id: 9330 size: 6 cleaned lines of code in 2 files: - toolkits/pretrain_data_preprocessing/preprocess_data_megatron.py (387:394) - toolkits/pretrain_data_preprocessing/preprocess_data_megatron.py (403:410) duplicated block id: 9331 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/multi_latent_attention.py (202:207) - megatron_patch/model/llama3/transformer/attention.py (528:533) duplicated block id: 9332 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (849:856) - megatron_patch/model/chatglm/transformer.py (510:518) duplicated block id: 9333 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/transformer/attention.py (619:624) - megatron_patch/model/qwen/transformer.py (680:685) duplicated block id: 9334 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/language_model.py (659:665) - megatron_patch/model/qwen_vl/language_model.py (674:680) duplicated block id: 9335 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/qwen1_5/transformer/attention.py (384:389) - megatron_patch/model/qwen2/transformer/attention.py (381:386) duplicated block id: 9336 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (1203:1208) - megatron_patch/model/mistral/transformer.py (1737:1742) duplicated block id: 9337 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/falcon/transformer.py (1074:1079) - megatron_patch/model/llama2/transformer.py (1409:1414) duplicated block id: 9338 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/glm/checkpoint_reshaping_and_interoperability.py (565:572) - toolkits/model_checkpoints_convertor/yi/checkpoint_reshaping_and_interoperability.py (666:672) duplicated block id: 9339 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/transformer/mlp.py (96:101) - megatron_patch/model/qwen2/transformer/attention.py (103:108) duplicated block id: 9340 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/mixtral_bak/transformer/attention.py (366:371) - megatron_patch/model/qwen/transformer.py (680:685) duplicated block id: 9341 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/qwen_vl/language_model.py (674:680) - megatron_patch/model/starcoder/language_model.py (563:569) duplicated block id: 9342 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/transformer/attention.py (496:501) - megatron_patch/model/qwen1_5/transformer/attention.py (528:533) duplicated block id: 9343 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore.py (416:421) - toolkits/model_checkpoints_convertor/qwen/hf2megablocks_qwen1.5.py (446:467) duplicated block id: 9344 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/multi_latent_attention.py (85:90) - megatron_patch/model/qwen2/transformer/attention.py (89:94) duplicated block id: 9345 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore_llama3_1.py (554:562) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_dense_and_moe_gqa.py (515:523) duplicated block id: 9346 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (60:66) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_moe.py (54:60) duplicated block id: 9347 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer_legacy.py (1240:1245) - megatron_patch/model/llava/transformer.py (1099:1104) duplicated block id: 9348 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/mistral/transformer.py (765:770) - megatron_patch/model/qwen1_5_megablocks/transformer.py (658:663) duplicated block id: 9349 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/qwen1_5/moe/moe_layer.py (95:100) - megatron_patch/model/qwen3_moe/moe/moe_layer.py (50:55) duplicated block id: 9350 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/llama/transformer.py (656:663) - megatron_patch/model/llama3/transformer_legacy.py (856:863) duplicated block id: 9351 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (299:304) - toolkits/model_checkpoints_convertor/baichuan/hf2te.py (172:177) duplicated block id: 9352 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/transformer.py (504:510) - megatron_patch/model/llama2/transformer.py (640:646) duplicated block id: 9353 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/transformer/mlp.py (96:101) - megatron_patch/model/qwen2_vl/attention_vision.py (121:126) duplicated block id: 9354 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (981:987) - megatron_patch/model/llama2/transformer.py (1440:1446) duplicated block id: 9355 size: 6 cleaned lines of code in 2 files: - megatron_patch/arguments.py (418:423) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (64:69) duplicated block id: 9356 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/llama/language_model.py (197:203) - megatron_patch/model/llama3/language_model.py (173:179) duplicated block id: 9357 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/llava/transformer.py (706:711) - megatron_patch/model/qwen2/transformer/attention.py (478:483) duplicated block id: 9358 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/transformer.py (906:912) - megatron_patch/model/qwen/transformer.py (1382:1388) duplicated block id: 9359 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon40b/checkpoint_reshaping_and_interoperability.py (411:416) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (360:365) duplicated block id: 9360 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (1740:1745) - megatron_patch/model/llama2/transformer.py (1201:1206) duplicated block id: 9361 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/qwen1_5_megablocks/language_model.py (397:402) - megatron_patch/model/qwen_vl/language_model.py (406:411) duplicated block id: 9362 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore_llama3_1.py (648:653) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_vl.py (328:333) duplicated block id: 9363 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/mistral/hf2mcore.py (407:412) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_dense_and_moe_gqa.py (546:551) duplicated block id: 9364 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/multi_latent_attention.py (85:90) - megatron_patch/model/qwen2_vl/attention_vision.py (106:111) duplicated block id: 9365 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (703:708) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (360:365) duplicated block id: 9366 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/transformer_block.py (300:305) - megatron_patch/model/qwen2_5_vl/transformer_block.py (357:362) duplicated block id: 9367 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (362:371) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (490:499) duplicated block id: 9368 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/checkpoint_reshaping_and_interoperability.py (491:498) - toolkits/model_checkpoints_convertor/falcon40b/checkpoint_reshaping_and_interoperability.py (774:781) duplicated block id: 9369 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_dense_and_moe_gqa.py (605:610) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_vl.py (612:617) duplicated block id: 9370 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/gpt_model.py (134:141) - megatron_patch/model/bloom/gpt_model.py (101:108) duplicated block id: 9371 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (164:170) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (123:129) duplicated block id: 9372 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (402:411) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (278:287) duplicated block id: 9373 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/llava/transformer.py (1211:1218) - megatron_patch/model/qwen/transformer.py (1156:1163) duplicated block id: 9374 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (703:708) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (520:525) duplicated block id: 9375 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/language_model.py (623:630) - megatron_patch/model/qwen/language_model.py (627:633) duplicated block id: 9376 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon40b/checkpoint_reshaping_and_interoperability.py (393:398) - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (374:379) duplicated block id: 9377 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/layer_specs.py (78:83) - megatron_patch/model/qwen2_vl/layer_specs.py (76:81) duplicated block id: 9378 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/mistral/transformer.py (1306:1319) - megatron_patch/model/starcoder/transformer.py (1008:1021) duplicated block id: 9379 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/transformer.py (33:41) - megatron_patch/model/glm130b/transformer.py (40:48) duplicated block id: 9380 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/bloom/language_model.py (593:599) - megatron_patch/model/qwen_vl/language_model.py (674:680) duplicated block id: 9381 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore.py (571:576) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_dense_and_moe_gqa.py (619:624) duplicated block id: 9382 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/language_model.py (722:728) - megatron_patch/model/starcoder/language_model.py (563:569) duplicated block id: 9383 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/galactica/language_model.py (228:234) - megatron_patch/model/llama3/language_model.py (173:179) duplicated block id: 9384 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/galactica/language_model.py (387:392) - megatron_patch/model/mistral/language_model.py (353:358) duplicated block id: 9385 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/qwen1_5_megablocks/transformer.py (598:603) - megatron_patch/model/qwen2_vl/attention_vision.py (619:624) duplicated block id: 9386 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon40b/checkpoint_reshaping_and_interoperability.py (411:416) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (603:608) duplicated block id: 9387 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/qwen2_5_vl/transformer_block.py (326:331) - megatron_patch/model/qwen2_5_vl/transformer_block.py (544:549) duplicated block id: 9388 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (635:641) - megatron_patch/model/qwen/transformer.py (866:873) duplicated block id: 9389 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/mixtral_bak/transformer/mlp.py (72:77) - megatron_patch/model/qwen1_5/transformer/attention.py (107:112) duplicated block id: 9390 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (1161:1169) - megatron_patch/model/starcoder/transformer.py (995:1003) duplicated block id: 9391 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/mlp.py (72:77) - megatron_patch/model/mixtral_bak/transformer/attention.py (81:86) duplicated block id: 9392 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/mixtral_bak/moe/token_dispatcher.py (264:270) - megatron_patch/model/qwen1_5/moe/token_dispatcher.py (227:233) duplicated block id: 9393 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/language_model.py (441:452) - megatron_patch/model/falcon/language_model.py (414:425) duplicated block id: 9394 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer_legacy.py (1152:1157) - megatron_patch/model/qwen_vl/transformer.py (1739:1744) duplicated block id: 9395 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/galactica/language_model.py (166:183) - megatron_patch/model/qwen_vl/language_model.py (131:149) duplicated block id: 9396 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/qwen_vl/language_model.py (353:358) - megatron_patch/model/starcoder/language_model.py (352:357) duplicated block id: 9397 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (495:500) - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (374:379) duplicated block id: 9398 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (1357:1362) - megatron_patch/model/qwen/transformer.py (1432:1437) duplicated block id: 9399 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/falcon/language_model.py (356:361) - megatron_patch/model/llama3/language_model.py (347:352) duplicated block id: 9400 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/hf2te.py (286:291) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (751:756) duplicated block id: 9401 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2megablocks_qwen1.5.py (123:131) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (152:159) duplicated block id: 9402 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/falcon/language_model.py (356:361) - megatron_patch/model/qwen_vl/language_model.py (353:358) duplicated block id: 9403 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (599:604) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (360:365) duplicated block id: 9404 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/galactica/language_model.py (445:456) - megatron_patch/model/llama3/language_model.py (425:436) duplicated block id: 9405 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer/attention.py (391:396) - megatron_patch/model/qwen2_vl/attention.py (531:536) duplicated block id: 9406 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/moe/router.py (170:175) - megatron_patch/model/qwen2/moe/router.py (248:253) duplicated block id: 9407 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/galactica/language_model.py (166:183) - megatron_patch/model/llama3/language_model.py (119:137) duplicated block id: 9408 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/transformer_block.py (279:284) - megatron_patch/model/deepseek_v2/transformer_block.py (497:502) duplicated block id: 9409 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (830:837) - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (1139:1146) duplicated block id: 9410 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/llama3_1/layer_specs.py (43:48) - megatron_patch/model/qwen2_vl/layer_specs.py (76:81) duplicated block id: 9411 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/llava/transformer.py (766:771) - megatron_patch/model/mistral/transformer.py (765:770) duplicated block id: 9412 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/hf2te.py (171:176) - toolkits/model_checkpoints_convertor/falcon40b/checkpoint_reshaping_and_interoperability.py (255:260) duplicated block id: 9413 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/falcon/transformer.py (1074:1079) - megatron_patch/model/llama3/transformer_legacy.py (1367:1372) duplicated block id: 9414 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2.5_vl.py (331:336) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_moe.py (380:385) duplicated block id: 9415 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/mistral/transformer.py (706:711) - megatron_patch/model/qwen2_vl/attention.py (620:625) duplicated block id: 9416 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/llama/language_model.py (414:425) - megatron_patch/model/mistral/language_model.py (436:447) duplicated block id: 9417 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (278:287) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (490:499) duplicated block id: 9418 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon/checkpoint_reshaping_and_interoperability.py (843:848) - toolkits/model_checkpoints_convertor/yi/checkpoint_reshaping_and_interoperability.py (669:674) duplicated block id: 9419 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (540:547) - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (615:622) duplicated block id: 9420 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/transformer.py (37:45) - megatron_patch/model/falcon/transformer.py (37:45) duplicated block id: 9421 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore_llama3_1.py (724:729) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_moe.py (380:385) duplicated block id: 9422 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/layer_specs.py (166:173) - megatron_patch/model/qwen2/layer_specs.py (127:134) duplicated block id: 9423 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/hf2te.py (289:298) - toolkits/model_checkpoints_convertor/galactica/checkpoint_reshaping_and_interoperability.py (432:441) duplicated block id: 9424 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (183:189) - megatron_patch/model/llama2/transformer.py (339:345) duplicated block id: 9425 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/transformer.py (40:48) - megatron_patch/model/llama/transformer.py (38:46) duplicated block id: 9426 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (311:316) - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (495:500) duplicated block id: 9427 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/galactica/transformer.py (154:160) - megatron_patch/model/qwen1_5_megablocks/transformer.py (230:236) duplicated block id: 9428 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (539:548) - megatron_patch/model/glm130b/transformer.py (585:594) duplicated block id: 9429 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/qwen2/transformer/attention.py (103:108) - megatron_patch/model/qwen2/transformer/mlp.py (109:114) duplicated block id: 9430 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/language_model.py (49:54) - megatron_patch/model/qwen_vl/language_model.py (48:53) duplicated block id: 9431 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (490:495) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (751:756) duplicated block id: 9432 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/language_model.py (391:397) - megatron_patch/model/mistral/language_model.py (374:380) duplicated block id: 9433 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/bloom/language_model.py (464:469) - megatron_patch/model/falcon40b/language_model.py (542:547) duplicated block id: 9434 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/qwen/transformer.py (1045:1050) - megatron_patch/model/qwen/transformer.py (1224:1229) duplicated block id: 9435 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/chatglm/checkpoint_reshaping_and_interoperability.py (367:372) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (837:842) duplicated block id: 9436 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/qwen2/moe/experts.py (324:329) - megatron_patch/model/qwen2/moe/experts.py (392:397) duplicated block id: 9437 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/galactica/transformer.py (542:550) - megatron_patch/model/llama/transformer.py (637:645) duplicated block id: 9438 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (374:379) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (472:477) duplicated block id: 9439 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/qwen1_5_megablocks/transformer.py (1133:1138) - megatron_patch/model/starcoder/transformer.py (962:967) duplicated block id: 9440 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/language_model.py (367:372) - megatron_patch/model/qwen/language_model.py (352:357) duplicated block id: 9441 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (414:419) - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (311:316) duplicated block id: 9442 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/mixtral_bak/transformer/mlp.py (72:77) - megatron_patch/model/qwen2/transformer/attention.py (103:108) duplicated block id: 9443 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (919:926) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (854:861) duplicated block id: 9444 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/qwen1_5_megablocks/transformer.py (81:87) - megatron_patch/model/starcoder/transformer.py (83:90) duplicated block id: 9445 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer/mlp.py (163:173) - megatron_patch/model/mixtral_bak/transformer/mlp.py (135:145) duplicated block id: 9446 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/llava/transformer.py (1248:1253) - megatron_patch/model/starcoder/transformer.py (962:967) duplicated block id: 9447 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/bloom/gpt_model.py (89:96) - megatron_patch/model/galactica/gpt_model.py (94:100) duplicated block id: 9448 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (673:678) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (603:608) duplicated block id: 9449 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/moe/moe_layer.py (31:42) - megatron_patch/model/qwen1_5/moe/moe_layer.py (31:42) duplicated block id: 9450 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (673:678) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (360:365) duplicated block id: 9451 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/llava/transformer.py (1281:1286) - megatron_patch/model/qwen_vl/transformer.py (1099:1104) duplicated block id: 9452 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/moe/experts.py (703:708) - megatron_patch/model/mixtral/moe/experts.py (220:225) duplicated block id: 9453 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (442:447) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (290:295) duplicated block id: 9454 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/llava/language_model.py (414:419) - megatron_patch/model/mistral/language_model.py (400:405) duplicated block id: 9455 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/hf2te.py (187:192) - toolkits/model_checkpoints_convertor/falcon40b/checkpoint_reshaping_and_interoperability.py (255:260) duplicated block id: 9456 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore_llama3_1.py (135:140) - toolkits/model_checkpoints_convertor/llava/hf2mcore_llava.py (159:164) duplicated block id: 9457 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (614:620) - toolkits/model_checkpoints_convertor/falcon/checkpoint_reshaping_and_interoperability.py (556:562) duplicated block id: 9458 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (123:129) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (201:207) duplicated block id: 9459 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/falcon/transformer.py (596:601) - megatron_patch/model/llama/transformer.py (427:432) duplicated block id: 9460 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/gpt_model.py (107:114) - megatron_patch/model/glm130b/gpt_model.py (100:107) duplicated block id: 9461 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/transformer.py (645:653) - megatron_patch/model/glm130b/transformer.py (698:706) duplicated block id: 9462 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (494:499) - toolkits/model_checkpoints_convertor/chatglm/checkpoint_reshaping_and_interoperability.py (431:436) duplicated block id: 9463 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (513:518) - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (816:821) duplicated block id: 9464 size: 6 cleaned lines of code in 2 files: - megatron_patch/arguments.py (418:423) - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (37:42) duplicated block id: 9465 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (990:997) - megatron_patch/model/llama2/transformer.py (1120:1127) duplicated block id: 9466 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/gpt_model.py (100:107) - megatron_patch/model/qwen1_5_megablocks/gpt_model.py (108:115) duplicated block id: 9467 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/checkpoint_reshaping_and_interoperability.py (379:384) - toolkits/model_checkpoints_convertor/glm/checkpoint_reshaping_and_interoperability.py (358:363) duplicated block id: 9468 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (925:935) - megatron_patch/model/falcon40b/transformer.py (781:791) duplicated block id: 9469 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (460:465) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (751:756) duplicated block id: 9470 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/llama2/language_model.py (645:651) - megatron_patch/model/starcoder/language_model.py (563:569) duplicated block id: 9471 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/llama2/transformer.py (1740:1745) - megatron_patch/model/llama3/transformer_legacy.py (1152:1157) duplicated block id: 9472 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon/checkpoint_reshaping_and_interoperability.py (422:427) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (837:842) duplicated block id: 9473 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (464:469) - megatron_patch/model/llama/transformer.py (463:468) duplicated block id: 9474 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/transformer_layer.py (309:314) - megatron_patch/model/mixtral/model.py (205:210) duplicated block id: 9475 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/transformer.py (595:600) - megatron_patch/model/llama/transformer.py (795:800) duplicated block id: 9476 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/moe/experts.py (703:708) - megatron_patch/model/mixtral/moe/experts.py (798:803) duplicated block id: 9477 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer_legacy.py (1700:1705) - megatron_patch/model/qwen_vl/transformer.py (1200:1205) duplicated block id: 9478 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2.5_vl.py (331:336) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_vl.py (221:226) duplicated block id: 9479 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/llava/rotary_pos_embedding.py (42:47) - megatron_patch/model/llava/rotary_pos_embedding.py (63:68) duplicated block id: 9480 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/transformer_block.py (310:315) - megatron_patch/model/qwen2_5_vl/transformer_block.py (347:352) duplicated block id: 9481 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/llama/transformer.py (868:873) - megatron_patch/model/llava/transformer.py (1408:1413) duplicated block id: 9482 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (1276:1281) - megatron_patch/model/chatglm/transformer.py (695:700) duplicated block id: 9483 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/chatglm/checkpoint_reshaping_and_interoperability.py (256:261) - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (322:327) duplicated block id: 9484 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/falcon/transformer.py (211:217) - megatron_patch/model/qwen/transformer.py (333:339) duplicated block id: 9485 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (698:703) - megatron_patch/model/qwen2_vl/attention_vision.py (619:624) duplicated block id: 9486 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (925:935) - megatron_patch/model/chatglm/transformer.py (663:673) duplicated block id: 9487 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (460:465) - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (373:378) duplicated block id: 9488 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/bloom/gpt_model.py (101:108) - megatron_patch/model/galactica/gpt_model.py (107:114) duplicated block id: 9489 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer/attention.py (476:481) - megatron_patch/model/mistral/transformer.py (706:711) duplicated block id: 9490 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/language_model.py (414:425) - megatron_patch/model/mistral/language_model.py (436:447) duplicated block id: 9491 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/deepseek/hf2mcore_deepseek_v3_moe.py (139:146) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (70:77) duplicated block id: 9492 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (1016:1021) - megatron_patch/model/llama/transformer.py (912:917) duplicated block id: 9493 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2.5_vl.py (221:226) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_vl.py (328:333) duplicated block id: 9494 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer_legacy.py (1172:1179) - megatron_patch/model/llava/transformer.py (1211:1218) duplicated block id: 9495 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon/checkpoint_reshaping_and_interoperability.py (637:643) - toolkits/model_checkpoints_convertor/starcoder/checkpoint_reshaping_and_interoperability.py (361:367) duplicated block id: 9496 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/transformer.py (906:912) - megatron_patch/model/llava/transformer.py (1439:1445) duplicated block id: 9497 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/language_model.py (362:367) - megatron_patch/model/llama/language_model.py (356:361) duplicated block id: 9498 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (1016:1021) - megatron_patch/model/falcon/transformer.py (1118:1123) duplicated block id: 9499 size: 6 cleaned lines of code in 2 files: - megatron_patch/arguments.py (418:423) - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (36:41) duplicated block id: 9500 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/chatglm/checkpoint_reshaping_and_interoperability.py (367:372) - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (599:604) duplicated block id: 9501 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/qwen/transformer.py (680:685) - megatron_patch/model/qwen2_vl/attention.py (620:625) duplicated block id: 9502 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/language_model.py (367:372) - megatron_patch/model/mistral/language_model.py (353:358) duplicated block id: 9503 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (464:469) - megatron_patch/model/falcon40b/transformer.py (453:458) duplicated block id: 9504 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/chatglm/checkpoint_reshaping_and_interoperability.py (407:414) - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (522:529) duplicated block id: 9505 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/qwen1_5/layer_specs.py (42:47) - megatron_patch/model/qwen2_vl/layer_specs.py (76:81) duplicated block id: 9506 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (1210:1218) - megatron_patch/model/falcon/transformer.py (1308:1316) duplicated block id: 9507 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/hf2te.py (187:192) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (318:323) duplicated block id: 9508 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon40b/checkpoint_reshaping_and_interoperability.py (124:131) - toolkits/model_checkpoints_convertor/mistral/hf2mcore.py (97:105) duplicated block id: 9509 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (816:821) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (490:495) duplicated block id: 9510 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/transformer.py (735:741) - megatron_patch/model/llama2/transformer.py (921:928) duplicated block id: 9511 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/chatglm/checkpoint_reshaping_and_interoperability.py (367:372) - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (373:378) duplicated block id: 9512 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/llama/language_model.py (356:361) - megatron_patch/model/qwen/language_model.py (352:357) duplicated block id: 9513 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/language_model.py (389:394) - megatron_patch/model/glm130b/language_model.py (406:411) duplicated block id: 9514 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/llama/language_model.py (356:361) - megatron_patch/model/llama3/language_model.py (347:352) duplicated block id: 9515 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/qwen2/transformer/mlp.py (109:114) - megatron_patch/model/qwen2_vl/attention_vision.py (121:126) duplicated block id: 9516 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/llama/transformer.py (192:198) - megatron_patch/model/qwen_vl/transformer.py (339:345) duplicated block id: 9517 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/falcon/transformer.py (1118:1123) - megatron_patch/model/starcoder/transformer.py (1174:1179) duplicated block id: 9518 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/deepseek/hf2mcore_deepseek_v2_moe.py (186:191) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_dense_and_moe_gqa.py (299:304) duplicated block id: 9519 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/transformer.py (663:673) - megatron_patch/model/starcoder/transformer.py (1088:1098) duplicated block id: 9520 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/checkpoint_reshaping_and_interoperability.py (491:498) - toolkits/model_checkpoints_convertor/falcon/checkpoint_reshaping_and_interoperability.py (798:805) duplicated block id: 9521 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/language_model.py (518:523) - megatron_patch/model/llama/language_model.py (542:547) duplicated block id: 9522 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/language_model.py (367:372) - megatron_patch/model/llama2/language_model.py (352:357) duplicated block id: 9523 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/transformer.py (695:700) - megatron_patch/model/mistral/transformer.py (1406:1411) duplicated block id: 9524 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/transformer_config.py (490:497) - megatron_patch/model/mixtral_bak/transformer_config.py (231:238) duplicated block id: 9525 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/multi_latent_attention.py (85:90) - megatron_patch/model/mixtral_bak/transformer/attention.py (67:72) duplicated block id: 9526 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/galactica/checkpoint_reshaping_and_interoperability.py (598:606) - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (479:487) duplicated block id: 9527 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/qwen2/transformer/attention.py (533:538) - megatron_patch/model/qwen2_vl/attention.py (497:502) duplicated block id: 9528 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer_legacy.py (1050:1055) - megatron_patch/model/mistral/transformer.py (1279:1284) duplicated block id: 9529 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/falcon/language_model.py (542:547) - megatron_patch/model/glm130b/language_model.py (486:491) duplicated block id: 9530 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (442:447) - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (374:379) duplicated block id: 9531 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/llava/transformer.py (1099:1104) - megatron_patch/model/qwen1_5_megablocks/transformer.py (1166:1171) duplicated block id: 9532 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (82:88) - megatron_patch/model/starcoder/transformer.py (83:90) duplicated block id: 9533 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/hf2te.py (289:298) - toolkits/model_checkpoints_convertor/falcon40b/checkpoint_reshaping_and_interoperability.py (384:393) duplicated block id: 9534 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/transformer/attention.py (677:682) - megatron_patch/model/qwen1_5/transformer/attention.py (362:367) duplicated block id: 9535 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/mistral/transformer.py (1209:1216) - megatron_patch/model/qwen/transformer.py (1156:1163) duplicated block id: 9536 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/qwen/transformer.py (1146:1151) - megatron_patch/model/qwen_vl/transformer.py (1739:1744) duplicated block id: 9537 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon/checkpoint_reshaping_and_interoperability.py (422:427) - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (599:604) duplicated block id: 9538 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (460:465) - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (599:604) duplicated block id: 9539 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/moe/experts.py (797:802) - megatron_patch/model/qwen2/moe/experts.py (324:329) duplicated block id: 9540 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/transformer_block.py (497:502) - megatron_patch/model/qwen2/transformer_block.py (231:236) duplicated block id: 9541 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/hf2te.py (314:319) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (837:842) duplicated block id: 9542 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/transformer.py (181:188) - megatron_patch/model/qwen1_5_megablocks/transformer.py (159:166) duplicated block id: 9543 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (786:795) - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (1083:1092) duplicated block id: 9544 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (1203:1208) - megatron_patch/model/qwen_vl/transformer.py (1739:1744) duplicated block id: 9545 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (211:219) - megatron_patch/model/falcon/transformer.py (234:242) duplicated block id: 9546 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer/mlp.py (197:204) - megatron_patch/model/mixtral_bak/transformer/mlp.py (170:177) duplicated block id: 9547 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (703:708) - toolkits/model_checkpoints_convertor/yi/checkpoint_reshaping_and_interoperability.py (435:440) duplicated block id: 9548 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/mistral/hf2mcore.py (299:304) - toolkits/model_checkpoints_convertor/qwen/hf2megablocks_qwen1.5.py (350:355) duplicated block id: 9549 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (613:621) - megatron_patch/model/llama2/transformer.py (886:893) duplicated block id: 9550 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon/checkpoint_reshaping_and_interoperability.py (422:427) - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (373:378) duplicated block id: 9551 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/transformer.py (992:997) - megatron_patch/model/llama/transformer.py (795:800) duplicated block id: 9552 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/mistral/hf2mcore.py (213:218) - toolkits/model_checkpoints_convertor/mistral/hf2mcore.py (417:422) duplicated block id: 9553 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore_llama3_1.py (648:653) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_dense_and_moe_gqa.py (426:431) duplicated block id: 9554 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore.py (571:576) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_moe.py (367:372) duplicated block id: 9555 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (1176:1189) - megatron_patch/model/starcoder/transformer.py (1008:1021) duplicated block id: 9556 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore.py (275:280) - toolkits/model_checkpoints_convertor/llama/hf2mcore.py (510:515) duplicated block id: 9557 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/moe/router.py (283:288) - megatron_patch/model/qwen3_moe/moe/router.py (95:100) duplicated block id: 9558 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/moe/router.py (283:288) - megatron_patch/model/qwen3_moe/moe/router.py (52:57) duplicated block id: 9559 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/deepseek/hf2mcore_deepseek_v2_moe.py (443:448) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_mha_to_moe.py (181:186) duplicated block id: 9560 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/transformer.py (504:510) - megatron_patch/model/llama3/transformer_legacy.py (618:624) duplicated block id: 9561 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer/attention.py (528:533) - megatron_patch/model/qwen2_vl/attention.py (497:502) duplicated block id: 9562 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/transformer/mlp.py (96:101) - megatron_patch/model/mixtral_bak/transformer/attention.py (81:86) duplicated block id: 9563 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (786:795) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (978:987) duplicated block id: 9564 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (925:935) - megatron_patch/model/llama/transformer.py (836:846) duplicated block id: 9565 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/mistral/transformer.py (1737:1742) - megatron_patch/model/qwen/transformer.py (1146:1151) duplicated block id: 9566 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/falcon/transformer.py (211:217) - megatron_patch/model/llama3/transformer_legacy.py (338:344) duplicated block id: 9567 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (1135:1140) - toolkits/model_checkpoints_convertor/yi/checkpoint_reshaping_and_interoperability.py (669:674) duplicated block id: 9568 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (990:997) - megatron_patch/model/baichuan2/transformer.py (1122:1129) duplicated block id: 9569 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/mixtral_bak/transformer/attention.py (335:340) - megatron_patch/model/qwen2_vl/attention.py (678:683) duplicated block id: 9570 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (460:465) - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (816:821) duplicated block id: 9571 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (700:705) - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (373:378) duplicated block id: 9572 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (664:670) - toolkits/model_checkpoints_convertor/falcon40b/checkpoint_reshaping_and_interoperability.py (640:646) duplicated block id: 9573 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/hf2te.py (171:176) - toolkits/model_checkpoints_convertor/galactica/checkpoint_reshaping_and_interoperability.py (273:278) duplicated block id: 9574 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/galactica/checkpoint_reshaping_and_interoperability.py (459:464) - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (599:604) duplicated block id: 9575 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/transformer.py (386:391) - megatron_patch/model/starcoder/transformer.py (614:619) duplicated block id: 9576 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore.py (518:523) - toolkits/model_checkpoints_convertor/qwen/hf2megablocks_qwen1.5.py (528:533) duplicated block id: 9577 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/moe/experts.py (219:224) - megatron_patch/model/mixtral/moe/experts.py (704:709) duplicated block id: 9578 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (242:248) - megatron_patch/model/glm130b/transformer.py (217:223) duplicated block id: 9579 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (698:703) - megatron_patch/model/qwen1_5/transformer/attention.py (476:481) duplicated block id: 9580 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/llama/language_model.py (414:425) - megatron_patch/model/llama3/language_model.py (425:436) duplicated block id: 9581 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/hf2te.py (314:319) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (603:608) duplicated block id: 9582 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/transformer.py (813:818) - megatron_patch/model/qwen1_5_megablocks/transformer.py (1294:1299) duplicated block id: 9583 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (430:439) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (278:287) duplicated block id: 9584 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/language_model.py (362:367) - megatron_patch/model/chatglm/language_model.py (367:372) duplicated block id: 9585 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/llama2/language_model.py (398:403) - megatron_patch/model/qwen_vl/language_model.py (406:411) duplicated block id: 9586 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (603:608) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (520:525) duplicated block id: 9587 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (942:947) - toolkits/model_checkpoints_convertor/chatglm/checkpoint_reshaping_and_interoperability.py (583:588) duplicated block id: 9588 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/transformer/mlp.py (96:101) - megatron_patch/model/qwen1_5/transformer/attention.py (107:112) duplicated block id: 9589 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon/checkpoint_reshaping_and_interoperability.py (849:855) - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (1139:1146) duplicated block id: 9590 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/llava/language_model.py (705:711) - megatron_patch/model/starcoder/language_model.py (563:569) duplicated block id: 9591 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (700:705) - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (599:604) duplicated block id: 9592 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/chatglm/checkpoint_reshaping_and_interoperability.py (583:588) - toolkits/model_checkpoints_convertor/falcon/checkpoint_reshaping_and_interoperability.py (843:848) duplicated block id: 9593 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/galactica/checkpoint_reshaping_and_interoperability.py (459:464) - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (373:378) duplicated block id: 9594 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (588:593) - megatron_patch/model/qwen1_5/transformer/attention.py (476:481) duplicated block id: 9595 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (360:365) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (490:495) duplicated block id: 9596 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (123:130) - toolkits/model_checkpoints_convertor/qwen/hf2megablocks_qwen1.5.py (123:131) duplicated block id: 9597 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (299:308) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (490:499) duplicated block id: 9598 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (362:371) - toolkits/model_checkpoints_convertor/yi/checkpoint_reshaping_and_interoperability.py (405:414) duplicated block id: 9599 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (603:608) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (837:842) duplicated block id: 9600 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/language_model.py (423:428) - megatron_patch/model/qwen/language_model.py (362:367) duplicated block id: 9601 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/qwen3_moe/gpt_layer_specs.py (70:75) - megatron_patch/model/qwen3_moe/gpt_layer_specs.py (168:173) duplicated block id: 9602 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (1071:1076) - megatron_patch/model/mistral/transformer.py (1198:1203) duplicated block id: 9603 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/moe/experts.py (219:224) - megatron_patch/model/qwen2/moe/experts.py (324:329) duplicated block id: 9604 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (336:342) - megatron_patch/model/glm130b/transformer.py (217:223) duplicated block id: 9605 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/hf2te.py (314:319) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (360:365) duplicated block id: 9606 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/transformer.py (134:140) - megatron_patch/model/llama3/transformer_legacy.py (338:344) duplicated block id: 9607 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (990:997) - megatron_patch/model/qwen_vl/transformer.py (1119:1126) duplicated block id: 9608 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (360:365) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (751:756) duplicated block id: 9609 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/galactica/transformer.py (154:160) - megatron_patch/model/llava/transformer.py (339:345) duplicated block id: 9610 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (1516:1523) - megatron_patch/model/glm130b/transformer.py (1060:1067) duplicated block id: 9611 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon40b/checkpoint_reshaping_and_interoperability.py (682:687) - toolkits/model_checkpoints_convertor/starcoder/checkpoint_reshaping_and_interoperability.py (637:642) duplicated block id: 9612 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (147:154) - toolkits/model_checkpoints_convertor/qwen/hf2megablocks_qwen1.5.py (123:131) duplicated block id: 9613 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (240:247) - megatron_patch/model/starcoder/transformer.py (281:288) duplicated block id: 9614 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/qwen_vl/visual.py (316:321) - megatron_patch/model/qwen_vl/visual.py (329:334) duplicated block id: 9615 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer/attention.py (362:367) - megatron_patch/model/qwen2_vl/attention_vision.py (677:682) duplicated block id: 9616 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/galactica/checkpoint_reshaping_and_interoperability.py (459:464) - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (442:447) duplicated block id: 9617 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/llava_mcore/llava_spec.py (80:85) - megatron_patch/model/qwen2_moe/layer_specs.py (228:233) duplicated block id: 9618 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (196:202) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (123:129) duplicated block id: 9619 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (240:247) - megatron_patch/model/starcoder/transformer.py (404:411) duplicated block id: 9620 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer_legacy.py (1172:1179) - megatron_patch/model/qwen_vl/transformer.py (1211:1218) duplicated block id: 9621 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon/checkpoint_reshaping_and_interoperability.py (404:409) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (290:295) duplicated block id: 9622 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/bloom/gpt_model.py (101:108) - megatron_patch/model/qwen/gpt_model.py (108:115) duplicated block id: 9623 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer/attention.py (107:112) - megatron_patch/model/llama3/transformer/mlp.py (94:99) duplicated block id: 9624 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (673:678) - toolkits/model_checkpoints_convertor/falcon40b/checkpoint_reshaping_and_interoperability.py (411:416) duplicated block id: 9625 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore_llama3_1.py (724:729) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_vl.py (221:226) duplicated block id: 9626 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (1610:1615) - megatron_patch/model/qwen1_5_megablocks/transformer.py (1085:1090) duplicated block id: 9627 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/falcon/language_model.py (533:538) - megatron_patch/model/glm130b/language_model.py (478:483) duplicated block id: 9628 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/qwen2/moe/router.py (236:241) - megatron_patch/model/qwen2/moe/router.py (353:358) duplicated block id: 9629 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (1071:1076) - megatron_patch/model/qwen_vl/transformer.py (1739:1744) duplicated block id: 9630 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (1071:1076) - megatron_patch/model/llama2/transformer.py (1201:1206) duplicated block id: 9631 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/galactica/checkpoint_reshaping_and_interoperability.py (459:464) - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (703:708) duplicated block id: 9632 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (373:378) - toolkits/model_checkpoints_convertor/yi/checkpoint_reshaping_and_interoperability.py (435:440) duplicated block id: 9633 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (164:170) - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (145:151) duplicated block id: 9634 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer/attention.py (391:396) - megatron_patch/model/qwen2/transformer/attention.py (391:396) duplicated block id: 9635 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/transformer_block.py (279:284) - megatron_patch/model/qwen2/transformer_block.py (402:407) duplicated block id: 9636 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (700:705) - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (432:437) duplicated block id: 9637 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/language_model.py (408:413) - megatron_patch/model/starcoder/language_model.py (352:357) duplicated block id: 9638 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (1440:1446) - megatron_patch/model/glm130b/transformer.py (906:912) duplicated block id: 9639 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (229:234) - megatron_patch/model/llama/transformer.py (231:236) duplicated block id: 9640 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/llama2/transformer.py (707:712) - megatron_patch/model/llama3/transformer/attention.py (476:481) duplicated block id: 9641 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/multi_latent_attention.py (85:90) - megatron_patch/model/mixtral/transformer/attention.py (120:125) duplicated block id: 9642 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/qwen2/transformer/attention.py (358:363) - megatron_patch/model/qwen2_vl/attention.py (678:683) duplicated block id: 9643 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/qwen2/transformer/mlp.py (109:114) - megatron_patch/model/qwen2_vl/attention.py (122:127) duplicated block id: 9644 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon40b/checkpoint_reshaping_and_interoperability.py (682:687) - toolkits/model_checkpoints_convertor/glm/checkpoint_reshaping_and_interoperability.py (358:363) duplicated block id: 9645 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/hf2te.py (286:295) - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (299:308) duplicated block id: 9646 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (70:77) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_moe.py (142:149) duplicated block id: 9647 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (700:705) - toolkits/model_checkpoints_convertor/baichuan/hf2te.py (314:319) duplicated block id: 9648 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (183:189) - megatron_patch/model/mistral/transformer.py (339:345) duplicated block id: 9649 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/mistral/hf2mcore.py (8:19) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_dense_and_moe_gqa.py (9:22) duplicated block id: 9650 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/language_model.py (423:428) - megatron_patch/model/qwen1_5_megablocks/language_model.py (361:366) duplicated block id: 9651 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (701:712) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (943:954) duplicated block id: 9652 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/llama2/transformer.py (1100:1105) - megatron_patch/model/qwen/transformer.py (1224:1229) duplicated block id: 9653 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (1139:1146) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (1024:1031) duplicated block id: 9654 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/hf2te.py (171:176) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (341:346) duplicated block id: 9655 size: 6 cleaned lines of code in 2 files: - toolkits/sft_data_preprocessing/build_idxmap_sft_dataset.py (79:84) - toolkits/sft_data_preprocessing/build_idxmap_sft_dataset.py (97:102) duplicated block id: 9656 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (664:670) - toolkits/model_checkpoints_convertor/falcon/checkpoint_reshaping_and_interoperability.py (637:643) duplicated block id: 9657 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon40b/checkpoint_reshaping_and_interoperability.py (411:416) - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (599:604) duplicated block id: 9658 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (633:638) - megatron_patch/model/qwen1_5_megablocks/transformer.py (533:538) duplicated block id: 9659 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (919:926) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (1024:1031) duplicated block id: 9660 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llava/hf2mcore_llava.py (32:37) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2.5_vl.py (47:52) duplicated block id: 9661 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/mistral/transformer.py (765:770) - megatron_patch/model/qwen_vl/transformer.py (766:771) duplicated block id: 9662 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer_legacy.py (1240:1245) - megatron_patch/model/qwen1_5_megablocks/transformer.py (984:989) duplicated block id: 9663 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/bloom/language_model.py (368:373) - megatron_patch/model/llama3/language_model.py (347:352) duplicated block id: 9664 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (603:608) - toolkits/model_checkpoints_convertor/yi/checkpoint_reshaping_and_interoperability.py (435:440) duplicated block id: 9665 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/llava/language_model.py (414:419) - megatron_patch/model/qwen1_5_megablocks/language_model.py (397:402) duplicated block id: 9666 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon40b/checkpoint_reshaping_and_interoperability.py (411:416) - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (373:378) duplicated block id: 9667 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (691:697) - toolkits/model_checkpoints_convertor/bloom/checkpoint_reshaping_and_interoperability.py (337:343) duplicated block id: 9668 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/llava_mcore/layer_specs.py (27:40) - megatron_patch/model/llava_mcore/llava_spec.py (18:31) duplicated block id: 9669 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer/mlp.py (94:99) - megatron_patch/model/mixtral_bak/transformer/attention.py (81:86) duplicated block id: 9670 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/gpt_model.py (102:109) - megatron_patch/model/qwen/gpt_model.py (108:115) duplicated block id: 9671 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer/attention.py (107:112) - megatron_patch/model/qwen1_5/transformer/mlp.py (110:115) duplicated block id: 9672 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/bloom/language_model.py (464:469) - megatron_patch/model/falcon/language_model.py (542:547) duplicated block id: 9673 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/llava_mcore/llava_model.py (658:676) - megatron_patch/model/qwen2_5_vl/model.py (279:297) duplicated block id: 9674 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/transformer.py (204:210) - megatron_patch/model/qwen_vl/transformer.py (339:345) duplicated block id: 9675 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/transformer.py (1158:1163) - megatron_patch/model/glm130b/transformer.py (1165:1170) duplicated block id: 9676 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2.5_vl.py (566:572) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_dense_and_moe_gqa.py (508:514) duplicated block id: 9677 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (284:289) - toolkits/model_checkpoints_convertor/chatglm/checkpoint_reshaping_and_interoperability.py (256:261) duplicated block id: 9678 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (633:638) - megatron_patch/model/llava/transformer.py (641:646) duplicated block id: 9679 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/transformer.py (906:912) - megatron_patch/model/mistral/transformer.py (1437:1443) duplicated block id: 9680 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/bloom/language_model.py (368:373) - megatron_patch/model/qwen_vl/language_model.py (353:358) duplicated block id: 9681 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/bloom/gpt_model.py (89:96) - megatron_patch/model/falcon40b/gpt_model.py (94:100) duplicated block id: 9682 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/deepseek/hf2mcore_deepseek_v2_moe.py (230:235) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_dense_and_moe_gqa.py (255:260) duplicated block id: 9683 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/hf2te.py (286:291) - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (373:378) duplicated block id: 9684 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/falcon/transformer.py (583:589) - megatron_patch/model/glm130b/transformer.py (438:445) duplicated block id: 9685 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (615:622) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (543:550) duplicated block id: 9686 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer/attention.py (391:396) - megatron_patch/model/mixtral/transformer/attention.py (530:535) duplicated block id: 9687 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/falcon/transformer.py (884:891) - megatron_patch/model/llama3/transformer_legacy.py (856:863) duplicated block id: 9688 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/transformer.py (438:445) - megatron_patch/model/llama/transformer.py (393:400) duplicated block id: 9689 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/language_model.py (722:728) - megatron_patch/model/glm130b/language_model.py (623:630) duplicated block id: 9690 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (998:1004) - megatron_patch/model/glm130b/transformer.py (924:930) duplicated block id: 9691 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/chatglm/checkpoint_reshaping_and_interoperability.py (583:588) - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (1123:1128) duplicated block id: 9692 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/llava/transformer.py (1281:1286) - megatron_patch/model/qwen1_5_megablocks/transformer.py (984:989) duplicated block id: 9693 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/qwen2_5_vl/transformer_block.py (347:352) - megatron_patch/model/qwen2_5_vl/transformer_block.py (357:362) duplicated block id: 9694 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (1276:1281) - megatron_patch/model/llama/transformer.py (868:873) duplicated block id: 9695 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/deepseek/hf2mcore_deepseek_v3_moe.py (550:558) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2.5_vl.py (572:581) duplicated block id: 9696 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/deepseek/hf2mcore_deepseek_v3_moe.py (550:558) - toolkits/model_checkpoints_convertor/mistral/hf2mcore.py (380:388) duplicated block id: 9697 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/mistral/hf2mcore.py (97:105) - toolkits/model_checkpoints_convertor/yi/checkpoint_reshaping_and_interoperability.py (124:131) duplicated block id: 9698 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (703:708) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (837:842) duplicated block id: 9699 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/llama3/language_model.py (347:352) - megatron_patch/model/starcoder/language_model.py (352:357) duplicated block id: 9700 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/reward_model_to_megatron.py (490:497) - toolkits/model_checkpoints_convertor/falcon40b/checkpoint_reshaping_and_interoperability.py (774:781) duplicated block id: 9701 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/transformer_block.py (300:305) - megatron_patch/model/deepseek_v2/transformer_block.py (310:315) duplicated block id: 9702 size: 6 cleaned lines of code in 2 files: - megatron_patch/arguments.py (418:423) - toolkits/model_checkpoints_convertor/falcon/checkpoint_reshaping_and_interoperability.py (37:42) duplicated block id: 9703 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/transformer_block.py (212:217) - megatron_patch/model/qwen2/transformer_block.py (142:147) duplicated block id: 9704 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (1740:1745) - megatron_patch/model/mistral/transformer.py (1198:1203) duplicated block id: 9705 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/hf2te.py (258:267) - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (299:308) duplicated block id: 9706 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/falcon/language_model.py (356:361) - megatron_patch/model/qwen1_5_megablocks/language_model.py (351:356) duplicated block id: 9707 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/chatglm/checkpoint_reshaping_and_interoperability.py (256:261) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (341:346) duplicated block id: 9708 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (673:678) - toolkits/model_checkpoints_convertor/galactica/checkpoint_reshaping_and_interoperability.py (459:464) duplicated block id: 9709 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/language_model.py (423:428) - megatron_patch/model/mistral/language_model.py (364:369) duplicated block id: 9710 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer/attention.py (528:533) - megatron_patch/model/mixtral/transformer/attention.py (496:501) duplicated block id: 9711 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/transformer.py (217:223) - megatron_patch/model/qwen1_5_megablocks/transformer.py (230:236) duplicated block id: 9712 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/qwen/transformer.py (1251:1264) - megatron_patch/model/starcoder/transformer.py (1008:1021) duplicated block id: 9713 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/bloom/gpt_model.py (101:108) - megatron_patch/model/falcon40b/gpt_model.py (107:114) duplicated block id: 9714 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/qwen_vl/transformer.py (1099:1104) - megatron_patch/model/qwen_vl/transformer.py (1281:1286) duplicated block id: 9715 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/transformer/attention.py (677:682) - megatron_patch/model/mixtral_bak/transformer/attention.py (335:340) duplicated block id: 9716 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/language_model.py (356:361) - megatron_patch/model/qwen/language_model.py (352:357) duplicated block id: 9717 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (513:518) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (360:365) duplicated block id: 9718 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/llama/positional_embeddings.py (60:65) - megatron_patch/model/qwen1_5_megablocks/rotary_pos_embedding.py (44:49) duplicated block id: 9719 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/galactica/language_model.py (387:392) - megatron_patch/model/llama2/language_model.py (352:357) duplicated block id: 9720 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/deepseek/hf2mcore_deepseek_v2_moe.py (146:152) - toolkits/model_checkpoints_convertor/deepseek/hf2mcore_deepseek_v2_moe.py (405:411) duplicated block id: 9721 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/transformer.py (890:895) - megatron_patch/model/mistral/transformer.py (1406:1411) duplicated block id: 9722 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/checkpoint_reshaping_and_interoperability.py (337:343) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (828:834) duplicated block id: 9723 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (513:518) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (603:608) duplicated block id: 9724 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/transformer.py (1046:1053) - megatron_patch/model/starcoder/transformer.py (1351:1358) duplicated block id: 9725 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/transformer_block.py (279:284) - megatron_patch/model/qwen2_5_vl/transformer_block.py (563:568) duplicated block id: 9726 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/language_model.py (49:54) - megatron_patch/model/qwen/language_model.py (47:52) duplicated block id: 9727 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer_legacy.py (1240:1245) - megatron_patch/model/qwen_vl/transformer.py (1099:1104) duplicated block id: 9728 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/moe/moe_layer.py (58:63) - megatron_patch/model/qwen1_5/moe/moe_layer.py (48:53) duplicated block id: 9729 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (899:908) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (1091:1100) duplicated block id: 9730 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer/attention.py (384:389) - megatron_patch/model/qwen2_vl/attention_vision.py (520:525) duplicated block id: 9731 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (990:997) - megatron_patch/model/qwen1_5_megablocks/transformer.py (1004:1011) duplicated block id: 9732 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/hf2te.py (261:270) - toolkits/model_checkpoints_convertor/chatglm/checkpoint_reshaping_and_interoperability.py (330:339) duplicated block id: 9733 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore_llama3_1.py (724:729) - toolkits/model_checkpoints_convertor/llava/hf2mcore_llava.py (454:459) duplicated block id: 9734 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/transformer_config.py (27:32) - megatron_patch/model/qwen2_moe/transformer_config.py (47:52) duplicated block id: 9735 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (830:837) - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (919:926) duplicated block id: 9736 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/llama2/transformer.py (1282:1287) - megatron_patch/model/llama3/transformer_legacy.py (1050:1055) duplicated block id: 9737 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/llama2/transformer.py (1212:1219) - megatron_patch/model/qwen/transformer.py (1156:1163) duplicated block id: 9738 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (707:712) - megatron_patch/model/starcoder/transformer.py (917:922) duplicated block id: 9739 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (290:295) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (472:477) duplicated block id: 9740 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/mistral/transformer.py (706:711) - megatron_patch/model/qwen2/transformer/attention.py (478:483) duplicated block id: 9741 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/galactica/transformer.py (35:43) - megatron_patch/model/glm130b/transformer.py (40:48) duplicated block id: 9742 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (1309:1322) - megatron_patch/model/starcoder/transformer.py (1008:1021) duplicated block id: 9743 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/falcon/language_model.py (197:203) - megatron_patch/model/llama3/language_model.py (173:179) duplicated block id: 9744 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon/checkpoint_reshaping_and_interoperability.py (697:702) - toolkits/model_checkpoints_convertor/starcoder/checkpoint_reshaping_and_interoperability.py (637:642) duplicated block id: 9745 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/galactica/language_model.py (445:456) - megatron_patch/model/qwen1_5_megablocks/language_model.py (433:444) duplicated block id: 9746 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (698:703) - megatron_patch/model/qwen2/transformer/attention.py (478:483) duplicated block id: 9747 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (635:641) - megatron_patch/model/mistral/transformer.py (918:925) duplicated block id: 9748 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (1016:1021) - megatron_patch/model/chatglm/transformer.py (742:747) duplicated block id: 9749 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (211:219) - megatron_patch/model/glm130b/transformer.py (245:253) duplicated block id: 9750 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/mistral/hf2mcore.py (519:526) - toolkits/model_checkpoints_convertor/qwen/hf2megablocks_qwen1.5.py (645:651) duplicated block id: 9751 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2.5_vl.py (346:351) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_vl.py (237:242) duplicated block id: 9752 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/llama/transformer.py (192:198) - megatron_patch/model/qwen1_5_megablocks/transformer.py (230:236) duplicated block id: 9753 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/llama/positional_embeddings.py (39:44) - megatron_patch/model/qwen1_5_megablocks/rotary_pos_embedding.py (65:70) duplicated block id: 9754 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/falcon/gpt_model.py (107:114) - megatron_patch/model/glm130b/gpt_model.py (100:107) duplicated block id: 9755 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/galactica/transformer.py (154:160) - megatron_patch/model/llama2/transformer.py (339:345) duplicated block id: 9756 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer/attention.py (476:481) - megatron_patch/model/qwen1_5_megablocks/transformer.py (598:603) duplicated block id: 9757 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon/checkpoint_reshaping_and_interoperability.py (637:643) - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (807:813) duplicated block id: 9758 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (633:638) - megatron_patch/model/mistral/transformer.py (641:646) duplicated block id: 9759 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/llama2/transformer.py (707:712) - megatron_patch/model/qwen2_vl/attention.py (620:625) duplicated block id: 9760 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/language_model.py (140:158) - megatron_patch/model/galactica/language_model.py (166:183) duplicated block id: 9761 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/falcon/transformer.py (647:653) - megatron_patch/model/glm130b/transformer.py (485:492) duplicated block id: 9762 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/hf2te.py (286:291) - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (599:604) duplicated block id: 9763 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/transformer.py (425:430) - megatron_patch/model/falcon40b/transformer.py (534:539) duplicated block id: 9764 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (299:308) - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (483:492) duplicated block id: 9765 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/falcon/language_model.py (414:425) - megatron_patch/model/qwen/language_model.py (431:442) duplicated block id: 9766 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon40b/checkpoint_reshaping_and_interoperability.py (682:687) - toolkits/model_checkpoints_convertor/glm130b/checkpoint_reshaping_and_interoperability.py (353:358) duplicated block id: 9767 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/language_model.py (623:630) - megatron_patch/model/mistral/language_model.py (657:663) duplicated block id: 9768 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (38:46) - megatron_patch/model/glm130b/transformer.py (40:48) duplicated block id: 9769 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (1317:1322) - megatron_patch/model/glm130b/transformer.py (916:921) duplicated block id: 9770 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/transformer.py (813:818) - megatron_patch/model/mistral/transformer.py (1406:1411) duplicated block id: 9771 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/mixtral_bak/transformer/attention.py (335:340) - megatron_patch/model/qwen2_vl/attention_vision.py (677:682) duplicated block id: 9772 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/transformer.py (510:518) - megatron_patch/model/falcon/transformer.py (865:872) duplicated block id: 9773 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (414:419) - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (374:379) duplicated block id: 9774 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/galactica/checkpoint_reshaping_and_interoperability.py (459:464) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (360:365) duplicated block id: 9775 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/language_model.py (486:491) - megatron_patch/model/llama/language_model.py (542:547) duplicated block id: 9776 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/language_model.py (722:728) - megatron_patch/model/bloom/language_model.py (593:599) duplicated block id: 9777 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/chatglm/checkpoint_reshaping_and_interoperability.py (407:414) - toolkits/model_checkpoints_convertor/falcon40b/checkpoint_reshaping_and_interoperability.py (420:427) duplicated block id: 9778 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore_llama3_1.py (648:653) - toolkits/model_checkpoints_convertor/llava/hf2mcore_llava.py (530:535) duplicated block id: 9779 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (1083:1092) - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (899:908) duplicated block id: 9780 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/gpt_model.py (102:109) - megatron_patch/model/llava/gpt_model.py (109:116) duplicated block id: 9781 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/language_model.py (423:428) - megatron_patch/model/baichuan2/language_model.py (372:377) duplicated block id: 9782 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/multi_latent_attention.py (85:90) - megatron_patch/model/qwen2_vl/attention.py (106:111) duplicated block id: 9783 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (902:911) - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (899:908) duplicated block id: 9784 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/llama3/layer_specs.py (69:74) - megatron_patch/model/mixtral/layer_specs.py (114:119) duplicated block id: 9785 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/transformer.py (645:653) - megatron_patch/model/qwen1_5_megablocks/transformer.py (777:784) duplicated block id: 9786 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/transformer/attention.py (695:700) - megatron_patch/model/mixtral_bak/transformer/attention.py (348:353) duplicated block id: 9787 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/galactica/language_model.py (166:183) - megatron_patch/model/qwen1_5_megablocks/language_model.py (129:147) duplicated block id: 9788 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/deepseek/hf2mcore_deepseek_v2_moe.py (420:427) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_moe.py (585:590) duplicated block id: 9789 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon/checkpoint_reshaping_and_interoperability.py (124:131) - toolkits/model_checkpoints_convertor/mistral/hf2mcore.py (97:105) duplicated block id: 9790 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/language_model.py (362:367) - megatron_patch/model/glm130b/language_model.py (370:375) duplicated block id: 9791 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (673:678) - toolkits/model_checkpoints_convertor/falcon/checkpoint_reshaping_and_interoperability.py (422:427) duplicated block id: 9792 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/transformer.py (645:653) - megatron_patch/model/llama3/transformer_legacy.py (837:844) duplicated block id: 9793 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/falcon/language_model.py (356:361) - megatron_patch/model/qwen/language_model.py (352:357) duplicated block id: 9794 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/falcon/transformer.py (865:872) - megatron_patch/model/falcon40b/transformer.py (645:653) duplicated block id: 9795 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/layer_specs.py (84:89) - megatron_patch/model/qwen3_moe/gpt_layer_specs.py (109:114) duplicated block id: 9796 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/transformer.py (181:188) - megatron_patch/model/qwen/transformer.py (155:162) duplicated block id: 9797 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/moe/experts.py (301:306) - megatron_patch/model/mixtral/moe/experts.py (287:292) duplicated block id: 9798 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/moe/experts.py (301:306) - megatron_patch/model/mixtral/moe/experts.py (274:279) duplicated block id: 9799 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/mixtral_bak/moe/experts.py (51:57) - megatron_patch/model/qwen2/moe/experts.py (73:79) duplicated block id: 9800 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (172:178) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (123:129) duplicated block id: 9801 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/galactica/language_model.py (166:183) - megatron_patch/model/llava/language_model.py (133:151) duplicated block id: 9802 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/transformer.py (1140:1145) - megatron_patch/model/glm130b/transformer.py (1165:1170) duplicated block id: 9803 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/transformer.py (1140:1145) - megatron_patch/model/glm130b/transformer.py (1158:1163) duplicated block id: 9804 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/transformer.py (1140:1145) - megatron_patch/model/glm130b/transformer.py (1149:1154) duplicated block id: 9805 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llava/hf2mcore_llava.py (32:37) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_vl.py (47:52) duplicated block id: 9806 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore.py (381:386) - toolkits/model_checkpoints_convertor/llama/hf2mcore.py (388:393) duplicated block id: 9807 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (286:291) - toolkits/model_checkpoints_convertor/baichuan2/hf2te.py (171:176) duplicated block id: 9808 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/qwen1_5/transformer/attention.py (391:396) - megatron_patch/model/qwen2_vl/attention_vision.py (530:535) duplicated block id: 9809 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/qwen_vl/transformer.py (1308:1321) - megatron_patch/model/starcoder/transformer.py (1008:1021) duplicated block id: 9810 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (1740:1745) - megatron_patch/model/llava/transformer.py (1200:1205) duplicated block id: 9811 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/llama/transformer.py (868:873) - megatron_patch/model/llama2/transformer.py (1409:1414) duplicated block id: 9812 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (460:465) - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (673:678) duplicated block id: 9813 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (483:492) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (278:287) duplicated block id: 9814 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (183:189) - megatron_patch/model/qwen_vl/transformer.py (339:345) duplicated block id: 9815 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/reward_model_to_megatron.py (623:628) - toolkits/model_checkpoints_convertor/starcoder/checkpoint_reshaping_and_interoperability.py (401:406) duplicated block id: 9816 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/language_model.py (367:372) - megatron_patch/model/qwen1_5_megablocks/language_model.py (351:356) duplicated block id: 9817 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/moe/experts.py (703:708) - megatron_patch/model/deepseek_v2/moe/experts.py (797:802) duplicated block id: 9818 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/transformer.py (81:86) - megatron_patch/model/llama/transformer.py (132:137) duplicated block id: 9819 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (698:703) - megatron_patch/model/mixtral_bak/transformer/attention.py (366:371) duplicated block id: 9820 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/galactica/checkpoint_reshaping_and_interoperability.py (459:464) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (603:608) duplicated block id: 9821 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer/mlp.py (94:99) - megatron_patch/model/mixtral/transformer/attention.py (135:140) duplicated block id: 9822 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/language_model.py (414:425) - megatron_patch/model/llama2/language_model.py (434:445) duplicated block id: 9823 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (633:638) - megatron_patch/model/llama3/transformer_legacy.py (620:625) duplicated block id: 9824 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/gpt_model.py (100:107) - megatron_patch/model/llama/gpt_model.py (107:114) duplicated block id: 9825 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/transformer.py (645:653) - megatron_patch/model/mistral/transformer.py (883:890) duplicated block id: 9826 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/llama/positional_embeddings.py (60:65) - megatron_patch/model/llava/rotary_pos_embedding.py (42:47) duplicated block id: 9827 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/transformer.py (735:741) - megatron_patch/model/llama3/transformer_legacy.py (872:878) duplicated block id: 9828 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore.py (10:25) - toolkits/model_checkpoints_convertor/mistral/hf2mcore.py (8:19) duplicated block id: 9829 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/transformer.py (735:741) - megatron_patch/model/starcoder/transformer.py (849:855) duplicated block id: 9830 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/reward_model_to_megatron.py (337:343) - toolkits/model_checkpoints_convertor/starcoder/checkpoint_reshaping_and_interoperability.py (361:367) duplicated block id: 9831 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/glm130b/checkpoint_reshaping_and_interoperability.py (314:321) - toolkits/model_checkpoints_convertor/starcoder/checkpoint_reshaping_and_interoperability.py (599:606) duplicated block id: 9832 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (613:621) - megatron_patch/model/mistral/transformer.py (883:890) duplicated block id: 9833 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/qwen1_5/transformer/attention.py (107:112) - megatron_patch/model/qwen2/transformer/mlp.py (109:114) duplicated block id: 9834 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/gpt_model.py (100:107) - megatron_patch/model/llama2/gpt_model.py (108:115) duplicated block id: 9835 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (633:638) - megatron_patch/model/qwen_vl/transformer.py (641:646) duplicated block id: 9836 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/deepseek/hf2mcore_deepseek_v2_moe.py (22:34) - toolkits/model_checkpoints_convertor/llama/hf2mcore.py (10:25) duplicated block id: 9837 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/layer_specs.py (114:119) - megatron_patch/model/qwen2/layer_specs.py (79:84) duplicated block id: 9838 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/deepseek/hf2mcore_deepseek_v3_moe.py (207:212) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_moe.py (579:584) duplicated block id: 9839 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/transformer.py (645:653) - megatron_patch/model/llama2/transformer.py (886:893) duplicated block id: 9840 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/moe/moe_layer.py (120:125) - megatron_patch/model/qwen2/moe/moe_layer.py (120:125) duplicated block id: 9841 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/galactica/checkpoint_reshaping_and_interoperability.py (123:130) - toolkits/model_checkpoints_convertor/qwen/hf2megablocks_qwen1.5.py (123:131) duplicated block id: 9842 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/transformer.py (735:741) - megatron_patch/model/llava/transformer.py (920:927) duplicated block id: 9843 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/transformer.py (204:210) - megatron_patch/model/llama3/transformer_legacy.py (338:344) duplicated block id: 9844 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer_legacy.py (1050:1055) - megatron_patch/model/qwen_vl/transformer.py (1281:1286) duplicated block id: 9845 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/moe/experts.py (797:802) - megatron_patch/model/mixtral/moe/experts.py (704:709) duplicated block id: 9846 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/qwen2/layer_specs.py (52:57) - megatron_patch/model/qwen2_vl/layer_specs.py (76:81) duplicated block id: 9847 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/llama2/transformer.py (1282:1287) - megatron_patch/model/mistral/transformer.py (1097:1102) duplicated block id: 9848 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (70:77) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2.5_vl.py (78:85) duplicated block id: 9849 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/deepseek/hf2mcore_deepseek_v3_moe.py (701:708) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_moe.py (686:693) duplicated block id: 9850 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (299:304) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (341:346) duplicated block id: 9851 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (374:379) - toolkits/model_checkpoints_convertor/yi/checkpoint_reshaping_and_interoperability.py (417:422) duplicated block id: 9852 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/language_model.py (408:413) - megatron_patch/model/galactica/language_model.py (387:392) duplicated block id: 9853 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/deepseek/hf2mcore_deepseek_v2_moe.py (22:34) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_mha.py (6:18) duplicated block id: 9854 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/transformer.py (645:653) - megatron_patch/model/qwen/transformer.py (831:838) duplicated block id: 9855 size: 6 cleaned lines of code in 2 files: - megatron_patch/arguments.py (418:423) - toolkits/model_checkpoints_convertor/yi/checkpoint_reshaping_and_interoperability.py (37:42) duplicated block id: 9856 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon40b/checkpoint_reshaping_and_interoperability.py (825:832) - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (1139:1146) duplicated block id: 9857 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/llava_mcore/llava_spec.py (66:71) - megatron_patch/model/mixtral_bak/layer_specs.py (92:97) duplicated block id: 9858 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/hf2te.py (286:295) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (278:287) duplicated block id: 9859 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/transformer.py (695:700) - megatron_patch/model/qwen_vl/transformer.py (1408:1413) duplicated block id: 9860 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/deepseek/hf2mcore_deepseek_v3_moe.py (23:34) - toolkits/model_checkpoints_convertor/mistral/hf2mcore.py (8:19) duplicated block id: 9861 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/checkpoint_reshaping_and_interoperability.py (700:705) - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (513:518) duplicated block id: 9862 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon/checkpoint_reshaping_and_interoperability.py (261:277) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (338:354) duplicated block id: 9863 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (483:492) - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (362:371) duplicated block id: 9864 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/language_model.py (49:54) - megatron_patch/model/mistral/language_model.py (48:53) duplicated block id: 9865 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/llava/transformer.py (1099:1104) - megatron_patch/model/qwen/transformer.py (1224:1229) duplicated block id: 9866 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (183:189) - megatron_patch/model/qwen/transformer.py (333:339) duplicated block id: 9867 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (599:604) - toolkits/model_checkpoints_convertor/yi/checkpoint_reshaping_and_interoperability.py (435:440) duplicated block id: 9868 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/transformer.py (37:45) - megatron_patch/model/glm130b/transformer.py (40:48) duplicated block id: 9869 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (738:745) - megatron_patch/model/galactica/transformer.py (542:550) duplicated block id: 9870 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/galactica/checkpoint_reshaping_and_interoperability.py (653:658) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (1135:1140) duplicated block id: 9871 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/llama2/language_model.py (352:357) - megatron_patch/model/starcoder/language_model.py (352:357) duplicated block id: 9872 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/transformer.py (504:510) - megatron_patch/model/mistral/transformer.py (639:645) duplicated block id: 9873 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (787:792) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (800:805) duplicated block id: 9874 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer/attention.py (476:481) - megatron_patch/model/qwen/transformer.py (680:685) duplicated block id: 9875 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/qwen2/layer_specs.py (25:31) - megatron_patch/model/qwen3_moe/moe_module_specs.py (18:24) duplicated block id: 9876 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (698:703) - megatron_patch/model/mixtral/transformer/attention.py (619:624) duplicated block id: 9877 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/transformer.py (204:210) - megatron_patch/model/llama2/transformer.py (339:345) duplicated block id: 9878 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (1102:1107) - megatron_patch/model/qwen1_5_megablocks/transformer.py (1166:1171) duplicated block id: 9879 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (1071:1076) - megatron_patch/model/mistral/transformer.py (1737:1742) duplicated block id: 9880 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer_legacy.py (1207:1212) - megatron_patch/model/starcoder/transformer.py (962:967) duplicated block id: 9881 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (922:928) - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (1139:1146) duplicated block id: 9882 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/qwen1_5/transformer/attention.py (528:533) - megatron_patch/model/qwen2_vl/attention.py (497:502) duplicated block id: 9883 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (673:678) - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (599:604) duplicated block id: 9884 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/llava/language_model.py (355:360) - megatron_patch/model/starcoder/language_model.py (352:357) duplicated block id: 9885 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/galactica/language_model.py (166:183) - megatron_patch/model/mistral/language_model.py (131:149) duplicated block id: 9886 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/mistral/language_model.py (353:358) - megatron_patch/model/starcoder/language_model.py (352:357) duplicated block id: 9887 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (442:447) - toolkits/model_checkpoints_convertor/yi/checkpoint_reshaping_and_interoperability.py (435:440) duplicated block id: 9888 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/llama2/transformer.py (1100:1105) - megatron_patch/model/llama2/transformer.py (1282:1287) duplicated block id: 9889 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (673:678) - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (373:378) duplicated block id: 9890 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/transformer_block.py (497:502) - megatron_patch/model/qwen2_5_vl/transformer_block.py (325:330) duplicated block id: 9891 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (432:437) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (837:842) duplicated block id: 9892 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/language_model.py (533:538) - megatron_patch/model/glm130b/language_model.py (478:483) duplicated block id: 9893 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer/attention.py (107:112) - megatron_patch/model/mixtral_bak/transformer/mlp.py (72:77) duplicated block id: 9894 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/gpt_model.py (104:110) - megatron_patch/model/starcoder/gpt_model.py (108:114) duplicated block id: 9895 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (373:378) - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (599:604) duplicated block id: 9896 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer/attention.py (476:481) - megatron_patch/model/qwen_vl/transformer.py (706:711) duplicated block id: 9897 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (624:634) - megatron_patch/model/falcon/transformer.py (891:901) duplicated block id: 9898 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (816:821) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (360:365) duplicated block id: 9899 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (123:130) - toolkits/model_checkpoints_convertor/mistral/hf2mcore.py (97:105) duplicated block id: 9900 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/glm/checkpoint_reshaping_and_interoperability.py (319:326) - toolkits/model_checkpoints_convertor/starcoder/checkpoint_reshaping_and_interoperability.py (599:606) duplicated block id: 9901 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/falcon/checkpoint_reshaping_and_interoperability.py (697:702) - toolkits/model_checkpoints_convertor/glm/checkpoint_reshaping_and_interoperability.py (358:363) duplicated block id: 9902 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/galactica/checkpoint_reshaping_and_interoperability.py (459:464) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (751:756) duplicated block id: 9903 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (599:604) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (520:525) duplicated block id: 9904 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer/attention.py (476:481) - megatron_patch/model/llava/transformer.py (706:711) duplicated block id: 9905 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/galactica/checkpoint_reshaping_and_interoperability.py (653:658) - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (1123:1128) duplicated block id: 9906 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/mixtral_bak/transformer/mlp.py (170:177) - megatron_patch/model/qwen1_5/transformer/mlp.py (213:220) duplicated block id: 9907 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/deepseek/hf2mcore_deepseek_v3_moe.py (207:212) - toolkits/model_checkpoints_convertor/deepseek/hf2mcore_deepseek_v3_moe.py (579:584) duplicated block id: 9908 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/transformer.py (37:45) - megatron_patch/model/falcon40b/transformer.py (33:41) duplicated block id: 9909 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/transformer.py (695:700) - megatron_patch/model/llama3/transformer_legacy.py (1367:1372) duplicated block id: 9910 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (405:412) - megatron_patch/model/starcoder/transformer.py (537:544) duplicated block id: 9911 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (441:448) - toolkits/model_checkpoints_convertor/chatglm/checkpoint_reshaping_and_interoperability.py (407:414) duplicated block id: 9912 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (599:604) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (837:842) duplicated block id: 9913 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/hf2te.py (258:267) - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (362:371) duplicated block id: 9914 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/multi_latent_attention.py (101:106) - megatron_patch/model/qwen1_5/transformer/mlp.py (110:115) duplicated block id: 9915 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/deepseek/hf2mcore_deepseek_v3_moe.py (550:558) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_vl.py (579:588) duplicated block id: 9916 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/galactica/checkpoint_reshaping_and_interoperability.py (441:446) - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (374:379) duplicated block id: 9917 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore.py (584:589) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_moe.py (388:393) duplicated block id: 9918 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/deepseek/hf2mcore_deepseek_v3_moe.py (491:496) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_moe.py (497:502) duplicated block id: 9919 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (919:926) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (1141:1148) duplicated block id: 9920 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/galactica/language_model.py (559:564) - megatron_patch/model/glm130b/language_model.py (478:483) duplicated block id: 9921 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/llama3/layer_specs.py (69:74) - megatron_patch/model/mixtral_bak/layer_specs.py (92:97) duplicated block id: 9922 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/mistral/hf2mcore.py (97:105) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (152:159) duplicated block id: 9923 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (673:678) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (520:525) duplicated block id: 9924 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/transformer_block.py (270:277) - megatron_patch/model/qwen2_5_vl/transformer_block.py (312:319) duplicated block id: 9925 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/llama3/language_model.py (393:401) - megatron_patch/model/starcoder/language_model.py (378:386) duplicated block id: 9926 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/transformer.py (204:210) - megatron_patch/model/qwen1_5_megablocks/transformer.py (230:236) duplicated block id: 9927 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/deepseek/hf2mcore_deepseek_v2_moe.py (443:448) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_dense_and_moe_gqa.py (548:553) duplicated block id: 9928 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer_legacy.py (685:690) - megatron_patch/model/qwen1_5/transformer/attention.py (476:481) duplicated block id: 9929 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2.5_vl.py (331:336) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_dense_and_moe_gqa.py (352:357) duplicated block id: 9930 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (876:885) - toolkits/model_checkpoints_convertor/mistral/hf2mcore_mixtral.py (899:908) duplicated block id: 9931 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2.5_vl.py (237:242) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_vl.py (343:348) duplicated block id: 9932 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/llama/transformer.py (795:800) - megatron_patch/model/llama/transformer.py (1047:1052) duplicated block id: 9933 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/chatglm/checkpoint_reshaping_and_interoperability.py (583:588) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (1135:1140) duplicated block id: 9934 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/language_model.py (441:452) - megatron_patch/model/galactica/language_model.py (445:456) duplicated block id: 9935 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/llama3_1/model.py (113:118) - megatron_patch/model/mixtral/model.py (92:97) duplicated block id: 9936 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/transformer.py (735:741) - megatron_patch/model/qwen/transformer.py (866:873) duplicated block id: 9937 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/llava_mcore/llava_spec.py (80:85) - megatron_patch/model/qwen3_moe/gpt_layer_specs.py (249:254) duplicated block id: 9938 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/chatglm/checkpoint_reshaping_and_interoperability.py (407:414) - toolkits/model_checkpoints_convertor/falcon/checkpoint_reshaping_and_interoperability.py (431:438) duplicated block id: 9939 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/starcoder/transformer.py (958:963) - megatron_patch/model/starcoder/transformer.py (984:989) duplicated block id: 9940 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/qwen1_5/transformer/attention.py (391:396) - megatron_patch/model/qwen2/transformer/attention.py (391:396) duplicated block id: 9941 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (211:219) - megatron_patch/model/falcon40b/transformer.py (227:235) duplicated block id: 9942 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/moe/experts.py (704:709) - megatron_patch/model/qwen1_5/moe/experts.py (220:225) duplicated block id: 9943 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/transformer.py (217:223) - megatron_patch/model/llama2/transformer.py (339:345) duplicated block id: 9944 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/llama2/transformer.py (1309:1322) - megatron_patch/model/starcoder/transformer.py (1008:1021) duplicated block id: 9945 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/qwen1_5/moe/experts.py (220:225) - megatron_patch/model/qwen2/moe/experts.py (324:329) duplicated block id: 9946 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/chatglm/gpt_model.py (102:109) - megatron_patch/model/falcon/gpt_model.py (107:114) duplicated block id: 9947 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/chatglm/checkpoint_reshaping_and_interoperability.py (256:261) - toolkits/model_checkpoints_convertor/yi/checkpoint_reshaping_and_interoperability.py (271:276) duplicated block id: 9948 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer_legacy.py (685:690) - megatron_patch/model/qwen2_vl/attention_vision.py (619:624) duplicated block id: 9949 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/hf2te.py (172:177) - toolkits/model_checkpoints_convertor/baichuan/hf2te.py (187:192) duplicated block id: 9950 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/transformer.py (813:818) - megatron_patch/model/llava/transformer.py (1408:1413) duplicated block id: 9951 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/falcon/transformer.py (211:217) - megatron_patch/model/mistral/transformer.py (339:345) duplicated block id: 9952 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/mixtral_bak/moe/token_dispatcher.py (246:252) - megatron_patch/model/qwen2/moe/token_dispatcher.py (215:221) duplicated block id: 9953 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/llava/transformer.py (1739:1744) - megatron_patch/model/qwen/transformer.py (1146:1151) duplicated block id: 9954 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (816:821) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_dense_gqa.py (603:608) duplicated block id: 9955 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/llama3_1/layer_specs.py (69:74) - megatron_patch/model/mixtral/layer_specs.py (114:119) duplicated block id: 9956 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/language_model.py (492:503) - megatron_patch/model/galactica/language_model.py (445:456) duplicated block id: 9957 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (816:821) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.5.py (520:525) duplicated block id: 9958 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/multi_latent_attention.py (85:90) - megatron_patch/model/qwen1_5/transformer/attention.py (93:98) duplicated block id: 9959 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/falcon/transformer.py (1253:1258) - megatron_patch/model/llama/transformer.py (795:800) duplicated block id: 9960 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/moe/router.py (155:160) - megatron_patch/model/qwen3_moe/moe/router.py (156:161) duplicated block id: 9961 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/language_model.py (370:375) - megatron_patch/model/llava/language_model.py (355:360) duplicated block id: 9962 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (1357:1362) - megatron_patch/model/baichuan2/transformer.py (1490:1495) duplicated block id: 9963 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/llava_mcore/llava_model.py (256:261) - megatron_patch/model/llava_mcore/llava_model.py (631:636) duplicated block id: 9964 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/bloom/gpt_model.py (101:108) - megatron_patch/model/falcon/gpt_model.py (107:114) duplicated block id: 9965 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore.py (500:505) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_dense_and_moe_gqa.py (546:551) duplicated block id: 9966 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/llama3/transformer/attention.py (384:389) - megatron_patch/model/qwen2_vl/attention.py (521:526) duplicated block id: 9967 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan/hf2te.py (314:319) - toolkits/model_checkpoints_convertor/baichuan2/checkpoint_reshaping_and_interoperability.py (673:678) duplicated block id: 9968 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/bloom/checkpoint_reshaping_and_interoperability.py (379:384) - toolkits/model_checkpoints_convertor/starcoder/checkpoint_reshaping_and_interoperability.py (637:642) duplicated block id: 9969 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_dense_and_moe_gqa.py (426:431) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_vl.py (221:226) duplicated block id: 9970 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/mistral/hf2mcore.py (350:355) - toolkits/model_checkpoints_convertor/qwen/hf2megablocks_qwen1.5.py (446:467) duplicated block id: 9971 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/llava_mcore/llava_model.py (658:676) - megatron_patch/model/qwen2_vl/model.py (279:297) duplicated block id: 9972 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore.py (766:774) - toolkits/model_checkpoints_convertor/mistral/hf2mcore.py (519:526) duplicated block id: 9973 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/galactica/transformer.py (542:550) - megatron_patch/model/llava/transformer.py (885:892) duplicated block id: 9974 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/mixtral_bak/transformer/attention.py (348:353) - megatron_patch/model/mixtral_bak/transformer/attention.py (431:436) duplicated block id: 9975 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/transformer.py (812:818) - megatron_patch/model/starcoder/transformer.py (930:936) duplicated block id: 9976 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/mixtral_bak/transformer/attention.py (348:353) - megatron_patch/model/mixtral_bak/transformer/attention.py (443:448) duplicated block id: 9977 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/galactica/transformer.py (670:680) - megatron_patch/model/starcoder/transformer.py (1088:1098) duplicated block id: 9978 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/bloom/language_model.py (593:599) - megatron_patch/model/llama2/language_model.py (645:651) duplicated block id: 9979 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/glm130b/transformer.py (924:930) - megatron_patch/model/starcoder/transformer.py (1159:1165) duplicated block id: 9980 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/falcon40b/language_model.py (356:361) - megatron_patch/model/llama2/language_model.py (352:357) duplicated block id: 9981 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (1102:1107) - megatron_patch/model/qwen/transformer.py (1224:1229) duplicated block id: 9982 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/bloom/gpt_model.py (101:108) - megatron_patch/model/llama3/gpt_model.py (110:117) duplicated block id: 9983 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/falcon/transformer.py (1074:1079) - megatron_patch/model/mistral/transformer.py (1406:1411) duplicated block id: 9984 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (156:163) - megatron_patch/model/glm130b/transformer.py (181:188) duplicated block id: 9985 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/transformer.py (1610:1615) - megatron_patch/model/mistral/transformer.py (1198:1203) duplicated block id: 9986 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llava/hf2mcore_llava.py (530:535) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_dense_and_moe_gqa.py (352:357) duplicated block id: 9987 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/deepseek_v2/moe/experts.py (286:291) - megatron_patch/model/deepseek_v2/moe/experts.py (301:306) duplicated block id: 9988 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (311:316) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (472:477) duplicated block id: 9989 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen1.5_moe.py (7:20) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_dense_and_moe_gqa.py (9:22) duplicated block id: 9990 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/bloom/transformer.py (613:621) - megatron_patch/model/falcon/transformer.py (865:872) duplicated block id: 9991 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/llama/hf2mcore_70b.py (145:151) - toolkits/model_checkpoints_convertor/qwen/hf2megatron_qwen1.0.py (201:207) duplicated block id: 9992 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/bloom/language_model.py (588:593) - megatron_patch/model/falcon/language_model.py (680:685) duplicated block id: 9993 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (336:342) - megatron_patch/model/bloom/transformer.py (183:189) duplicated block id: 9994 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/baichuan2/transformer.py (1102:1107) - megatron_patch/model/llama2/transformer.py (1282:1287) duplicated block id: 9995 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/baichuan2/hf2te.py (286:291) - toolkits/model_checkpoints_convertor/llama/hf2megatron.py (816:821) duplicated block id: 9996 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/galactica/language_model.py (445:456) - megatron_patch/model/mistral/language_model.py (436:447) duplicated block id: 9997 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/baichuan/language_model.py (492:503) - megatron_patch/model/falcon/language_model.py (414:425) duplicated block id: 9998 size: 6 cleaned lines of code in 2 files: - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_dense_and_moe_gqa.py (107:112) - toolkits/model_checkpoints_convertor/qwen/hf2mcore_qwen2_moe.py (187:194) duplicated block id: 9999 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/llama/transformer.py (192:198) - megatron_patch/model/qwen/transformer.py (333:339) duplicated block id: 10000 size: 6 cleaned lines of code in 2 files: - megatron_patch/model/mixtral/transformer/attention.py (530:535) - megatron_patch/model/qwen1_5/transformer/attention.py (391:396)