Delete run_data
Browse files- run_data/gpu_telemetry.csv +0 -0
- run_data/qwen_full.log +0 -196
- run_data/train_metrics.json +0 -286
run_data/gpu_telemetry.csv
DELETED
|
The diff for this file is too large to render.
See raw diff
|
|
|
run_data/qwen_full.log
DELETED
|
@@ -1,196 +0,0 @@
|
|
| 1 |
-
NCCL version 2.21.5+cuda12.4
|
| 2 |
-
`torch_dtype` is deprecated! Use `dtype` instead!
|
| 3 |
-
`torch_dtype` is deprecated! Use `dtype` instead!
|
| 4 |
-
`torch_dtype` is deprecated! Use `dtype` instead!
|
| 5 |
-
`torch_dtype` is deprecated! Use `dtype` instead!
|
| 6 |
-
`torch_dtype` is deprecated! Use `dtype` instead!
|
| 7 |
-
`torch_dtype` is deprecated! Use `dtype` instead!
|
| 8 |
-
`torch_dtype` is deprecated! Use `dtype` instead!
|
| 9 |
-
`torch_dtype` is deprecated! Use `dtype` instead!
|
| 10 |
-
|
| 11 |
-
|
| 12 |
-
|
| 13 |
-
|
| 14 |
-
|
| 15 |
-
|
| 16 |
-
|
| 17 |
-
|
| 18 |
-
Too many dataloader workers: 32 (max is dataset.num_shards=7). Stopping 25 dataloader workers.
|
| 19 |
-
Too many dataloader workers: 32 (max is dataset.num_shards=7). Stopping 25 dataloader workers.
|
| 20 |
-
Too many dataloader workers: 32 (max is dataset.num_shards=6). Stopping 26 dataloader workers.
|
| 21 |
-
Too many dataloader workers: 32 (max is dataset.num_shards=6). Stopping 26 dataloader workers.
|
| 22 |
-
Too many dataloader workers: 32 (max is dataset.num_shards=6). Stopping 26 dataloader workers.
|
| 23 |
-
Too many dataloader workers: 32 (max is dataset.num_shards=6). Stopping 26 dataloader workers.
|
| 24 |
-
[rank1]:W0307 03:26:40.690000 29391 torch/_logging/_internal.py:1089] [0/0] Profiler function <class 'torch.autograd.profiler.record_function'> will be ignored
|
| 25 |
-
Too many dataloader workers: 32 (max is dataset.num_shards=6). Stopping 26 dataloader workers.
|
| 26 |
-
Too many dataloader workers: 32 (max is dataset.num_shards=6). Stopping 26 dataloader workers.
|
| 27 |
-
[rank7]:W0307 03:26:43.017000 29397 torch/_logging/_internal.py:1089] [0/0] Profiler function <class 'torch.autograd.profiler.record_function'> will be ignored
|
| 28 |
-
[rank0]:W0307 03:26:43.105000 29390 torch/_logging/_internal.py:1089] [0/0] Profiler function <class 'torch.autograd.profiler.record_function'> will be ignored
|
| 29 |
-
[rank3]:W0307 03:26:43.220000 29393 torch/_logging/_internal.py:1089] [0/0] Profiler function <class 'torch.autograd.profiler.record_function'> will be ignored
|
| 30 |
-
[rank5]:W0307 03:26:44.141000 29395 torch/_logging/_internal.py:1089] [0/0] Profiler function <class 'torch.autograd.profiler.record_function'> will be ignored
|
| 31 |
-
[rank4]:W0307 03:26:44.231000 29394 torch/_logging/_internal.py:1089] [0/0] Profiler function <class 'torch.autograd.profiler.record_function'> will be ignored
|
| 32 |
-
[rank6]:W0307 03:26:46.461000 29396 torch/_logging/_internal.py:1089] [0/0] Profiler function <class 'torch.autograd.profiler.record_function'> will be ignored
|
| 33 |
-
[rank2]:W0307 03:26:49.647000 29392 torch/_logging/_internal.py:1089] [0/0] Profiler function <class 'torch.autograd.profiler.record_function'> will be ignored
|
| 34 |
-
/root/PyC/.venv/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:194: UserWarning: TensorFloat32 tensor cores for float32 matrix multiplication available but not enabled. Consider setting `torch.set_float32_matmul_precision('high')` for better performance.
|
| 35 |
-
warnings.warn(
|
| 36 |
-
/root/PyC/.venv/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:194: UserWarning: TensorFloat32 tensor cores for float32 matrix multiplication available but not enabled. Consider setting `torch.set_float32_matmul_precision('high')` for better performance.
|
| 37 |
-
warnings.warn(
|
| 38 |
-
/root/PyC/.venv/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:194: UserWarning: TensorFloat32 tensor cores for float32 matrix multiplication available but not enabled. Consider setting `torch.set_float32_matmul_precision('high')` for better performance.
|
| 39 |
-
warnings.warn(
|
| 40 |
-
/root/PyC/.venv/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:194: UserWarning: TensorFloat32 tensor cores for float32 matrix multiplication available but not enabled. Consider setting `torch.set_float32_matmul_precision('high')` for better performance.
|
| 41 |
-
warnings.warn(
|
| 42 |
-
/root/PyC/.venv/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:194: UserWarning: TensorFloat32 tensor cores for float32 matrix multiplication available but not enabled. Consider setting `torch.set_float32_matmul_precision('high')` for better performance.
|
| 43 |
-
warnings.warn(
|
| 44 |
-
/root/PyC/.venv/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:194: UserWarning: TensorFloat32 tensor cores for float32 matrix multiplication available but not enabled. Consider setting `torch.set_float32_matmul_precision('high')` for better performance.
|
| 45 |
-
warnings.warn(
|
| 46 |
-
/root/PyC/.venv/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:194: UserWarning: TensorFloat32 tensor cores for float32 matrix multiplication available but not enabled. Consider setting `torch.set_float32_matmul_precision('high')` for better performance.
|
| 47 |
-
warnings.warn(
|
| 48 |
-
/root/PyC/.venv/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:194: UserWarning: TensorFloat32 tensor cores for float32 matrix multiplication available but not enabled. Consider setting `torch.set_float32_matmul_precision('high')` for better performance.
|
| 49 |
-
warnings.warn(
|
| 50 |
-
[train_sft] step=10/2000 loss=0.3106 tokens=66019 samples_per_sec=0.42 tokens_per_sec=172.88
|
| 51 |
-
[train_sft] step=20/2000 loss=0.2507 tokens=130469 samples_per_sec=0.51 tokens_per_sec=209.38
|
| 52 |
-
[train_sft] step=30/2000 loss=0.2291 tokens=196381 samples_per_sec=0.55 tokens_per_sec=226.84
|
| 53 |
-
[train_sft] step=40/2000 loss=0.2799 tokens=260313 samples_per_sec=0.58 tokens_per_sec=234.92
|
| 54 |
-
[train_sft] step=50/2000 loss=0.0738 tokens=326884 samples_per_sec=0.59 tokens_per_sec=241.97
|
| 55 |
-
[train_sft] step=60/2000 loss=0.2171 tokens=392579 samples_per_sec=0.60 tokens_per_sec=246.20
|
| 56 |
-
[train_sft] step=70/2000 loss=0.1810 tokens=458551 samples_per_sec=0.61 tokens_per_sec=249.57
|
| 57 |
-
[train_sft] step=80/2000 loss=0.1329 tokens=527838 samples_per_sec=0.62 tokens_per_sec=253.71
|
| 58 |
-
[train_sft] step=90/2000 loss=0.1080 tokens=598351 samples_per_sec=0.62 tokens_per_sec=257.55
|
| 59 |
-
[train_sft] step=100/2000 loss=0.1271 tokens=665669 samples_per_sec=0.62 tokens_per_sec=259.31
|
| 60 |
-
[train_sft] step=110/2000 loss=0.1596 tokens=727376 samples_per_sec=0.63 tokens_per_sec=258.71
|
| 61 |
-
[train_sft] step=120/2000 loss=0.1730 tokens=789579 samples_per_sec=0.63 tokens_per_sec=258.39
|
| 62 |
-
[train_sft] step=130/2000 loss=0.1666 tokens=852405 samples_per_sec=0.63 tokens_per_sec=258.28
|
| 63 |
-
[train_sft] step=140/2000 loss=0.1658 tokens=922146 samples_per_sec=0.63 tokens_per_sec=260.12
|
| 64 |
-
[train_sft] step=150/2000 loss=0.0985 tokens=986385 samples_per_sec=0.63 tokens_per_sec=260.33
|
| 65 |
-
[train_sft] step=160/2000 loss=0.0701 tokens=1049899 samples_per_sec=0.63 tokens_per_sec=260.30
|
| 66 |
-
[train_sft] step=170/2000 loss=0.0789 tokens=1114566 samples_per_sec=0.64 tokens_per_sec=260.56
|
| 67 |
-
[train_sft] step=180/2000 loss=0.1549 tokens=1184173 samples_per_sec=0.64 tokens_per_sec=261.97
|
| 68 |
-
[train_sft] step=190/2000 loss=0.1046 tokens=1248076 samples_per_sec=0.64 tokens_per_sec=261.99
|
| 69 |
-
[train_sft] step=200/2000 loss=0.2644 tokens=1312805 samples_per_sec=0.64 tokens_per_sec=262.13
|
| 70 |
-
[train_sft] eval step=200 loss=n/a
|
| 71 |
-
[train_sft] step=210/2000 loss=0.1345 tokens=1378046 samples_per_sec=0.64 tokens_per_sec=262.34
|
| 72 |
-
[train_sft] step=220/2000 loss=0.1363 tokens=1440118 samples_per_sec=0.64 tokens_per_sec=261.97
|
| 73 |
-
[train_sft] step=230/2000 loss=0.1488 tokens=1506320 samples_per_sec=0.64 tokens_per_sec=262.35
|
| 74 |
-
[train_sft] step=240/2000 loss=0.1060 tokens=1571030 samples_per_sec=0.64 tokens_per_sec=262.45
|
| 75 |
-
[train_sft] step=250/2000 loss=0.1710 tokens=1636503 samples_per_sec=0.64 tokens_per_sec=262.64
|
| 76 |
-
[train_sft] step=260/2000 loss=0.1923 tokens=1702374 samples_per_sec=0.64 tokens_per_sec=262.90
|
| 77 |
-
[train_sft] step=270/2000 loss=0.1278 tokens=1765267 samples_per_sec=0.64 tokens_per_sec=262.69
|
| 78 |
-
[train_sft] step=280/2000 loss=0.1669 tokens=1834329 samples_per_sec=0.64 tokens_per_sec=263.40
|
| 79 |
-
[train_sft] step=290/2000 loss=0.1825 tokens=1901061 samples_per_sec=0.64 tokens_per_sec=263.73
|
| 80 |
-
[train_sft] step=300/2000 loss=0.1071 tokens=1965869 samples_per_sec=0.64 tokens_per_sec=263.76
|
| 81 |
-
[train_sft] step=310/2000 loss=0.1100 tokens=2026328 samples_per_sec=0.64 tokens_per_sec=263.24
|
| 82 |
-
[train_sft] step=320/2000 loss=0.1569 tokens=2091438 samples_per_sec=0.64 tokens_per_sec=263.35
|
| 83 |
-
[train_sft] step=330/2000 loss=0.0981 tokens=2159388 samples_per_sec=0.64 tokens_per_sec=263.79
|
| 84 |
-
[train_sft] step=340/2000 loss=0.1418 tokens=2224937 samples_per_sec=0.65 tokens_per_sec=263.91
|
| 85 |
-
[train_sft] step=350/2000 loss=0.1125 tokens=2290801 samples_per_sec=0.65 tokens_per_sec=264.05
|
| 86 |
-
[train_sft] step=360/2000 loss=0.0595 tokens=2356921 samples_per_sec=0.65 tokens_per_sec=264.21
|
| 87 |
-
[train_sft] step=370/2000 loss=0.0877 tokens=2419567 samples_per_sec=0.65 tokens_per_sec=264.01
|
| 88 |
-
[train_sft] step=380/2000 loss=0.2055 tokens=2484959 samples_per_sec=0.65 tokens_per_sec=264.08
|
| 89 |
-
[train_sft] step=390/2000 loss=0.2391 tokens=2548421 samples_per_sec=0.65 tokens_per_sec=263.97
|
| 90 |
-
[train_sft] step=400/2000 loss=0.1708 tokens=2611454 samples_per_sec=0.65 tokens_per_sec=263.81
|
| 91 |
-
[train_sft] eval step=400 loss=n/a
|
| 92 |
-
[train_sft] step=410/2000 loss=0.1243 tokens=2673912 samples_per_sec=0.65 tokens_per_sec=263.61
|
| 93 |
-
[train_sft] step=420/2000 loss=0.1155 tokens=2741532 samples_per_sec=0.65 tokens_per_sec=263.91
|
| 94 |
-
[train_sft] step=430/2000 loss=0.1285 tokens=2803400 samples_per_sec=0.65 tokens_per_sec=263.66
|
| 95 |
-
[train_sft] step=440/2000 loss=0.0763 tokens=2870087 samples_per_sec=0.65 tokens_per_sec=263.86
|
| 96 |
-
[train_sft] step=450/2000 loss=0.1212 tokens=2938147 samples_per_sec=0.65 tokens_per_sec=264.18
|
| 97 |
-
[train_sft] step=460/2000 loss=0.1697 tokens=3006919 samples_per_sec=0.65 tokens_per_sec=264.55
|
| 98 |
-
[train_sft] step=470/2000 loss=0.0634 tokens=3074418 samples_per_sec=0.65 tokens_per_sec=264.79
|
| 99 |
-
[train_sft] step=480/2000 loss=0.1071 tokens=3140612 samples_per_sec=0.65 tokens_per_sec=264.92
|
| 100 |
-
[train_sft] step=490/2000 loss=0.0840 tokens=3207555 samples_per_sec=0.65 tokens_per_sec=265.09
|
| 101 |
-
[train_sft] step=500/2000 loss=0.1685 tokens=3272736 samples_per_sec=0.65 tokens_per_sec=265.12
|
| 102 |
-
|
| 103 |
-
[train_sft] step=510/2000 loss=11.9358 tokens=3333255 samples_per_sec=0.64 tokens_per_sec=261.82
|
| 104 |
-
[train_sft] step=520/2000 loss=11.9305 tokens=3400813 samples_per_sec=0.64 tokens_per_sec=260.02
|
| 105 |
-
[train_sft] step=530/2000 loss=11.9318 tokens=3468065 samples_per_sec=0.63 tokens_per_sec=258.30
|
| 106 |
-
[train_sft] step=540/2000 loss=11.9336 tokens=3535022 samples_per_sec=0.63 tokens_per_sec=256.63
|
| 107 |
-
[train_sft] step=550/2000 loss=11.9334 tokens=3600853 samples_per_sec=0.62 tokens_per_sec=254.97
|
| 108 |
-
[train_sft] step=560/2000 loss=11.9341 tokens=3662071 samples_per_sec=0.62 tokens_per_sec=253.07
|
| 109 |
-
[train_sft] step=570/2000 loss=11.9347 tokens=3730017 samples_per_sec=0.62 tokens_per_sec=251.71
|
| 110 |
-
[train_sft] step=580/2000 loss=11.9316 tokens=3796693 samples_per_sec=0.61 tokens_per_sec=250.33
|
| 111 |
-
[train_sft] step=590/2000 loss=11.9386 tokens=3858933 samples_per_sec=0.61 tokens_per_sec=248.73
|
| 112 |
-
[train_sft] step=600/2000 loss=11.9369 tokens=3920622 samples_per_sec=0.61 tokens_per_sec=247.17
|
| 113 |
-
[train_sft] eval step=600 loss=n/a
|
| 114 |
-
[train_sft] step=610/2000 loss=11.9311 tokens=3984325 samples_per_sec=0.60 tokens_per_sec=245.79
|
| 115 |
-
[train_sft] step=620/2000 loss=11.9305 tokens=4051226 samples_per_sec=0.60 tokens_per_sec=244.67
|
| 116 |
-
[train_sft] step=630/2000 loss=11.9329 tokens=4117533 samples_per_sec=0.60 tokens_per_sec=243.55
|
| 117 |
-
[train_sft] step=640/2000 loss=11.9343 tokens=4183625 samples_per_sec=0.59 tokens_per_sec=242.47
|
| 118 |
-
[train_sft] step=650/2000 loss=11.9327 tokens=4252154 samples_per_sec=0.59 tokens_per_sec=241.57
|
| 119 |
-
[train_sft] step=660/2000 loss=11.9336 tokens=4313556 samples_per_sec=0.59 tokens_per_sec=240.31
|
| 120 |
-
[train_sft] step=670/2000 loss=11.9328 tokens=4376047 samples_per_sec=0.59 tokens_per_sec=239.16
|
| 121 |
-
[train_sft] step=680/2000 loss=11.9349 tokens=4443065 samples_per_sec=0.58 tokens_per_sec=238.29
|
| 122 |
-
[train_sft] step=690/2000 loss=11.9312 tokens=4508609 samples_per_sec=0.58 tokens_per_sec=237.37
|
| 123 |
-
[train_sft] step=700/2000 loss=11.9365 tokens=4571767 samples_per_sec=0.58 tokens_per_sec=236.37
|
| 124 |
-
[train_sft] step=710/2000 loss=11.9332 tokens=4635156 samples_per_sec=0.58 tokens_per_sec=235.42
|
| 125 |
-
[train_sft] step=720/2000 loss=11.9340 tokens=4700197 samples_per_sec=0.57 tokens_per_sec=234.57
|
| 126 |
-
[train_sft] step=730/2000 loss=11.9302 tokens=4764290 samples_per_sec=0.57 tokens_per_sec=233.71
|
| 127 |
-
[train_sft] step=740/2000 loss=11.9358 tokens=4831552 samples_per_sec=0.57 tokens_per_sec=233.03
|
| 128 |
-
[train_sft] step=750/2000 loss=11.9338 tokens=4896926 samples_per_sec=0.57 tokens_per_sec=232.29
|
| 129 |
-
[train_sft] step=760/2000 loss=11.9299 tokens=4963790 samples_per_sec=0.57 tokens_per_sec=231.64
|
| 130 |
-
[train_sft] step=770/2000 loss=11.9375 tokens=5030112 samples_per_sec=0.57 tokens_per_sec=230.98
|
| 131 |
-
[train_sft] step=780/2000 loss=11.9328 tokens=5093917 samples_per_sec=0.56 tokens_per_sec=230.23
|
| 132 |
-
[train_sft] step=790/2000 loss=11.9376 tokens=5160155 samples_per_sec=0.56 tokens_per_sec=229.61
|
| 133 |
-
[train_sft] step=800/2000 loss=11.9353 tokens=5229774 samples_per_sec=0.56 tokens_per_sec=229.16
|
| 134 |
-
[train_sft] eval step=800 loss=n/a
|
| 135 |
-
[train_sft] step=810/2000 loss=11.9334 tokens=5295154 samples_per_sec=0.56 tokens_per_sec=228.54
|
| 136 |
-
[train_sft] step=820/2000 loss=11.9345 tokens=5357311 samples_per_sec=0.56 tokens_per_sec=227.80
|
| 137 |
-
[train_sft] step=830/2000 loss=11.9387 tokens=5422927 samples_per_sec=0.56 tokens_per_sec=227.23
|
| 138 |
-
[train_sft] step=840/2000 loss=11.9358 tokens=5490488 samples_per_sec=0.56 tokens_per_sec=226.75
|
| 139 |
-
[train_sft] step=850/2000 loss=11.9340 tokens=5551808 samples_per_sec=0.55 tokens_per_sec=226.04
|
| 140 |
-
[train_sft] step=860/2000 loss=11.9372 tokens=5617359 samples_per_sec=0.55 tokens_per_sec=225.52
|
| 141 |
-
[train_sft] step=870/2000 loss=11.9311 tokens=5685346 samples_per_sec=0.55 tokens_per_sec=225.10
|
| 142 |
-
[train_sft] step=880/2000 loss=11.9340 tokens=5750741 samples_per_sec=0.55 tokens_per_sec=224.59
|
| 143 |
-
[train_sft] step=890/2000 loss=11.9300 tokens=5816629 samples_per_sec=0.55 tokens_per_sec=224.12
|
| 144 |
-
[train_sft] step=900/2000 loss=11.9343 tokens=5882089 samples_per_sec=0.55 tokens_per_sec=223.64
|
| 145 |
-
[train_sft] step=910/2000 loss=11.9320 tokens=5948442 samples_per_sec=0.55 tokens_per_sec=223.21
|
| 146 |
-
[train_sft] step=920/2000 loss=11.9363 tokens=6009051 samples_per_sec=0.55 tokens_per_sec=222.58
|
| 147 |
-
[train_sft] step=930/2000 loss=11.9309 tokens=6078081 samples_per_sec=0.54 tokens_per_sec=222.27
|
| 148 |
-
[train_sft] step=940/2000 loss=11.9353 tokens=6138758 samples_per_sec=0.54 tokens_per_sec=221.66
|
| 149 |
-
[train_sft] step=950/2000 loss=11.9371 tokens=6204982 samples_per_sec=0.54 tokens_per_sec=221.27
|
| 150 |
-
[train_sft] step=960/2000 loss=11.9341 tokens=6272247 samples_per_sec=0.54 tokens_per_sec=220.93
|
| 151 |
-
[train_sft] step=970/2000 loss=11.9314 tokens=6337243 samples_per_sec=0.54 tokens_per_sec=220.51
|
| 152 |
-
[train_sft] step=980/2000 loss=11.9361 tokens=6401913 samples_per_sec=0.54 tokens_per_sec=220.10
|
| 153 |
-
[train_sft] step=990/2000 loss=11.9325 tokens=6465635 samples_per_sec=0.54 tokens_per_sec=219.66
|
| 154 |
-
[train_sft] step=1000/2000 loss=11.9347 tokens=6532582 samples_per_sec=0.54 tokens_per_sec=219.34
|
| 155 |
-
[train_sft] eval step=1000 loss=n/a
|
| 156 |
-
|
| 157 |
-
[train_sft] step=1010/2000 loss=nan tokens=6598011 samples_per_sec=0.54 tokens_per_sec=219.25
|
| 158 |
-
W0307 11:46:34.841000 29377 torch/distributed/elastic/agent/server/api.py:719] Received Signals.SIGINT death signal, shutting down workers
|
| 159 |
-
W0307 11:46:34.843000 29377 torch/distributed/elastic/multiprocessing/api.py:897] Sending process 29390 closing signal SIGINT
|
| 160 |
-
W0307 11:46:34.843000 29377 torch/distributed/elastic/multiprocessing/api.py:897] Sending process 29391 closing signal SIGINT
|
| 161 |
-
W0307 11:46:34.844000 29377 torch/distributed/elastic/multiprocessing/api.py:897] Sending process 29392 closing signal SIGINT
|
| 162 |
-
W0307 11:46:34.844000 29377 torch/distributed/elastic/multiprocessing/api.py:897] Sending process 29393 closing signal SIGINT
|
| 163 |
-
W0307 11:46:34.844000 29377 torch/distributed/elastic/multiprocessing/api.py:897] Sending process 29394 closing signal SIGINT
|
| 164 |
-
W0307 11:46:34.844000 29377 torch/distributed/elastic/multiprocessing/api.py:897] Sending process 29395 closing signal SIGINT
|
| 165 |
-
W0307 11:46:34.845000 29377 torch/distributed/elastic/multiprocessing/api.py:897] Sending process 29396 closing signal SIGINT
|
| 166 |
-
W0307 11:46:34.845000 29377 torch/distributed/elastic/multiprocessing/api.py:897] Sending process 29397 closing signal SIGINT
|
| 167 |
-
W0307 11:47:04.846000 29377 torch/distributed/elastic/multiprocessing/api.py:916] Unable to shutdown process 29390 via Signals.SIGINT, forcefully exiting via Signals.SIGKILL
|
| 168 |
-
W0307 11:47:05.104000 29377 torch/distributed/elastic/multiprocessing/api.py:916] Unable to shutdown process 29391 via Signals.SIGINT, forcefully exiting via Signals.SIGKILL
|
| 169 |
-
W0307 11:47:05.500000 29377 torch/distributed/elastic/multiprocessing/api.py:916] Unable to shutdown process 29392 via Signals.SIGINT, forcefully exiting via Signals.SIGKILL
|
| 170 |
-
W0307 11:47:05.889000 29377 torch/distributed/elastic/multiprocessing/api.py:916] Unable to shutdown process 29393 via Signals.SIGINT, forcefully exiting via Signals.SIGKILL
|
| 171 |
-
W0307 11:47:06.314000 29377 torch/distributed/elastic/multiprocessing/api.py:916] Unable to shutdown process 29394 via Signals.SIGINT, forcefully exiting via Signals.SIGKILL
|
| 172 |
-
W0307 11:47:06.691000 29377 torch/distributed/elastic/multiprocessing/api.py:916] Unable to shutdown process 29395 via Signals.SIGINT, forcefully exiting via Signals.SIGKILL
|
| 173 |
-
W0307 11:47:07.074000 29377 torch/distributed/elastic/multiprocessing/api.py:916] Unable to shutdown process 29396 via Signals.SIGINT, forcefully exiting via Signals.SIGKILL
|
| 174 |
-
W0307 11:47:07.463000 29377 torch/distributed/elastic/multiprocessing/api.py:916] Unable to shutdown process 29397 via Signals.SIGINT, forcefully exiting via Signals.SIGKILL
|
| 175 |
-
Traceback (most recent call last):
|
| 176 |
-
File "/root/PyC/.venv/bin/torchrun", line 6, in <module>
|
| 177 |
-
sys.exit(main())
|
| 178 |
-
File "/root/PyC/.venv/lib/python3.10/site-packages/torch/distributed/elastic/multiprocessing/errors/__init__.py", line 355, in wrapper
|
| 179 |
-
return f(*args, **kwargs)
|
| 180 |
-
File "/root/PyC/.venv/lib/python3.10/site-packages/torch/distributed/run.py", line 918, in main
|
| 181 |
-
run(args)
|
| 182 |
-
File "/root/PyC/.venv/lib/python3.10/site-packages/torch/distributed/run.py", line 909, in run
|
| 183 |
-
elastic_launch(
|
| 184 |
-
File "/root/PyC/.venv/lib/python3.10/site-packages/torch/distributed/launcher/api.py", line 138, in __call__
|
| 185 |
-
return launch_agent(self._config, self._entrypoint, list(args))
|
| 186 |
-
File "/root/PyC/.venv/lib/python3.10/site-packages/torch/distributed/launcher/api.py", line 260, in launch_agent
|
| 187 |
-
result = agent.run()
|
| 188 |
-
File "/root/PyC/.venv/lib/python3.10/site-packages/torch/distributed/elastic/metrics/api.py", line 137, in wrapper
|
| 189 |
-
result = f(*args, **kwargs)
|
| 190 |
-
File "/root/PyC/.venv/lib/python3.10/site-packages/torch/distributed/elastic/agent/server/api.py", line 711, in run
|
| 191 |
-
result = self._invoke_run(role)
|
| 192 |
-
File "/root/PyC/.venv/lib/python3.10/site-packages/torch/distributed/elastic/agent/server/api.py", line 870, in _invoke_run
|
| 193 |
-
time.sleep(monitor_interval)
|
| 194 |
-
File "/root/PyC/.venv/lib/python3.10/site-packages/torch/distributed/elastic/multiprocessing/api.py", line 84, in _terminate_process_handler
|
| 195 |
-
raise SignalException(f"Process {os.getpid()} got signal: {sigval}", sigval=sigval)
|
| 196 |
-
torch.distributed.elastic.multiprocessing.api.SignalException: Process 29377 got signal: 2
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
run_data/train_metrics.json
DELETED
|
@@ -1,286 +0,0 @@
|
|
| 1 |
-
{
|
| 2 |
-
"run_id": "qwen14b_codeinstruct",
|
| 3 |
-
"timestamp_utc": "2026-03-07T03:24:36Z",
|
| 4 |
-
"host": "5cea9450de15",
|
| 5 |
-
"train_runtime_sec": 30093.7884,
|
| 6 |
-
"samples_per_sec": 0.537,
|
| 7 |
-
"steps_per_sec": 0.0336,
|
| 8 |
-
"tokens_per_sec": 219.2483,
|
| 9 |
-
"loss_final": NaN,
|
| 10 |
-
"loss_curve": [
|
| 11 |
-
11.933758,
|
| 12 |
-
11.935425,
|
| 13 |
-
11.936735,
|
| 14 |
-
11.934477,
|
| 15 |
-
11.936081,
|
| 16 |
-
11.934277,
|
| 17 |
-
11.933105,
|
| 18 |
-
11.934149,
|
| 19 |
-
11.931114,
|
| 20 |
-
11.931803,
|
| 21 |
-
11.931644,
|
| 22 |
-
11.933023,
|
| 23 |
-
11.935471,
|
| 24 |
-
11.934968,
|
| 25 |
-
11.935336,
|
| 26 |
-
11.936069,
|
| 27 |
-
11.931381,
|
| 28 |
-
11.929764,
|
| 29 |
-
11.933843,
|
| 30 |
-
11.932031,
|
| 31 |
-
11.934188,
|
| 32 |
-
11.933489,
|
| 33 |
-
11.932564,
|
| 34 |
-
11.933738,
|
| 35 |
-
11.933002,
|
| 36 |
-
11.935679,
|
| 37 |
-
11.930973,
|
| 38 |
-
11.933707,
|
| 39 |
-
11.935345,
|
| 40 |
-
11.932213,
|
| 41 |
-
11.931738,
|
| 42 |
-
11.938071,
|
| 43 |
-
11.934389,
|
| 44 |
-
11.935558,
|
| 45 |
-
11.933566,
|
| 46 |
-
11.931173,
|
| 47 |
-
11.933773,
|
| 48 |
-
11.930953,
|
| 49 |
-
11.933154,
|
| 50 |
-
11.935999,
|
| 51 |
-
11.934997,
|
| 52 |
-
11.934099,
|
| 53 |
-
11.93278,
|
| 54 |
-
11.938663,
|
| 55 |
-
11.933143,
|
| 56 |
-
11.932686,
|
| 57 |
-
11.934217,
|
| 58 |
-
11.934953,
|
| 59 |
-
11.936878,
|
| 60 |
-
11.933104,
|
| 61 |
-
11.930088,
|
| 62 |
-
11.929935,
|
| 63 |
-
11.932352,
|
| 64 |
-
11.932141,
|
| 65 |
-
11.934212,
|
| 66 |
-
11.937968,
|
| 67 |
-
11.93185,
|
| 68 |
-
11.935164,
|
| 69 |
-
11.933124,
|
| 70 |
-
11.934265,
|
| 71 |
-
11.931581,
|
| 72 |
-
11.93028,
|
| 73 |
-
11.935074,
|
| 74 |
-
11.930362,
|
| 75 |
-
11.93231,
|
| 76 |
-
11.934386,
|
| 77 |
-
11.931125,
|
| 78 |
-
11.933617,
|
| 79 |
-
11.934947,
|
| 80 |
-
11.935481,
|
| 81 |
-
11.931095,
|
| 82 |
-
11.936635,
|
| 83 |
-
11.934564,
|
| 84 |
-
11.934128,
|
| 85 |
-
11.933428,
|
| 86 |
-
11.933599,
|
| 87 |
-
11.93398,
|
| 88 |
-
11.930263,
|
| 89 |
-
11.934587,
|
| 90 |
-
11.933071,
|
| 91 |
-
11.932242,
|
| 92 |
-
11.933808,
|
| 93 |
-
11.931849,
|
| 94 |
-
11.936655,
|
| 95 |
-
11.932829,
|
| 96 |
-
11.931339,
|
| 97 |
-
11.936092,
|
| 98 |
-
11.935149,
|
| 99 |
-
11.934838,
|
| 100 |
-
11.932211,
|
| 101 |
-
11.939731,
|
| 102 |
-
11.935974,
|
| 103 |
-
11.93283,
|
| 104 |
-
11.933028,
|
| 105 |
-
11.938888,
|
| 106 |
-
11.932542,
|
| 107 |
-
11.935485,
|
| 108 |
-
11.93673,
|
| 109 |
-
11.930267,
|
| 110 |
-
11.933331,
|
| 111 |
-
11.934375,
|
| 112 |
-
11.932673,
|
| 113 |
-
11.93618,
|
| 114 |
-
11.933468,
|
| 115 |
-
11.932549,
|
| 116 |
-
11.936689,
|
| 117 |
-
11.932197,
|
| 118 |
-
11.934677,
|
| 119 |
-
11.934959,
|
| 120 |
-
11.932065,
|
| 121 |
-
11.934155,
|
| 122 |
-
11.9322,
|
| 123 |
-
11.935388,
|
| 124 |
-
11.935803,
|
| 125 |
-
11.936669,
|
| 126 |
-
11.929192,
|
| 127 |
-
11.937821,
|
| 128 |
-
11.933041,
|
| 129 |
-
11.933342,
|
| 130 |
-
11.935324,
|
| 131 |
-
11.934742,
|
| 132 |
-
11.934669,
|
| 133 |
-
11.935793,
|
| 134 |
-
11.937335,
|
| 135 |
-
11.931247,
|
| 136 |
-
11.932714,
|
| 137 |
-
11.933646,
|
| 138 |
-
11.935673,
|
| 139 |
-
11.932477,
|
| 140 |
-
11.931051,
|
| 141 |
-
11.935923,
|
| 142 |
-
11.935831,
|
| 143 |
-
11.935593,
|
| 144 |
-
11.937632,
|
| 145 |
-
11.933506,
|
| 146 |
-
11.933863,
|
| 147 |
-
11.936568,
|
| 148 |
-
11.93511,
|
| 149 |
-
11.929846,
|
| 150 |
-
11.928335,
|
| 151 |
-
11.935637,
|
| 152 |
-
11.931632,
|
| 153 |
-
11.933343,
|
| 154 |
-
11.932386,
|
| 155 |
-
11.935142,
|
| 156 |
-
11.934553,
|
| 157 |
-
11.937015,
|
| 158 |
-
11.930168,
|
| 159 |
-
11.937183,
|
| 160 |
-
11.934591,
|
| 161 |
-
11.937156,
|
| 162 |
-
11.934317,
|
| 163 |
-
11.933546,
|
| 164 |
-
11.937324,
|
| 165 |
-
11.92873,
|
| 166 |
-
11.934661,
|
| 167 |
-
11.935086,
|
| 168 |
-
11.934525,
|
| 169 |
-
11.934164,
|
| 170 |
-
11.933876,
|
| 171 |
-
11.932452,
|
| 172 |
-
11.932648,
|
| 173 |
-
11.934206,
|
| 174 |
-
11.932563,
|
| 175 |
-
11.934017,
|
| 176 |
-
11.934988,
|
| 177 |
-
11.933892,
|
| 178 |
-
11.933537,
|
| 179 |
-
11.929354,
|
| 180 |
-
11.935569,
|
| 181 |
-
11.93031,
|
| 182 |
-
11.934333,
|
| 183 |
-
11.934294,
|
| 184 |
-
11.931088,
|
| 185 |
-
11.936719,
|
| 186 |
-
11.934738,
|
| 187 |
-
11.932435,
|
| 188 |
-
11.934471,
|
| 189 |
-
11.934482,
|
| 190 |
-
11.93504,
|
| 191 |
-
11.934249,
|
| 192 |
-
11.931784,
|
| 193 |
-
11.932494,
|
| 194 |
-
11.932414,
|
| 195 |
-
11.933865,
|
| 196 |
-
11.931005,
|
| 197 |
-
11.93541,
|
| 198 |
-
11.932836,
|
| 199 |
-
11.935291,
|
| 200 |
-
11.933075,
|
| 201 |
-
11.931273,
|
| 202 |
-
11.931405,
|
| 203 |
-
11.93127,
|
| 204 |
-
11.931889,
|
| 205 |
-
11.935592,
|
| 206 |
-
11.930029,
|
| 207 |
-
11.935498,
|
| 208 |
-
11.934025,
|
| 209 |
-
11.935754,
|
| 210 |
-
11.932719,
|
| 211 |
-
11.933537,
|
| 212 |
-
11.935834,
|
| 213 |
-
11.933773,
|
| 214 |
-
11.931642,
|
| 215 |
-
11.931832,
|
| 216 |
-
11.934774,
|
| 217 |
-
11.932487,
|
| 218 |
-
11.932528,
|
| 219 |
-
11.932735,
|
| 220 |
-
11.934239,
|
| 221 |
-
11.93345,
|
| 222 |
-
11.933942,
|
| 223 |
-
11.931937,
|
| 224 |
-
11.937059,
|
| 225 |
-
11.938506,
|
| 226 |
-
11.932685,
|
| 227 |
-
11.932717,
|
| 228 |
-
11.936692,
|
| 229 |
-
11.9363,
|
| 230 |
-
11.936151,
|
| 231 |
-
11.931432,
|
| 232 |
-
11.933336,
|
| 233 |
-
11.932334,
|
| 234 |
-
11.932056,
|
| 235 |
-
11.934895,
|
| 236 |
-
11.933721,
|
| 237 |
-
11.934547,
|
| 238 |
-
11.933809,
|
| 239 |
-
11.933655,
|
| 240 |
-
11.933565,
|
| 241 |
-
11.933448,
|
| 242 |
-
11.932833,
|
| 243 |
-
11.936954,
|
| 244 |
-
11.935034,
|
| 245 |
-
11.931088,
|
| 246 |
-
11.935302,
|
| 247 |
-
11.933675,
|
| 248 |
-
11.935065,
|
| 249 |
-
11.935748,
|
| 250 |
-
11.932056,
|
| 251 |
-
11.938281,
|
| 252 |
-
11.934908,
|
| 253 |
-
11.932071,
|
| 254 |
-
11.933029,
|
| 255 |
-
11.934909,
|
| 256 |
-
11.933907,
|
| 257 |
-
11.933957,
|
| 258 |
-
NaN,
|
| 259 |
-
NaN,
|
| 260 |
-
NaN,
|
| 261 |
-
NaN,
|
| 262 |
-
NaN,
|
| 263 |
-
NaN,
|
| 264 |
-
NaN,
|
| 265 |
-
NaN,
|
| 266 |
-
NaN
|
| 267 |
-
],
|
| 268 |
-
"eval_loss": null,
|
| 269 |
-
"gpu_util_mean": 99.2073,
|
| 270 |
-
"gpu_util_p95": 100.0,
|
| 271 |
-
"h2d_time_ms_mean": 0.0726,
|
| 272 |
-
"compute_time_ms_mean": 3697.3475,
|
| 273 |
-
"comm_time_ms_mean": 6.5765,
|
| 274 |
-
"idle_gap_ms_mean": 11.6392,
|
| 275 |
-
"idle_gap_ms_p95": 0.3205,
|
| 276 |
-
"mode": "sft",
|
| 277 |
-
"dist": "fsdp",
|
| 278 |
-
"backend": "nccl",
|
| 279 |
-
"world_size": 8,
|
| 280 |
-
"model_name": "Qwen/Qwen2.5-14B",
|
| 281 |
-
"dataset_name": "nvidia/OpenCodeInstruct",
|
| 282 |
-
"dataset_config": "",
|
| 283 |
-
"precision": "bf16",
|
| 284 |
-
"seq_length": 4096,
|
| 285 |
-
"gradient_accumulation_steps": 8
|
| 286 |
-
}
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|