Allanatrix commited on
Commit
e0eb9c6
·
verified ·
1 Parent(s): 7270a66

Delete run_data

Browse files
run_data/gpu_telemetry.csv DELETED
The diff for this file is too large to render. See raw diff
 
run_data/qwen_full.log DELETED
@@ -1,196 +0,0 @@
1
- NCCL version 2.21.5+cuda12.4
2
- `torch_dtype` is deprecated! Use `dtype` instead!
3
- `torch_dtype` is deprecated! Use `dtype` instead!
4
- `torch_dtype` is deprecated! Use `dtype` instead!
5
- `torch_dtype` is deprecated! Use `dtype` instead!
6
- `torch_dtype` is deprecated! Use `dtype` instead!
7
- `torch_dtype` is deprecated! Use `dtype` instead!
8
- `torch_dtype` is deprecated! Use `dtype` instead!
9
- `torch_dtype` is deprecated! Use `dtype` instead!
10
-
11
-
12
-
13
-
14
-
15
-
16
-
17
-
18
- Too many dataloader workers: 32 (max is dataset.num_shards=7). Stopping 25 dataloader workers.
19
- Too many dataloader workers: 32 (max is dataset.num_shards=7). Stopping 25 dataloader workers.
20
- Too many dataloader workers: 32 (max is dataset.num_shards=6). Stopping 26 dataloader workers.
21
- Too many dataloader workers: 32 (max is dataset.num_shards=6). Stopping 26 dataloader workers.
22
- Too many dataloader workers: 32 (max is dataset.num_shards=6). Stopping 26 dataloader workers.
23
- Too many dataloader workers: 32 (max is dataset.num_shards=6). Stopping 26 dataloader workers.
24
- [rank1]:W0307 03:26:40.690000 29391 torch/_logging/_internal.py:1089] [0/0] Profiler function <class 'torch.autograd.profiler.record_function'> will be ignored
25
- Too many dataloader workers: 32 (max is dataset.num_shards=6). Stopping 26 dataloader workers.
26
- Too many dataloader workers: 32 (max is dataset.num_shards=6). Stopping 26 dataloader workers.
27
- [rank7]:W0307 03:26:43.017000 29397 torch/_logging/_internal.py:1089] [0/0] Profiler function <class 'torch.autograd.profiler.record_function'> will be ignored
28
- [rank0]:W0307 03:26:43.105000 29390 torch/_logging/_internal.py:1089] [0/0] Profiler function <class 'torch.autograd.profiler.record_function'> will be ignored
29
- [rank3]:W0307 03:26:43.220000 29393 torch/_logging/_internal.py:1089] [0/0] Profiler function <class 'torch.autograd.profiler.record_function'> will be ignored
30
- [rank5]:W0307 03:26:44.141000 29395 torch/_logging/_internal.py:1089] [0/0] Profiler function <class 'torch.autograd.profiler.record_function'> will be ignored
31
- [rank4]:W0307 03:26:44.231000 29394 torch/_logging/_internal.py:1089] [0/0] Profiler function <class 'torch.autograd.profiler.record_function'> will be ignored
32
- [rank6]:W0307 03:26:46.461000 29396 torch/_logging/_internal.py:1089] [0/0] Profiler function <class 'torch.autograd.profiler.record_function'> will be ignored
33
- [rank2]:W0307 03:26:49.647000 29392 torch/_logging/_internal.py:1089] [0/0] Profiler function <class 'torch.autograd.profiler.record_function'> will be ignored
34
- /root/PyC/.venv/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:194: UserWarning: TensorFloat32 tensor cores for float32 matrix multiplication available but not enabled. Consider setting `torch.set_float32_matmul_precision('high')` for better performance.
35
- warnings.warn(
36
- /root/PyC/.venv/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:194: UserWarning: TensorFloat32 tensor cores for float32 matrix multiplication available but not enabled. Consider setting `torch.set_float32_matmul_precision('high')` for better performance.
37
- warnings.warn(
38
- /root/PyC/.venv/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:194: UserWarning: TensorFloat32 tensor cores for float32 matrix multiplication available but not enabled. Consider setting `torch.set_float32_matmul_precision('high')` for better performance.
39
- warnings.warn(
40
- /root/PyC/.venv/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:194: UserWarning: TensorFloat32 tensor cores for float32 matrix multiplication available but not enabled. Consider setting `torch.set_float32_matmul_precision('high')` for better performance.
41
- warnings.warn(
42
- /root/PyC/.venv/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:194: UserWarning: TensorFloat32 tensor cores for float32 matrix multiplication available but not enabled. Consider setting `torch.set_float32_matmul_precision('high')` for better performance.
43
- warnings.warn(
44
- /root/PyC/.venv/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:194: UserWarning: TensorFloat32 tensor cores for float32 matrix multiplication available but not enabled. Consider setting `torch.set_float32_matmul_precision('high')` for better performance.
45
- warnings.warn(
46
- /root/PyC/.venv/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:194: UserWarning: TensorFloat32 tensor cores for float32 matrix multiplication available but not enabled. Consider setting `torch.set_float32_matmul_precision('high')` for better performance.
47
- warnings.warn(
48
- /root/PyC/.venv/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:194: UserWarning: TensorFloat32 tensor cores for float32 matrix multiplication available but not enabled. Consider setting `torch.set_float32_matmul_precision('high')` for better performance.
49
- warnings.warn(
50
- [train_sft] step=10/2000 loss=0.3106 tokens=66019 samples_per_sec=0.42 tokens_per_sec=172.88
51
- [train_sft] step=20/2000 loss=0.2507 tokens=130469 samples_per_sec=0.51 tokens_per_sec=209.38
52
- [train_sft] step=30/2000 loss=0.2291 tokens=196381 samples_per_sec=0.55 tokens_per_sec=226.84
53
- [train_sft] step=40/2000 loss=0.2799 tokens=260313 samples_per_sec=0.58 tokens_per_sec=234.92
54
- [train_sft] step=50/2000 loss=0.0738 tokens=326884 samples_per_sec=0.59 tokens_per_sec=241.97
55
- [train_sft] step=60/2000 loss=0.2171 tokens=392579 samples_per_sec=0.60 tokens_per_sec=246.20
56
- [train_sft] step=70/2000 loss=0.1810 tokens=458551 samples_per_sec=0.61 tokens_per_sec=249.57
57
- [train_sft] step=80/2000 loss=0.1329 tokens=527838 samples_per_sec=0.62 tokens_per_sec=253.71
58
- [train_sft] step=90/2000 loss=0.1080 tokens=598351 samples_per_sec=0.62 tokens_per_sec=257.55
59
- [train_sft] step=100/2000 loss=0.1271 tokens=665669 samples_per_sec=0.62 tokens_per_sec=259.31
60
- [train_sft] step=110/2000 loss=0.1596 tokens=727376 samples_per_sec=0.63 tokens_per_sec=258.71
61
- [train_sft] step=120/2000 loss=0.1730 tokens=789579 samples_per_sec=0.63 tokens_per_sec=258.39
62
- [train_sft] step=130/2000 loss=0.1666 tokens=852405 samples_per_sec=0.63 tokens_per_sec=258.28
63
- [train_sft] step=140/2000 loss=0.1658 tokens=922146 samples_per_sec=0.63 tokens_per_sec=260.12
64
- [train_sft] step=150/2000 loss=0.0985 tokens=986385 samples_per_sec=0.63 tokens_per_sec=260.33
65
- [train_sft] step=160/2000 loss=0.0701 tokens=1049899 samples_per_sec=0.63 tokens_per_sec=260.30
66
- [train_sft] step=170/2000 loss=0.0789 tokens=1114566 samples_per_sec=0.64 tokens_per_sec=260.56
67
- [train_sft] step=180/2000 loss=0.1549 tokens=1184173 samples_per_sec=0.64 tokens_per_sec=261.97
68
- [train_sft] step=190/2000 loss=0.1046 tokens=1248076 samples_per_sec=0.64 tokens_per_sec=261.99
69
- [train_sft] step=200/2000 loss=0.2644 tokens=1312805 samples_per_sec=0.64 tokens_per_sec=262.13
70
- [train_sft] eval step=200 loss=n/a
71
- [train_sft] step=210/2000 loss=0.1345 tokens=1378046 samples_per_sec=0.64 tokens_per_sec=262.34
72
- [train_sft] step=220/2000 loss=0.1363 tokens=1440118 samples_per_sec=0.64 tokens_per_sec=261.97
73
- [train_sft] step=230/2000 loss=0.1488 tokens=1506320 samples_per_sec=0.64 tokens_per_sec=262.35
74
- [train_sft] step=240/2000 loss=0.1060 tokens=1571030 samples_per_sec=0.64 tokens_per_sec=262.45
75
- [train_sft] step=250/2000 loss=0.1710 tokens=1636503 samples_per_sec=0.64 tokens_per_sec=262.64
76
- [train_sft] step=260/2000 loss=0.1923 tokens=1702374 samples_per_sec=0.64 tokens_per_sec=262.90
77
- [train_sft] step=270/2000 loss=0.1278 tokens=1765267 samples_per_sec=0.64 tokens_per_sec=262.69
78
- [train_sft] step=280/2000 loss=0.1669 tokens=1834329 samples_per_sec=0.64 tokens_per_sec=263.40
79
- [train_sft] step=290/2000 loss=0.1825 tokens=1901061 samples_per_sec=0.64 tokens_per_sec=263.73
80
- [train_sft] step=300/2000 loss=0.1071 tokens=1965869 samples_per_sec=0.64 tokens_per_sec=263.76
81
- [train_sft] step=310/2000 loss=0.1100 tokens=2026328 samples_per_sec=0.64 tokens_per_sec=263.24
82
- [train_sft] step=320/2000 loss=0.1569 tokens=2091438 samples_per_sec=0.64 tokens_per_sec=263.35
83
- [train_sft] step=330/2000 loss=0.0981 tokens=2159388 samples_per_sec=0.64 tokens_per_sec=263.79
84
- [train_sft] step=340/2000 loss=0.1418 tokens=2224937 samples_per_sec=0.65 tokens_per_sec=263.91
85
- [train_sft] step=350/2000 loss=0.1125 tokens=2290801 samples_per_sec=0.65 tokens_per_sec=264.05
86
- [train_sft] step=360/2000 loss=0.0595 tokens=2356921 samples_per_sec=0.65 tokens_per_sec=264.21
87
- [train_sft] step=370/2000 loss=0.0877 tokens=2419567 samples_per_sec=0.65 tokens_per_sec=264.01
88
- [train_sft] step=380/2000 loss=0.2055 tokens=2484959 samples_per_sec=0.65 tokens_per_sec=264.08
89
- [train_sft] step=390/2000 loss=0.2391 tokens=2548421 samples_per_sec=0.65 tokens_per_sec=263.97
90
- [train_sft] step=400/2000 loss=0.1708 tokens=2611454 samples_per_sec=0.65 tokens_per_sec=263.81
91
- [train_sft] eval step=400 loss=n/a
92
- [train_sft] step=410/2000 loss=0.1243 tokens=2673912 samples_per_sec=0.65 tokens_per_sec=263.61
93
- [train_sft] step=420/2000 loss=0.1155 tokens=2741532 samples_per_sec=0.65 tokens_per_sec=263.91
94
- [train_sft] step=430/2000 loss=0.1285 tokens=2803400 samples_per_sec=0.65 tokens_per_sec=263.66
95
- [train_sft] step=440/2000 loss=0.0763 tokens=2870087 samples_per_sec=0.65 tokens_per_sec=263.86
96
- [train_sft] step=450/2000 loss=0.1212 tokens=2938147 samples_per_sec=0.65 tokens_per_sec=264.18
97
- [train_sft] step=460/2000 loss=0.1697 tokens=3006919 samples_per_sec=0.65 tokens_per_sec=264.55
98
- [train_sft] step=470/2000 loss=0.0634 tokens=3074418 samples_per_sec=0.65 tokens_per_sec=264.79
99
- [train_sft] step=480/2000 loss=0.1071 tokens=3140612 samples_per_sec=0.65 tokens_per_sec=264.92
100
- [train_sft] step=490/2000 loss=0.0840 tokens=3207555 samples_per_sec=0.65 tokens_per_sec=265.09
101
- [train_sft] step=500/2000 loss=0.1685 tokens=3272736 samples_per_sec=0.65 tokens_per_sec=265.12
102
-
103
- [train_sft] step=510/2000 loss=11.9358 tokens=3333255 samples_per_sec=0.64 tokens_per_sec=261.82
104
- [train_sft] step=520/2000 loss=11.9305 tokens=3400813 samples_per_sec=0.64 tokens_per_sec=260.02
105
- [train_sft] step=530/2000 loss=11.9318 tokens=3468065 samples_per_sec=0.63 tokens_per_sec=258.30
106
- [train_sft] step=540/2000 loss=11.9336 tokens=3535022 samples_per_sec=0.63 tokens_per_sec=256.63
107
- [train_sft] step=550/2000 loss=11.9334 tokens=3600853 samples_per_sec=0.62 tokens_per_sec=254.97
108
- [train_sft] step=560/2000 loss=11.9341 tokens=3662071 samples_per_sec=0.62 tokens_per_sec=253.07
109
- [train_sft] step=570/2000 loss=11.9347 tokens=3730017 samples_per_sec=0.62 tokens_per_sec=251.71
110
- [train_sft] step=580/2000 loss=11.9316 tokens=3796693 samples_per_sec=0.61 tokens_per_sec=250.33
111
- [train_sft] step=590/2000 loss=11.9386 tokens=3858933 samples_per_sec=0.61 tokens_per_sec=248.73
112
- [train_sft] step=600/2000 loss=11.9369 tokens=3920622 samples_per_sec=0.61 tokens_per_sec=247.17
113
- [train_sft] eval step=600 loss=n/a
114
- [train_sft] step=610/2000 loss=11.9311 tokens=3984325 samples_per_sec=0.60 tokens_per_sec=245.79
115
- [train_sft] step=620/2000 loss=11.9305 tokens=4051226 samples_per_sec=0.60 tokens_per_sec=244.67
116
- [train_sft] step=630/2000 loss=11.9329 tokens=4117533 samples_per_sec=0.60 tokens_per_sec=243.55
117
- [train_sft] step=640/2000 loss=11.9343 tokens=4183625 samples_per_sec=0.59 tokens_per_sec=242.47
118
- [train_sft] step=650/2000 loss=11.9327 tokens=4252154 samples_per_sec=0.59 tokens_per_sec=241.57
119
- [train_sft] step=660/2000 loss=11.9336 tokens=4313556 samples_per_sec=0.59 tokens_per_sec=240.31
120
- [train_sft] step=670/2000 loss=11.9328 tokens=4376047 samples_per_sec=0.59 tokens_per_sec=239.16
121
- [train_sft] step=680/2000 loss=11.9349 tokens=4443065 samples_per_sec=0.58 tokens_per_sec=238.29
122
- [train_sft] step=690/2000 loss=11.9312 tokens=4508609 samples_per_sec=0.58 tokens_per_sec=237.37
123
- [train_sft] step=700/2000 loss=11.9365 tokens=4571767 samples_per_sec=0.58 tokens_per_sec=236.37
124
- [train_sft] step=710/2000 loss=11.9332 tokens=4635156 samples_per_sec=0.58 tokens_per_sec=235.42
125
- [train_sft] step=720/2000 loss=11.9340 tokens=4700197 samples_per_sec=0.57 tokens_per_sec=234.57
126
- [train_sft] step=730/2000 loss=11.9302 tokens=4764290 samples_per_sec=0.57 tokens_per_sec=233.71
127
- [train_sft] step=740/2000 loss=11.9358 tokens=4831552 samples_per_sec=0.57 tokens_per_sec=233.03
128
- [train_sft] step=750/2000 loss=11.9338 tokens=4896926 samples_per_sec=0.57 tokens_per_sec=232.29
129
- [train_sft] step=760/2000 loss=11.9299 tokens=4963790 samples_per_sec=0.57 tokens_per_sec=231.64
130
- [train_sft] step=770/2000 loss=11.9375 tokens=5030112 samples_per_sec=0.57 tokens_per_sec=230.98
131
- [train_sft] step=780/2000 loss=11.9328 tokens=5093917 samples_per_sec=0.56 tokens_per_sec=230.23
132
- [train_sft] step=790/2000 loss=11.9376 tokens=5160155 samples_per_sec=0.56 tokens_per_sec=229.61
133
- [train_sft] step=800/2000 loss=11.9353 tokens=5229774 samples_per_sec=0.56 tokens_per_sec=229.16
134
- [train_sft] eval step=800 loss=n/a
135
- [train_sft] step=810/2000 loss=11.9334 tokens=5295154 samples_per_sec=0.56 tokens_per_sec=228.54
136
- [train_sft] step=820/2000 loss=11.9345 tokens=5357311 samples_per_sec=0.56 tokens_per_sec=227.80
137
- [train_sft] step=830/2000 loss=11.9387 tokens=5422927 samples_per_sec=0.56 tokens_per_sec=227.23
138
- [train_sft] step=840/2000 loss=11.9358 tokens=5490488 samples_per_sec=0.56 tokens_per_sec=226.75
139
- [train_sft] step=850/2000 loss=11.9340 tokens=5551808 samples_per_sec=0.55 tokens_per_sec=226.04
140
- [train_sft] step=860/2000 loss=11.9372 tokens=5617359 samples_per_sec=0.55 tokens_per_sec=225.52
141
- [train_sft] step=870/2000 loss=11.9311 tokens=5685346 samples_per_sec=0.55 tokens_per_sec=225.10
142
- [train_sft] step=880/2000 loss=11.9340 tokens=5750741 samples_per_sec=0.55 tokens_per_sec=224.59
143
- [train_sft] step=890/2000 loss=11.9300 tokens=5816629 samples_per_sec=0.55 tokens_per_sec=224.12
144
- [train_sft] step=900/2000 loss=11.9343 tokens=5882089 samples_per_sec=0.55 tokens_per_sec=223.64
145
- [train_sft] step=910/2000 loss=11.9320 tokens=5948442 samples_per_sec=0.55 tokens_per_sec=223.21
146
- [train_sft] step=920/2000 loss=11.9363 tokens=6009051 samples_per_sec=0.55 tokens_per_sec=222.58
147
- [train_sft] step=930/2000 loss=11.9309 tokens=6078081 samples_per_sec=0.54 tokens_per_sec=222.27
148
- [train_sft] step=940/2000 loss=11.9353 tokens=6138758 samples_per_sec=0.54 tokens_per_sec=221.66
149
- [train_sft] step=950/2000 loss=11.9371 tokens=6204982 samples_per_sec=0.54 tokens_per_sec=221.27
150
- [train_sft] step=960/2000 loss=11.9341 tokens=6272247 samples_per_sec=0.54 tokens_per_sec=220.93
151
- [train_sft] step=970/2000 loss=11.9314 tokens=6337243 samples_per_sec=0.54 tokens_per_sec=220.51
152
- [train_sft] step=980/2000 loss=11.9361 tokens=6401913 samples_per_sec=0.54 tokens_per_sec=220.10
153
- [train_sft] step=990/2000 loss=11.9325 tokens=6465635 samples_per_sec=0.54 tokens_per_sec=219.66
154
- [train_sft] step=1000/2000 loss=11.9347 tokens=6532582 samples_per_sec=0.54 tokens_per_sec=219.34
155
- [train_sft] eval step=1000 loss=n/a
156
-
157
- [train_sft] step=1010/2000 loss=nan tokens=6598011 samples_per_sec=0.54 tokens_per_sec=219.25
158
- W0307 11:46:34.841000 29377 torch/distributed/elastic/agent/server/api.py:719] Received Signals.SIGINT death signal, shutting down workers
159
- W0307 11:46:34.843000 29377 torch/distributed/elastic/multiprocessing/api.py:897] Sending process 29390 closing signal SIGINT
160
- W0307 11:46:34.843000 29377 torch/distributed/elastic/multiprocessing/api.py:897] Sending process 29391 closing signal SIGINT
161
- W0307 11:46:34.844000 29377 torch/distributed/elastic/multiprocessing/api.py:897] Sending process 29392 closing signal SIGINT
162
- W0307 11:46:34.844000 29377 torch/distributed/elastic/multiprocessing/api.py:897] Sending process 29393 closing signal SIGINT
163
- W0307 11:46:34.844000 29377 torch/distributed/elastic/multiprocessing/api.py:897] Sending process 29394 closing signal SIGINT
164
- W0307 11:46:34.844000 29377 torch/distributed/elastic/multiprocessing/api.py:897] Sending process 29395 closing signal SIGINT
165
- W0307 11:46:34.845000 29377 torch/distributed/elastic/multiprocessing/api.py:897] Sending process 29396 closing signal SIGINT
166
- W0307 11:46:34.845000 29377 torch/distributed/elastic/multiprocessing/api.py:897] Sending process 29397 closing signal SIGINT
167
- W0307 11:47:04.846000 29377 torch/distributed/elastic/multiprocessing/api.py:916] Unable to shutdown process 29390 via Signals.SIGINT, forcefully exiting via Signals.SIGKILL
168
- W0307 11:47:05.104000 29377 torch/distributed/elastic/multiprocessing/api.py:916] Unable to shutdown process 29391 via Signals.SIGINT, forcefully exiting via Signals.SIGKILL
169
- W0307 11:47:05.500000 29377 torch/distributed/elastic/multiprocessing/api.py:916] Unable to shutdown process 29392 via Signals.SIGINT, forcefully exiting via Signals.SIGKILL
170
- W0307 11:47:05.889000 29377 torch/distributed/elastic/multiprocessing/api.py:916] Unable to shutdown process 29393 via Signals.SIGINT, forcefully exiting via Signals.SIGKILL
171
- W0307 11:47:06.314000 29377 torch/distributed/elastic/multiprocessing/api.py:916] Unable to shutdown process 29394 via Signals.SIGINT, forcefully exiting via Signals.SIGKILL
172
- W0307 11:47:06.691000 29377 torch/distributed/elastic/multiprocessing/api.py:916] Unable to shutdown process 29395 via Signals.SIGINT, forcefully exiting via Signals.SIGKILL
173
- W0307 11:47:07.074000 29377 torch/distributed/elastic/multiprocessing/api.py:916] Unable to shutdown process 29396 via Signals.SIGINT, forcefully exiting via Signals.SIGKILL
174
- W0307 11:47:07.463000 29377 torch/distributed/elastic/multiprocessing/api.py:916] Unable to shutdown process 29397 via Signals.SIGINT, forcefully exiting via Signals.SIGKILL
175
- Traceback (most recent call last):
176
- File "/root/PyC/.venv/bin/torchrun", line 6, in <module>
177
- sys.exit(main())
178
- File "/root/PyC/.venv/lib/python3.10/site-packages/torch/distributed/elastic/multiprocessing/errors/__init__.py", line 355, in wrapper
179
- return f(*args, **kwargs)
180
- File "/root/PyC/.venv/lib/python3.10/site-packages/torch/distributed/run.py", line 918, in main
181
- run(args)
182
- File "/root/PyC/.venv/lib/python3.10/site-packages/torch/distributed/run.py", line 909, in run
183
- elastic_launch(
184
- File "/root/PyC/.venv/lib/python3.10/site-packages/torch/distributed/launcher/api.py", line 138, in __call__
185
- return launch_agent(self._config, self._entrypoint, list(args))
186
- File "/root/PyC/.venv/lib/python3.10/site-packages/torch/distributed/launcher/api.py", line 260, in launch_agent
187
- result = agent.run()
188
- File "/root/PyC/.venv/lib/python3.10/site-packages/torch/distributed/elastic/metrics/api.py", line 137, in wrapper
189
- result = f(*args, **kwargs)
190
- File "/root/PyC/.venv/lib/python3.10/site-packages/torch/distributed/elastic/agent/server/api.py", line 711, in run
191
- result = self._invoke_run(role)
192
- File "/root/PyC/.venv/lib/python3.10/site-packages/torch/distributed/elastic/agent/server/api.py", line 870, in _invoke_run
193
- time.sleep(monitor_interval)
194
- File "/root/PyC/.venv/lib/python3.10/site-packages/torch/distributed/elastic/multiprocessing/api.py", line 84, in _terminate_process_handler
195
- raise SignalException(f"Process {os.getpid()} got signal: {sigval}", sigval=sigval)
196
- torch.distributed.elastic.multiprocessing.api.SignalException: Process 29377 got signal: 2
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
run_data/train_metrics.json DELETED
@@ -1,286 +0,0 @@
1
- {
2
- "run_id": "qwen14b_codeinstruct",
3
- "timestamp_utc": "2026-03-07T03:24:36Z",
4
- "host": "5cea9450de15",
5
- "train_runtime_sec": 30093.7884,
6
- "samples_per_sec": 0.537,
7
- "steps_per_sec": 0.0336,
8
- "tokens_per_sec": 219.2483,
9
- "loss_final": NaN,
10
- "loss_curve": [
11
- 11.933758,
12
- 11.935425,
13
- 11.936735,
14
- 11.934477,
15
- 11.936081,
16
- 11.934277,
17
- 11.933105,
18
- 11.934149,
19
- 11.931114,
20
- 11.931803,
21
- 11.931644,
22
- 11.933023,
23
- 11.935471,
24
- 11.934968,
25
- 11.935336,
26
- 11.936069,
27
- 11.931381,
28
- 11.929764,
29
- 11.933843,
30
- 11.932031,
31
- 11.934188,
32
- 11.933489,
33
- 11.932564,
34
- 11.933738,
35
- 11.933002,
36
- 11.935679,
37
- 11.930973,
38
- 11.933707,
39
- 11.935345,
40
- 11.932213,
41
- 11.931738,
42
- 11.938071,
43
- 11.934389,
44
- 11.935558,
45
- 11.933566,
46
- 11.931173,
47
- 11.933773,
48
- 11.930953,
49
- 11.933154,
50
- 11.935999,
51
- 11.934997,
52
- 11.934099,
53
- 11.93278,
54
- 11.938663,
55
- 11.933143,
56
- 11.932686,
57
- 11.934217,
58
- 11.934953,
59
- 11.936878,
60
- 11.933104,
61
- 11.930088,
62
- 11.929935,
63
- 11.932352,
64
- 11.932141,
65
- 11.934212,
66
- 11.937968,
67
- 11.93185,
68
- 11.935164,
69
- 11.933124,
70
- 11.934265,
71
- 11.931581,
72
- 11.93028,
73
- 11.935074,
74
- 11.930362,
75
- 11.93231,
76
- 11.934386,
77
- 11.931125,
78
- 11.933617,
79
- 11.934947,
80
- 11.935481,
81
- 11.931095,
82
- 11.936635,
83
- 11.934564,
84
- 11.934128,
85
- 11.933428,
86
- 11.933599,
87
- 11.93398,
88
- 11.930263,
89
- 11.934587,
90
- 11.933071,
91
- 11.932242,
92
- 11.933808,
93
- 11.931849,
94
- 11.936655,
95
- 11.932829,
96
- 11.931339,
97
- 11.936092,
98
- 11.935149,
99
- 11.934838,
100
- 11.932211,
101
- 11.939731,
102
- 11.935974,
103
- 11.93283,
104
- 11.933028,
105
- 11.938888,
106
- 11.932542,
107
- 11.935485,
108
- 11.93673,
109
- 11.930267,
110
- 11.933331,
111
- 11.934375,
112
- 11.932673,
113
- 11.93618,
114
- 11.933468,
115
- 11.932549,
116
- 11.936689,
117
- 11.932197,
118
- 11.934677,
119
- 11.934959,
120
- 11.932065,
121
- 11.934155,
122
- 11.9322,
123
- 11.935388,
124
- 11.935803,
125
- 11.936669,
126
- 11.929192,
127
- 11.937821,
128
- 11.933041,
129
- 11.933342,
130
- 11.935324,
131
- 11.934742,
132
- 11.934669,
133
- 11.935793,
134
- 11.937335,
135
- 11.931247,
136
- 11.932714,
137
- 11.933646,
138
- 11.935673,
139
- 11.932477,
140
- 11.931051,
141
- 11.935923,
142
- 11.935831,
143
- 11.935593,
144
- 11.937632,
145
- 11.933506,
146
- 11.933863,
147
- 11.936568,
148
- 11.93511,
149
- 11.929846,
150
- 11.928335,
151
- 11.935637,
152
- 11.931632,
153
- 11.933343,
154
- 11.932386,
155
- 11.935142,
156
- 11.934553,
157
- 11.937015,
158
- 11.930168,
159
- 11.937183,
160
- 11.934591,
161
- 11.937156,
162
- 11.934317,
163
- 11.933546,
164
- 11.937324,
165
- 11.92873,
166
- 11.934661,
167
- 11.935086,
168
- 11.934525,
169
- 11.934164,
170
- 11.933876,
171
- 11.932452,
172
- 11.932648,
173
- 11.934206,
174
- 11.932563,
175
- 11.934017,
176
- 11.934988,
177
- 11.933892,
178
- 11.933537,
179
- 11.929354,
180
- 11.935569,
181
- 11.93031,
182
- 11.934333,
183
- 11.934294,
184
- 11.931088,
185
- 11.936719,
186
- 11.934738,
187
- 11.932435,
188
- 11.934471,
189
- 11.934482,
190
- 11.93504,
191
- 11.934249,
192
- 11.931784,
193
- 11.932494,
194
- 11.932414,
195
- 11.933865,
196
- 11.931005,
197
- 11.93541,
198
- 11.932836,
199
- 11.935291,
200
- 11.933075,
201
- 11.931273,
202
- 11.931405,
203
- 11.93127,
204
- 11.931889,
205
- 11.935592,
206
- 11.930029,
207
- 11.935498,
208
- 11.934025,
209
- 11.935754,
210
- 11.932719,
211
- 11.933537,
212
- 11.935834,
213
- 11.933773,
214
- 11.931642,
215
- 11.931832,
216
- 11.934774,
217
- 11.932487,
218
- 11.932528,
219
- 11.932735,
220
- 11.934239,
221
- 11.93345,
222
- 11.933942,
223
- 11.931937,
224
- 11.937059,
225
- 11.938506,
226
- 11.932685,
227
- 11.932717,
228
- 11.936692,
229
- 11.9363,
230
- 11.936151,
231
- 11.931432,
232
- 11.933336,
233
- 11.932334,
234
- 11.932056,
235
- 11.934895,
236
- 11.933721,
237
- 11.934547,
238
- 11.933809,
239
- 11.933655,
240
- 11.933565,
241
- 11.933448,
242
- 11.932833,
243
- 11.936954,
244
- 11.935034,
245
- 11.931088,
246
- 11.935302,
247
- 11.933675,
248
- 11.935065,
249
- 11.935748,
250
- 11.932056,
251
- 11.938281,
252
- 11.934908,
253
- 11.932071,
254
- 11.933029,
255
- 11.934909,
256
- 11.933907,
257
- 11.933957,
258
- NaN,
259
- NaN,
260
- NaN,
261
- NaN,
262
- NaN,
263
- NaN,
264
- NaN,
265
- NaN,
266
- NaN
267
- ],
268
- "eval_loss": null,
269
- "gpu_util_mean": 99.2073,
270
- "gpu_util_p95": 100.0,
271
- "h2d_time_ms_mean": 0.0726,
272
- "compute_time_ms_mean": 3697.3475,
273
- "comm_time_ms_mean": 6.5765,
274
- "idle_gap_ms_mean": 11.6392,
275
- "idle_gap_ms_p95": 0.3205,
276
- "mode": "sft",
277
- "dist": "fsdp",
278
- "backend": "nccl",
279
- "world_size": 8,
280
- "model_name": "Qwen/Qwen2.5-14B",
281
- "dataset_name": "nvidia/OpenCodeInstruct",
282
- "dataset_config": "",
283
- "precision": "bf16",
284
- "seq_length": 4096,
285
- "gradient_accumulation_steps": 8
286
- }