llama3-3b-coding-gpt4o-100k2 / train_results.json
chansung's picture
Model save
3da6413 verified
raw
history blame contribute delete
251 Bytes
{
"epoch": 9.85925925925926,
"total_flos": 2.9601022627828204e+18,
"train_loss": 0.9062140895359552,
"train_runtime": 3484.2972,
"train_samples": 116368,
"train_samples_per_second": 49.516,
"train_steps_per_second": 0.192
}