config.json: 0%| | 0.00/560 [00:00 [2025-11-05 14:40:58,349] [DEBUG] [axolotl.loaders.tokenizer.load_tokenizer:279] [PID:989] BOS: 1 / [2025-11-05 14:40:58,349] [DEBUG] [axolotl.loaders.tokenizer.load_tokenizer:280] [PID:989] PAD: 2 / [2025-11-05 14:40:58,349] [DEBUG] [axolotl.loaders.tokenizer.load_tokenizer:281] [PID:989] UNK: 0 / [2025-11-05 14:40:58,349] [INFO] [axolotl.loaders.tokenizer.load_tokenizer:295] [PID:989] No Chat template selected. Consider adding a chat template for easier inference. [2025-11-05 14:40:58,350] [INFO] [axolotl.utils.data.shared.load_preprocessed_dataset:476] [PID:989] Unable to find prepared dataset in last_run_prepared/0ab988fe8f81ce9dc593498716abd51f [2025-11-05 14:40:58,350] [INFO] [axolotl.utils.data.sft._load_raw_datasets:320] [PID:989] Loading raw datasets... [2025-11-05 14:40:58,350] [WARNING] [axolotl.utils.data.sft._load_raw_datasets:322] [PID:989] Processing datasets during training can lead to VRAM instability. Please pre-process your dataset using `axolotl preprocess path/to/config.yml`. README.md: 0%| | 0.00/31.0 [00:004096) (num_proc=26): 0%| | 0/1000 [00:004096) (num_proc=26): 4%|████▍ | 39/1000 [00:00<00:10, 88.06 examples/s] Dropping Long Sequences (>4096) (num_proc=26): 100%|███████████████████████████████████████████████████████████████████████████████████████████████████████████████| 1000/1000 [00:00<00:00, 1491.76 examples/s] Saving the dataset (0/3 shards): 0%| | 0/1000 [00:00 [2025-11-05 14:41:05,471] [DEBUG] [axolotl.loaders.tokenizer.load_tokenizer:279] [PID:989] BOS: 1 / [2025-11-05 14:41:05,471] [DEBUG] [axolotl.loaders.tokenizer.load_tokenizer:280] [PID:989] PAD: 2 / [2025-11-05 14:41:05,471] [DEBUG] [axolotl.loaders.tokenizer.load_tokenizer:281] [PID:989] UNK: 0 / [2025-11-05 14:41:05,471] [INFO] [axolotl.loaders.tokenizer.load_tokenizer:295] [PID:989] No Chat template selected. Consider adding a chat template for easier inference. [2025-11-05 14:41:05,471] [DEBUG] [axolotl.train.setup_model_and_tokenizer:74] [PID:989] Loading model [2025-11-05 14:41:05,685] [DEBUG] [axolotl.monkeypatch.transformers.trainer_loss_calc.patch_evaluation_loop:87] [PID:989] Patched Trainer.evaluation_loop with nanmean loss calculation [2025-11-05 14:41:05,688] [DEBUG] [axolotl.monkeypatch.transformers.trainer_loss_calc.patch_maybe_log_save_evaluate:138] [PID:989] Patched Trainer._maybe_log_save_evaluate with nanmean loss calculation pytorch_model.bin: 0%| | 0.00/4.40G [00:00