wjbmattingly commited on
Commit
8086dc2
·
verified ·
1 Parent(s): 135c6c6

Upload folder using huggingface_hub

Browse files
README.md CHANGED
@@ -3,68 +3,75 @@ base_model:
3
  - LiquidAI/LFM2-VL-450M
4
  ---
5
 
6
- # model_step_300
7
 
8
  ## Model Description
9
 
10
  This model is a fine-tuned version of **LiquidAI/LFM2-VL-450M** using the brute-force-training package.
11
 
12
  - **Base Model**: LiquidAI/LFM2-VL-450M
13
- - **Training Status**: 🔄 In Progress
14
- - **Generated**: 2025-08-14 21:03:04
15
- - **Training Steps**: 300
16
 
17
  ## Training Details
18
 
19
  ### Dataset
20
- - **Dataset**: wjbmattingly/medieval-letters-htr-synthetic
21
  - **Training Examples**: 148,000
22
  - **Validation Examples**: 1,999
23
 
24
  ### Training Configuration
25
- - **Max Steps**: 1,000
26
- - **Batch Size**: 10
27
- - **Learning Rate**: 1e-05
28
- - **Gradient Accumulation**: 4 steps
29
- - **Evaluation Frequency**: Every 100 steps
30
 
31
  ### Current Performance
32
- - **Training Loss**: 4.575191
33
- - **Evaluation Loss**: 4.492388
34
 
35
  ## Pre-Training Evaluation
36
 
37
  **Initial Model Performance (before training):**
38
- - **Loss**: 5.419081
39
- - **Perplexity**: 225.67
40
- - **Character Accuracy**: 11.4%
41
- - **Word Accuracy**: 6.9%
42
 
43
  ## Evaluation History
44
 
45
  ### All Checkpoint Evaluations
46
  | Step | Checkpoint Type | Loss | Perplexity | Char Acc | Word Acc | Improvement vs Pre |
47
  |------|----------------|------|------------|----------|----------|--------------------|
48
- | Pre | pre_training | 5.4191 | 225.67 | 11.4% | 6.9% | +0.0% |
49
- | 100 | checkpoint | 4.8218 | 124.19 | 10.6% | 6.4% | +11.0% |
50
- | 200 | checkpoint | 4.6127 | 100.76 | 10.8% | 6.4% | +14.9% |
51
- | 300 | checkpoint | 4.4924 | 89.33 | 10.6% | 6.4% | +17.1% |
 
 
 
 
 
 
 
52
 
53
  ## Training Progress
54
 
55
  ### Recent Training Steps (Loss Only)
56
  | Step | Training Loss | Timestamp |
57
  |------|---------------|-----------|
58
- | 291 | 4.513980 | 2025-08-14T20:59 |
59
- | 292 | 4.423133 | 2025-08-14T20:59 |
60
- | 293 | 4.640058 | 2025-08-14T20:59 |
61
- | 294 | 4.488780 | 2025-08-14T20:59 |
62
- | 295 | 4.027132 | 2025-08-14T20:59 |
63
- | 296 | 4.805581 | 2025-08-14T20:59 |
64
- | 297 | 4.652530 | 2025-08-14T21:00 |
65
- | 298 | 4.494508 | 2025-08-14T21:00 |
66
- | 299 | 4.449580 | 2025-08-14T21:00 |
67
- | 300 | 4.575191 | 2025-08-14T21:00 |
68
 
69
  ## Training Visualizations
70
 
@@ -88,8 +95,8 @@ This model is a fine-tuned version of **LiquidAI/LFM2-VL-450M** using the brute-
88
  from transformers import AutoModelForCausalLM, AutoTokenizer
89
  # For vision-language models, use appropriate imports
90
 
91
- model = AutoModelForCausalLM.from_pretrained("./model_step_300")
92
- tokenizer = AutoTokenizer.from_pretrained("./model_step_300")
93
 
94
  # Your inference code here
95
  ```
@@ -98,23 +105,23 @@ tokenizer = AutoTokenizer.from_pretrained("./model_step_300")
98
 
99
  ```json
100
  {
101
- "dataset_name": "wjbmattingly/medieval-letters-htr-synthetic",
102
  "model_name": "LiquidAI/LFM2-VL-450M",
103
- "max_steps": 1000,
104
- "eval_steps": 100,
105
- "num_accumulation_steps": 4,
106
- "learning_rate": 1e-05,
107
- "train_batch_size": 10,
108
- "val_batch_size": 10,
109
  "train_select_start": 0,
110
  "train_select_end": 148000,
111
  "val_select_start": 148001,
112
  "val_select_end": 150000,
113
  "train_field": "train",
114
  "val_field": "train",
115
- "image_column": "image",
116
- "text_column": "transcription",
117
- "user_text": "Transcribe this medieval manuscript line",
118
  "max_image_size": 200
119
  }
120
  ```
@@ -129,4 +136,4 @@ tokenizer = AutoTokenizer.from_pretrained("./model_step_300")
129
 
130
  ---
131
 
132
- *This model card was automatically generated by brute-force-training on 2025-08-14 21:03:04*
 
3
  - LiquidAI/LFM2-VL-450M
4
  ---
5
 
6
+ # final
7
 
8
  ## Model Description
9
 
10
  This model is a fine-tuned version of **LiquidAI/LFM2-VL-450M** using the brute-force-training package.
11
 
12
  - **Base Model**: LiquidAI/LFM2-VL-450M
13
+ - **Training Status**: Complete
14
+ - **Generated**: 2025-08-18 19:51:23
15
+ - **Training Steps**: 10,000
16
 
17
  ## Training Details
18
 
19
  ### Dataset
20
+ - **Dataset**: CATMuS/medieval
21
  - **Training Examples**: 148,000
22
  - **Validation Examples**: 1,999
23
 
24
  ### Training Configuration
25
+ - **Max Steps**: 10,000
26
+ - **Batch Size**: 2
27
+ - **Learning Rate**: 5e-06
28
+ - **Gradient Accumulation**: 2 steps
29
+ - **Evaluation Frequency**: Every 1,000 steps
30
 
31
  ### Current Performance
32
+ - **Training Loss**: 0.918973
33
+ - **Evaluation Loss**: 3.461486
34
 
35
  ## Pre-Training Evaluation
36
 
37
  **Initial Model Performance (before training):**
38
+ - **Loss**: 6.255835
39
+ - **Perplexity**: 521.04
40
+ - **Character Accuracy**: 27.7%
41
+ - **Word Accuracy**: 11.6%
42
 
43
  ## Evaluation History
44
 
45
  ### All Checkpoint Evaluations
46
  | Step | Checkpoint Type | Loss | Perplexity | Char Acc | Word Acc | Improvement vs Pre |
47
  |------|----------------|------|------------|----------|----------|--------------------|
48
+ | Pre | pre_training | 6.2558 | 521.04 | 27.7% | 11.6% | +0.0% |
49
+ | 1,000 | checkpoint | 4.4533 | 85.91 | 33.8% | 16.1% | +28.8% |
50
+ | 2,000 | checkpoint | 4.1024 | 60.49 | 32.5% | 15.1% | +34.4% |
51
+ | 3,000 | checkpoint | 3.9043 | 49.62 | 33.2% | 16.9% | +37.6% |
52
+ | 4,000 | checkpoint | 3.7561 | 42.78 | 30.9% | 14.2% | +40.0% |
53
+ | 5,000 | checkpoint | 3.6675 | 39.15 | 33.5% | 17.0% | +41.4% |
54
+ | 6,000 | checkpoint | 3.6180 | 37.26 | 31.8% | 15.1% | +42.2% |
55
+ | 7,000 | checkpoint | 3.5651 | 35.34 | 32.2% | 15.6% | +43.0% |
56
+ | 8,000 | checkpoint | 3.5113 | 33.49 | 30.6% | 14.2% | +43.9% |
57
+ | 9,000 | checkpoint | 3.4908 | 32.81 | 33.4% | 16.9% | +44.2% |
58
+ | 10,000 | final | 3.4615 | 31.86 | 32.5% | 16.5% | +44.7% |
59
 
60
  ## Training Progress
61
 
62
  ### Recent Training Steps (Loss Only)
63
  | Step | Training Loss | Timestamp |
64
  |------|---------------|-----------|
65
+ | 9,991 | 0.764521 | 2025-08-18T19:50 |
66
+ | 9,992 | 0.948460 | 2025-08-18T19:50 |
67
+ | 9,993 | 0.758166 | 2025-08-18T19:50 |
68
+ | 9,994 | 0.898506 | 2025-08-18T19:50 |
69
+ | 9,995 | 0.784889 | 2025-08-18T19:50 |
70
+ | 9,996 | 0.786168 | 2025-08-18T19:50 |
71
+ | 9,997 | 0.674831 | 2025-08-18T19:50 |
72
+ | 9,998 | 0.950868 | 2025-08-18T19:50 |
73
+ | 9,999 | 0.960045 | 2025-08-18T19:50 |
74
+ | 10,000 | 0.918973 | 2025-08-18T19:50 |
75
 
76
  ## Training Visualizations
77
 
 
95
  from transformers import AutoModelForCausalLM, AutoTokenizer
96
  # For vision-language models, use appropriate imports
97
 
98
+ model = AutoModelForCausalLM.from_pretrained("./final")
99
+ tokenizer = AutoTokenizer.from_pretrained("./final")
100
 
101
  # Your inference code here
102
  ```
 
105
 
106
  ```json
107
  {
108
+ "dataset_name": "CATMuS/medieval",
109
  "model_name": "LiquidAI/LFM2-VL-450M",
110
+ "max_steps": 10000,
111
+ "eval_steps": 1000,
112
+ "num_accumulation_steps": 2,
113
+ "learning_rate": 5e-06,
114
+ "train_batch_size": 2,
115
+ "val_batch_size": 2,
116
  "train_select_start": 0,
117
  "train_select_end": 148000,
118
  "val_select_start": 148001,
119
  "val_select_end": 150000,
120
  "train_field": "train",
121
  "val_field": "train",
122
+ "image_column": "im",
123
+ "text_column": "text",
124
+ "user_text": "Transcribe this medieval manuscript line.",
125
  "max_image_size": 200
126
  }
127
  ```
 
136
 
137
  ---
138
 
139
+ *This model card was automatically generated by brute-force-training on 2025-08-18 19:51:23*
evaluation_comparison.png CHANGED

Git LFS Details

  • SHA256: ee921359acbb367ea0f72a32b1f4bde433e6e38fd1828773286eb1b540acb5f4
  • Pointer size: 131 Bytes
  • Size of remote file: 321 kB

Git LFS Details

  • SHA256: b3328db6b598bc3ef5ccff86e00b1281d3aa556b8e0121c6871bdd7d6320a148
  • Pointer size: 131 Bytes
  • Size of remote file: 472 kB
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:94846411f2c1171e1ec1a3aff354d161896539bd14085e6bd518449a8c04eaf4
3
  size 901692416
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7d1c7bf9a67a65d0305165d7d04b3a5275ece0b8040925556407941f0b971da8
3
  size 901692416
model_card_metadata.json CHANGED
@@ -1,31 +1,31 @@
1
  {
2
  "base_model": "LiquidAI/LFM2-VL-450M",
3
  "training_framework": "brute-force-training",
4
- "training_date": "2025-08-14T21:03:04.407697",
5
- "training_steps": 300,
6
- "dataset": "wjbmattingly/medieval-letters-htr-synthetic",
7
  "training_config": {
8
- "dataset_name": "wjbmattingly/medieval-letters-htr-synthetic",
9
  "model_name": "LiquidAI/LFM2-VL-450M",
10
- "max_steps": 1000,
11
- "eval_steps": 100,
12
- "num_accumulation_steps": 4,
13
- "learning_rate": 1e-05,
14
- "train_batch_size": 10,
15
- "val_batch_size": 10,
16
  "train_select_start": 0,
17
  "train_select_end": 148000,
18
  "val_select_start": 148001,
19
  "val_select_end": 150000,
20
  "train_field": "train",
21
  "val_field": "train",
22
- "image_column": "image",
23
- "text_column": "transcription",
24
- "user_text": "Transcribe this medieval manuscript line",
25
  "max_image_size": 200
26
  },
27
- "final_training_loss": 4.575191020965576,
28
- "final_evaluation_loss": 4.492388168970744,
29
- "final_char_accuracy": 0.1061965003999563,
30
- "final_word_accuracy": 0.06438001131166525
31
  }
 
1
  {
2
  "base_model": "LiquidAI/LFM2-VL-450M",
3
  "training_framework": "brute-force-training",
4
+ "training_date": "2025-08-18T19:51:23.156386",
5
+ "training_steps": 10000,
6
+ "dataset": "CATMuS/medieval",
7
  "training_config": {
8
+ "dataset_name": "CATMuS/medieval",
9
  "model_name": "LiquidAI/LFM2-VL-450M",
10
+ "max_steps": 10000,
11
+ "eval_steps": 1000,
12
+ "num_accumulation_steps": 2,
13
+ "learning_rate": 5e-06,
14
+ "train_batch_size": 2,
15
+ "val_batch_size": 2,
16
  "train_select_start": 0,
17
  "train_select_end": 148000,
18
  "val_select_start": 148001,
19
  "val_select_end": 150000,
20
  "train_field": "train",
21
  "val_field": "train",
22
+ "image_column": "im",
23
+ "text_column": "text",
24
+ "user_text": "Transcribe this medieval manuscript line.",
25
  "max_image_size": 200
26
  },
27
+ "final_training_loss": 0.9189727306365967,
28
+ "final_evaluation_loss": 3.461486041545868,
29
+ "final_char_accuracy": 0.3254617025073297,
30
+ "final_word_accuracy": 0.16456672762721147
31
  }
training_curves.png CHANGED

Git LFS Details

  • SHA256: 93448b67f97f101190f7e32ed6381fbbb9028e12d05ae0db3cbf70b3921b379f
  • Pointer size: 131 Bytes
  • Size of remote file: 587 kB

Git LFS Details

  • SHA256: 1d301c3edc0615f896061d9b927c4988efda607eb1418c8b9538aba173501aaf
  • Pointer size: 131 Bytes
  • Size of remote file: 447 kB
training_metrics.json CHANGED
The diff for this file is too large to render. See raw diff