wjbmattingly commited on
Commit
135c6c6
·
verified ·
1 Parent(s): 7add64a

Upload folder using huggingface_hub

Browse files
README.md CHANGED
@@ -3,7 +3,7 @@ base_model:
3
  - LiquidAI/LFM2-VL-450M
4
  ---
5
 
6
- # model_step_100
7
 
8
  ## Model Description
9
 
@@ -11,8 +11,8 @@ This model is a fine-tuned version of **LiquidAI/LFM2-VL-450M** using the brute-
11
 
12
  - **Base Model**: LiquidAI/LFM2-VL-450M
13
  - **Training Status**: 🔄 In Progress
14
- - **Generated**: 2025-08-14 20:55:09
15
- - **Training Steps**: 100
16
 
17
  ## Training Details
18
 
@@ -29,8 +29,8 @@ This model is a fine-tuned version of **LiquidAI/LFM2-VL-450M** using the brute-
29
  - **Evaluation Frequency**: Every 100 steps
30
 
31
  ### Current Performance
32
- - **Training Loss**: 4.847933
33
- - **Evaluation Loss**: 4.821800
34
 
35
  ## Pre-Training Evaluation
36
 
@@ -47,22 +47,24 @@ This model is a fine-tuned version of **LiquidAI/LFM2-VL-450M** using the brute-
47
  |------|----------------|------|------------|----------|----------|--------------------|
48
  | Pre | pre_training | 5.4191 | 225.67 | 11.4% | 6.9% | +0.0% |
49
  | 100 | checkpoint | 4.8218 | 124.19 | 10.6% | 6.4% | +11.0% |
 
 
50
 
51
  ## Training Progress
52
 
53
  ### Recent Training Steps (Loss Only)
54
  | Step | Training Loss | Timestamp |
55
  |------|---------------|-----------|
56
- | 91 | 5.012662 | 2025-08-14T20:51 |
57
- | 92 | 4.833282 | 2025-08-14T20:52 |
58
- | 93 | 4.884547 | 2025-08-14T20:52 |
59
- | 94 | 4.841271 | 2025-08-14T20:52 |
60
- | 95 | 4.992236 | 2025-08-14T20:52 |
61
- | 96 | 4.861138 | 2025-08-14T20:52 |
62
- | 97 | 4.886461 | 2025-08-14T20:52 |
63
- | 98 | 4.913207 | 2025-08-14T20:52 |
64
- | 99 | 5.008471 | 2025-08-14T20:52 |
65
- | 100 | 4.847933 | 2025-08-14T20:52 |
66
 
67
  ## Training Visualizations
68
 
@@ -86,8 +88,8 @@ This model is a fine-tuned version of **LiquidAI/LFM2-VL-450M** using the brute-
86
  from transformers import AutoModelForCausalLM, AutoTokenizer
87
  # For vision-language models, use appropriate imports
88
 
89
- model = AutoModelForCausalLM.from_pretrained("./model_step_100")
90
- tokenizer = AutoTokenizer.from_pretrained("./model_step_100")
91
 
92
  # Your inference code here
93
  ```
@@ -127,4 +129,4 @@ tokenizer = AutoTokenizer.from_pretrained("./model_step_100")
127
 
128
  ---
129
 
130
- *This model card was automatically generated by brute-force-training on 2025-08-14 20:55:09*
 
3
  - LiquidAI/LFM2-VL-450M
4
  ---
5
 
6
+ # model_step_300
7
 
8
  ## Model Description
9
 
 
11
 
12
  - **Base Model**: LiquidAI/LFM2-VL-450M
13
  - **Training Status**: 🔄 In Progress
14
+ - **Generated**: 2025-08-14 21:03:04
15
+ - **Training Steps**: 300
16
 
17
  ## Training Details
18
 
 
29
  - **Evaluation Frequency**: Every 100 steps
30
 
31
  ### Current Performance
32
+ - **Training Loss**: 4.575191
33
+ - **Evaluation Loss**: 4.492388
34
 
35
  ## Pre-Training Evaluation
36
 
 
47
  |------|----------------|------|------------|----------|----------|--------------------|
48
  | Pre | pre_training | 5.4191 | 225.67 | 11.4% | 6.9% | +0.0% |
49
  | 100 | checkpoint | 4.8218 | 124.19 | 10.6% | 6.4% | +11.0% |
50
+ | 200 | checkpoint | 4.6127 | 100.76 | 10.8% | 6.4% | +14.9% |
51
+ | 300 | checkpoint | 4.4924 | 89.33 | 10.6% | 6.4% | +17.1% |
52
 
53
  ## Training Progress
54
 
55
  ### Recent Training Steps (Loss Only)
56
  | Step | Training Loss | Timestamp |
57
  |------|---------------|-----------|
58
+ | 291 | 4.513980 | 2025-08-14T20:59 |
59
+ | 292 | 4.423133 | 2025-08-14T20:59 |
60
+ | 293 | 4.640058 | 2025-08-14T20:59 |
61
+ | 294 | 4.488780 | 2025-08-14T20:59 |
62
+ | 295 | 4.027132 | 2025-08-14T20:59 |
63
+ | 296 | 4.805581 | 2025-08-14T20:59 |
64
+ | 297 | 4.652530 | 2025-08-14T21:00 |
65
+ | 298 | 4.494508 | 2025-08-14T21:00 |
66
+ | 299 | 4.449580 | 2025-08-14T21:00 |
67
+ | 300 | 4.575191 | 2025-08-14T21:00 |
68
 
69
  ## Training Visualizations
70
 
 
88
  from transformers import AutoModelForCausalLM, AutoTokenizer
89
  # For vision-language models, use appropriate imports
90
 
91
+ model = AutoModelForCausalLM.from_pretrained("./model_step_300")
92
+ tokenizer = AutoTokenizer.from_pretrained("./model_step_300")
93
 
94
  # Your inference code here
95
  ```
 
129
 
130
  ---
131
 
132
+ *This model card was automatically generated by brute-force-training on 2025-08-14 21:03:04*
evaluation_comparison.png CHANGED

Git LFS Details

  • SHA256: bbcddeeddaaac7719a489972dec2a9a3e0b40c3dcf425ed1b4dd42eebb2094f1
  • Pointer size: 131 Bytes
  • Size of remote file: 294 kB

Git LFS Details

  • SHA256: ee921359acbb367ea0f72a32b1f4bde433e6e38fd1828773286eb1b540acb5f4
  • Pointer size: 131 Bytes
  • Size of remote file: 321 kB
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:73bd67087ca08096ebdeca7389f53fe631c41e9ed1f1a5e286109ea97e089fea
3
  size 901692416
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:94846411f2c1171e1ec1a3aff354d161896539bd14085e6bd518449a8c04eaf4
3
  size 901692416
model_card_metadata.json CHANGED
@@ -1,8 +1,8 @@
1
  {
2
  "base_model": "LiquidAI/LFM2-VL-450M",
3
  "training_framework": "brute-force-training",
4
- "training_date": "2025-08-14T20:55:09.352395",
5
- "training_steps": 100,
6
  "dataset": "wjbmattingly/medieval-letters-htr-synthetic",
7
  "training_config": {
8
  "dataset_name": "wjbmattingly/medieval-letters-htr-synthetic",
@@ -24,8 +24,8 @@
24
  "user_text": "Transcribe this medieval manuscript line",
25
  "max_image_size": 200
26
  },
27
- "final_training_loss": 4.847932815551758,
28
- "final_evaluation_loss": 4.8217997550964355,
29
- "final_char_accuracy": 0.10633391201404206,
30
- "final_word_accuracy": 0.06352046181580152
31
  }
 
1
  {
2
  "base_model": "LiquidAI/LFM2-VL-450M",
3
  "training_framework": "brute-force-training",
4
+ "training_date": "2025-08-14T21:03:04.407697",
5
+ "training_steps": 300,
6
  "dataset": "wjbmattingly/medieval-letters-htr-synthetic",
7
  "training_config": {
8
  "dataset_name": "wjbmattingly/medieval-letters-htr-synthetic",
 
24
  "user_text": "Transcribe this medieval manuscript line",
25
  "max_image_size": 200
26
  },
27
+ "final_training_loss": 4.575191020965576,
28
+ "final_evaluation_loss": 4.492388168970744,
29
+ "final_char_accuracy": 0.1061965003999563,
30
+ "final_word_accuracy": 0.06438001131166525
31
  }
training_curves.png CHANGED

Git LFS Details

  • SHA256: a1d36844f9c39c3d06c4c5ffe60533d926168a83872039a99cb845552f2ade05
  • Pointer size: 131 Bytes
  • Size of remote file: 533 kB

Git LFS Details

  • SHA256: 93448b67f97f101190f7e32ed6381fbbb9028e12d05ae0db3cbf70b3921b379f
  • Pointer size: 131 Bytes
  • Size of remote file: 587 kB
training_metrics.json CHANGED
The diff for this file is too large to render. See raw diff