Upload folder using huggingface_hub
Browse files- README.md +20 -18
- evaluation_comparison.png +2 -2
- model.safetensors +1 -1
- model_card_metadata.json +6 -6
- training_curves.png +2 -2
- training_metrics.json +0 -0
README.md
CHANGED
|
@@ -3,7 +3,7 @@ base_model:
|
|
| 3 |
- LiquidAI/LFM2-VL-450M
|
| 4 |
---
|
| 5 |
|
| 6 |
-
#
|
| 7 |
|
| 8 |
## Model Description
|
| 9 |
|
|
@@ -11,8 +11,8 @@ This model is a fine-tuned version of **LiquidAI/LFM2-VL-450M** using the brute-
|
|
| 11 |
|
| 12 |
- **Base Model**: LiquidAI/LFM2-VL-450M
|
| 13 |
- **Training Status**: 🔄 In Progress
|
| 14 |
-
- **Generated**: 2025-08-14
|
| 15 |
-
- **Training Steps**:
|
| 16 |
|
| 17 |
## Training Details
|
| 18 |
|
|
@@ -29,8 +29,8 @@ This model is a fine-tuned version of **LiquidAI/LFM2-VL-450M** using the brute-
|
|
| 29 |
- **Evaluation Frequency**: Every 100 steps
|
| 30 |
|
| 31 |
### Current Performance
|
| 32 |
-
- **Training Loss**: 4.
|
| 33 |
-
- **Evaluation Loss**: 4.
|
| 34 |
|
| 35 |
## Pre-Training Evaluation
|
| 36 |
|
|
@@ -47,22 +47,24 @@ This model is a fine-tuned version of **LiquidAI/LFM2-VL-450M** using the brute-
|
|
| 47 |
|------|----------------|------|------------|----------|----------|--------------------|
|
| 48 |
| Pre | pre_training | 5.4191 | 225.67 | 11.4% | 6.9% | +0.0% |
|
| 49 |
| 100 | checkpoint | 4.8218 | 124.19 | 10.6% | 6.4% | +11.0% |
|
|
|
|
|
|
|
| 50 |
|
| 51 |
## Training Progress
|
| 52 |
|
| 53 |
### Recent Training Steps (Loss Only)
|
| 54 |
| Step | Training Loss | Timestamp |
|
| 55 |
|------|---------------|-----------|
|
| 56 |
-
|
|
| 57 |
-
|
|
| 58 |
-
|
|
| 59 |
-
|
|
| 60 |
-
|
|
| 61 |
-
|
|
| 62 |
-
|
|
| 63 |
-
|
|
| 64 |
-
|
|
| 65 |
-
|
|
| 66 |
|
| 67 |
## Training Visualizations
|
| 68 |
|
|
@@ -86,8 +88,8 @@ This model is a fine-tuned version of **LiquidAI/LFM2-VL-450M** using the brute-
|
|
| 86 |
from transformers import AutoModelForCausalLM, AutoTokenizer
|
| 87 |
# For vision-language models, use appropriate imports
|
| 88 |
|
| 89 |
-
model = AutoModelForCausalLM.from_pretrained("./
|
| 90 |
-
tokenizer = AutoTokenizer.from_pretrained("./
|
| 91 |
|
| 92 |
# Your inference code here
|
| 93 |
```
|
|
@@ -127,4 +129,4 @@ tokenizer = AutoTokenizer.from_pretrained("./model_step_100")
|
|
| 127 |
|
| 128 |
---
|
| 129 |
|
| 130 |
-
*This model card was automatically generated by brute-force-training on 2025-08-14
|
|
|
|
| 3 |
- LiquidAI/LFM2-VL-450M
|
| 4 |
---
|
| 5 |
|
| 6 |
+
# model_step_300
|
| 7 |
|
| 8 |
## Model Description
|
| 9 |
|
|
|
|
| 11 |
|
| 12 |
- **Base Model**: LiquidAI/LFM2-VL-450M
|
| 13 |
- **Training Status**: 🔄 In Progress
|
| 14 |
+
- **Generated**: 2025-08-14 21:03:04
|
| 15 |
+
- **Training Steps**: 300
|
| 16 |
|
| 17 |
## Training Details
|
| 18 |
|
|
|
|
| 29 |
- **Evaluation Frequency**: Every 100 steps
|
| 30 |
|
| 31 |
### Current Performance
|
| 32 |
+
- **Training Loss**: 4.575191
|
| 33 |
+
- **Evaluation Loss**: 4.492388
|
| 34 |
|
| 35 |
## Pre-Training Evaluation
|
| 36 |
|
|
|
|
| 47 |
|------|----------------|------|------------|----------|----------|--------------------|
|
| 48 |
| Pre | pre_training | 5.4191 | 225.67 | 11.4% | 6.9% | +0.0% |
|
| 49 |
| 100 | checkpoint | 4.8218 | 124.19 | 10.6% | 6.4% | +11.0% |
|
| 50 |
+
| 200 | checkpoint | 4.6127 | 100.76 | 10.8% | 6.4% | +14.9% |
|
| 51 |
+
| 300 | checkpoint | 4.4924 | 89.33 | 10.6% | 6.4% | +17.1% |
|
| 52 |
|
| 53 |
## Training Progress
|
| 54 |
|
| 55 |
### Recent Training Steps (Loss Only)
|
| 56 |
| Step | Training Loss | Timestamp |
|
| 57 |
|------|---------------|-----------|
|
| 58 |
+
| 291 | 4.513980 | 2025-08-14T20:59 |
|
| 59 |
+
| 292 | 4.423133 | 2025-08-14T20:59 |
|
| 60 |
+
| 293 | 4.640058 | 2025-08-14T20:59 |
|
| 61 |
+
| 294 | 4.488780 | 2025-08-14T20:59 |
|
| 62 |
+
| 295 | 4.027132 | 2025-08-14T20:59 |
|
| 63 |
+
| 296 | 4.805581 | 2025-08-14T20:59 |
|
| 64 |
+
| 297 | 4.652530 | 2025-08-14T21:00 |
|
| 65 |
+
| 298 | 4.494508 | 2025-08-14T21:00 |
|
| 66 |
+
| 299 | 4.449580 | 2025-08-14T21:00 |
|
| 67 |
+
| 300 | 4.575191 | 2025-08-14T21:00 |
|
| 68 |
|
| 69 |
## Training Visualizations
|
| 70 |
|
|
|
|
| 88 |
from transformers import AutoModelForCausalLM, AutoTokenizer
|
| 89 |
# For vision-language models, use appropriate imports
|
| 90 |
|
| 91 |
+
model = AutoModelForCausalLM.from_pretrained("./model_step_300")
|
| 92 |
+
tokenizer = AutoTokenizer.from_pretrained("./model_step_300")
|
| 93 |
|
| 94 |
# Your inference code here
|
| 95 |
```
|
|
|
|
| 129 |
|
| 130 |
---
|
| 131 |
|
| 132 |
+
*This model card was automatically generated by brute-force-training on 2025-08-14 21:03:04*
|
evaluation_comparison.png
CHANGED
|
Git LFS Details
|
|
Git LFS Details
|
model.safetensors
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 901692416
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:94846411f2c1171e1ec1a3aff354d161896539bd14085e6bd518449a8c04eaf4
|
| 3 |
size 901692416
|
model_card_metadata.json
CHANGED
|
@@ -1,8 +1,8 @@
|
|
| 1 |
{
|
| 2 |
"base_model": "LiquidAI/LFM2-VL-450M",
|
| 3 |
"training_framework": "brute-force-training",
|
| 4 |
-
"training_date": "2025-08-
|
| 5 |
-
"training_steps":
|
| 6 |
"dataset": "wjbmattingly/medieval-letters-htr-synthetic",
|
| 7 |
"training_config": {
|
| 8 |
"dataset_name": "wjbmattingly/medieval-letters-htr-synthetic",
|
|
@@ -24,8 +24,8 @@
|
|
| 24 |
"user_text": "Transcribe this medieval manuscript line",
|
| 25 |
"max_image_size": 200
|
| 26 |
},
|
| 27 |
-
"final_training_loss": 4.
|
| 28 |
-
"final_evaluation_loss": 4.
|
| 29 |
-
"final_char_accuracy": 0.
|
| 30 |
-
"final_word_accuracy": 0.
|
| 31 |
}
|
|
|
|
| 1 |
{
|
| 2 |
"base_model": "LiquidAI/LFM2-VL-450M",
|
| 3 |
"training_framework": "brute-force-training",
|
| 4 |
+
"training_date": "2025-08-14T21:03:04.407697",
|
| 5 |
+
"training_steps": 300,
|
| 6 |
"dataset": "wjbmattingly/medieval-letters-htr-synthetic",
|
| 7 |
"training_config": {
|
| 8 |
"dataset_name": "wjbmattingly/medieval-letters-htr-synthetic",
|
|
|
|
| 24 |
"user_text": "Transcribe this medieval manuscript line",
|
| 25 |
"max_image_size": 200
|
| 26 |
},
|
| 27 |
+
"final_training_loss": 4.575191020965576,
|
| 28 |
+
"final_evaluation_loss": 4.492388168970744,
|
| 29 |
+
"final_char_accuracy": 0.1061965003999563,
|
| 30 |
+
"final_word_accuracy": 0.06438001131166525
|
| 31 |
}
|
training_curves.png
CHANGED
|
Git LFS Details
|
|
Git LFS Details
|
training_metrics.json
CHANGED
|
The diff for this file is too large to render.
See raw diff
|
|
|