jise commited on
Commit
4bed7fd
·
verified ·
1 Parent(s): 45b0681

Upload Qwen2_5_VLForConditionalGeneration

Browse files
config.json CHANGED
@@ -44,7 +44,7 @@
44
  "rope_theta": 1000000.0,
45
  "sliding_window": 32768,
46
  "tie_word_embeddings": false,
47
- "torch_dtype": "bfloat16",
48
  "transformers_version": "4.51.3",
49
  "use_cache": true,
50
  "use_sliding_window": false,
 
44
  "rope_theta": 1000000.0,
45
  "sliding_window": 32768,
46
  "tie_word_embeddings": false,
47
+ "torch_dtype": "float32",
48
  "transformers_version": "4.51.3",
49
  "use_cache": true,
50
  "use_sliding_window": false,
model-00001-of-00002.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:f8ef7bec75e337ef6eab96b84e740ba2f261d75eb8ea89e08b5845531232d7c9
3
- size 4809612533
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8027ed4558648a8104ebd39e32d3a5453de8b3e2b2318311ae9f9fcb5588ffb3
3
+ size 4992162636
model-00002-of-00002.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:0c859795ad3a627a9b95bcb762e059d5b768a4a36fdd4affeff269d93fdecc67
3
- size 1089994880
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a36f454c4f66110e46750ee133e71b32db2a4f226ace3802f55ff66483092915
3
+ size 3092067506
model.safetensors.index.json CHANGED
@@ -1,6 +1,6 @@
1
  {
2
  "metadata": {
3
- "total_size": 5899313093
4
  },
5
  "weight_map": {
6
  "lm_head.weight": "model-00002-of-00002.safetensors",
@@ -616,26 +616,26 @@
616
  "model.layers.2.self_attn.v_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
617
  "model.layers.2.self_attn.v_proj.weight.quant_map": "model-00001-of-00002.safetensors",
618
  "model.layers.2.self_attn.v_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
619
- "model.layers.20.input_layernorm.weight": "model-00001-of-00002.safetensors",
620
- "model.layers.20.mlp.down_proj.weight": "model-00001-of-00002.safetensors",
621
- "model.layers.20.mlp.down_proj.weight.absmax": "model-00001-of-00002.safetensors",
622
- "model.layers.20.mlp.down_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
623
- "model.layers.20.mlp.down_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
624
- "model.layers.20.mlp.down_proj.weight.quant_map": "model-00001-of-00002.safetensors",
625
- "model.layers.20.mlp.down_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
626
  "model.layers.20.mlp.gate_proj.weight": "model-00001-of-00002.safetensors",
627
  "model.layers.20.mlp.gate_proj.weight.absmax": "model-00001-of-00002.safetensors",
628
  "model.layers.20.mlp.gate_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
629
  "model.layers.20.mlp.gate_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
630
  "model.layers.20.mlp.gate_proj.weight.quant_map": "model-00001-of-00002.safetensors",
631
  "model.layers.20.mlp.gate_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
632
- "model.layers.20.mlp.up_proj.weight": "model-00001-of-00002.safetensors",
633
- "model.layers.20.mlp.up_proj.weight.absmax": "model-00001-of-00002.safetensors",
634
- "model.layers.20.mlp.up_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
635
- "model.layers.20.mlp.up_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
636
- "model.layers.20.mlp.up_proj.weight.quant_map": "model-00001-of-00002.safetensors",
637
- "model.layers.20.mlp.up_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
638
- "model.layers.20.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
639
  "model.layers.20.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
640
  "model.layers.20.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
641
  "model.layers.20.self_attn.k_proj.weight.absmax": "model-00001-of-00002.safetensors",
@@ -663,335 +663,335 @@
663
  "model.layers.20.self_attn.v_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
664
  "model.layers.20.self_attn.v_proj.weight.quant_map": "model-00001-of-00002.safetensors",
665
  "model.layers.20.self_attn.v_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
666
- "model.layers.21.input_layernorm.weight": "model-00001-of-00002.safetensors",
667
- "model.layers.21.mlp.down_proj.weight": "model-00001-of-00002.safetensors",
668
- "model.layers.21.mlp.down_proj.weight.absmax": "model-00001-of-00002.safetensors",
669
- "model.layers.21.mlp.down_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
670
- "model.layers.21.mlp.down_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
671
- "model.layers.21.mlp.down_proj.weight.quant_map": "model-00001-of-00002.safetensors",
672
- "model.layers.21.mlp.down_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
673
- "model.layers.21.mlp.gate_proj.weight": "model-00001-of-00002.safetensors",
674
- "model.layers.21.mlp.gate_proj.weight.absmax": "model-00001-of-00002.safetensors",
675
- "model.layers.21.mlp.gate_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
676
- "model.layers.21.mlp.gate_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
677
- "model.layers.21.mlp.gate_proj.weight.quant_map": "model-00001-of-00002.safetensors",
678
- "model.layers.21.mlp.gate_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
679
- "model.layers.21.mlp.up_proj.weight": "model-00001-of-00002.safetensors",
680
- "model.layers.21.mlp.up_proj.weight.absmax": "model-00001-of-00002.safetensors",
681
- "model.layers.21.mlp.up_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
682
- "model.layers.21.mlp.up_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
683
- "model.layers.21.mlp.up_proj.weight.quant_map": "model-00001-of-00002.safetensors",
684
- "model.layers.21.mlp.up_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
685
- "model.layers.21.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
686
- "model.layers.21.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
687
- "model.layers.21.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
688
- "model.layers.21.self_attn.k_proj.weight.absmax": "model-00001-of-00002.safetensors",
689
- "model.layers.21.self_attn.k_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
690
- "model.layers.21.self_attn.k_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
691
- "model.layers.21.self_attn.k_proj.weight.quant_map": "model-00001-of-00002.safetensors",
692
- "model.layers.21.self_attn.k_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
693
- "model.layers.21.self_attn.o_proj.weight": "model-00001-of-00002.safetensors",
694
- "model.layers.21.self_attn.o_proj.weight.absmax": "model-00001-of-00002.safetensors",
695
- "model.layers.21.self_attn.o_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
696
- "model.layers.21.self_attn.o_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
697
- "model.layers.21.self_attn.o_proj.weight.quant_map": "model-00001-of-00002.safetensors",
698
- "model.layers.21.self_attn.o_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
699
- "model.layers.21.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
700
- "model.layers.21.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
701
- "model.layers.21.self_attn.q_proj.weight.absmax": "model-00001-of-00002.safetensors",
702
- "model.layers.21.self_attn.q_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
703
- "model.layers.21.self_attn.q_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
704
- "model.layers.21.self_attn.q_proj.weight.quant_map": "model-00001-of-00002.safetensors",
705
- "model.layers.21.self_attn.q_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
706
- "model.layers.21.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
707
- "model.layers.21.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
708
- "model.layers.21.self_attn.v_proj.weight.absmax": "model-00001-of-00002.safetensors",
709
- "model.layers.21.self_attn.v_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
710
- "model.layers.21.self_attn.v_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
711
- "model.layers.21.self_attn.v_proj.weight.quant_map": "model-00001-of-00002.safetensors",
712
- "model.layers.21.self_attn.v_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
713
- "model.layers.22.input_layernorm.weight": "model-00001-of-00002.safetensors",
714
- "model.layers.22.mlp.down_proj.weight": "model-00001-of-00002.safetensors",
715
- "model.layers.22.mlp.down_proj.weight.absmax": "model-00001-of-00002.safetensors",
716
- "model.layers.22.mlp.down_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
717
- "model.layers.22.mlp.down_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
718
- "model.layers.22.mlp.down_proj.weight.quant_map": "model-00001-of-00002.safetensors",
719
- "model.layers.22.mlp.down_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
720
- "model.layers.22.mlp.gate_proj.weight": "model-00001-of-00002.safetensors",
721
- "model.layers.22.mlp.gate_proj.weight.absmax": "model-00001-of-00002.safetensors",
722
- "model.layers.22.mlp.gate_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
723
- "model.layers.22.mlp.gate_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
724
- "model.layers.22.mlp.gate_proj.weight.quant_map": "model-00001-of-00002.safetensors",
725
- "model.layers.22.mlp.gate_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
726
- "model.layers.22.mlp.up_proj.weight": "model-00001-of-00002.safetensors",
727
- "model.layers.22.mlp.up_proj.weight.absmax": "model-00001-of-00002.safetensors",
728
- "model.layers.22.mlp.up_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
729
- "model.layers.22.mlp.up_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
730
- "model.layers.22.mlp.up_proj.weight.quant_map": "model-00001-of-00002.safetensors",
731
- "model.layers.22.mlp.up_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
732
- "model.layers.22.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
733
- "model.layers.22.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
734
- "model.layers.22.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
735
- "model.layers.22.self_attn.k_proj.weight.absmax": "model-00001-of-00002.safetensors",
736
- "model.layers.22.self_attn.k_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
737
- "model.layers.22.self_attn.k_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
738
- "model.layers.22.self_attn.k_proj.weight.quant_map": "model-00001-of-00002.safetensors",
739
- "model.layers.22.self_attn.k_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
740
- "model.layers.22.self_attn.o_proj.weight": "model-00001-of-00002.safetensors",
741
- "model.layers.22.self_attn.o_proj.weight.absmax": "model-00001-of-00002.safetensors",
742
- "model.layers.22.self_attn.o_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
743
- "model.layers.22.self_attn.o_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
744
- "model.layers.22.self_attn.o_proj.weight.quant_map": "model-00001-of-00002.safetensors",
745
- "model.layers.22.self_attn.o_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
746
- "model.layers.22.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
747
- "model.layers.22.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
748
- "model.layers.22.self_attn.q_proj.weight.absmax": "model-00001-of-00002.safetensors",
749
- "model.layers.22.self_attn.q_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
750
- "model.layers.22.self_attn.q_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
751
- "model.layers.22.self_attn.q_proj.weight.quant_map": "model-00001-of-00002.safetensors",
752
- "model.layers.22.self_attn.q_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
753
- "model.layers.22.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
754
- "model.layers.22.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
755
- "model.layers.22.self_attn.v_proj.weight.absmax": "model-00001-of-00002.safetensors",
756
- "model.layers.22.self_attn.v_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
757
- "model.layers.22.self_attn.v_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
758
- "model.layers.22.self_attn.v_proj.weight.quant_map": "model-00001-of-00002.safetensors",
759
- "model.layers.22.self_attn.v_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
760
- "model.layers.23.input_layernorm.weight": "model-00001-of-00002.safetensors",
761
- "model.layers.23.mlp.down_proj.weight": "model-00001-of-00002.safetensors",
762
- "model.layers.23.mlp.down_proj.weight.absmax": "model-00001-of-00002.safetensors",
763
- "model.layers.23.mlp.down_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
764
- "model.layers.23.mlp.down_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
765
- "model.layers.23.mlp.down_proj.weight.quant_map": "model-00001-of-00002.safetensors",
766
- "model.layers.23.mlp.down_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
767
- "model.layers.23.mlp.gate_proj.weight": "model-00001-of-00002.safetensors",
768
- "model.layers.23.mlp.gate_proj.weight.absmax": "model-00001-of-00002.safetensors",
769
- "model.layers.23.mlp.gate_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
770
- "model.layers.23.mlp.gate_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
771
- "model.layers.23.mlp.gate_proj.weight.quant_map": "model-00001-of-00002.safetensors",
772
- "model.layers.23.mlp.gate_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
773
- "model.layers.23.mlp.up_proj.weight": "model-00001-of-00002.safetensors",
774
- "model.layers.23.mlp.up_proj.weight.absmax": "model-00001-of-00002.safetensors",
775
- "model.layers.23.mlp.up_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
776
- "model.layers.23.mlp.up_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
777
- "model.layers.23.mlp.up_proj.weight.quant_map": "model-00001-of-00002.safetensors",
778
- "model.layers.23.mlp.up_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
779
- "model.layers.23.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
780
- "model.layers.23.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
781
- "model.layers.23.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
782
- "model.layers.23.self_attn.k_proj.weight.absmax": "model-00001-of-00002.safetensors",
783
- "model.layers.23.self_attn.k_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
784
- "model.layers.23.self_attn.k_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
785
- "model.layers.23.self_attn.k_proj.weight.quant_map": "model-00001-of-00002.safetensors",
786
- "model.layers.23.self_attn.k_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
787
- "model.layers.23.self_attn.o_proj.weight": "model-00001-of-00002.safetensors",
788
- "model.layers.23.self_attn.o_proj.weight.absmax": "model-00001-of-00002.safetensors",
789
- "model.layers.23.self_attn.o_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
790
- "model.layers.23.self_attn.o_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
791
- "model.layers.23.self_attn.o_proj.weight.quant_map": "model-00001-of-00002.safetensors",
792
- "model.layers.23.self_attn.o_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
793
- "model.layers.23.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
794
- "model.layers.23.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
795
- "model.layers.23.self_attn.q_proj.weight.absmax": "model-00001-of-00002.safetensors",
796
- "model.layers.23.self_attn.q_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
797
- "model.layers.23.self_attn.q_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
798
- "model.layers.23.self_attn.q_proj.weight.quant_map": "model-00001-of-00002.safetensors",
799
- "model.layers.23.self_attn.q_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
800
- "model.layers.23.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
801
- "model.layers.23.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
802
- "model.layers.23.self_attn.v_proj.weight.absmax": "model-00001-of-00002.safetensors",
803
- "model.layers.23.self_attn.v_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
804
- "model.layers.23.self_attn.v_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
805
- "model.layers.23.self_attn.v_proj.weight.quant_map": "model-00001-of-00002.safetensors",
806
- "model.layers.23.self_attn.v_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
807
- "model.layers.24.input_layernorm.weight": "model-00001-of-00002.safetensors",
808
- "model.layers.24.mlp.down_proj.weight": "model-00001-of-00002.safetensors",
809
- "model.layers.24.mlp.down_proj.weight.absmax": "model-00001-of-00002.safetensors",
810
- "model.layers.24.mlp.down_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
811
- "model.layers.24.mlp.down_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
812
- "model.layers.24.mlp.down_proj.weight.quant_map": "model-00001-of-00002.safetensors",
813
- "model.layers.24.mlp.down_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
814
- "model.layers.24.mlp.gate_proj.weight": "model-00001-of-00002.safetensors",
815
- "model.layers.24.mlp.gate_proj.weight.absmax": "model-00001-of-00002.safetensors",
816
- "model.layers.24.mlp.gate_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
817
- "model.layers.24.mlp.gate_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
818
- "model.layers.24.mlp.gate_proj.weight.quant_map": "model-00001-of-00002.safetensors",
819
- "model.layers.24.mlp.gate_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
820
- "model.layers.24.mlp.up_proj.weight": "model-00001-of-00002.safetensors",
821
- "model.layers.24.mlp.up_proj.weight.absmax": "model-00001-of-00002.safetensors",
822
- "model.layers.24.mlp.up_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
823
- "model.layers.24.mlp.up_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
824
- "model.layers.24.mlp.up_proj.weight.quant_map": "model-00001-of-00002.safetensors",
825
- "model.layers.24.mlp.up_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
826
- "model.layers.24.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
827
- "model.layers.24.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
828
- "model.layers.24.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
829
- "model.layers.24.self_attn.k_proj.weight.absmax": "model-00001-of-00002.safetensors",
830
- "model.layers.24.self_attn.k_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
831
- "model.layers.24.self_attn.k_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
832
- "model.layers.24.self_attn.k_proj.weight.quant_map": "model-00001-of-00002.safetensors",
833
- "model.layers.24.self_attn.k_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
834
- "model.layers.24.self_attn.o_proj.weight": "model-00001-of-00002.safetensors",
835
- "model.layers.24.self_attn.o_proj.weight.absmax": "model-00001-of-00002.safetensors",
836
- "model.layers.24.self_attn.o_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
837
- "model.layers.24.self_attn.o_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
838
- "model.layers.24.self_attn.o_proj.weight.quant_map": "model-00001-of-00002.safetensors",
839
- "model.layers.24.self_attn.o_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
840
- "model.layers.24.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
841
- "model.layers.24.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
842
- "model.layers.24.self_attn.q_proj.weight.absmax": "model-00001-of-00002.safetensors",
843
- "model.layers.24.self_attn.q_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
844
- "model.layers.24.self_attn.q_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
845
- "model.layers.24.self_attn.q_proj.weight.quant_map": "model-00001-of-00002.safetensors",
846
- "model.layers.24.self_attn.q_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
847
- "model.layers.24.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
848
- "model.layers.24.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
849
- "model.layers.24.self_attn.v_proj.weight.absmax": "model-00001-of-00002.safetensors",
850
- "model.layers.24.self_attn.v_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
851
- "model.layers.24.self_attn.v_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
852
- "model.layers.24.self_attn.v_proj.weight.quant_map": "model-00001-of-00002.safetensors",
853
- "model.layers.24.self_attn.v_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
854
- "model.layers.25.input_layernorm.weight": "model-00001-of-00002.safetensors",
855
- "model.layers.25.mlp.down_proj.weight": "model-00001-of-00002.safetensors",
856
- "model.layers.25.mlp.down_proj.weight.absmax": "model-00001-of-00002.safetensors",
857
- "model.layers.25.mlp.down_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
858
- "model.layers.25.mlp.down_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
859
- "model.layers.25.mlp.down_proj.weight.quant_map": "model-00001-of-00002.safetensors",
860
- "model.layers.25.mlp.down_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
861
- "model.layers.25.mlp.gate_proj.weight": "model-00001-of-00002.safetensors",
862
- "model.layers.25.mlp.gate_proj.weight.absmax": "model-00001-of-00002.safetensors",
863
- "model.layers.25.mlp.gate_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
864
- "model.layers.25.mlp.gate_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
865
- "model.layers.25.mlp.gate_proj.weight.quant_map": "model-00001-of-00002.safetensors",
866
- "model.layers.25.mlp.gate_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
867
- "model.layers.25.mlp.up_proj.weight": "model-00001-of-00002.safetensors",
868
- "model.layers.25.mlp.up_proj.weight.absmax": "model-00001-of-00002.safetensors",
869
- "model.layers.25.mlp.up_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
870
- "model.layers.25.mlp.up_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
871
- "model.layers.25.mlp.up_proj.weight.quant_map": "model-00001-of-00002.safetensors",
872
- "model.layers.25.mlp.up_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
873
- "model.layers.25.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
874
- "model.layers.25.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
875
- "model.layers.25.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
876
- "model.layers.25.self_attn.k_proj.weight.absmax": "model-00001-of-00002.safetensors",
877
- "model.layers.25.self_attn.k_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
878
- "model.layers.25.self_attn.k_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
879
- "model.layers.25.self_attn.k_proj.weight.quant_map": "model-00001-of-00002.safetensors",
880
- "model.layers.25.self_attn.k_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
881
- "model.layers.25.self_attn.o_proj.weight": "model-00001-of-00002.safetensors",
882
- "model.layers.25.self_attn.o_proj.weight.absmax": "model-00001-of-00002.safetensors",
883
- "model.layers.25.self_attn.o_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
884
- "model.layers.25.self_attn.o_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
885
- "model.layers.25.self_attn.o_proj.weight.quant_map": "model-00001-of-00002.safetensors",
886
- "model.layers.25.self_attn.o_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
887
- "model.layers.25.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
888
- "model.layers.25.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
889
- "model.layers.25.self_attn.q_proj.weight.absmax": "model-00001-of-00002.safetensors",
890
- "model.layers.25.self_attn.q_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
891
- "model.layers.25.self_attn.q_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
892
- "model.layers.25.self_attn.q_proj.weight.quant_map": "model-00001-of-00002.safetensors",
893
- "model.layers.25.self_attn.q_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
894
- "model.layers.25.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
895
- "model.layers.25.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
896
- "model.layers.25.self_attn.v_proj.weight.absmax": "model-00001-of-00002.safetensors",
897
- "model.layers.25.self_attn.v_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
898
- "model.layers.25.self_attn.v_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
899
- "model.layers.25.self_attn.v_proj.weight.quant_map": "model-00001-of-00002.safetensors",
900
- "model.layers.25.self_attn.v_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
901
- "model.layers.26.input_layernorm.weight": "model-00001-of-00002.safetensors",
902
- "model.layers.26.mlp.down_proj.weight": "model-00001-of-00002.safetensors",
903
- "model.layers.26.mlp.down_proj.weight.absmax": "model-00001-of-00002.safetensors",
904
- "model.layers.26.mlp.down_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
905
- "model.layers.26.mlp.down_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
906
- "model.layers.26.mlp.down_proj.weight.quant_map": "model-00001-of-00002.safetensors",
907
- "model.layers.26.mlp.down_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
908
- "model.layers.26.mlp.gate_proj.weight": "model-00001-of-00002.safetensors",
909
- "model.layers.26.mlp.gate_proj.weight.absmax": "model-00001-of-00002.safetensors",
910
- "model.layers.26.mlp.gate_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
911
- "model.layers.26.mlp.gate_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
912
- "model.layers.26.mlp.gate_proj.weight.quant_map": "model-00001-of-00002.safetensors",
913
- "model.layers.26.mlp.gate_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
914
- "model.layers.26.mlp.up_proj.weight": "model-00001-of-00002.safetensors",
915
- "model.layers.26.mlp.up_proj.weight.absmax": "model-00001-of-00002.safetensors",
916
- "model.layers.26.mlp.up_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
917
- "model.layers.26.mlp.up_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
918
- "model.layers.26.mlp.up_proj.weight.quant_map": "model-00001-of-00002.safetensors",
919
- "model.layers.26.mlp.up_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
920
- "model.layers.26.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
921
- "model.layers.26.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
922
- "model.layers.26.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
923
- "model.layers.26.self_attn.k_proj.weight.absmax": "model-00001-of-00002.safetensors",
924
- "model.layers.26.self_attn.k_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
925
- "model.layers.26.self_attn.k_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
926
- "model.layers.26.self_attn.k_proj.weight.quant_map": "model-00001-of-00002.safetensors",
927
- "model.layers.26.self_attn.k_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
928
- "model.layers.26.self_attn.o_proj.weight": "model-00001-of-00002.safetensors",
929
- "model.layers.26.self_attn.o_proj.weight.absmax": "model-00001-of-00002.safetensors",
930
- "model.layers.26.self_attn.o_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
931
- "model.layers.26.self_attn.o_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
932
- "model.layers.26.self_attn.o_proj.weight.quant_map": "model-00001-of-00002.safetensors",
933
- "model.layers.26.self_attn.o_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
934
- "model.layers.26.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
935
- "model.layers.26.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
936
- "model.layers.26.self_attn.q_proj.weight.absmax": "model-00001-of-00002.safetensors",
937
- "model.layers.26.self_attn.q_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
938
- "model.layers.26.self_attn.q_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
939
- "model.layers.26.self_attn.q_proj.weight.quant_map": "model-00001-of-00002.safetensors",
940
- "model.layers.26.self_attn.q_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
941
- "model.layers.26.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
942
- "model.layers.26.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
943
- "model.layers.26.self_attn.v_proj.weight.absmax": "model-00001-of-00002.safetensors",
944
- "model.layers.26.self_attn.v_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
945
- "model.layers.26.self_attn.v_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
946
- "model.layers.26.self_attn.v_proj.weight.quant_map": "model-00001-of-00002.safetensors",
947
- "model.layers.26.self_attn.v_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
948
- "model.layers.27.input_layernorm.weight": "model-00001-of-00002.safetensors",
949
- "model.layers.27.mlp.down_proj.weight": "model-00001-of-00002.safetensors",
950
- "model.layers.27.mlp.down_proj.weight.absmax": "model-00001-of-00002.safetensors",
951
- "model.layers.27.mlp.down_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
952
- "model.layers.27.mlp.down_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
953
- "model.layers.27.mlp.down_proj.weight.quant_map": "model-00001-of-00002.safetensors",
954
- "model.layers.27.mlp.down_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
955
- "model.layers.27.mlp.gate_proj.weight": "model-00001-of-00002.safetensors",
956
- "model.layers.27.mlp.gate_proj.weight.absmax": "model-00001-of-00002.safetensors",
957
- "model.layers.27.mlp.gate_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
958
- "model.layers.27.mlp.gate_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
959
- "model.layers.27.mlp.gate_proj.weight.quant_map": "model-00001-of-00002.safetensors",
960
- "model.layers.27.mlp.gate_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
961
- "model.layers.27.mlp.up_proj.weight": "model-00001-of-00002.safetensors",
962
- "model.layers.27.mlp.up_proj.weight.absmax": "model-00001-of-00002.safetensors",
963
- "model.layers.27.mlp.up_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
964
- "model.layers.27.mlp.up_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
965
- "model.layers.27.mlp.up_proj.weight.quant_map": "model-00001-of-00002.safetensors",
966
- "model.layers.27.mlp.up_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
967
- "model.layers.27.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
968
- "model.layers.27.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
969
- "model.layers.27.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
970
- "model.layers.27.self_attn.k_proj.weight.absmax": "model-00001-of-00002.safetensors",
971
- "model.layers.27.self_attn.k_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
972
- "model.layers.27.self_attn.k_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
973
- "model.layers.27.self_attn.k_proj.weight.quant_map": "model-00001-of-00002.safetensors",
974
- "model.layers.27.self_attn.k_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
975
- "model.layers.27.self_attn.o_proj.weight": "model-00001-of-00002.safetensors",
976
- "model.layers.27.self_attn.o_proj.weight.absmax": "model-00001-of-00002.safetensors",
977
- "model.layers.27.self_attn.o_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
978
- "model.layers.27.self_attn.o_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
979
- "model.layers.27.self_attn.o_proj.weight.quant_map": "model-00001-of-00002.safetensors",
980
- "model.layers.27.self_attn.o_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
981
- "model.layers.27.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
982
- "model.layers.27.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
983
- "model.layers.27.self_attn.q_proj.weight.absmax": "model-00001-of-00002.safetensors",
984
- "model.layers.27.self_attn.q_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
985
- "model.layers.27.self_attn.q_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
986
- "model.layers.27.self_attn.q_proj.weight.quant_map": "model-00001-of-00002.safetensors",
987
- "model.layers.27.self_attn.q_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
988
- "model.layers.27.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
989
- "model.layers.27.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
990
- "model.layers.27.self_attn.v_proj.weight.absmax": "model-00001-of-00002.safetensors",
991
- "model.layers.27.self_attn.v_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
992
- "model.layers.27.self_attn.v_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
993
- "model.layers.27.self_attn.v_proj.weight.quant_map": "model-00001-of-00002.safetensors",
994
- "model.layers.27.self_attn.v_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
995
  "model.layers.3.input_layernorm.weight": "model-00001-of-00002.safetensors",
996
  "model.layers.3.mlp.down_proj.weight": "model-00001-of-00002.safetensors",
997
  "model.layers.3.mlp.down_proj.weight.absmax": "model-00001-of-00002.safetensors",
@@ -1321,7 +1321,7 @@
1321
  "model.layers.9.self_attn.v_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
1322
  "model.layers.9.self_attn.v_proj.weight.quant_map": "model-00001-of-00002.safetensors",
1323
  "model.layers.9.self_attn.v_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
1324
- "model.norm.weight": "model-00001-of-00002.safetensors",
1325
  "visual.blocks.0.attn.proj.bias": "model-00001-of-00002.safetensors",
1326
  "visual.blocks.0.attn.proj.weight": "model-00001-of-00002.safetensors",
1327
  "visual.blocks.0.attn.proj.weight.absmax": "model-00001-of-00002.safetensors",
 
1
  {
2
  "metadata": {
3
+ "total_size": 8083929022
4
  },
5
  "weight_map": {
6
  "lm_head.weight": "model-00002-of-00002.safetensors",
 
616
  "model.layers.2.self_attn.v_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
617
  "model.layers.2.self_attn.v_proj.weight.quant_map": "model-00001-of-00002.safetensors",
618
  "model.layers.2.self_attn.v_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
619
+ "model.layers.20.input_layernorm.weight": "model-00002-of-00002.safetensors",
620
+ "model.layers.20.mlp.down_proj.weight": "model-00002-of-00002.safetensors",
621
+ "model.layers.20.mlp.down_proj.weight.absmax": "model-00002-of-00002.safetensors",
622
+ "model.layers.20.mlp.down_proj.weight.nested_absmax": "model-00002-of-00002.safetensors",
623
+ "model.layers.20.mlp.down_proj.weight.nested_quant_map": "model-00002-of-00002.safetensors",
624
+ "model.layers.20.mlp.down_proj.weight.quant_map": "model-00002-of-00002.safetensors",
625
+ "model.layers.20.mlp.down_proj.weight.quant_state.bitsandbytes__nf4": "model-00002-of-00002.safetensors",
626
  "model.layers.20.mlp.gate_proj.weight": "model-00001-of-00002.safetensors",
627
  "model.layers.20.mlp.gate_proj.weight.absmax": "model-00001-of-00002.safetensors",
628
  "model.layers.20.mlp.gate_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
629
  "model.layers.20.mlp.gate_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
630
  "model.layers.20.mlp.gate_proj.weight.quant_map": "model-00001-of-00002.safetensors",
631
  "model.layers.20.mlp.gate_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
632
+ "model.layers.20.mlp.up_proj.weight": "model-00002-of-00002.safetensors",
633
+ "model.layers.20.mlp.up_proj.weight.absmax": "model-00002-of-00002.safetensors",
634
+ "model.layers.20.mlp.up_proj.weight.nested_absmax": "model-00002-of-00002.safetensors",
635
+ "model.layers.20.mlp.up_proj.weight.nested_quant_map": "model-00002-of-00002.safetensors",
636
+ "model.layers.20.mlp.up_proj.weight.quant_map": "model-00002-of-00002.safetensors",
637
+ "model.layers.20.mlp.up_proj.weight.quant_state.bitsandbytes__nf4": "model-00002-of-00002.safetensors",
638
+ "model.layers.20.post_attention_layernorm.weight": "model-00002-of-00002.safetensors",
639
  "model.layers.20.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
640
  "model.layers.20.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
641
  "model.layers.20.self_attn.k_proj.weight.absmax": "model-00001-of-00002.safetensors",
 
663
  "model.layers.20.self_attn.v_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
664
  "model.layers.20.self_attn.v_proj.weight.quant_map": "model-00001-of-00002.safetensors",
665
  "model.layers.20.self_attn.v_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
666
+ "model.layers.21.input_layernorm.weight": "model-00002-of-00002.safetensors",
667
+ "model.layers.21.mlp.down_proj.weight": "model-00002-of-00002.safetensors",
668
+ "model.layers.21.mlp.down_proj.weight.absmax": "model-00002-of-00002.safetensors",
669
+ "model.layers.21.mlp.down_proj.weight.nested_absmax": "model-00002-of-00002.safetensors",
670
+ "model.layers.21.mlp.down_proj.weight.nested_quant_map": "model-00002-of-00002.safetensors",
671
+ "model.layers.21.mlp.down_proj.weight.quant_map": "model-00002-of-00002.safetensors",
672
+ "model.layers.21.mlp.down_proj.weight.quant_state.bitsandbytes__nf4": "model-00002-of-00002.safetensors",
673
+ "model.layers.21.mlp.gate_proj.weight": "model-00002-of-00002.safetensors",
674
+ "model.layers.21.mlp.gate_proj.weight.absmax": "model-00002-of-00002.safetensors",
675
+ "model.layers.21.mlp.gate_proj.weight.nested_absmax": "model-00002-of-00002.safetensors",
676
+ "model.layers.21.mlp.gate_proj.weight.nested_quant_map": "model-00002-of-00002.safetensors",
677
+ "model.layers.21.mlp.gate_proj.weight.quant_map": "model-00002-of-00002.safetensors",
678
+ "model.layers.21.mlp.gate_proj.weight.quant_state.bitsandbytes__nf4": "model-00002-of-00002.safetensors",
679
+ "model.layers.21.mlp.up_proj.weight": "model-00002-of-00002.safetensors",
680
+ "model.layers.21.mlp.up_proj.weight.absmax": "model-00002-of-00002.safetensors",
681
+ "model.layers.21.mlp.up_proj.weight.nested_absmax": "model-00002-of-00002.safetensors",
682
+ "model.layers.21.mlp.up_proj.weight.nested_quant_map": "model-00002-of-00002.safetensors",
683
+ "model.layers.21.mlp.up_proj.weight.quant_map": "model-00002-of-00002.safetensors",
684
+ "model.layers.21.mlp.up_proj.weight.quant_state.bitsandbytes__nf4": "model-00002-of-00002.safetensors",
685
+ "model.layers.21.post_attention_layernorm.weight": "model-00002-of-00002.safetensors",
686
+ "model.layers.21.self_attn.k_proj.bias": "model-00002-of-00002.safetensors",
687
+ "model.layers.21.self_attn.k_proj.weight": "model-00002-of-00002.safetensors",
688
+ "model.layers.21.self_attn.k_proj.weight.absmax": "model-00002-of-00002.safetensors",
689
+ "model.layers.21.self_attn.k_proj.weight.nested_absmax": "model-00002-of-00002.safetensors",
690
+ "model.layers.21.self_attn.k_proj.weight.nested_quant_map": "model-00002-of-00002.safetensors",
691
+ "model.layers.21.self_attn.k_proj.weight.quant_map": "model-00002-of-00002.safetensors",
692
+ "model.layers.21.self_attn.k_proj.weight.quant_state.bitsandbytes__nf4": "model-00002-of-00002.safetensors",
693
+ "model.layers.21.self_attn.o_proj.weight": "model-00002-of-00002.safetensors",
694
+ "model.layers.21.self_attn.o_proj.weight.absmax": "model-00002-of-00002.safetensors",
695
+ "model.layers.21.self_attn.o_proj.weight.nested_absmax": "model-00002-of-00002.safetensors",
696
+ "model.layers.21.self_attn.o_proj.weight.nested_quant_map": "model-00002-of-00002.safetensors",
697
+ "model.layers.21.self_attn.o_proj.weight.quant_map": "model-00002-of-00002.safetensors",
698
+ "model.layers.21.self_attn.o_proj.weight.quant_state.bitsandbytes__nf4": "model-00002-of-00002.safetensors",
699
+ "model.layers.21.self_attn.q_proj.bias": "model-00002-of-00002.safetensors",
700
+ "model.layers.21.self_attn.q_proj.weight": "model-00002-of-00002.safetensors",
701
+ "model.layers.21.self_attn.q_proj.weight.absmax": "model-00002-of-00002.safetensors",
702
+ "model.layers.21.self_attn.q_proj.weight.nested_absmax": "model-00002-of-00002.safetensors",
703
+ "model.layers.21.self_attn.q_proj.weight.nested_quant_map": "model-00002-of-00002.safetensors",
704
+ "model.layers.21.self_attn.q_proj.weight.quant_map": "model-00002-of-00002.safetensors",
705
+ "model.layers.21.self_attn.q_proj.weight.quant_state.bitsandbytes__nf4": "model-00002-of-00002.safetensors",
706
+ "model.layers.21.self_attn.v_proj.bias": "model-00002-of-00002.safetensors",
707
+ "model.layers.21.self_attn.v_proj.weight": "model-00002-of-00002.safetensors",
708
+ "model.layers.21.self_attn.v_proj.weight.absmax": "model-00002-of-00002.safetensors",
709
+ "model.layers.21.self_attn.v_proj.weight.nested_absmax": "model-00002-of-00002.safetensors",
710
+ "model.layers.21.self_attn.v_proj.weight.nested_quant_map": "model-00002-of-00002.safetensors",
711
+ "model.layers.21.self_attn.v_proj.weight.quant_map": "model-00002-of-00002.safetensors",
712
+ "model.layers.21.self_attn.v_proj.weight.quant_state.bitsandbytes__nf4": "model-00002-of-00002.safetensors",
713
+ "model.layers.22.input_layernorm.weight": "model-00002-of-00002.safetensors",
714
+ "model.layers.22.mlp.down_proj.weight": "model-00002-of-00002.safetensors",
715
+ "model.layers.22.mlp.down_proj.weight.absmax": "model-00002-of-00002.safetensors",
716
+ "model.layers.22.mlp.down_proj.weight.nested_absmax": "model-00002-of-00002.safetensors",
717
+ "model.layers.22.mlp.down_proj.weight.nested_quant_map": "model-00002-of-00002.safetensors",
718
+ "model.layers.22.mlp.down_proj.weight.quant_map": "model-00002-of-00002.safetensors",
719
+ "model.layers.22.mlp.down_proj.weight.quant_state.bitsandbytes__nf4": "model-00002-of-00002.safetensors",
720
+ "model.layers.22.mlp.gate_proj.weight": "model-00002-of-00002.safetensors",
721
+ "model.layers.22.mlp.gate_proj.weight.absmax": "model-00002-of-00002.safetensors",
722
+ "model.layers.22.mlp.gate_proj.weight.nested_absmax": "model-00002-of-00002.safetensors",
723
+ "model.layers.22.mlp.gate_proj.weight.nested_quant_map": "model-00002-of-00002.safetensors",
724
+ "model.layers.22.mlp.gate_proj.weight.quant_map": "model-00002-of-00002.safetensors",
725
+ "model.layers.22.mlp.gate_proj.weight.quant_state.bitsandbytes__nf4": "model-00002-of-00002.safetensors",
726
+ "model.layers.22.mlp.up_proj.weight": "model-00002-of-00002.safetensors",
727
+ "model.layers.22.mlp.up_proj.weight.absmax": "model-00002-of-00002.safetensors",
728
+ "model.layers.22.mlp.up_proj.weight.nested_absmax": "model-00002-of-00002.safetensors",
729
+ "model.layers.22.mlp.up_proj.weight.nested_quant_map": "model-00002-of-00002.safetensors",
730
+ "model.layers.22.mlp.up_proj.weight.quant_map": "model-00002-of-00002.safetensors",
731
+ "model.layers.22.mlp.up_proj.weight.quant_state.bitsandbytes__nf4": "model-00002-of-00002.safetensors",
732
+ "model.layers.22.post_attention_layernorm.weight": "model-00002-of-00002.safetensors",
733
+ "model.layers.22.self_attn.k_proj.bias": "model-00002-of-00002.safetensors",
734
+ "model.layers.22.self_attn.k_proj.weight": "model-00002-of-00002.safetensors",
735
+ "model.layers.22.self_attn.k_proj.weight.absmax": "model-00002-of-00002.safetensors",
736
+ "model.layers.22.self_attn.k_proj.weight.nested_absmax": "model-00002-of-00002.safetensors",
737
+ "model.layers.22.self_attn.k_proj.weight.nested_quant_map": "model-00002-of-00002.safetensors",
738
+ "model.layers.22.self_attn.k_proj.weight.quant_map": "model-00002-of-00002.safetensors",
739
+ "model.layers.22.self_attn.k_proj.weight.quant_state.bitsandbytes__nf4": "model-00002-of-00002.safetensors",
740
+ "model.layers.22.self_attn.o_proj.weight": "model-00002-of-00002.safetensors",
741
+ "model.layers.22.self_attn.o_proj.weight.absmax": "model-00002-of-00002.safetensors",
742
+ "model.layers.22.self_attn.o_proj.weight.nested_absmax": "model-00002-of-00002.safetensors",
743
+ "model.layers.22.self_attn.o_proj.weight.nested_quant_map": "model-00002-of-00002.safetensors",
744
+ "model.layers.22.self_attn.o_proj.weight.quant_map": "model-00002-of-00002.safetensors",
745
+ "model.layers.22.self_attn.o_proj.weight.quant_state.bitsandbytes__nf4": "model-00002-of-00002.safetensors",
746
+ "model.layers.22.self_attn.q_proj.bias": "model-00002-of-00002.safetensors",
747
+ "model.layers.22.self_attn.q_proj.weight": "model-00002-of-00002.safetensors",
748
+ "model.layers.22.self_attn.q_proj.weight.absmax": "model-00002-of-00002.safetensors",
749
+ "model.layers.22.self_attn.q_proj.weight.nested_absmax": "model-00002-of-00002.safetensors",
750
+ "model.layers.22.self_attn.q_proj.weight.nested_quant_map": "model-00002-of-00002.safetensors",
751
+ "model.layers.22.self_attn.q_proj.weight.quant_map": "model-00002-of-00002.safetensors",
752
+ "model.layers.22.self_attn.q_proj.weight.quant_state.bitsandbytes__nf4": "model-00002-of-00002.safetensors",
753
+ "model.layers.22.self_attn.v_proj.bias": "model-00002-of-00002.safetensors",
754
+ "model.layers.22.self_attn.v_proj.weight": "model-00002-of-00002.safetensors",
755
+ "model.layers.22.self_attn.v_proj.weight.absmax": "model-00002-of-00002.safetensors",
756
+ "model.layers.22.self_attn.v_proj.weight.nested_absmax": "model-00002-of-00002.safetensors",
757
+ "model.layers.22.self_attn.v_proj.weight.nested_quant_map": "model-00002-of-00002.safetensors",
758
+ "model.layers.22.self_attn.v_proj.weight.quant_map": "model-00002-of-00002.safetensors",
759
+ "model.layers.22.self_attn.v_proj.weight.quant_state.bitsandbytes__nf4": "model-00002-of-00002.safetensors",
760
+ "model.layers.23.input_layernorm.weight": "model-00002-of-00002.safetensors",
761
+ "model.layers.23.mlp.down_proj.weight": "model-00002-of-00002.safetensors",
762
+ "model.layers.23.mlp.down_proj.weight.absmax": "model-00002-of-00002.safetensors",
763
+ "model.layers.23.mlp.down_proj.weight.nested_absmax": "model-00002-of-00002.safetensors",
764
+ "model.layers.23.mlp.down_proj.weight.nested_quant_map": "model-00002-of-00002.safetensors",
765
+ "model.layers.23.mlp.down_proj.weight.quant_map": "model-00002-of-00002.safetensors",
766
+ "model.layers.23.mlp.down_proj.weight.quant_state.bitsandbytes__nf4": "model-00002-of-00002.safetensors",
767
+ "model.layers.23.mlp.gate_proj.weight": "model-00002-of-00002.safetensors",
768
+ "model.layers.23.mlp.gate_proj.weight.absmax": "model-00002-of-00002.safetensors",
769
+ "model.layers.23.mlp.gate_proj.weight.nested_absmax": "model-00002-of-00002.safetensors",
770
+ "model.layers.23.mlp.gate_proj.weight.nested_quant_map": "model-00002-of-00002.safetensors",
771
+ "model.layers.23.mlp.gate_proj.weight.quant_map": "model-00002-of-00002.safetensors",
772
+ "model.layers.23.mlp.gate_proj.weight.quant_state.bitsandbytes__nf4": "model-00002-of-00002.safetensors",
773
+ "model.layers.23.mlp.up_proj.weight": "model-00002-of-00002.safetensors",
774
+ "model.layers.23.mlp.up_proj.weight.absmax": "model-00002-of-00002.safetensors",
775
+ "model.layers.23.mlp.up_proj.weight.nested_absmax": "model-00002-of-00002.safetensors",
776
+ "model.layers.23.mlp.up_proj.weight.nested_quant_map": "model-00002-of-00002.safetensors",
777
+ "model.layers.23.mlp.up_proj.weight.quant_map": "model-00002-of-00002.safetensors",
778
+ "model.layers.23.mlp.up_proj.weight.quant_state.bitsandbytes__nf4": "model-00002-of-00002.safetensors",
779
+ "model.layers.23.post_attention_layernorm.weight": "model-00002-of-00002.safetensors",
780
+ "model.layers.23.self_attn.k_proj.bias": "model-00002-of-00002.safetensors",
781
+ "model.layers.23.self_attn.k_proj.weight": "model-00002-of-00002.safetensors",
782
+ "model.layers.23.self_attn.k_proj.weight.absmax": "model-00002-of-00002.safetensors",
783
+ "model.layers.23.self_attn.k_proj.weight.nested_absmax": "model-00002-of-00002.safetensors",
784
+ "model.layers.23.self_attn.k_proj.weight.nested_quant_map": "model-00002-of-00002.safetensors",
785
+ "model.layers.23.self_attn.k_proj.weight.quant_map": "model-00002-of-00002.safetensors",
786
+ "model.layers.23.self_attn.k_proj.weight.quant_state.bitsandbytes__nf4": "model-00002-of-00002.safetensors",
787
+ "model.layers.23.self_attn.o_proj.weight": "model-00002-of-00002.safetensors",
788
+ "model.layers.23.self_attn.o_proj.weight.absmax": "model-00002-of-00002.safetensors",
789
+ "model.layers.23.self_attn.o_proj.weight.nested_absmax": "model-00002-of-00002.safetensors",
790
+ "model.layers.23.self_attn.o_proj.weight.nested_quant_map": "model-00002-of-00002.safetensors",
791
+ "model.layers.23.self_attn.o_proj.weight.quant_map": "model-00002-of-00002.safetensors",
792
+ "model.layers.23.self_attn.o_proj.weight.quant_state.bitsandbytes__nf4": "model-00002-of-00002.safetensors",
793
+ "model.layers.23.self_attn.q_proj.bias": "model-00002-of-00002.safetensors",
794
+ "model.layers.23.self_attn.q_proj.weight": "model-00002-of-00002.safetensors",
795
+ "model.layers.23.self_attn.q_proj.weight.absmax": "model-00002-of-00002.safetensors",
796
+ "model.layers.23.self_attn.q_proj.weight.nested_absmax": "model-00002-of-00002.safetensors",
797
+ "model.layers.23.self_attn.q_proj.weight.nested_quant_map": "model-00002-of-00002.safetensors",
798
+ "model.layers.23.self_attn.q_proj.weight.quant_map": "model-00002-of-00002.safetensors",
799
+ "model.layers.23.self_attn.q_proj.weight.quant_state.bitsandbytes__nf4": "model-00002-of-00002.safetensors",
800
+ "model.layers.23.self_attn.v_proj.bias": "model-00002-of-00002.safetensors",
801
+ "model.layers.23.self_attn.v_proj.weight": "model-00002-of-00002.safetensors",
802
+ "model.layers.23.self_attn.v_proj.weight.absmax": "model-00002-of-00002.safetensors",
803
+ "model.layers.23.self_attn.v_proj.weight.nested_absmax": "model-00002-of-00002.safetensors",
804
+ "model.layers.23.self_attn.v_proj.weight.nested_quant_map": "model-00002-of-00002.safetensors",
805
+ "model.layers.23.self_attn.v_proj.weight.quant_map": "model-00002-of-00002.safetensors",
806
+ "model.layers.23.self_attn.v_proj.weight.quant_state.bitsandbytes__nf4": "model-00002-of-00002.safetensors",
807
+ "model.layers.24.input_layernorm.weight": "model-00002-of-00002.safetensors",
808
+ "model.layers.24.mlp.down_proj.weight": "model-00002-of-00002.safetensors",
809
+ "model.layers.24.mlp.down_proj.weight.absmax": "model-00002-of-00002.safetensors",
810
+ "model.layers.24.mlp.down_proj.weight.nested_absmax": "model-00002-of-00002.safetensors",
811
+ "model.layers.24.mlp.down_proj.weight.nested_quant_map": "model-00002-of-00002.safetensors",
812
+ "model.layers.24.mlp.down_proj.weight.quant_map": "model-00002-of-00002.safetensors",
813
+ "model.layers.24.mlp.down_proj.weight.quant_state.bitsandbytes__nf4": "model-00002-of-00002.safetensors",
814
+ "model.layers.24.mlp.gate_proj.weight": "model-00002-of-00002.safetensors",
815
+ "model.layers.24.mlp.gate_proj.weight.absmax": "model-00002-of-00002.safetensors",
816
+ "model.layers.24.mlp.gate_proj.weight.nested_absmax": "model-00002-of-00002.safetensors",
817
+ "model.layers.24.mlp.gate_proj.weight.nested_quant_map": "model-00002-of-00002.safetensors",
818
+ "model.layers.24.mlp.gate_proj.weight.quant_map": "model-00002-of-00002.safetensors",
819
+ "model.layers.24.mlp.gate_proj.weight.quant_state.bitsandbytes__nf4": "model-00002-of-00002.safetensors",
820
+ "model.layers.24.mlp.up_proj.weight": "model-00002-of-00002.safetensors",
821
+ "model.layers.24.mlp.up_proj.weight.absmax": "model-00002-of-00002.safetensors",
822
+ "model.layers.24.mlp.up_proj.weight.nested_absmax": "model-00002-of-00002.safetensors",
823
+ "model.layers.24.mlp.up_proj.weight.nested_quant_map": "model-00002-of-00002.safetensors",
824
+ "model.layers.24.mlp.up_proj.weight.quant_map": "model-00002-of-00002.safetensors",
825
+ "model.layers.24.mlp.up_proj.weight.quant_state.bitsandbytes__nf4": "model-00002-of-00002.safetensors",
826
+ "model.layers.24.post_attention_layernorm.weight": "model-00002-of-00002.safetensors",
827
+ "model.layers.24.self_attn.k_proj.bias": "model-00002-of-00002.safetensors",
828
+ "model.layers.24.self_attn.k_proj.weight": "model-00002-of-00002.safetensors",
829
+ "model.layers.24.self_attn.k_proj.weight.absmax": "model-00002-of-00002.safetensors",
830
+ "model.layers.24.self_attn.k_proj.weight.nested_absmax": "model-00002-of-00002.safetensors",
831
+ "model.layers.24.self_attn.k_proj.weight.nested_quant_map": "model-00002-of-00002.safetensors",
832
+ "model.layers.24.self_attn.k_proj.weight.quant_map": "model-00002-of-00002.safetensors",
833
+ "model.layers.24.self_attn.k_proj.weight.quant_state.bitsandbytes__nf4": "model-00002-of-00002.safetensors",
834
+ "model.layers.24.self_attn.o_proj.weight": "model-00002-of-00002.safetensors",
835
+ "model.layers.24.self_attn.o_proj.weight.absmax": "model-00002-of-00002.safetensors",
836
+ "model.layers.24.self_attn.o_proj.weight.nested_absmax": "model-00002-of-00002.safetensors",
837
+ "model.layers.24.self_attn.o_proj.weight.nested_quant_map": "model-00002-of-00002.safetensors",
838
+ "model.layers.24.self_attn.o_proj.weight.quant_map": "model-00002-of-00002.safetensors",
839
+ "model.layers.24.self_attn.o_proj.weight.quant_state.bitsandbytes__nf4": "model-00002-of-00002.safetensors",
840
+ "model.layers.24.self_attn.q_proj.bias": "model-00002-of-00002.safetensors",
841
+ "model.layers.24.self_attn.q_proj.weight": "model-00002-of-00002.safetensors",
842
+ "model.layers.24.self_attn.q_proj.weight.absmax": "model-00002-of-00002.safetensors",
843
+ "model.layers.24.self_attn.q_proj.weight.nested_absmax": "model-00002-of-00002.safetensors",
844
+ "model.layers.24.self_attn.q_proj.weight.nested_quant_map": "model-00002-of-00002.safetensors",
845
+ "model.layers.24.self_attn.q_proj.weight.quant_map": "model-00002-of-00002.safetensors",
846
+ "model.layers.24.self_attn.q_proj.weight.quant_state.bitsandbytes__nf4": "model-00002-of-00002.safetensors",
847
+ "model.layers.24.self_attn.v_proj.bias": "model-00002-of-00002.safetensors",
848
+ "model.layers.24.self_attn.v_proj.weight": "model-00002-of-00002.safetensors",
849
+ "model.layers.24.self_attn.v_proj.weight.absmax": "model-00002-of-00002.safetensors",
850
+ "model.layers.24.self_attn.v_proj.weight.nested_absmax": "model-00002-of-00002.safetensors",
851
+ "model.layers.24.self_attn.v_proj.weight.nested_quant_map": "model-00002-of-00002.safetensors",
852
+ "model.layers.24.self_attn.v_proj.weight.quant_map": "model-00002-of-00002.safetensors",
853
+ "model.layers.24.self_attn.v_proj.weight.quant_state.bitsandbytes__nf4": "model-00002-of-00002.safetensors",
854
+ "model.layers.25.input_layernorm.weight": "model-00002-of-00002.safetensors",
855
+ "model.layers.25.mlp.down_proj.weight": "model-00002-of-00002.safetensors",
856
+ "model.layers.25.mlp.down_proj.weight.absmax": "model-00002-of-00002.safetensors",
857
+ "model.layers.25.mlp.down_proj.weight.nested_absmax": "model-00002-of-00002.safetensors",
858
+ "model.layers.25.mlp.down_proj.weight.nested_quant_map": "model-00002-of-00002.safetensors",
859
+ "model.layers.25.mlp.down_proj.weight.quant_map": "model-00002-of-00002.safetensors",
860
+ "model.layers.25.mlp.down_proj.weight.quant_state.bitsandbytes__nf4": "model-00002-of-00002.safetensors",
861
+ "model.layers.25.mlp.gate_proj.weight": "model-00002-of-00002.safetensors",
862
+ "model.layers.25.mlp.gate_proj.weight.absmax": "model-00002-of-00002.safetensors",
863
+ "model.layers.25.mlp.gate_proj.weight.nested_absmax": "model-00002-of-00002.safetensors",
864
+ "model.layers.25.mlp.gate_proj.weight.nested_quant_map": "model-00002-of-00002.safetensors",
865
+ "model.layers.25.mlp.gate_proj.weight.quant_map": "model-00002-of-00002.safetensors",
866
+ "model.layers.25.mlp.gate_proj.weight.quant_state.bitsandbytes__nf4": "model-00002-of-00002.safetensors",
867
+ "model.layers.25.mlp.up_proj.weight": "model-00002-of-00002.safetensors",
868
+ "model.layers.25.mlp.up_proj.weight.absmax": "model-00002-of-00002.safetensors",
869
+ "model.layers.25.mlp.up_proj.weight.nested_absmax": "model-00002-of-00002.safetensors",
870
+ "model.layers.25.mlp.up_proj.weight.nested_quant_map": "model-00002-of-00002.safetensors",
871
+ "model.layers.25.mlp.up_proj.weight.quant_map": "model-00002-of-00002.safetensors",
872
+ "model.layers.25.mlp.up_proj.weight.quant_state.bitsandbytes__nf4": "model-00002-of-00002.safetensors",
873
+ "model.layers.25.post_attention_layernorm.weight": "model-00002-of-00002.safetensors",
874
+ "model.layers.25.self_attn.k_proj.bias": "model-00002-of-00002.safetensors",
875
+ "model.layers.25.self_attn.k_proj.weight": "model-00002-of-00002.safetensors",
876
+ "model.layers.25.self_attn.k_proj.weight.absmax": "model-00002-of-00002.safetensors",
877
+ "model.layers.25.self_attn.k_proj.weight.nested_absmax": "model-00002-of-00002.safetensors",
878
+ "model.layers.25.self_attn.k_proj.weight.nested_quant_map": "model-00002-of-00002.safetensors",
879
+ "model.layers.25.self_attn.k_proj.weight.quant_map": "model-00002-of-00002.safetensors",
880
+ "model.layers.25.self_attn.k_proj.weight.quant_state.bitsandbytes__nf4": "model-00002-of-00002.safetensors",
881
+ "model.layers.25.self_attn.o_proj.weight": "model-00002-of-00002.safetensors",
882
+ "model.layers.25.self_attn.o_proj.weight.absmax": "model-00002-of-00002.safetensors",
883
+ "model.layers.25.self_attn.o_proj.weight.nested_absmax": "model-00002-of-00002.safetensors",
884
+ "model.layers.25.self_attn.o_proj.weight.nested_quant_map": "model-00002-of-00002.safetensors",
885
+ "model.layers.25.self_attn.o_proj.weight.quant_map": "model-00002-of-00002.safetensors",
886
+ "model.layers.25.self_attn.o_proj.weight.quant_state.bitsandbytes__nf4": "model-00002-of-00002.safetensors",
887
+ "model.layers.25.self_attn.q_proj.bias": "model-00002-of-00002.safetensors",
888
+ "model.layers.25.self_attn.q_proj.weight": "model-00002-of-00002.safetensors",
889
+ "model.layers.25.self_attn.q_proj.weight.absmax": "model-00002-of-00002.safetensors",
890
+ "model.layers.25.self_attn.q_proj.weight.nested_absmax": "model-00002-of-00002.safetensors",
891
+ "model.layers.25.self_attn.q_proj.weight.nested_quant_map": "model-00002-of-00002.safetensors",
892
+ "model.layers.25.self_attn.q_proj.weight.quant_map": "model-00002-of-00002.safetensors",
893
+ "model.layers.25.self_attn.q_proj.weight.quant_state.bitsandbytes__nf4": "model-00002-of-00002.safetensors",
894
+ "model.layers.25.self_attn.v_proj.bias": "model-00002-of-00002.safetensors",
895
+ "model.layers.25.self_attn.v_proj.weight": "model-00002-of-00002.safetensors",
896
+ "model.layers.25.self_attn.v_proj.weight.absmax": "model-00002-of-00002.safetensors",
897
+ "model.layers.25.self_attn.v_proj.weight.nested_absmax": "model-00002-of-00002.safetensors",
898
+ "model.layers.25.self_attn.v_proj.weight.nested_quant_map": "model-00002-of-00002.safetensors",
899
+ "model.layers.25.self_attn.v_proj.weight.quant_map": "model-00002-of-00002.safetensors",
900
+ "model.layers.25.self_attn.v_proj.weight.quant_state.bitsandbytes__nf4": "model-00002-of-00002.safetensors",
901
+ "model.layers.26.input_layernorm.weight": "model-00002-of-00002.safetensors",
902
+ "model.layers.26.mlp.down_proj.weight": "model-00002-of-00002.safetensors",
903
+ "model.layers.26.mlp.down_proj.weight.absmax": "model-00002-of-00002.safetensors",
904
+ "model.layers.26.mlp.down_proj.weight.nested_absmax": "model-00002-of-00002.safetensors",
905
+ "model.layers.26.mlp.down_proj.weight.nested_quant_map": "model-00002-of-00002.safetensors",
906
+ "model.layers.26.mlp.down_proj.weight.quant_map": "model-00002-of-00002.safetensors",
907
+ "model.layers.26.mlp.down_proj.weight.quant_state.bitsandbytes__nf4": "model-00002-of-00002.safetensors",
908
+ "model.layers.26.mlp.gate_proj.weight": "model-00002-of-00002.safetensors",
909
+ "model.layers.26.mlp.gate_proj.weight.absmax": "model-00002-of-00002.safetensors",
910
+ "model.layers.26.mlp.gate_proj.weight.nested_absmax": "model-00002-of-00002.safetensors",
911
+ "model.layers.26.mlp.gate_proj.weight.nested_quant_map": "model-00002-of-00002.safetensors",
912
+ "model.layers.26.mlp.gate_proj.weight.quant_map": "model-00002-of-00002.safetensors",
913
+ "model.layers.26.mlp.gate_proj.weight.quant_state.bitsandbytes__nf4": "model-00002-of-00002.safetensors",
914
+ "model.layers.26.mlp.up_proj.weight": "model-00002-of-00002.safetensors",
915
+ "model.layers.26.mlp.up_proj.weight.absmax": "model-00002-of-00002.safetensors",
916
+ "model.layers.26.mlp.up_proj.weight.nested_absmax": "model-00002-of-00002.safetensors",
917
+ "model.layers.26.mlp.up_proj.weight.nested_quant_map": "model-00002-of-00002.safetensors",
918
+ "model.layers.26.mlp.up_proj.weight.quant_map": "model-00002-of-00002.safetensors",
919
+ "model.layers.26.mlp.up_proj.weight.quant_state.bitsandbytes__nf4": "model-00002-of-00002.safetensors",
920
+ "model.layers.26.post_attention_layernorm.weight": "model-00002-of-00002.safetensors",
921
+ "model.layers.26.self_attn.k_proj.bias": "model-00002-of-00002.safetensors",
922
+ "model.layers.26.self_attn.k_proj.weight": "model-00002-of-00002.safetensors",
923
+ "model.layers.26.self_attn.k_proj.weight.absmax": "model-00002-of-00002.safetensors",
924
+ "model.layers.26.self_attn.k_proj.weight.nested_absmax": "model-00002-of-00002.safetensors",
925
+ "model.layers.26.self_attn.k_proj.weight.nested_quant_map": "model-00002-of-00002.safetensors",
926
+ "model.layers.26.self_attn.k_proj.weight.quant_map": "model-00002-of-00002.safetensors",
927
+ "model.layers.26.self_attn.k_proj.weight.quant_state.bitsandbytes__nf4": "model-00002-of-00002.safetensors",
928
+ "model.layers.26.self_attn.o_proj.weight": "model-00002-of-00002.safetensors",
929
+ "model.layers.26.self_attn.o_proj.weight.absmax": "model-00002-of-00002.safetensors",
930
+ "model.layers.26.self_attn.o_proj.weight.nested_absmax": "model-00002-of-00002.safetensors",
931
+ "model.layers.26.self_attn.o_proj.weight.nested_quant_map": "model-00002-of-00002.safetensors",
932
+ "model.layers.26.self_attn.o_proj.weight.quant_map": "model-00002-of-00002.safetensors",
933
+ "model.layers.26.self_attn.o_proj.weight.quant_state.bitsandbytes__nf4": "model-00002-of-00002.safetensors",
934
+ "model.layers.26.self_attn.q_proj.bias": "model-00002-of-00002.safetensors",
935
+ "model.layers.26.self_attn.q_proj.weight": "model-00002-of-00002.safetensors",
936
+ "model.layers.26.self_attn.q_proj.weight.absmax": "model-00002-of-00002.safetensors",
937
+ "model.layers.26.self_attn.q_proj.weight.nested_absmax": "model-00002-of-00002.safetensors",
938
+ "model.layers.26.self_attn.q_proj.weight.nested_quant_map": "model-00002-of-00002.safetensors",
939
+ "model.layers.26.self_attn.q_proj.weight.quant_map": "model-00002-of-00002.safetensors",
940
+ "model.layers.26.self_attn.q_proj.weight.quant_state.bitsandbytes__nf4": "model-00002-of-00002.safetensors",
941
+ "model.layers.26.self_attn.v_proj.bias": "model-00002-of-00002.safetensors",
942
+ "model.layers.26.self_attn.v_proj.weight": "model-00002-of-00002.safetensors",
943
+ "model.layers.26.self_attn.v_proj.weight.absmax": "model-00002-of-00002.safetensors",
944
+ "model.layers.26.self_attn.v_proj.weight.nested_absmax": "model-00002-of-00002.safetensors",
945
+ "model.layers.26.self_attn.v_proj.weight.nested_quant_map": "model-00002-of-00002.safetensors",
946
+ "model.layers.26.self_attn.v_proj.weight.quant_map": "model-00002-of-00002.safetensors",
947
+ "model.layers.26.self_attn.v_proj.weight.quant_state.bitsandbytes__nf4": "model-00002-of-00002.safetensors",
948
+ "model.layers.27.input_layernorm.weight": "model-00002-of-00002.safetensors",
949
+ "model.layers.27.mlp.down_proj.weight": "model-00002-of-00002.safetensors",
950
+ "model.layers.27.mlp.down_proj.weight.absmax": "model-00002-of-00002.safetensors",
951
+ "model.layers.27.mlp.down_proj.weight.nested_absmax": "model-00002-of-00002.safetensors",
952
+ "model.layers.27.mlp.down_proj.weight.nested_quant_map": "model-00002-of-00002.safetensors",
953
+ "model.layers.27.mlp.down_proj.weight.quant_map": "model-00002-of-00002.safetensors",
954
+ "model.layers.27.mlp.down_proj.weight.quant_state.bitsandbytes__nf4": "model-00002-of-00002.safetensors",
955
+ "model.layers.27.mlp.gate_proj.weight": "model-00002-of-00002.safetensors",
956
+ "model.layers.27.mlp.gate_proj.weight.absmax": "model-00002-of-00002.safetensors",
957
+ "model.layers.27.mlp.gate_proj.weight.nested_absmax": "model-00002-of-00002.safetensors",
958
+ "model.layers.27.mlp.gate_proj.weight.nested_quant_map": "model-00002-of-00002.safetensors",
959
+ "model.layers.27.mlp.gate_proj.weight.quant_map": "model-00002-of-00002.safetensors",
960
+ "model.layers.27.mlp.gate_proj.weight.quant_state.bitsandbytes__nf4": "model-00002-of-00002.safetensors",
961
+ "model.layers.27.mlp.up_proj.weight": "model-00002-of-00002.safetensors",
962
+ "model.layers.27.mlp.up_proj.weight.absmax": "model-00002-of-00002.safetensors",
963
+ "model.layers.27.mlp.up_proj.weight.nested_absmax": "model-00002-of-00002.safetensors",
964
+ "model.layers.27.mlp.up_proj.weight.nested_quant_map": "model-00002-of-00002.safetensors",
965
+ "model.layers.27.mlp.up_proj.weight.quant_map": "model-00002-of-00002.safetensors",
966
+ "model.layers.27.mlp.up_proj.weight.quant_state.bitsandbytes__nf4": "model-00002-of-00002.safetensors",
967
+ "model.layers.27.post_attention_layernorm.weight": "model-00002-of-00002.safetensors",
968
+ "model.layers.27.self_attn.k_proj.bias": "model-00002-of-00002.safetensors",
969
+ "model.layers.27.self_attn.k_proj.weight": "model-00002-of-00002.safetensors",
970
+ "model.layers.27.self_attn.k_proj.weight.absmax": "model-00002-of-00002.safetensors",
971
+ "model.layers.27.self_attn.k_proj.weight.nested_absmax": "model-00002-of-00002.safetensors",
972
+ "model.layers.27.self_attn.k_proj.weight.nested_quant_map": "model-00002-of-00002.safetensors",
973
+ "model.layers.27.self_attn.k_proj.weight.quant_map": "model-00002-of-00002.safetensors",
974
+ "model.layers.27.self_attn.k_proj.weight.quant_state.bitsandbytes__nf4": "model-00002-of-00002.safetensors",
975
+ "model.layers.27.self_attn.o_proj.weight": "model-00002-of-00002.safetensors",
976
+ "model.layers.27.self_attn.o_proj.weight.absmax": "model-00002-of-00002.safetensors",
977
+ "model.layers.27.self_attn.o_proj.weight.nested_absmax": "model-00002-of-00002.safetensors",
978
+ "model.layers.27.self_attn.o_proj.weight.nested_quant_map": "model-00002-of-00002.safetensors",
979
+ "model.layers.27.self_attn.o_proj.weight.quant_map": "model-00002-of-00002.safetensors",
980
+ "model.layers.27.self_attn.o_proj.weight.quant_state.bitsandbytes__nf4": "model-00002-of-00002.safetensors",
981
+ "model.layers.27.self_attn.q_proj.bias": "model-00002-of-00002.safetensors",
982
+ "model.layers.27.self_attn.q_proj.weight": "model-00002-of-00002.safetensors",
983
+ "model.layers.27.self_attn.q_proj.weight.absmax": "model-00002-of-00002.safetensors",
984
+ "model.layers.27.self_attn.q_proj.weight.nested_absmax": "model-00002-of-00002.safetensors",
985
+ "model.layers.27.self_attn.q_proj.weight.nested_quant_map": "model-00002-of-00002.safetensors",
986
+ "model.layers.27.self_attn.q_proj.weight.quant_map": "model-00002-of-00002.safetensors",
987
+ "model.layers.27.self_attn.q_proj.weight.quant_state.bitsandbytes__nf4": "model-00002-of-00002.safetensors",
988
+ "model.layers.27.self_attn.v_proj.bias": "model-00002-of-00002.safetensors",
989
+ "model.layers.27.self_attn.v_proj.weight": "model-00002-of-00002.safetensors",
990
+ "model.layers.27.self_attn.v_proj.weight.absmax": "model-00002-of-00002.safetensors",
991
+ "model.layers.27.self_attn.v_proj.weight.nested_absmax": "model-00002-of-00002.safetensors",
992
+ "model.layers.27.self_attn.v_proj.weight.nested_quant_map": "model-00002-of-00002.safetensors",
993
+ "model.layers.27.self_attn.v_proj.weight.quant_map": "model-00002-of-00002.safetensors",
994
+ "model.layers.27.self_attn.v_proj.weight.quant_state.bitsandbytes__nf4": "model-00002-of-00002.safetensors",
995
  "model.layers.3.input_layernorm.weight": "model-00001-of-00002.safetensors",
996
  "model.layers.3.mlp.down_proj.weight": "model-00001-of-00002.safetensors",
997
  "model.layers.3.mlp.down_proj.weight.absmax": "model-00001-of-00002.safetensors",
 
1321
  "model.layers.9.self_attn.v_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
1322
  "model.layers.9.self_attn.v_proj.weight.quant_map": "model-00001-of-00002.safetensors",
1323
  "model.layers.9.self_attn.v_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
1324
+ "model.norm.weight": "model-00002-of-00002.safetensors",
1325
  "visual.blocks.0.attn.proj.bias": "model-00001-of-00002.safetensors",
1326
  "visual.blocks.0.attn.proj.weight": "model-00001-of-00002.safetensors",
1327
  "visual.blocks.0.attn.proj.weight.absmax": "model-00001-of-00002.safetensors",