| { | |
| "_attn_implementation_autoset": true, | |
| "ae_mode": "token", | |
| "attn_implementation": null, | |
| "auto_map": { | |
| "AutoConfig": "modelling_pisco.COCOMConfig", | |
| "AutoModel": "modelling_pisco.COCOM" | |
| }, | |
| "compr_base_model_name": "mistralai/Mistral-7B-Instruct-v0.2", | |
| "compr_every_n_layer": null, | |
| "compr_linear_type": "concat", | |
| "compr_mlp_hidden_dim": 8096, | |
| "compr_model_name": null, | |
| "compr_n_layers": null, | |
| "compr_rate": 16, | |
| "compr_rms_norm": false, | |
| "compr_use_mlp": true, | |
| "decoder_model_name": "Upstage/SOLAR-10.7B-Instruct-v1.0", | |
| "device_map": null, | |
| "different_mem_tokens": true, | |
| "doc_max_length": 128, | |
| "generation_top_k": 1, | |
| "kbtc_training": false, | |
| "load_adapters": true, | |
| "lora": true, | |
| "lora_compressor": false, | |
| "lora_r": 16, | |
| "lora_r_compressor": 16, | |
| "max_new_tokens": 128, | |
| "model_type": "COCOM", | |
| "optimize_mem_tokens": true, | |
| "quantization": "no", | |
| "sep": true, | |
| "training_form": "both_separately", | |
| "transformers_version": "4.48.0" | |
| } | |