--- base_model: - yamatazen/EtherealAurora-12B - Sicarius-Prototyping/Impish_Longtail_12B - allura-org/MN-Lyrebird-12B library_name: transformers tags: - mergekit - merge license: apache-2.0 model-index: - name: Chrysologus-12B results: - task: type: text-generation name: Text Generation dataset: name: IFEval (0-Shot) type: HuggingFaceH4/ifeval args: num_few_shot: 0 metrics: - type: prompt_level_strict_acc value: 65.62 name: strict accuracy comment: self-reported --- # Chrysologus-12B Storytelling with decent instruction following. Has better instruction following than [Retreatcost/Impish-LongPen-12B](https://huggingface.co/Retreatcost/Impish-LongPen-12B) This is a merge of pre-trained language models created using [mergekit](https://github.com/cg123/mergekit). ## Merge Details ### Merge Method This model was merged using the [Karcher Mean](https://en.wikipedia.org/wiki/Karcher_mean) merge method. ### Models Merged The following models were included in the merge: * [yamatazen/EtherealAurora-12B](https://huggingface.co/yamatazen/EtherealAurora-12B) * [Sicarius-Prototyping/Impish_Longtail_12B](https://huggingface.co/Sicarius-Prototyping/Impish_Longtail_12B) * [allura-org/MN-Lyrebird-12B](https://huggingface.co/allura-org/MN-Lyrebird-12B) ### Configuration The following YAML configuration was used to produce this model: ```yaml merge_method: karcher models: - model: allura-org/MN-Lyrebird-12B - model: Sicarius-Prototyping/Impish_Longtail_12B - model: yamatazen/EtherealAurora-12B parameters: max_iter: 100000 tol: 1e-9 dtype: bfloat16 ```