| | --- |
| | base_model: |
| | - nbeerbower/Llama3.1-Gutenberg-Doppel-70B |
| | - Nexesenex/Llama_3.x_70b_SmarTricks_V1.01 |
| | - Nexesenex/Llama_3.1_70b_HighPriestess_R1_V1 |
| | - Steelskull/L3.3-Electra-R1-70b |
| | - Nexesenex/Llama_3.3_70b_DarkHorse |
| | - Nexesenex/Llama_3.1_70b_FLDx2-Tess3_abliterated_fusion_norm |
| | library_name: transformers |
| | tags: |
| | - mergekit |
| | - merge |
| |
|
| | --- |
| | # about |
| |
|
| | V3.0 changes (relatively minor update) : |
| | - DarkHorse replace DoppelGangerR1 to add a bit of Negative Llama at the expense of a bit of Fallen Llama R1. |
| | - A bit of Fallen Llama is recovered by using Smartricks instead of Smartracks as a base. |
| | - Priestess is upgraded with Lumitron Lorablated. |
| | - Tess is merged with Hitachi FLDx2 in the perplexity dropper model. |
| |
|
| | Electra R1 and GutenbergDoppel are kept as they were. |
| |
|
| | If you have V2 already, this model is quite similar, and the difference might not be worth a download. |
| |
|
| | --- |
| | # merge |
| |
|
| | This is a merge of pre-trained language models created using [mergekit](https://github.com/cg123/mergekit). |
| |
|
| | ## Merge Details |
| | ### Merge Method |
| |
|
| | This model was merged using the [Model Stock](https://arxiv.org/abs/2403.19522) merge method using [Nexesenex/Llama_3.x_70b_SmarTricks_V1.01](https://huggingface.co/Nexesenex/Llama_3.x_70b_SmarTricks_V1.01) as a base. |
| |
|
| | ### Models Merged |
| |
|
| | The following models were included in the merge: |
| | * [nbeerbower/Llama3.1-Gutenberg-Doppel-70B](https://huggingface.co/nbeerbower/Llama3.1-Gutenberg-Doppel-70B) |
| | * [Nexesenex/Llama_3.1_70b_HighPriestess_R1_V1](https://huggingface.co/Nexesenex/Llama_3.1_70b_HighPriestess_R1_V1) |
| | * [Steelskull/L3.3-Electra-R1-70b](https://huggingface.co/Steelskull/L3.3-Electra-R1-70b) |
| | * [Nexesenex/Llama_3.3_70b_DarkHorse](https://huggingface.co/Nexesenex/Llama_3.3_70b_DarkHorse) |
| | * [Nexesenex/Llama_3.1_70b_FLDx2-Tess3_abliterated_fusion_norm](https://huggingface.co/Nexesenex/Llama_3.1_70b_FLDx2-Tess3_abliterated_fusion_norm) |
| |
|
| | ### Configuration |
| |
|
| | The following YAML configuration was used to produce this model: |
| |
|
| | ```yaml |
| | merge_method: model_stock |
| | models: |
| | - model: Nexesenex/Llama_3.1_70b_FLDx2-Tess3_abliterated_fusion_norm |
| | parameters: |
| | weight: 1.0 |
| | - model: nbeerbower/Llama3.1-Gutenberg-Doppel-70B |
| | parameters: |
| | weight: 1.0 |
| | - model: Nexesenex/Llama_3.1_70b_HighPriestess_R1_V1 |
| | parameters: |
| | weight: 1.0 |
| | - model: Steelskull/L3.3-Electra-R1-70b |
| | parameters: |
| | weight: 1.0 |
| | - model: Nexesenex/Llama_3.3_70b_DarkHorse |
| | parameters: |
| | weight: 1.0 |
| | base_model: Nexesenex/Llama_3.x_70b_SmarTricks_V1.01 |
| | dtype: bfloat16 |
| | out_dtype: bfloat16 |
| | parameters: |
| | int8_mask: true |
| | normalize: true |
| | rescale: false |
| | chat_template: auto |
| | tokenizer: |
| | source: union |
| | ``` |
| |
|