--- base_model: - CultriX/Llama3-8B-DPO - hiieu/Meta-Llama-3-8B-Instruct-function-calling-json-mode - Orenguteng/Llama-3.1-8B-Lexi-Uncensored-V2 - UCLA-AGI/Llama-3-Instruct-8B-SPPO-Iter3 library_name: transformers tags: - mergekit - merge --- # merge This is a merge of pre-trained language models created using [mergekit](https://github.com/cg123/mergekit). ## Merge Details ### Merge Method This model was merged using the [task arithmetic](https://arxiv.org/abs/2212.04089) merge method using [UCLA-AGI/Llama-3-Instruct-8B-SPPO-Iter3](https://huggingface.co/UCLA-AGI/Llama-3-Instruct-8B-SPPO-Iter3) as a base. ### Models Merged The following models were included in the merge: * [CultriX/Llama3-8B-DPO](https://huggingface.co/CultriX/Llama3-8B-DPO) * [hiieu/Meta-Llama-3-8B-Instruct-function-calling-json-mode](https://huggingface.co/hiieu/Meta-Llama-3-8B-Instruct-function-calling-json-mode) * [Orenguteng/Llama-3.1-8B-Lexi-Uncensored-V2](https://huggingface.co/Orenguteng/Llama-3.1-8B-Lexi-Uncensored-V2) ### Configuration The following YAML configuration was used to produce this model: ```yaml slices: - sources: - model: hiieu/Meta-Llama-3-8B-Instruct-function-calling-json-mode parameters: weight: 1 layer_range: [0, 32] - model: Orenguteng/Llama-3.1-8B-Lexi-Uncensored-V2 parameters: weight: 1 layer_range: [0, 32] - model: CultriX/Llama3-8B-DPO parameters: weight: 0.3 layer_range: [0, 32] - model: UCLA-AGI/Llama-3-Instruct-8B-SPPO-Iter3 parameters: weight: 0.7 layer_range: [0, 32] merge_method: task_arithmetic base_model: UCLA-AGI/Llama-3-Instruct-8B-SPPO-Iter3 parameters: t: - filter: self_attn value: [0, 0.5, 0.3, 0.7, 1] - filter: mlp value: [1, 0.5, 0.7, 0.3, 0] - value: 0.5 dtype: bfloat16 ```