xxx777xxxASD's picture
Update README.md
3be88fe verified
|
raw
history blame
2.3 kB
metadata
license: llama3
tags:
  - moe
language:
  - en

(No waifu image yet)

Experimental RP-oriented MoE, the idea was to get a model that would be equal to or better than the Mixtral 8x7B and it's finetunes in RP/ERP tasks.

Please feedback me if it's more stable than the previous version

Llama 3 ChaoticSoliloquy-v2-4x8B test

base_model: L3_ChaosMaid_8B
gate_mode: random
dtype: bfloat16
experts_per_token: 2
experts:
  - source_model: ChaoticNeutrals_Poppy_Porpoise-0.72-L3-8B
  - source_model: L3_ChaosMaid_8B
  - source_model: openlynn_Llama-3-Soliloquy-8B-v2
  - source_model: Sao10K_L3-Solana-8B-v1

Models used

Difference

L3 ChaosMaid-8B

models:
  - model: jeiku_Chaos_RP_l3_8B
    # No parameters necessary for base model
  - model: NeverSleep_Llama-3-Lumimaid-8B-v0.1
    parameters:
      density: 0.5
      weight: 0.5
merge_method: dare_ties
base_model: jeiku_Chaos_RP_l3_8B
parameters:
  int8_mask: true
dtype: bfloat16

Vision

llama3_mmproj

image/png

Prompt format: Llama 3