kuotient commited on
Commit
cc96ade
1 Parent(s): 6ddd3ef

Upload folder using huggingface_hub

Browse files
.ipynb_checkpoints/quantize_config-checkpoint.json ADDED
@@ -0,0 +1,13 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "bits": 4,
3
+ "group_size": 128,
4
+ "damp_percent": 0.01,
5
+ "desc_act": false,
6
+ "static_groups": false,
7
+ "sym": true,
8
+ "true_sequential": true,
9
+ "model_name_or_path": null,
10
+ "model_file_base_name": null,
11
+ "is_marlin_format": true,
12
+ "quant_method": "gptq"
13
+ }
config.json CHANGED
@@ -1,5 +1,5 @@
1
  {
2
- "_name_or_path": "/data/unsloth/final/Llama-3-Alpha-ORPO-gpt4o-v1",
3
  "architectures": [
4
  "LlamaForCausalLM"
5
  ],
@@ -12,6 +12,7 @@
12
  "initializer_range": 0.02,
13
  "intermediate_size": 14336,
14
  "max_position_embeddings": 8192,
 
15
  "model_type": "llama",
16
  "num_attention_heads": 32,
17
  "num_hidden_layers": 32,
@@ -20,10 +21,10 @@
20
  "pretraining_tp": 1,
21
  "quantization_config": {
22
  "bits": 4,
23
- "checkpoint_format": "marlin",
24
  "damp_percent": 0.01,
25
  "desc_act": false,
26
  "group_size": 128,
 
27
  "model_file_base_name": null,
28
  "model_name_or_path": null,
29
  "quant_method": "gptq",
@@ -36,7 +37,7 @@
36
  "rope_theta": 500000.0,
37
  "tie_word_embeddings": false,
38
  "torch_dtype": "float16",
39
- "transformers_version": "4.40.0",
40
  "unsloth_version": "2024.5",
41
  "use_cache": true,
42
  "vocab_size": 128256
 
1
  {
2
+ "_name_or_path": "allganize/Llama-3-Alpha-Ko-8B-Instruct",
3
  "architectures": [
4
  "LlamaForCausalLM"
5
  ],
 
12
  "initializer_range": 0.02,
13
  "intermediate_size": 14336,
14
  "max_position_embeddings": 8192,
15
+ "mlp_bias": false,
16
  "model_type": "llama",
17
  "num_attention_heads": 32,
18
  "num_hidden_layers": 32,
 
21
  "pretraining_tp": 1,
22
  "quantization_config": {
23
  "bits": 4,
 
24
  "damp_percent": 0.01,
25
  "desc_act": false,
26
  "group_size": 128,
27
+ "is_marlin_format": true,
28
  "model_file_base_name": null,
29
  "model_name_or_path": null,
30
  "quant_method": "gptq",
 
37
  "rope_theta": 500000.0,
38
  "tie_word_embeddings": false,
39
  "torch_dtype": "float16",
40
+ "transformers_version": "4.41.0",
41
  "unsloth_version": "2024.5",
42
  "use_cache": true,
43
  "vocab_size": 128256
gptq_model-4bit-128g.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:7499a142bdd73bf148d9668a576714f17ae9c06fe8679776a33fe009de470350
3
- size 5700648184
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e467556a53e8e6781e59d02734976ccebb8961b8040f911fa9d746e9749422ee
3
+ size 5735720552
quantize_config.json CHANGED
@@ -8,7 +8,6 @@
8
  "true_sequential": true,
9
  "model_name_or_path": null,
10
  "model_file_base_name": null,
11
- "quant_method": "gptq",
12
- "checkpoint_format": "marlin",
13
- "is_marlin_format": true
14
  }
 
8
  "true_sequential": true,
9
  "model_name_or_path": null,
10
  "model_file_base_name": null,
11
+ "is_marlin_format": true,
12
+ "quant_method": "gptq"
 
13
  }