TheBloke commited on
Commit
d99f20a
1 Parent(s): e61fe8f

Upload README.md

Browse files
Files changed (1) hide show
  1. README.md +190 -0
README.md ADDED
@@ -0,0 +1,190 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ base_model: nRuaif/Rose-Kimiko-20B
3
+ inference: false
4
+ library_name: peft
5
+ license: llama2
6
+ model-index:
7
+ - name: Rose-Kimiko
8
+ results: []
9
+ model_creator: nRuaif
10
+ model_name: Rose Kimiko 20B
11
+ model_type: llama
12
+ prompt_template: 'Below is an instruction that describes a task. Write a response
13
+ that appropriately completes the request.
14
+
15
+
16
+ ### Instruction:
17
+
18
+ {prompt}
19
+
20
+
21
+ ### Response:
22
+
23
+ '
24
+ quantized_by: TheBloke
25
+ tags:
26
+ - generated_from_trainer
27
+ ---
28
+
29
+ <!-- header start -->
30
+ <!-- 200823 -->
31
+ <div style="width: auto; margin-left: auto; margin-right: auto">
32
+ <img src="https://i.imgur.com/EBdldam.jpg" alt="TheBlokeAI" style="width: 100%; min-width: 400px; display: block; margin: auto;">
33
+ </div>
34
+ <div style="display: flex; justify-content: space-between; width: 100%;">
35
+ <div style="display: flex; flex-direction: column; align-items: flex-start;">
36
+ <p style="margin-top: 0.5em; margin-bottom: 0em;"><a href="https://discord.gg/theblokeai">Chat & support: TheBloke's Discord server</a></p>
37
+ </div>
38
+ <div style="display: flex; flex-direction: column; align-items: flex-end;">
39
+ <p style="margin-top: 0.5em; margin-bottom: 0em;"><a href="https://www.patreon.com/TheBlokeAI">Want to contribute? TheBloke's Patreon page</a></p>
40
+ </div>
41
+ </div>
42
+ <div style="text-align:center; margin-top: 0em; margin-bottom: 0em"><p style="margin-top: 0.25em; margin-bottom: 0em;">TheBloke's LLM work is generously supported by a grant from <a href="https://a16z.com">andreessen horowitz (a16z)</a></p></div>
43
+ <hr style="margin-top: 1.0em; margin-bottom: 1.0em;">
44
+ <!-- header end -->
45
+
46
+ # Rose Kimiko 20B - FP16
47
+ - Model creator: [nRuaif](https://huggingface.co/nRuaif)
48
+ - Original model: [Rose Kimiko 20B](nRuaif/Rose-Kimiko-20B)
49
+
50
+ <!-- description start -->
51
+ ## Description
52
+
53
+ This repo contains pytorch format fp16 model files for [nRuaif's Rose Kimiko 20B](nRuaif/Rose-Kimiko-20B).
54
+
55
+ It is the result of either merging a LoRA, or converting the source repository to float16.
56
+
57
+ These files were quantised using hardware kindly provided by [Massed Compute](https://massedcompute.com/).
58
+
59
+ <!-- description end -->
60
+ <!-- repositories-available start -->
61
+ ## Repositories available
62
+
63
+ * [AWQ model(s) for GPU inference.](https://huggingface.co/TheBloke/Rose-Kimiko-20B-AWQ)
64
+ * [GPTQ models for GPU inference, with multiple quantisation parameter options.](https://huggingface.co/TheBloke/Rose-Kimiko-20B-GPTQ)
65
+ * [2, 3, 4, 5, 6 and 8-bit GGUF models for CPU+GPU inference](https://huggingface.co/TheBloke/Rose-Kimiko-20B-GGUF)
66
+ * [Unquantised fp16 model in pytorch format, for GPU inference and for further conversions](https://huggingface.co/TheBloke/Rose-Kimiko-20B-fp16)
67
+ * [nRuaif's original LoRA adapter, which can be merged on to the base model.](https://huggingface.co/nRuaif/Rose-Kimiko-20B)
68
+
69
+ <!-- repositories-available start -->
70
+
71
+ <!-- prompt-template start -->
72
+ ## Prompt template: Alpaca
73
+
74
+ ```
75
+ Below is an instruction that describes a task. Write a response that appropriately completes the request.
76
+
77
+ ### Instruction:
78
+ {prompt}
79
+
80
+ ### Response:
81
+
82
+ ```
83
+
84
+ <!-- prompt-template end -->
85
+
86
+
87
+
88
+
89
+ <!-- footer start -->
90
+ <!-- 200823 -->
91
+ ## Discord
92
+
93
+ For further support, and discussions on these models and AI in general, join us at:
94
+
95
+ [TheBloke AI's Discord server](https://discord.gg/theblokeai)
96
+
97
+ ## Thanks, and how to contribute
98
+
99
+ Thanks to the [chirper.ai](https://chirper.ai) team!
100
+
101
+ Thanks to Clay from [gpus.llm-utils.org](llm-utils)!
102
+
103
+ I've had a lot of people ask if they can contribute. I enjoy providing models and helping people, and would love to be able to spend even more time doing it, as well as expanding into new projects like fine tuning/training.
104
+
105
+ If you're able and willing to contribute it will be most gratefully received and will help me to keep providing more models, and to start work on new AI projects.
106
+
107
+ Donaters will get priority support on any and all AI/LLM/model questions and requests, access to a private Discord room, plus other benefits.
108
+
109
+ * Patreon: https://patreon.com/TheBlokeAI
110
+ * Ko-Fi: https://ko-fi.com/TheBlokeAI
111
+
112
+ **Special thanks to**: Aemon Algiz.
113
+
114
+ **Patreon special mentions**: Michael Levine, 阿明, Trailburnt, Nikolai Manek, John Detwiler, Randy H, Will Dee, Sebastain Graf, NimbleBox.ai, Eugene Pentland, Emad Mostaque, Ai Maven, Jim Angel, Jeff Scroggin, Michael Davis, Manuel Alberto Morcote, Stephen Murray, Robert, Justin Joy, Luke @flexchar, Brandon Frisco, Elijah Stavena, S_X, Dan Guido, Undi ., Komninos Chatzipapas, Shadi, theTransient, Lone Striker, Raven Klaugh, jjj, Cap'n Zoog, Michel-Marie MAUDET (LINAGORA), Matthew Berman, David, Fen Risland, Omer Bin Jawed, Luke Pendergrass, Kalila, OG, Erik Bjäreholt, Rooh Singh, Joseph William Delisle, Dan Lewis, TL, John Villwock, AzureBlack, Brad, Pedro Madruga, Caitlyn Gatomon, K, jinyuan sun, Mano Prime, Alex, Jeffrey Morgan, Alicia Loh, Illia Dulskyi, Chadd, transmissions 11, fincy, Rainer Wilmers, ReadyPlayerEmma, knownsqashed, Mandus, biorpg, Deo Leter, Brandon Phillips, SuperWojo, Sean Connelly, Iucharbius, Jack West, Harry Royden McLaughlin, Nicholas, terasurfer, Vitor Caleffi, Duane Dunston, Johann-Peter Hartmann, David Ziegler, Olakabola, Ken Nordquist, Trenton Dambrowitz, Tom X Nguyen, Vadim, Ajan Kanaga, Leonard Tan, Clay Pascal, Alexandros Triantafyllidis, JM33133, Xule, vamX, ya boyyy, subjectnull, Talal Aujan, Alps Aficionado, wassieverse, Ari Malik, James Bentley, Woland, Spencer Kim, Michael Dempsey, Fred von Graf, Elle, zynix, William Richards, Stanislav Ovsiannikov, Edmond Seymore, Jonathan Leane, Martin Kemka, usrbinkat, Enrico Ros
115
+
116
+
117
+ Thank you to all my generous patrons and donaters!
118
+
119
+ And thank you again to a16z for their generous grant.
120
+
121
+ <!-- footer end -->
122
+
123
+ # Original model card: nRuaif's Rose Kimiko 20B
124
+
125
+
126
+ <!-- This model card has been generated automatically according to the information the Trainer had access to. You
127
+ should probably proofread and complete it, then remove this comment. -->
128
+
129
+ [<img src="https://raw.githubusercontent.com/OpenAccess-AI-Collective/axolotl/main/image/axolotl-badge-web.png" alt="Built with Axolotl" width="200" height="32"/>](https://github.com/OpenAccess-AI-Collective/axolotl)
130
+ # qlora-out
131
+
132
+ This model is a fine-tuned version of [tavtav/Rose-20B](https://huggingface.co/tavtav/Rose-20B) on the Kimiko dataset.
133
+
134
+ ## Model description
135
+
136
+ The prompt formats used is ShareGPT/Vicuna format.
137
+
138
+ ## Intended uses & limitations
139
+
140
+ Per many people requests, this LoRA is intended to fix spelling from Rose 20B.
141
+
142
+ ## Training and evaluation data
143
+
144
+ More information needed
145
+
146
+ ## Training procedure
147
+
148
+ ### Training hyperparameters
149
+
150
+ The following hyperparameters were used during training:
151
+ - learning_rate: 0.0002
152
+ - train_batch_size: 2
153
+ - eval_batch_size: 2
154
+ - seed: 42
155
+ - gradient_accumulation_steps: 4
156
+ - total_train_batch_size: 8
157
+ - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
158
+ - lr_scheduler_type: cosine
159
+ - lr_scheduler_warmup_steps: 10
160
+ - num_epochs: 2
161
+
162
+ ### Training results
163
+
164
+
165
+
166
+ ### Framework versions
167
+
168
+ - Transformers 4.36.0.dev0
169
+ - Pytorch 2.0.1+cu118
170
+ - Datasets 2.15.0
171
+ - Tokenizers 0.15.0
172
+ ## Training procedure
173
+
174
+
175
+ The following `bitsandbytes` quantization config was used during training:
176
+ - quant_method: bitsandbytes
177
+ - load_in_8bit: False
178
+ - load_in_4bit: True
179
+ - llm_int8_threshold: 6.0
180
+ - llm_int8_skip_modules: None
181
+ - llm_int8_enable_fp32_cpu_offload: False
182
+ - llm_int8_has_fp16_weight: False
183
+ - bnb_4bit_quant_type: nf4
184
+ - bnb_4bit_use_double_quant: True
185
+ - bnb_4bit_compute_dtype: bfloat16
186
+
187
+ ### Framework versions
188
+
189
+
190
+ - PEFT 0.6.0