lmzheng commited on
Commit
8fb9ac0
1 Parent(s): f238a00

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +28 -23
README.md CHANGED
@@ -6,43 +6,48 @@ inference: false
6
  Please check out a newer version of the weights [here](https://huggingface.co/lmsys/vicuna-7b-v1.3).
7
  If you still want to use this old version, please see the compatibility and difference between different versions [here](https://github.com/lm-sys/FastChat/blob/main/docs/vicuna_weights_version.md).
8
 
9
- <br>
10
  <br>
11
 
12
  # Vicuna Model Card
13
 
14
- ## Model details
15
-
16
- **Model type:**
17
- Vicuna is an open-source chatbot trained by fine-tuning LLaMA on user-shared conversations collected from ShareGPT.
18
- It is an auto-regressive language model, based on the transformer architecture.
19
 
20
- **Model date:**
21
- Vicuna was trained between March 2023 and April 2023.
22
 
23
- **Organizations developing the model:**
24
- The Vicuna team with members from UC Berkeley, CMU, Stanford, and UC San Diego.
 
 
25
 
26
- **Paper or resources for more information:**
27
- https://lmsys.org/blog/2023-03-30-vicuna/
28
 
29
- **Where to send questions or comments about the model:**
30
- https://github.com/lm-sys/FastChat/issues
 
 
31
 
 
32
 
33
- ## Intended use
34
- **Primary intended uses:**
35
  The primary use of Vicuna is research on large language models and chatbots.
36
-
37
- **Primary intended users:**
38
  The primary intended users of the model are researchers and hobbyists in natural language processing, machine learning, and artificial intelligence.
39
 
40
- ## Training dataset
41
- 70K conversations collected from ShareGPT.com.
 
 
 
 
 
 
 
 
 
 
 
 
42
 
43
- ## Evaluation dataset
44
- A preliminary evaluation of the model quality is conducted by creating a set of 80 diverse questions and utilizing GPT-4 to judge the model outputs.
45
- See https://lmsys.org/blog/2023-03-30-vicuna/ for more details.
46
 
47
  ## Acknowledgement
48
  Special thanks to [@TheBloke](https://huggingface.co/TheBloke) for hosting this merged version of weights earlier.
 
6
  Please check out a newer version of the weights [here](https://huggingface.co/lmsys/vicuna-7b-v1.3).
7
  If you still want to use this old version, please see the compatibility and difference between different versions [here](https://github.com/lm-sys/FastChat/blob/main/docs/vicuna_weights_version.md).
8
 
 
9
  <br>
10
 
11
  # Vicuna Model Card
12
 
13
+ ## Model Details
 
 
 
 
14
 
15
+ Vicuna is a chat assistant trained by fine-tuning LLaMA on user-shared conversations collected from ShareGPT.
 
16
 
17
+ - **Developed by:** [LMSYS](https://lmsys.org/)
18
+ - **Model type:** An auto-regressive language model based on the transformer architecture.
19
+ - **License:** Non-commercial license
20
+ - **Finetuned from model:** [LLaMA](https://arxiv.org/abs/2302.13971).
21
 
22
+ ### Model Sources
 
23
 
24
+ - **Repository:** https://github.com/lm-sys/FastChat
25
+ - **Blog:** https://lmsys.org/blog/2023-03-30-vicuna/
26
+ - **Paper:** https://arxiv.org/abs/2306.05685
27
+ - **Demo:** https://chat.lmsys.org/
28
 
29
+ ## Uses
30
 
 
 
31
  The primary use of Vicuna is research on large language models and chatbots.
 
 
32
  The primary intended users of the model are researchers and hobbyists in natural language processing, machine learning, and artificial intelligence.
33
 
34
+ ## How to Get Started with the Model
35
+
36
+ Command line interface: https://github.com/lm-sys/FastChat#vicuna-weights.
37
+ APIs (OpenAI API, Huggingface API): https://github.com/lm-sys/FastChat/tree/main#api.
38
+
39
+ ## Training Details
40
+
41
+ Vicuna v1.1 is fine-tuned from LLaMA with supervised instruction fine-tuning.
42
+ The training data is around 70K conversations collected from ShareGPT.com.
43
+ See more details in the "Training Details of Vicuna Models" section in the appendix of this [paper](https://arxiv.org/pdf/2306.05685.pdf).
44
+
45
+ ## Evaluation
46
+
47
+ Vicuna is evaluated with standard benchmarks, human preference, and LLM-as-a-judge. See more details in this [paper](https://arxiv.org/pdf/2306.05685.pdf) and [leaderboard](https://huggingface.co/spaces/lmsys/chatbot-arena-leaderboard).
48
 
49
+ ## Difference between different versions of Vicuna
50
+ See [vicuna_weights_version.md](https://github.com/lm-sys/FastChat/blob/main/docs/vicuna_weights_version.md)
 
51
 
52
  ## Acknowledgement
53
  Special thanks to [@TheBloke](https://huggingface.co/TheBloke) for hosting this merged version of weights earlier.