Update README.md
Browse files
README.md
CHANGED
@@ -84,7 +84,7 @@ Or you can install vLLM from [source](https://github.com/vllm-project/vllm/).
|
|
84 |
3. **Model Deployment**: Utilize vLLM to deploy your model. For instance, you can set up an openAI-like server using the command:
|
85 |
|
86 |
```bash
|
87 |
-
python -m vllm.entrypoints.openai.api_server --served-model-name
|
88 |
```
|
89 |
|
90 |
Then you can access the Chat API by:
|
@@ -101,9 +101,6 @@ Or you can install vLLM from [source](https://github.com/vllm-project/vllm/).
|
|
101 |
}'
|
102 |
```
|
103 |
|
104 |
-
For further usage instructions of vLLM, please refer to our [Github](https://github.com/QwenLM/Qwen2).
|
105 |
-
|
106 |
-
|
107 |
### Applications
|
108 |
This fine-tuned model is particularly suited for [mention applications, e.g., chatbots, question-answering systems, etc.]. Its enhanced capabilities ensure more accurate and contextually appropriate responses in these domains.
|
109 |
|
|
|
84 |
3. **Model Deployment**: Utilize vLLM to deploy your model. For instance, you can set up an openAI-like server using the command:
|
85 |
|
86 |
```bash
|
87 |
+
python -m vllm.entrypoints.openai.api_server --served-model-name CarrotAI/Rabbit-Ko-15B-Instruct --model path/to/weights
|
88 |
```
|
89 |
|
90 |
Then you can access the Chat API by:
|
|
|
101 |
}'
|
102 |
```
|
103 |
|
|
|
|
|
|
|
104 |
### Applications
|
105 |
This fine-tuned model is particularly suited for [mention applications, e.g., chatbots, question-answering systems, etc.]. Its enhanced capabilities ensure more accurate and contextually appropriate responses in these domains.
|
106 |
|