SushantGautam commited on
Commit
f134ad1
1 Parent(s): 7048b2a

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +1 -1
README.md CHANGED
@@ -111,7 +111,7 @@ If you try the vLLM examples below and get an error about `quantization` being u
111
  - When using vLLM as a server, pass the `--quantization awq` parameter, for example:
112
 
113
  ```shell
114
- python3 python -m vllm.entrypoints.api_server --model TheBloke/Mistral-7B-OpenOrca-AWQ --quantization awq --dtype half
115
  ```
116
 
117
  When using vLLM from Python code, pass the `quantization=awq` parameter, for example:
 
111
  - When using vLLM as a server, pass the `--quantization awq` parameter, for example:
112
 
113
  ```shell
114
+ python3 -m vllm.entrypoints.api_server --model TheBloke/Mistral-7B-OpenOrca-AWQ --quantization awq --dtype half
115
  ```
116
 
117
  When using vLLM from Python code, pass the `quantization=awq` parameter, for example: