feihu.hf commited on
Commit
22d9ef9
1 Parent(s): a66b26e

update README.md

Browse files
Files changed (1) hide show
  1. README.md +1 -1
README.md CHANGED
@@ -17,7 +17,7 @@ Compared with the state-of-the-art opensource language models, including the pre
17
 
18
  Qwen2-7B-Instruct-GPTQ-Int8 supports a context length of up to 131,072 tokens, enabling the processing of extensive inputs. Please refer to [this section](#processing-long-texts) for detailed instructions on how to deploy Qwen2 for handling long texts.
19
 
20
- For more details, please refer to our [blog](https://qwenlm.github.io/blog/qwen2/) and [GitHub](https://github.com/QwenLM/Qwen2).
21
 
22
  **Note**: If you encounter ``RuntimeError: probability tensor contains either `inf`, `nan` or element < 0`` during inference with ``transformer``, we recommand [deploying this model with vLLM](https://qwen.readthedocs.io/en/latest/deployment/vllm.html).
23
  <br>
 
17
 
18
  Qwen2-7B-Instruct-GPTQ-Int8 supports a context length of up to 131,072 tokens, enabling the processing of extensive inputs. Please refer to [this section](#processing-long-texts) for detailed instructions on how to deploy Qwen2 for handling long texts.
19
 
20
+ For more details, please refer to our [blog](https://qwenlm.github.io/blog/qwen2/), [GitHub](https://github.com/QwenLM/Qwen2), and [Documentation](https://qwen.readthedocs.io/en/latest/).
21
 
22
  **Note**: If you encounter ``RuntimeError: probability tensor contains either `inf`, `nan` or element < 0`` during inference with ``transformer``, we recommand [deploying this model with vLLM](https://qwen.readthedocs.io/en/latest/deployment/vllm.html).
23
  <br>