From 0efa58245d2c2298d05534709db80f3a9b5a0659 Mon Sep 17 00:00:00 2001
From: simonJJJ <821898965@qq.com>
Date: Mon, 25 Sep 2023 15:23:58 +0800
Subject: [PATCH] qwen.cpp link
---
README.md | 2 ++
README_CN.md | 2 ++
2 files changed, 4 insertions(+)
diff --git a/README.md b/README.md
index 83d00fc..2a45dbe 100644
--- a/README.md
+++ b/README.md
@@ -467,6 +467,8 @@ model = load_model_on_gpus('Qwen/Qwen-7B-Chat', num_gpus=2)
Then you can run the 7B chat model on 2 GPUs using the above scripts.
+We also provide pure C++ implementation of Qwen-LM and tiktoken, see [qwen.cpp](https://github.com/QwenLM/qwen.cpp) for details.
+
## Tool Usage
Qwen-Chat has been optimized for tool usage and function calling capabilities. Users can develop agents, LangChain applications, and even agument Qwen with a Python Code Interpreter.
diff --git a/README_CN.md b/README_CN.md
index b4ef71e..1ad7f29 100644
--- a/README_CN.md
+++ b/README_CN.md
@@ -456,6 +456,8 @@ model = load_model_on_gpus('Qwen/Qwen-7B-Chat', num_gpus=2)
你即可使用2张GPU进行推理。
+我们同时提供了Qwen-LM和tiktoken的C++实现, 更多细节请查看[qwen.cpp](https://github.com/QwenLM/qwen.cpp).
+
## 工具调用
Qwen-Chat针对工具使用、函数调用能力进行了优化。用户可以开发基于Qwen的Agent、LangChain应用、甚至Code Interpreter。