Skip to content

Commit

Permalink
Merge pull request THUDM#925 from zRzRzRzRzRzRzR/main
Browse files Browse the repository at this point in the history
fix OpenVINO Readme Problem
  • Loading branch information
zRzRzRzRzRzRzR authored Mar 5, 2024
2 parents 144a3e4 + 095eabe commit f4dfe41
Show file tree
Hide file tree
Showing 2 changed files with 2 additions and 2 deletions.
2 changes: 1 addition & 1 deletion README.md
Original file line number Diff line number Diff line change
Expand Up @@ -82,7 +82,7 @@ Huggingface 更新后一段时间内同步更新。
* [TensorRT-LLM](https://github.com/NVIDIA/TensorRT-LLM/tree/main):
NVIDIA开发的高性能 GPU 加速推理方案,可以参考此 [步骤](./tensorrt_llm_demo/README.md) 部署 ChatGLM3-6B 模型
* [OpenVINO](https://github.com/OpenVINO-dev-contest/chatglm3.openvino):
Intel 开发的高性能 CPU 和 GPU 加速推理方案,可以参考此 [步骤](./openvino_demo/README.md) 部署 ChatGLM3-6B 模型
Intel 开发的高性能 CPU 和 GPU 加速推理方案,可以参考此 [步骤](./Intel_device_demo/openvino_demo/README.md) 部署 ChatGLM3-6B 模型

高效微调:

Expand Down
2 changes: 1 addition & 1 deletion README_en.md
Original file line number Diff line number Diff line change
Expand Up @@ -63,7 +63,7 @@ Inference acceleration:
* [chatglm.cpp](https://github.com/li-plus/chatglm.cpp): Real-time inference on your laptop accelerated by quantization, similar to llama.cpp.
* [ChatGLM3-TPU](https://github.com/sophgo/ChatGLM3-TPU): Using the TPU accelerated inference solution, it runs about 7.5 token/s in real time on the end-side chip BM1684X (16T@FP16, 16G DDR).
* [TensorRT-LLM](https://github.com/NVIDIA/TensorRT-LLM/tree/main): A high-performance GPU-accelerated inference solution developed by NVIDIA, you can refer to these [steps](./tensorrt_llm_demo/README.md) to deploy ChatGLM3.
* [OpenVINO](https://github.com/OpenVINO-dev-contest/chatglm3.openvino): A high-performance CPU and GPU accelerated inference solution developed by Intel, you can refer to this [step](./openvino_demo/README.md) to deploy the ChatGLM3-6B model
* [OpenVINO](https://github.com/OpenVINO-dev-contest/chatglm3.openvino): A high-performance CPU and GPU accelerated inference solution developed by Intel, you can refer to this [step](./Intel_device_demo/openvino_demo/README.md) to deploy the ChatGLM3-6B model

Efficient fine-tuning:
* [LLaMA-Factory](https://github.com/hiyouga/LLaMA-Factory): An excellent, easy-to-use and efficient fine-tuning framework.
Expand Down

0 comments on commit f4dfe41

Please sign in to comment.