Skip to content

Commit 44d4265

Browse files
authored
Merge pull request #1521 from kvcache-ai/Atream-patch-8
Add citation section to README
2 parents d38bcc8 + 1e48eab commit 44d4265

File tree

1 file changed

+13
-0
lines changed

1 file changed

+13
-0
lines changed

README.md

Lines changed: 13 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -185,6 +185,19 @@ You can find example rule templates for optimizing DeepSeek-V2 and Qwen2-57B-A14
185185

186186
If you are interested in our design principles and the implementation of the injection framework, please refer to the [design document](doc/en/deepseek-v2-injection.md).
187187

188+
## Citation
189+
190+
If you use KTransformers for your research, please cite our [paper](https://madsys.cs.tsinghua.edu.cn/publication/ktransformers-unleashing-the-full-potential-of-cpu/gpu-hybrid-inference-for-moe-models/):
191+
192+
```
193+
@inproceedings{10.1145/3731569.3764843,
194+
title = {KTransformers: Unleashing the Full Potential of CPU/GPU Hybrid Inference for MoE Models},
195+
author = {Chen, Hongtao and Xie, Weiyu and Zhang, Boxin and Tang, Jingqi and Wang, Jiahao and Dong, Jianwei and Chen, Shaoyuan and Yuan, Ziwei and Lin, Chen and Qiu, Chengyu and Zhu, Yuening and Ou, Qingliang and Liao, Jiaqi and Chen, Xianglin and Ai, Zhiyuan and Wu, Yongwei and Zhang, Mingxing},
196+
booktitle = {Proceedings of the ACM SIGOPS 31st Symposium on Operating Systems Principles},
197+
year = {2025}
198+
}
199+
```
200+
188201
<h2 id="ack">Acknowledgment and Contributors</h2>
189202

190203
The development of KTransformers is based on the flexible and versatile framework provided by Transformers. We also benefit from advanced kernels such as GGUF/GGML, Llamafile, Marlin, sglang and flashinfer. We are planning to contribute back to the community by upstreaming our modifications.

0 commit comments

Comments
 (0)