Skip to content

Commit 561f646

Browse files
authored
misc: add bibtex reference (#712)
This pull request includes an update to the `README.md` file to add a new citation section. The most important change is the addition of a citation format for users who find FlashInfer helpful in their projects or research. Documentation update: * [`README.md`](diffhunk://#diff-b335630551682c19a781afebcf4d07bf978fb1f8ac04c6bf87428ed5106870f5R144-R169): Added a new "Citation" section with a BibTeX entry for citing the FlashInfer paper.
1 parent 644ccd7 commit 561f646

File tree

1 file changed

+26
-0
lines changed

1 file changed

+26
-0
lines changed

README.md

+26
Original file line numberDiff line numberDiff line change
@@ -141,3 +141,29 @@ We are thrilled to share that FlashInfer is being adopted by many cutting-edge p
141141
## Acknowledgement
142142

143143
FlashInfer is inspired by [FlashAttention 1&2](https://github.com/dao-AILab/flash-attention/), [vLLM](https://github.com/vllm-project/vllm), [stream-K](https://arxiv.org/abs/2301.03598), [cutlass](https://github.com/nvidia/cutlass) and [AITemplate](https://github.com/facebookincubator/AITemplate) projects.
144+
145+
## Citation
146+
147+
If you find FlashInfer helpful in your project or research, please consider citing our [paper](https://arxiv.org/abs/2501.01005):
148+
149+
```bibtex
150+
@article{ye2025flashinfer,
151+
title = {FlashInfer: Efficient and Customizable Attention Engine for LLM Inference Serving},
152+
author = {
153+
Ye, Zihao and
154+
Chen, Lequn and
155+
Lai, Ruihang and
156+
Lin, Wuwei and
157+
Zhang, Yineng and
158+
Wang, Stephanie and
159+
Chen, Tianqi and
160+
Kasikci, Baris and
161+
Grover, Vinod and
162+
Krishnamurthy, Arvind and
163+
Ceze, Luis
164+
},
165+
journal = {arXiv preprint arXiv:2501.01005},
166+
year = {2025},
167+
url = {https://arxiv.org/abs/2501.01005}
168+
}
169+
```

0 commit comments

Comments
 (0)