Update README.md
Browse files
README.md
CHANGED
|
@@ -112,17 +112,18 @@ outputs = model.generate(**inputs, max_new_tokens=100)
|
|
| 112 |
print(tokenizer.decode(outputs[0], skip_special_tokens=True))
|
| 113 |
```
|
| 114 |
|
|
|
|
| 115 |
|
|
|
|
| 116 |
## Citation
|
| 117 |
|
| 118 |
```bibtex
|
| 119 |
-
@article{
|
| 120 |
title={Scaling Latent Reasoning via Looped Language Models},
|
| 121 |
-
author={Zhu, Rui-Jie and Wang, Zixuan and Hua, Kai and Zhang, Tianyu and Li, Ziniu and Que, Haoran and Wei, Boyi and
|
| 122 |
-
journal={arXiv preprint},
|
| 123 |
year={2025}
|
| 124 |
}
|
| 125 |
-
```
|
| 126 |
|
| 127 |
## License
|
| 128 |
|
|
|
|
| 112 |
print(tokenizer.decode(outputs[0], skip_special_tokens=True))
|
| 113 |
```
|
| 114 |
|
| 115 |
+
## Acknowledgments
|
| 116 |
|
| 117 |
+
We thank [@Antizana](https://github.com/Antizana) for the KV cache fix merged from [ouro-cache-fix](https://github.com/Antizana/ouro-cache-fix), which resolved a critical compatibility issue with transformers>=4.56.0.
|
| 118 |
## Citation
|
| 119 |
|
| 120 |
```bibtex
|
| 121 |
+
@article{zhu2025scaling,
|
| 122 |
title={Scaling Latent Reasoning via Looped Language Models},
|
| 123 |
+
author={Zhu, Rui-Jie and Wang, Zixuan and Hua, Kai and Zhang, Tianyu and Li, Ziniu and Que, Haoran and Wei, Boyi and Wen, Zixin and Yin, Fan and Xing, He and others},
|
| 124 |
+
journal={arXiv preprint arXiv:2510.25741},
|
| 125 |
year={2025}
|
| 126 |
}
|
|
|
|
| 127 |
|
| 128 |
## License
|
| 129 |
|