|
Quantization made by Richard Erkhov. |
|
|
|
[Github](https://github.com/RichardErkhov) |
|
|
|
[Discord](https://discord.gg/pvy7H8DZMG) |
|
|
|
[Request more models](https://github.com/RichardErkhov/quant_request) |
|
|
|
|
|
TinyLlama-NoPE-HeadScale8k - AWQ |
|
- Model creator: https://huggingface.co/AntNLP/ |
|
- Original model: https://huggingface.co/AntNLP/TinyLlama-NoPE-HeadScale8k/ |
|
|
|
|
|
|
|
|
|
Original model description: |
|
--- |
|
license: mit |
|
--- |
|
|
|
# TinyLlama-NoPE-HeadScale8k |
|
|
|
## Citation |
|
|
|
``` |
|
@misc{wang2024length, |
|
title={Length Generalization of Causal Transformers without Position Encoding}, |
|
author={Jie Wang and Tao Ji and Yuanbin Wu and Hang Yan and Tao Gui and Qi Zhang and Xuanjing Huang and Xiaoling Wang}, |
|
year={2024}, |
|
eprint={2404.12224}, |
|
archivePrefix={arXiv}, |
|
primaryClass={cs.CL} |
|
} |
|
``` |
|
|
|
|