Update README.md
Browse files
README.md
CHANGED
@@ -28,7 +28,7 @@ The name `nekomata` comes from the Japanese word [`猫又/ねこまた/Nekomata`
|
|
28 |
|
29 |
* **Library**
|
30 |
|
31 |
-
The model was trained using code based on [
|
32 |
|
33 |
* **Model architecture**
|
34 |
|
@@ -126,19 +126,5 @@ We compared the `Qwen` tokenizer (as used in `nekomata`) and the `llama-2` token
|
|
126 |
~~~
|
127 |
---
|
128 |
|
129 |
-
# Citations
|
130 |
-
~~~
|
131 |
-
@software{gpt-neox-library,
|
132 |
-
title = {{GPT-NeoX: Large Scale Autoregressive Language Modeling in PyTorch}},
|
133 |
-
author = {Andonian, Alex and Anthony, Quentin and Biderman, Stella and Black, Sid and Gali, Preetham and Gao, Leo and Hallahan, Eric and Levy-Kramer, Josh and Leahy, Connor and Nestler, Lucas and Parker, Kip and Pieler, Michael and Purohit, Shivanshu and Songz, Tri and Phil, Wang and Weinbach, Samuel},
|
134 |
-
url = {https://www.github.com/eleutherai/gpt-neox},
|
135 |
-
doi = {10.5281/zenodo.5879544},
|
136 |
-
month = {8},
|
137 |
-
year = {2021},
|
138 |
-
version = {0.0.1},
|
139 |
-
}
|
140 |
-
~~~
|
141 |
-
---
|
142 |
-
|
143 |
# License
|
144 |
[Tongyi Qianwen LICENSE AGREEMENT](https://github.com/QwenLM/Qwen/blob/main/Tongyi%20Qianwen%20LICENSE%20AGREEMENT)
|
|
|
28 |
|
29 |
* **Library**
|
30 |
|
31 |
+
The model was trained using code based on [aws-neuron/neuronx-nemo-megatron](https://github.com/aws-neuron/neuronx-nemo-megatron/).
|
32 |
|
33 |
* **Model architecture**
|
34 |
|
|
|
126 |
~~~
|
127 |
---
|
128 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
129 |
# License
|
130 |
[Tongyi Qianwen LICENSE AGREEMENT](https://github.com/QwenLM/Qwen/blob/main/Tongyi%20Qianwen%20LICENSE%20AGREEMENT)
|