hassanzay commited on
Commit
a14f62d
1 Parent(s): 9f2d4e9

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +5 -2
README.md CHANGED
@@ -122,6 +122,8 @@ model-index:
122
  ---
123
  # `StableLM Zephyr 3B`
124
 
 
 
125
  ## Model Description
126
 
127
  `StableLM Zephyr 3B` is a 3 billion parameter instruction tuned inspired by [HugginFaceH4's Zephyr 7B](https://huggingface.co/HuggingFaceH4/zephyr-7b-beta) training pipeline this model was trained on a mix of publicly available datasets, synthetic datasets using [Direct Preference Optimization (DPO)](https://arxiv.org/abs/2305.18290), evaluation for this model based on
@@ -176,7 +178,8 @@ You can also see how to run a performance optimized version of this model [here]
176
  * **Language(s)**: English
177
  * **Library**: [Alignment Handbook](https://github.com/huggingface/alignment-handbook.git)
178
  * **Finetuned from model**: [stabilityai/stablelm-3b-4e1t](https://huggingface.co/stabilityai/stablelm-3b-4e1t)
179
- * **License**: [StabilityAI Non-Commercial Research Community License](https://huggingface.co/stabilityai/stablelm-zephyr-3b/raw/main/LICENSE). If you want to use this model for your commercial products or purposes, please contact us [here](https://stability.ai/contact) to learn more.
 
180
  * **Contact**: For questions and comments about the model, please email `[email protected]`
181
 
182
  ### Training Dataset
@@ -247,7 +250,7 @@ We have incorporated the findings of our malicious prompts evaluation and red te
247
 
248
  ### Intended Use
249
 
250
- The model is intended to be used as a foundational base model for application-specific fine-tuning. Developers must evaluate and fine-tune the model for safe performance in downstream applications.
251
 
252
  ### Limitations and Bias
253
 
 
122
  ---
123
  # `StableLM Zephyr 3B`
124
 
125
+ Please note: For commercial use, please refer to https://stability.ai/membership.
126
+
127
  ## Model Description
128
 
129
  `StableLM Zephyr 3B` is a 3 billion parameter instruction tuned inspired by [HugginFaceH4's Zephyr 7B](https://huggingface.co/HuggingFaceH4/zephyr-7b-beta) training pipeline this model was trained on a mix of publicly available datasets, synthetic datasets using [Direct Preference Optimization (DPO)](https://arxiv.org/abs/2305.18290), evaluation for this model based on
 
178
  * **Language(s)**: English
179
  * **Library**: [Alignment Handbook](https://github.com/huggingface/alignment-handbook.git)
180
  * **Finetuned from model**: [stabilityai/stablelm-3b-4e1t](https://huggingface.co/stabilityai/stablelm-3b-4e1t)
181
+ * **License**: [StabilityAI Non-Commercial Research Community License](https://huggingface.co/stabilityai/stablelm-zephyr-3b/raw/main/LICENSE).
182
+ * **Commercial License**: to use this model commercially, please refer to https://stability.ai/membership
183
  * **Contact**: For questions and comments about the model, please email `[email protected]`
184
 
185
  ### Training Dataset
 
250
 
251
  ### Intended Use
252
 
253
+ The model is intended to be used as a foundational base model for application-specific fine-tuning. Developers must evaluate and fine-tune the model for safe performance in downstream applications. For commercial use, please refer to https://stability.ai/membership.
254
 
255
  ### Limitations and Bias
256