|
--- |
|
|
|
|
|
{} |
|
--- |
|
|
|
# Model Card for Extended-Mind-MPT-7b |
|
|
|
<!-- Provide a quick summary of what the model is/does. --> |
|
|
|
Extended Mind MPT-7b, as described in [Supersizing Transformers](https://blog.normalcomputing.ai/posts/2023-09-12-supersizing-transformers/supersizing-transformers.html). |
|
|
|
### Model Description |
|
|
|
<!-- Provide a longer summary of what this model is. --> |
|
|
|
This model implements active externalism for MPT's 7b model. The model weights have not been edited. Original architecture and code by Mosaic ML. |
|
|
|
For more details on active externalism, check out our [blog](https://blog.normalcomputing.ai/posts/2023-09-12-supersizing-transformers/supersizing-transformers.html)! |
|
|
|
|
|
- **Developed by:** [Normal Computing](https://huggingface.co/normalcomputing), Adapted from [Mosacic ML](https://huggingface.co/mosaicml) |
|
- **License:** Apache 2.0 |
|
|
|
|
|
## Limitations |
|
|
|
<!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. --> |
|
|
|
This model is part of ongoing research at Normal Computing. |
|
|