alpayariyak
commited on
Commit
•
0fc98e3
1
Parent(s):
c8ac815
Update README.md
Browse files
README.md
CHANGED
@@ -46,6 +46,7 @@ OpenChat is an innovative library of open-source language models, fine-tuned wit
|
|
46 |
|
47 |
[![DOI](https://zenodo.org/badge/645397533.svg)](https://zenodo.org/badge/latestdoi/645397533)
|
48 |
|
|
|
49 |
## Usage
|
50 |
|
51 |
To use this model, we highly recommend installing the OpenChat package by following the [installation guide](https://github.com/imoneoi/openchat#installation) in our repository and using the OpenChat OpenAI-compatible API server by running the serving command from the table below. The server is optimized for high-throughput deployment using [vLLM](https://github.com/vllm-project/vllm) and can run on a consumer GPU with 24GB RAM. To enable tensor parallelism, append `--tensor-parallel-size N` to the serving command.
|
|
|
46 |
|
47 |
[![DOI](https://zenodo.org/badge/645397533.svg)](https://zenodo.org/badge/latestdoi/645397533)
|
48 |
|
49 |
+
|
50 |
## Usage
|
51 |
|
52 |
To use this model, we highly recommend installing the OpenChat package by following the [installation guide](https://github.com/imoneoi/openchat#installation) in our repository and using the OpenChat OpenAI-compatible API server by running the serving command from the table below. The server is optimized for high-throughput deployment using [vLLM](https://github.com/vllm-project/vllm) and can run on a consumer GPU with 24GB RAM. To enable tensor parallelism, append `--tensor-parallel-size N` to the serving command.
|