jieunhan commited on
Commit
e55184c
1 Parent(s): c5e5d0b

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +1 -61
README.md CHANGED
@@ -1,67 +1,7 @@
1
  ---
2
  license: apache-2.0
3
- tags:
4
- - moe
5
- - frankenmoe
6
- - merge
7
- - mergekit
8
- - lazymergekit
9
- - yanolja/EEVE-Korean-10.8B-v1.0
10
- - upstage/SOLAR-10.7B-v1.0
11
  base_model:
12
  - yanolja/EEVE-Korean-10.8B-v1.0
13
  - upstage/SOLAR-10.7B-v1.0
14
  ---
15
-
16
- # TEST_MODEL
17
-
18
- TEST_MODEL is a Mixture of Experts (MoE) made with the following models using [LazyMergekit](https://colab.research.google.com/drive/1obulZ1ROXHjYLn6PPZJwRR6GzgQogxxb?usp=sharing):
19
- * [yanolja/EEVE-Korean-10.8B-v1.0](https://huggingface.co/yanolja/EEVE-Korean-10.8B-v1.0)
20
- * [upstage/SOLAR-10.7B-v1.0](https://huggingface.co/upstage/SOLAR-10.7B-v1.0)
21
-
22
- ## 🧩 Configuration
23
-
24
- ```yaml
25
- base_model: yanolja/EEVE-Korean-10.8B-v1.0
26
- dtype: float16
27
- experts:
28
- - source_model: yanolja/EEVE-Korean-10.8B-v1.0
29
- positive_prompts: ["You are an helpful general-pupose assistant."]
30
- - source_model: upstage/SOLAR-10.7B-v1.0
31
- positive_prompts: ["You are helpful assistant."]
32
- merge_method: slerp
33
- gate_mode: cheap_embed
34
- parameters:
35
- t:
36
- - filter: self_attn
37
- value: [0, 0.5, 0.3, 0.7, 1]
38
- - filter: mlp
39
- value: [1, 0.5, 0.7, 0.3, 0]
40
- - value: 0.5
41
-
42
- tokenizer_source: base
43
- ```
44
-
45
- ## 💻 Usage
46
-
47
- ```python
48
- !pip install -qU transformers bitsandbytes accelerate
49
-
50
- from transformers import AutoTokenizer
51
- import transformers
52
- import torch
53
-
54
- model = "jieunhan/TEST_MODEL"
55
-
56
- tokenizer = AutoTokenizer.from_pretrained(model)
57
- pipeline = transformers.pipeline(
58
- "text-generation",
59
- model=model,
60
- model_kwargs={"torch_dtype": torch.float16, "load_in_4bit": True},
61
- )
62
-
63
- messages = [{"role": "user", "content": "Explain what a Mixture of Experts is in less than 100 words."}]
64
- prompt = pipeline.tokenizer.apply_chat_template(messages, tokenize=False, add_generation_prompt=True)
65
- outputs = pipeline(prompt, max_new_tokens=256, do_sample=True, temperature=0.7, top_k=50, top_p=0.95)
66
- print(outputs[0]["generated_text"])
67
- ```
 
1
  ---
2
  license: apache-2.0
 
 
 
 
 
 
 
 
3
  base_model:
4
  - yanolja/EEVE-Korean-10.8B-v1.0
5
  - upstage/SOLAR-10.7B-v1.0
6
  ---
7
+ TESTING