File size: 3,555 Bytes
40927c8
 
 
 
e0110ac
 
7c210e5
 
d8fd1db
e0110ac
215ecb7
c4eade6
 
 
 
 
a033654
d8fd1db
002a533
 
e0110ac
0aa06b3
215ecb7
d8ed2fe
 
 
40927c8
54384f9
 
d8fd1db
002a533
 
54384f9
0aa06b3
215ecb7
693d098
 
 
 
54384f9
 
4347362
d8fd1db
4347362
 
 
 
 
9a453cf
4347362
 
d8fd1db
215ecb7
 
 
 
 
5d10c99
ec897e6
 
d8fd1db
ec897e6
d8fd1db
ec897e6
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
---
license: apache-2.0
---

# Model Card for MediaTek Research Breeze-7B-FC-v1_0



## πŸ† Performance

| Models                                                                                     | #Parameters | Organization | License    | Function Calling? | Instrustion Following? |
|--------------------------------------------------------------------------------------------|-------------|------------|------------|-------------------|----------|
| [Breeze-7B-Instruct-v1_0](https://huggingface.co/MediaTek-Research/Breeze-7B-Instruct-v1_0)| 7B          | MediaTek Research | Apache 2.0 | No               | Yes       |
| [**Breeze-7B-FC-v1_0**](https://huggingface.co/MediaTek-Research/Breeze-7B-FC-v1_0)        | 7B          | MediaTek Research | Apache 2.0 | Yes               | Yes      |
| [Gorilla-OpenFunctions-v2](https://huggingface.co/MediaTek-Research/Breeze-7B-FC-v1_0)     | 7B          | Gorilla LLM | Apache 2.0 | Yes               | No       |
| [GPT-3.5-Turbo-0125](https://openai.com)                                                   |             | OpenAI | Proprietary| Yes               | Yes      |

**Evaluate function calling on EN benchmark**

Berkeley function-calling leaderboard

| Models                            | ↑ Overall | Irrelevance<br/>Detection | AST/<br/>Simple | AST/<br/>Multiple | AST/<br/>Parallel | AST/<br/>Parallel-Multiple  | Exec/<br/>Simple | Exec/<br/>Multiple | Exec/<br/>Parallel | Exec/<br/>Parallel-Multiple  | 
|-----------------------------------|----------|---------------------|------------|--------------|--------------|------------------------|--------------|---------------------|---------------------|-------------------------------|
| **Breeze-7B-FC-v1_0 (FC)**        | 86.01 |  74.58 | 90.00 | 93.00 | 82.00 | 83.00 | 98.00 | 92.00 | 88.00 | 75.00 |
| Gorilla-OpenFunctions-v2 (FC)     | 85.95 |  60.00 | 94.25 | 95.50 | 86.50 | 86.00 | 97.00 | 96.00 | 80.00 | 75.00 |
| GPT-3.5-Turbo-0125 (FC)           | 72.77 |  4.58  | 87.75 | 90.50 | 88.50 | 82.50 | 91.00 | 82.00 | 78.00 | 52.50 |

![](misc/radar_chart_en.png)

**Evaluate function calling on ZHTW benchmark**

function-calling-leaderboard-for-zhtw

| Models                            | ↑ Overall | Irrelevance<br/>Detection | AST/<br/>Simple | AST/<br/>Multiple | AST/<br/>Parallel | AST/<br/>Parallel-Multiple  | Exec/<br/>Simple | Exec/<br/>Multiple | Exec/<br/>Parallel | Exec/<br/>Parallel-Multiple  | 
|-----------------------------------|----------|---------------------|------------|--------------|--------------|------------------------|--------------|---------------------|---------------------|-------------------------------|
| **Breeze-7B-FC-v1_0 (FC)**        | 77.70 |  71.67 | 82.00 |	86.50 |	76.00 |	65.50 |	87.00 |	88.00 |	80.00 |	57.50 |
| Gorilla-OpenFunctions-v2 (FC)     | 75.68 |  53.75 | 84.75 |	86.50 |	72.50 |	68.00 |	92.00 |	92.00 |	62.00 |	72.50 |
| GPT-3.5-Turbo-0125 (FC)           | 66.15 |  7.50  | 83.75 |	83.50 |	73.00 |	65.50 |	88.00 |	84.00 |	72.00 |	40.00 |

![](misc/radar_chart_zhtw.png)


 **Evaluate instrustion following on EN benchmark**

MT-Bench

| | Win | Tie | Lose |
|---|---|---|---|
| **Breeze-7B-FC-v1_0** *v.s.* Breeze-7B-Instruct-v1_0 | 25 (15.6%) | 72 (45.0%) | 63 (39.4%) |


**Evaluate instrustion following on ZHTW benchmark**

MT-Bench-TC

| | Win | Tie | Lose |
|---|---|---|---|
| **Breeze-7B-FC-v1_0** *v.s.* Breeze-7B-Instruct-v1_0 | 36 (22.5%) | 81 (50.6%) | 43 (26.9%) |


## πŸ‘©β€πŸ’» How to use

**Dependiency**

```
pip install mtkresearch vllm
```