nehcgs commited on
Commit
55e7a8f
·
verified ·
1 Parent(s): 5b31888

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +12 -12
README.md CHANGED
@@ -2,7 +2,7 @@
2
  license: other
3
  license_name: katanemo-research
4
  license_link: >-
5
- https://huggingface.co/katanemolabs/Arch-Function-Calling-1.5B/blob/main/LICENSE
6
  base_model:
7
  - Qwen/Qwen2.5-1.5B-Instruct
8
  language:
@@ -11,12 +11,12 @@ pipeline_tag: text-generation
11
  library_name: transformers
12
  ---
13
 
14
- # Katanemo/Arch-Function-Calling-1.5B
15
 
16
  ## Overview
17
- The Katanemo Arch-Function-Calling collection of large language models (LLMs) is a collection state-of-the-art (SOTA) LLMs specifically designed for **function calling** tasks. The models are designed to understand complex function signatures, identify required parameters, and produce accurate function call outputs based on natural language prompts. Achieving performance on par with GPT-4, these models set a new benchmark in the domain of function-oriented tasks, making them suitable for scenarios where automated API interaction and function execution is crucial.
18
 
19
- In summary, the Katanemo Arch-Function-Calling collection demonstrates:
20
  - **State-of-the-art performance** in function calling
21
  - **Accurate parameter identification and suggestion**, even in ambiguous or incomplete inputs
22
  - **High generalization** across multiple function calling use cases, from API interactions to automated backend tasks.
@@ -49,11 +49,11 @@ In summary, the Katanemo Arch-Function-Calling collection demonstrates:
49
 
50
 
51
  ## Training Details
52
- Katanemo Arch-Function-Calling collection is built on top of the [Qwen 2.5](https://huggingface.co/collections/Qwen/qwen25-66e81a666513e518adb90d9e). A blog with technical details leading to our models will be published soon.
53
 
54
 
55
  ## Performance Benchmarks
56
- We evaluate Katanemo Arch-Function-Calling series on the [Berkeley Function-Calling Leaderboard (BFCL)](https://gorilla.cs.berkeley.edu/leaderboard.html#leaderboard). For each model family, we select the one with the highest rank. The results are shwon below:
57
 
58
  <table>
59
  <tr style="text-align: center; vertical-align: middle; font-weight: bold;">
@@ -96,7 +96,7 @@ We evaluate Katanemo Arch-Function-Calling series on the [Berkeley Function-Call
96
  </tr>
97
  <tr style="text-align: center; vertical-align: middle; font-weight: bold;">
98
  <td> </td>
99
- <td>Arch-Function-Calling-7B</td>
100
  <td>57.48%</td>
101
  <td>87.50%</td>
102
  <td>86.80%</td>
@@ -107,7 +107,7 @@ We evaluate Katanemo Arch-Function-Calling series on the [Berkeley Function-Call
107
  </tr>
108
  <tr style="text-align: center; vertical-align: middle; font-weight: bold;">
109
  <td> </td>
110
- <td>Arch-Function-Calling-3B</td>
111
  <td>56.23%</td>
112
  <td>85.10%</td>
113
  <td>89.16%</td>
@@ -141,7 +141,7 @@ We evaluate Katanemo Arch-Function-Calling series on the [Berkeley Function-Call
141
  </tr>
142
  <tr style="text-align: center; vertical-align: middle; font-weight: bold;">
143
  <td> </td>
144
- <td>Arch-Function-Calling-1.5B</td>
145
  <td>53.61%</td>
146
  <td>82.60%</td>
147
  <td>87.36%</td>
@@ -176,7 +176,7 @@ We evaluate Katanemo Arch-Function-Calling series on the [Berkeley Function-Call
176
 
177
 
178
  # Requirements
179
- The code of Arch-Function-Calling-1.5B has been in the Hugging Face `transformers` library and we advise you to install latest version:
180
  ```bash
181
  pip install transformers>=4.37.0
182
  ```
@@ -192,7 +192,7 @@ import json
192
  from typing import Any, Dict, List
193
  from transformers import AutoModelForCausalLM, AutoTokenizer
194
 
195
- model_name = "katanemolabs/Arch-Function-Calling-1.5B"
196
  model = AutoModelForCausalLM.from_pretrained(
197
  model_name, device_map="auto", torch_dtype="auto", trust_remote_code=True
198
  )
@@ -332,4 +332,4 @@ The current temperature in Seattle is 62 degrees in Fahrenheit.
332
 
333
 
334
  # License
335
- Katanemo Arch-Function-Calling collection is distributed under the [Katanemo license](https://huggingface.co/katanemolabs/Arch-Function-Calling-1.5B/blob/main/LICENSE).
 
2
  license: other
3
  license_name: katanemo-research
4
  license_link: >-
5
+ https://huggingface.co/katanemolabs/Arch-Function-1.5B/blob/main/LICENSE
6
  base_model:
7
  - Qwen/Qwen2.5-1.5B-Instruct
8
  language:
 
11
  library_name: transformers
12
  ---
13
 
14
+ # katanemolabs/Arch-Function-1.5B
15
 
16
  ## Overview
17
+ The Katanemo Arch-Function collection of large language models (LLMs) is a collection state-of-the-art (SOTA) LLMs specifically designed for **function calling** tasks. The models are designed to understand complex function signatures, identify required parameters, and produce accurate function call outputs based on natural language prompts. Achieving performance on par with GPT-4, these models set a new benchmark in the domain of function-oriented tasks, making them suitable for scenarios where automated API interaction and function execution is crucial.
18
 
19
+ In summary, the Katanemo Arch-Function collection demonstrates:
20
  - **State-of-the-art performance** in function calling
21
  - **Accurate parameter identification and suggestion**, even in ambiguous or incomplete inputs
22
  - **High generalization** across multiple function calling use cases, from API interactions to automated backend tasks.
 
49
 
50
 
51
  ## Training Details
52
+ Katanemo Arch-Function collection is built on top of the [Qwen 2.5](https://huggingface.co/collections/Qwen/qwen25-66e81a666513e518adb90d9e). A blog with technical details leading to our models will be published soon.
53
 
54
 
55
  ## Performance Benchmarks
56
+ We evaluate Katanemo Arch-Function series on the [Berkeley Function-Calling Leaderboard (BFCL)](https://gorilla.cs.berkeley.edu/leaderboard.html#leaderboard). For each model family, we select the one with the highest rank. The results are shwon below:
57
 
58
  <table>
59
  <tr style="text-align: center; vertical-align: middle; font-weight: bold;">
 
96
  </tr>
97
  <tr style="text-align: center; vertical-align: middle; font-weight: bold;">
98
  <td> </td>
99
+ <td>Arch-Function-7B</td>
100
  <td>57.48%</td>
101
  <td>87.50%</td>
102
  <td>86.80%</td>
 
107
  </tr>
108
  <tr style="text-align: center; vertical-align: middle; font-weight: bold;">
109
  <td> </td>
110
+ <td>Arch-Function-3B</td>
111
  <td>56.23%</td>
112
  <td>85.10%</td>
113
  <td>89.16%</td>
 
141
  </tr>
142
  <tr style="text-align: center; vertical-align: middle; font-weight: bold;">
143
  <td> </td>
144
+ <td>Arch-Function-1.5B</td>
145
  <td>53.61%</td>
146
  <td>82.60%</td>
147
  <td>87.36%</td>
 
176
 
177
 
178
  # Requirements
179
+ The code of Arch-Function-1.5B has been in the Hugging Face `transformers` library and we advise you to install latest version:
180
  ```bash
181
  pip install transformers>=4.37.0
182
  ```
 
192
  from typing import Any, Dict, List
193
  from transformers import AutoModelForCausalLM, AutoTokenizer
194
 
195
+ model_name = "katanemolabs/Arch-Function-1.5B"
196
  model = AutoModelForCausalLM.from_pretrained(
197
  model_name, device_map="auto", torch_dtype="auto", trust_remote_code=True
198
  )
 
332
 
333
 
334
  # License
335
+ Katanemo Arch-Function collection is distributed under the [Katanemo license](https://huggingface.co/katanemolabs/Arch-Function-1.5B/blob/main/LICENSE).