CometAPI commited on
Commit
37d6668
·
verified ·
1 Parent(s): b20f6dc

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +85 -3
README.md CHANGED
@@ -1,3 +1,85 @@
1
- ---
2
- license: apache-2.0
3
- ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: apache-2.0
3
+ ---
4
+ CometAPI Model Page: [GPT-5 mini](https://www.cometapi.com/gpt-5-mini-api/)
5
+
6
+ GPT-5 mini is a **lightweight**, cost-optimized variant of OpenAI’s flagship GPT-5 model, designed to deliver **high-quality** reasoning and multimodal capabilities at reduced latency and expense.
7
+
8
+ ## Basic information & key features
9
+
10
+ **GPT-5 mini** is OpenAI’s **cost- and latency-optimized** member of the GPT-5 family, intended to deliver much of GPT-5’s multimodal and instruction-following strengths at **substantially lower cost** for large-scale production use. It targets environments where **throughput**, **predictable per-token pricing**, and **fast responses** are the primary constraints while still providing strong general-purpose capabilities.
11
+
12
+ - **Model Name**: `gpt-5-mini`
13
+ - **Context Window**: 400 000 tokens
14
+ - **Max Output Tokens**: 128 000
15
+ - **Key features:** speed, throughput, cost-efficiency, deterministic outputs for concise prompts
16
+
17
+ ## Technical details — architecture, inference, and controls
18
+
19
+ **Optimized inference path & deployment.** Practical speedups come from **kernel fusion**, **tensor parallelism tuned for a smaller graph**, and an inference runtime that prefers **shorter internal “thinking” loops** unless the developer requests deeper reasoning. That is why mini achieves noticeably lower compute per call and predictable latency for high-volume traffic. This tradeoff is deliberate: **lower compute per forward pass → lower cost and lower average latency**.
20
+
21
+ **Developer controls.** GPT-5 mini exposes parameters such as **`verbosity`** (controls detail/length) and **`reasoning_effort`** (trade speed vs. depth), plus robust **tool-calling** support (function calls, parallel tool chains, and structured error handling), which lets production systems tune accuracy vs. cost precisely.
22
+
23
+ ## Benchmark performance — headline numbers and interpretation
24
+
25
+ GPT-5 mini typically sits **within ~85–95%** of GPT-5 high on general benchmarks while substantially improving latency/price. The platform launch materials indicate **very high absolute scores** for GPT-5 high (AIME ≈ **94.6%** reported for the top variant), with mini somewhat lower but still industry-leading for its price point.
26
+
27
+ Across a range of standardized and internal benchmarks, **GPT-5 mini** achieves:
28
+
29
+ - **Intelligence** (AIME ’25): 91.1% (vs. 94.6% for GPT-5 high)
30
+ - **Multimodal** (MMMU): 81.6% (vs. 84.2% for GPT-5 high)
31
+ - **Coding** (SWE-bench Verified): 71.0% (vs. 74.9% for GPT-5 high)
32
+ - **Instruction Following** (Scale MultiChallenge): 62.3% (vs. 69.6%)
33
+ - **Function Calling** (τ²-bench telecom): 74.1% (vs. 96.7%)
34
+ - **Hallucination Rates** (LongFact-Concepts): 0.7% (lower is better)([OpenAI][4])
35
+
36
+ These results demonstrate GPT-5 mini’s **robust** trade-offs between performance, cost, and speed.
37
+
38
+ ## Limitations
39
+
40
+ **Known limitations:** GPT-5 mini *reduced deep-reasoning capacity vs full GPT-5, higher sensitivity to ambiguous prompts, and remaining risks of hallucination.*
41
+
42
+ - **Reduced deep reasoning:** For multi-step, long-horizon reasoning tasks the full reasoning model or “thinking” variants outperform mini.
43
+ - **Hallucinations & overconfidence:** Mini reduces hallucination relative to very small models but does not eliminate it; outputs should be validated in high-stakes flows (legal, clinical, compliance).
44
+ - **Context sensitivity:** Very long, highly interdependent context chains are better served by the full GPT-5 variants with larger context windows or the “thinking” model.
45
+ - **Safety & policy limits:** Same safety guardrails and rate/usage limits that apply to other GPT-5 models apply to mini; sensitive tasks require human oversight.
46
+
47
+ ## Recommended use cases (where mini excels)
48
+
49
+ - **High-volume conversational agents:** low latency, predictable cost. **Keyword:** *throughput*.
50
+ - **Document & multimodal summarization:** long-context summarization, image+text reports. **Keyword:** *long context*.
51
+ - **Developer tooling at scale:** CI code checks, auto-review, lightweight code generation. **Keyword:** *cost-efficient coding*.
52
+ - **Agent orchestration:** tool-calling with parallel chains when deep reasoning is not required. **Keyword:** *tool calling*.
53
+
54
+ ## How to call ***\*`gpt-5-mini`\**** API from CometAPI
55
+
56
+ ## [GPT-5 mini](https://www.cometapi.com/gpt-5-mini-api/) API Pricing in CometAPI,20% off the official price:
57
+
58
+ | Input Tokens | $0.20 |
59
+ | ------------- | ----- |
60
+ | Output Tokens | $1.60 |
61
+
62
+ ### Required Steps
63
+
64
+ - Log in to [cometapi.com](http://cometapi.com/). If you are not our user yet, please register first
65
+ - Get the access credential API key of the interface. Click “Add Token” at the API token in the personal center, get the token key: sk-xxxxx and submit.
66
+ - Get the url of this site: https://api.cometapi.com/
67
+
68
+ ### Use Method
69
+
70
+ 1. Select the “`gpt-5-mini`“ / “`gpt-5-mini-2025-08-07`” endpoint to send the API request and set the request body. The request method and request body are obtained from our website API doc. Our website also provides Apifox test for your convenience.
71
+ 2. Replace <YOUR_API_KEY> with your actual CometAPI key from your account.
72
+ 3. Insert your question or request into the content field—this is what the model will respond to.
73
+ 4. . Process the API response to get the generated answer.
74
+
75
+ CometAPI provides a fully compatible REST API—for seamless migration. Key details to [API doc](https://apidoc.cometapi.com/api-13851472):
76
+
77
+ - **Core Parameters**: `prompt`, `max_tokens_to_sample`, `temperature`, `stop_sequences`
78
+ - **Endpoint:** https://api.cometapi.com/v1/chat/completions
79
+ - **Model Parameter:** “`gpt-5-mini`“ / “`gpt-5-mini-2025-08-07`“
80
+ - **Authentication:** ` Bearer YOUR_CometAPI_API_KEY`
81
+ - **Content-Type:** `application/json` .
82
+
83
+ API Call Instructions: gpt-5-chat-latest should be called using the standard `/v1/chat/completions forma`t. For other models (gpt-5, gpt-5-mini, gpt-5-nano, and their dated versions), using `the /v1/responses format` [is recommended](https://apidoc.cometapi.com/api-18535147). Currently two modes are available.
84
+
85
+ **See Also [GPT-5](https://www.cometapi.com/gpt-5-api/)** Model