Update README.md
Browse files
README.md
CHANGED
@@ -5,15 +5,23 @@ datasets:
|
|
5 |
- tensorplex-labs/Dojo-HumanFeedback-DPO
|
6 |
language:
|
7 |
- en
|
|
|
8 |
---
|
9 |
|
10 |
-
#
|
11 |
|
12 |
-
|
13 |
|
|
|
14 |
|
|
|
15 |
|
16 |
-
|
|
|
|
|
|
|
|
|
|
|
17 |
|
18 |
### Model Description
|
19 |
|
@@ -21,13 +29,11 @@ language:
|
|
21 |
|
22 |
|
23 |
|
24 |
-
- **Developed by:**
|
25 |
-
- **
|
26 |
-
- **
|
27 |
-
- **
|
28 |
-
- **
|
29 |
-
- **License:** [More Information Needed]
|
30 |
-
- **Finetuned from model [optional]:** [More Information Needed]
|
31 |
|
32 |
### Model Sources [optional]
|
33 |
|
|
|
5 |
- tensorplex-labs/Dojo-HumanFeedback-DPO
|
6 |
language:
|
7 |
- en
|
8 |
+
license: cc-by-4.0
|
9 |
---
|
10 |
|
11 |
+
# INTERFACE-CODER-7B: First-of-its-kind Interface Generation Model Trained On High Quality Synthetic Data Curated Using Distributed Human Feedback
|
12 |
|
13 |
+
We are thrilled to release INTERFACE-CODER-7B, a-first-of-its-kind Large Language Model (LLM) specialized in generating complex, interactive, and visually appealing frontend interfaces.
|
14 |
|
15 |
+
INTERFACE-CODER-7B is trained on high quality synthetic data generated by state-of-the-art AI models. Data quality is further guaranteed using code verifiers, LLM-as-judge, and distributed human feedback.
|
16 |
|
17 |
+
Leveraging Dojo's distributed human feedback infrastructure, we curated two datasets:
|
18 |
|
19 |
+
- Dojo-SFT: A comprehensive dataset for supervised fine-tuning (SFT), filtered using LLM-as-judge.
|
20 |
+
- Dojo-DPO: A preference dataset for Direct Preference Optimization (DPO), curated using human feedback scores to align the model's output with human aesthetic and functional preferences.
|
21 |
+
|
22 |
+
Our development process followed a two-stage post-training methodology. We began with the powerful **Qwen2.5-Coder-7B-Instruct** as our base model. This foundation was then elevated through a supervised fine-tuning phase with Dojo-SFT, followed by a direct preference optimization stage using Dojo-DPO. This produced the final, highly specialized INTERFACE-CODER-7B.
|
23 |
+
|
24 |
+
INTERFACE-CODER-7B is capable of generating functional and visually appealing frontend, far exceeding the interface generation capabilities of its base model. Beyond its primary use case, the model demonstrates remarkable generalization against other benchmarks beyond MMLU, GSM8k, and HumanEval.
|
25 |
|
26 |
### Model Description
|
27 |
|
|
|
29 |
|
30 |
|
31 |
|
32 |
+
- **Developed by:** Shi Jie Yu, Tensorplex Labs
|
33 |
+
- **Model type:** LoRA DPO
|
34 |
+
- **Language(s) (NLP):** English
|
35 |
+
- **License:** Creative Commons Attribution 4.0
|
36 |
+
- **Finetuned from model:** [tensorplex-labs/INTERFACE-CODER-7B-SFT](https://huggingface.co/tensorplex-labs/INTERFACE-CODER-7B-SFT)
|
|
|
|
|
37 |
|
38 |
### Model Sources [optional]
|
39 |
|