PEFT
Safetensors
English
tensorplex-labs commited on
Commit
5973efd
·
verified ·
1 Parent(s): 746bbf4

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +16 -10
README.md CHANGED
@@ -5,15 +5,23 @@ datasets:
5
  - tensorplex-labs/Dojo-HumanFeedback-DPO
6
  language:
7
  - en
 
8
  ---
9
 
10
- # Model Card for Model ID
11
 
12
- <!-- Provide a quick summary of what the model is/does. -->
13
 
 
14
 
 
15
 
16
- ## Model Details
 
 
 
 
 
17
 
18
  ### Model Description
19
 
@@ -21,13 +29,11 @@ language:
21
 
22
 
23
 
24
- - **Developed by:** [More Information Needed]
25
- - **Funded by [optional]:** [More Information Needed]
26
- - **Shared by [optional]:** [More Information Needed]
27
- - **Model type:** [More Information Needed]
28
- - **Language(s) (NLP):** [More Information Needed]
29
- - **License:** [More Information Needed]
30
- - **Finetuned from model [optional]:** [More Information Needed]
31
 
32
  ### Model Sources [optional]
33
 
 
5
  - tensorplex-labs/Dojo-HumanFeedback-DPO
6
  language:
7
  - en
8
+ license: cc-by-4.0
9
  ---
10
 
11
+ # INTERFACE-CODER-7B: First-of-its-kind Interface Generation Model Trained On High Quality Synthetic Data Curated Using Distributed Human Feedback
12
 
13
+ We are thrilled to release INTERFACE-CODER-7B, a-first-of-its-kind Large Language Model (LLM) specialized in generating complex, interactive, and visually appealing frontend interfaces.
14
 
15
+ INTERFACE-CODER-7B is trained on high quality synthetic data generated by state-of-the-art AI models. Data quality is further guaranteed using code verifiers, LLM-as-judge, and distributed human feedback.
16
 
17
+ Leveraging Dojo's distributed human feedback infrastructure, we curated two datasets:
18
 
19
+ - Dojo-SFT: A comprehensive dataset for supervised fine-tuning (SFT), filtered using LLM-as-judge.
20
+ - Dojo-DPO: A preference dataset for Direct Preference Optimization (DPO), curated using human feedback scores to align the model's output with human aesthetic and functional preferences.
21
+
22
+ Our development process followed a two-stage post-training methodology. We began with the powerful **Qwen2.5-Coder-7B-Instruct** as our base model. This foundation was then elevated through a supervised fine-tuning phase with Dojo-SFT, followed by a direct preference optimization stage using Dojo-DPO. This produced the final, highly specialized INTERFACE-CODER-7B.
23
+
24
+ INTERFACE-CODER-7B is capable of generating functional and visually appealing frontend, far exceeding the interface generation capabilities of its base model. Beyond its primary use case, the model demonstrates remarkable generalization against other benchmarks beyond MMLU, GSM8k, and HumanEval.
25
 
26
  ### Model Description
27
 
 
29
 
30
 
31
 
32
+ - **Developed by:** Shi Jie Yu, Tensorplex Labs
33
+ - **Model type:** LoRA DPO
34
+ - **Language(s) (NLP):** English
35
+ - **License:** Creative Commons Attribution 4.0
36
+ - **Finetuned from model:** [tensorplex-labs/INTERFACE-CODER-7B-SFT](https://huggingface.co/tensorplex-labs/INTERFACE-CODER-7B-SFT)
 
 
37
 
38
  ### Model Sources [optional]
39