Update README.md
Browse files
README.md
CHANGED
@@ -32,6 +32,13 @@ For example, if you prompt it with the "Two Sum" problem and the input `nums = [
|
|
32 |
|
33 |
**Finetuned from model:** `unsloth/Llama-3.2-1B-unsloth-bnb-4bit`
|
34 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
35 |
## Uses
|
36 |
|
37 |
### Direct Use
|
@@ -91,11 +98,4 @@ It is a specialized model designed only to replicate the input-output pairs from
|
|
91 |
|
92 |
The model was fine-tuned on a dataset by newfacade from here: https://huggingface.co/datasets/newfacade/LeetCodeDataset
|
93 |
|
94 |
-
I just used 5k samples from it and trained only for 1 epoch
|
95 |
-
|
96 |
-
# NOTE
|
97 |
-
|
98 |
-
I used only two target models ("q_proj", "v_proj") which focuses only on the attention blocks and kept rank value (r=8).
|
99 |
-
Why? I have neither money nor time to run the model.
|
100 |
-
|
101 |
-
If you like to waste your time on this, the notebook is available here: https://www.kaggle.com/code/yashasnadig/leetcode2output
|
|
|
32 |
|
33 |
**Finetuned from model:** `unsloth/Llama-3.2-1B-unsloth-bnb-4bit`
|
34 |
|
35 |
+
## NOTE
|
36 |
+
|
37 |
+
I used only two target models ("q_proj", "v_proj") which focuses only on the attention blocks and kept rank value (r=8).
|
38 |
+
Why? I have neither money nor time to run the model.
|
39 |
+
|
40 |
+
If you like to waste your time on this, the notebook is available here: https://www.kaggle.com/code/yashasnadig/leetcode2output
|
41 |
+
|
42 |
## Uses
|
43 |
|
44 |
### Direct Use
|
|
|
98 |
|
99 |
The model was fine-tuned on a dataset by newfacade from here: https://huggingface.co/datasets/newfacade/LeetCodeDataset
|
100 |
|
101 |
+
I just used 5k samples from it and trained only for 1 epoch.
|
|
|
|
|
|
|
|
|
|
|
|
|
|