kshabana commited on
Commit
461ec6b
·
verified ·
1 Parent(s): 5645d11

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +77 -5
README.md CHANGED
@@ -10,14 +10,86 @@ tags:
10
  - llama
11
  - trl
12
  - sft
 
 
 
 
 
13
  ---
 
14
 
15
- # Uploaded model
16
 
17
- - **Developed by:** kshabana
18
- - **License:** apache-2.0
19
- - **Finetuned from model :** unsloth/Meta-Llama-3.1-8B-Instruct-bnb-4bit
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
20
 
21
  This llama model was trained 2x faster with [Unsloth](https://github.com/unslothai/unsloth) and Huggingface's TRL library.
22
 
23
- [<img src="https://raw.githubusercontent.com/unslothai/unsloth/main/images/unsloth%20made%20with%20love.png" width="200"/>](https://github.com/unslothai/unsloth)
 
10
  - llama
11
  - trl
12
  - sft
13
+ - safetersors
14
+ - pytorch
15
+ - coding
16
+ datasets:
17
+ - Replete-AI/code_bagel
18
  ---
19
+ ![a-captivating-and-surreal-image-of-a-goat-with-wil-XIKQzKvDRjihmI3sKK7IoA-0fxZxe6tRAKlXRRc4S9EPA.jpeg](https://cdn-uploads.huggingface.co/production/uploads/650b3f38b5b6029d37298ae8/Mlfr8UeQmKMt8gtgDh_aP.jpeg)
20
 
 
21
 
22
+ # INTRO
23
+
24
+ we are happy to announce our frist coding model
25
+
26
+ # Model Card for Model ID
27
+
28
+ this is a finetune model of llama3.1 that can perform well in coding
29
+
30
+
31
+
32
+ - **Developed by: kshabana
33
+ - **Funded by no one
34
+ - **Shared by kshabana
35
+ - **Model type: safetensors and gguf
36
+ - **Language(s) English and coding
37
+ - **License:** Apache2.0
38
+ - **Finetuned from model [llama3.1-instruct]
39
+
40
+ ### Model Sources
41
+
42
+ https://huggingface.co/meta-llama/Meta-Llama-3.1-8B-Instruct
43
+
44
+ - **Dataset Repository:**
45
+
46
+ https://huggingface.co/datasets/Replete-AI/code_bagel
47
+
48
+ ## Uses
49
+
50
+ this model can be used for combanys that need multilagul support and it has a 131072 context lingth.
51
+
52
+
53
+
54
+ ## Bias, Risks, and Limitations
55
+
56
+ it can some times produce a wrong answer.
57
+
58
+
59
+
60
+ ### Recommendations
61
+
62
+ NOTE:you shold have lm studio or ollama to use this model
63
+
64
+ IN LM-STUDIO:it is recommended to use this model with the defult lm-studio configration.
65
+ IN OLLAMA: we will be pushing the model soon to ollama.
66
+
67
+
68
+
69
+ ## Training Details
70
+
71
+ model trained with unsloth.
72
+
73
+ ### Training Data
74
+
75
+ the training dataset that i used: https://huggingface.co/datasets/arbml/alpaca_arabic
76
+
77
+ ## IMPORTANT LINKS
78
+
79
+ OLLAMA: https://ollama.com
80
+
81
+ LM-STUDIO: https://lmstudio.ai
82
+
83
+ llama3.1: instruct: https://huggingface.co/meta-llama/Meta-Llama-3.1-8B-Instruct
84
+
85
+ dataset: https://huggingface.co/datasets/arbml/alpaca_arabic
86
+
87
+
88
+ ### Results
89
+
90
+
91
+ ![Screenshot 2024-08-14 141239.png](https://cdn-uploads.huggingface.co/production/uploads/650b3f38b5b6029d37298ae8/N-jYQm1uDwxeaAFZXZsPA.png)
92
 
93
  This llama model was trained 2x faster with [Unsloth](https://github.com/unslothai/unsloth) and Huggingface's TRL library.
94
 
95
+ [<img src="https://raw.githubusercontent.com/unslothai/unsloth/main/images/unsloth%20made%20with%20love.png" width="200"/>](https://github.com/unslothai/unsloth)