SivilTaram commited on
Commit
b3ec645
·
verified ·
1 Parent(s): 81bc6c4

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +37 -1
README.md CHANGED
@@ -7,4 +7,40 @@ sdk: static
7
  pinned: false
8
  ---
9
 
10
- The Sailor2 project is to build a large language model with 24B to 32B parameters, optimized for multiple South-East Asian languages, such as Cebuano, Indonesian, Khmer, Lao, Minangkabau, Malay, Burmese, Sundanese, Javanese, Thai, and Vietnamese. The model will be continually pre-trained on a base model proficient in both Chinese and English, and its performance is expected to be comparable to the most advanced business models for the above South-East Asian languages.
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
7
  pinned: false
8
  ---
9
 
10
+ # Sailor2
11
+
12
+ The Sailor2 community is to build open large language models optimized for multiple South-East Asian languages, such as Cebuano, Indonesian, Khmer, Lao, Minangkabau, Malay, Burmese, Sundanese, Javanese, Thai, and Vietnamese. The model will be continually pre-trained on a base model proficient in both Chinese and English, and its performance is expected to be comparable to the most advanced business models for the above South-East Asian languages.
13
+
14
+ Read more details about Sailor2 at https://sailorllm.github.io/blog/sailor2.
15
+
16
+ ---
17
+
18
+ <details>
19
+ <summary>
20
+ <b><font size="+1">🔱 Sailor2 Models</font></b>
21
+ </summary>
22
+
23
+ - [GitHub](https://github.com/sail-sg/sailor2): All you need to know about using or fine-tuning Sailor2.
24
+ - [Sailor2-1B](https://huggingface.co/sail/Sailor2-1B): 1B base model continually pre-trained on 500B tokens from Qwen2.5-0.5B.
25
+ - [Sailor2-8B](https://huggingface.co/sail/Sailor2-8B): 8B base model continually pre-trained on 500B tokens from Qwen2.5-7B.
26
+ - [Sailor2-20B](https://huggingface.co/sail/Sailor2-20B): 20B base model continually pre-trained on 500B tokens from Qwen2.5-14B.
27
+ </details>
28
+
29
+ ---
30
+ <details>
31
+ <summary>
32
+ <b><font size="+1">📚 Sailor2 Pre-training Dataset </font></b>
33
+ </summary>
34
+
35
+ </details>
36
+
37
+ ---
38
+
39
+ <details>
40
+ <summary>
41
+ <b><font size="+1">📑 Sailor2 Post-training Dataset </font></b>
42
+ </summary>
43
+
44
+ </details>
45
+
46
+ ---