Spaces:
Running
Running
Update README.md
Browse files
README.md
CHANGED
@@ -131,20 +131,43 @@ pinned: true
|
|
131 |
<li>Innovative data preparation and curation techniques</li>
|
132 |
<li>Focus on domain-specific excellence and versatility</li>
|
133 |
<li>Open collaboration and knowledge sharing within the AI community</li>
|
|
|
134 |
</ul>
|
135 |
<p></p>
|
136 |
<h2>Team:</h2>
|
137 |
<div class="team-member">
|
|
|
138 |
<p><strong>@elinas</strong> - <a href="https://huggingface.co/elinas" target="_blank">HuggingFace Profile</a></p>
|
139 |
</div>
|
140 |
<div class="team-member">
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
141 |
<p><strong>@SteelSkull</strong> - <a href="https://huggingface.co/Steelskull" target="_blank">HuggingFace Profile</a></p>
|
142 |
</div>
|
143 |
<p></p>
|
144 |
<h2>Notable Achievements</h2>
|
145 |
<div class="achievements">
|
146 |
<ul>
|
147 |
-
<li>
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
148 |
<li>Creation of the Aether-Lite-V1.8.1 dataset, a carefully curated dataset for AI training</li>
|
149 |
</ul>
|
150 |
</div>
|
|
|
131 |
<li>Innovative data preparation and curation techniques</li>
|
132 |
<li>Focus on domain-specific excellence and versatility</li>
|
133 |
<li>Open collaboration and knowledge sharing within the AI community</li>
|
134 |
+
<li>Advancing LLM research via novel techniques, which has been applied by all of our members.</li>
|
135 |
</ul>
|
136 |
<p></p>
|
137 |
<h2>Team:</h2>
|
138 |
<div class="team-member">
|
139 |
+
<p>Chief Engineer</p>
|
140 |
<p><strong>@elinas</strong> - <a href="https://huggingface.co/elinas" target="_blank">HuggingFace Profile</a></p>
|
141 |
</div>
|
142 |
<div class="team-member">
|
143 |
+
<p>Data Scientist</p>
|
144 |
+
<p><strong>@ToastyPigeon</strong> - <a href="https://huggingface.co/ToastyPigeon" target="_blank">HuggingFace Profile</a></p>
|
145 |
+
</div>
|
146 |
+
<div class="team-member">
|
147 |
+
<p>Ops Engineer</p>
|
148 |
+
<p><strong>@fizz</strong> - <a href="https://huggingface.co/Fizzarolli" target="_blank">HuggingFace Profile</a></p>
|
149 |
+
</div>
|
150 |
+
<div class="team-member">
|
151 |
+
<p>ML / DS Engineer</p>
|
152 |
<p><strong>@SteelSkull</strong> - <a href="https://huggingface.co/Steelskull" target="_blank">HuggingFace Profile</a></p>
|
153 |
</div>
|
154 |
<p></p>
|
155 |
<h2>Notable Achievements</h2>
|
156 |
<div class="achievements">
|
157 |
<ul>
|
158 |
+
<li>Revival of Llama 1 33B by training on over 500M tokens</li>
|
159 |
+
<li>We did this based on the original pretraining token count of 1.4T and decided to add another 500M tokens to it, to which our surprise ended up
|
160 |
+
surpassing expectations in both quality and length</li>
|
161 |
+
<li>
|
162 |
+
It was trained at 16384 context legth with an *effective* context legnth around 12k due to the nature of the samples, but exceeds in RP.
|
163 |
+
</li>
|
164 |
+
<li>
|
165 |
+
Our next goal is to apply GQA to it, but in the meantime, we will appreciate quanters who will help with running this model on less VRAM!
|
166 |
+
</li>
|
167 |
+
</ul>
|
168 |
+
<ul>
|
169 |
+
<li>Development of L3-Aethora-15B series, The first heavily fintuned 15b model that focuses in creative writing and general intelligence using a novel
|
170 |
+
technique known as "zeroing layers."</li>
|
171 |
<li>Creation of the Aether-Lite-V1.8.1 dataset, a carefully curated dataset for AI training</li>
|
172 |
</ul>
|
173 |
</div>
|