l-salewski commited on
Commit
36efefe
·
verified ·
1 Parent(s): 3f29c0c

Update README.md

Browse files

- fix missing `)` to repair one of the links

Files changed (1) hide show
  1. README.md +1 -1
README.md CHANGED
@@ -7,7 +7,7 @@ sdk: static
7
  pinned: false
8
  ---
9
 
10
- [Aleph Alpha](https://aleph-alpha.com/research/) is dedicated to building sovereign and trustworthy AI systems. Our research has produced state-of-the-art multi-modal models ([MAGMA](https://github.com/Aleph-Alpha-Research/magma)), explainability techniques for transformer-based models ([AtMan](https://github.com/Aleph-Alpha-Research/AtMan)), and a comprehensive [evaluation framework for large-scale model assessment](https://github.com/Aleph-Alpha-Research/eval-framework/. We have also researched how to [move beyond traditional tokenizers](https://arxiv.org/html/2406.19223v1). Our work on tokenizer-free architectures uses [byte-level trigrams](https://huggingface.co/Aleph-Alpha/tfree-research-vocab-32k-fineweb-steps-370k) to create more resilient and adaptable models in non-english languages and new domains. Key models demonstrating the effectiveness of our innovative [Hierarchical Autoregressive Transformer (HAT)](https://arxiv.org/pdf/2501.10322) architecture include:
11
 
12
  - llama-3_1-tfree-hat models: This model family replaces the Llama 3.1 tokenizer with our HAT architecture. The [8b-dpo model](https://huggingface.co/Aleph-Alpha/llama-3_1-8b-tfree-hat-dpo) is tuned for helpfulness and reduced refusal in sensitive applications, while the larger [70b-sft model](https://huggingface.co/Aleph-Alpha/llama-3_1-70b-tfree-hat-sft) is trained on English/German for improved text compression and adaptability.
13
 
 
7
  pinned: false
8
  ---
9
 
10
+ [Aleph Alpha](https://aleph-alpha.com/research/) is dedicated to building sovereign and trustworthy AI systems. Our research has produced state-of-the-art multi-modal models ([MAGMA](https://github.com/Aleph-Alpha-Research/magma)), explainability techniques for transformer-based models ([AtMan](https://github.com/Aleph-Alpha-Research/AtMan)), and a comprehensive [evaluation framework for large-scale model assessment](https://github.com/Aleph-Alpha-Research/eval-framework/). We have also researched how to [move beyond traditional tokenizers](https://arxiv.org/html/2406.19223v1). Our work on tokenizer-free architectures uses [byte-level trigrams](https://huggingface.co/Aleph-Alpha/tfree-research-vocab-32k-fineweb-steps-370k) to create more resilient and adaptable models in non-english languages and new domains. Key models demonstrating the effectiveness of our innovative [Hierarchical Autoregressive Transformer (HAT)](https://arxiv.org/pdf/2501.10322) architecture include:
11
 
12
  - llama-3_1-tfree-hat models: This model family replaces the Llama 3.1 tokenizer with our HAT architecture. The [8b-dpo model](https://huggingface.co/Aleph-Alpha/llama-3_1-8b-tfree-hat-dpo) is tuned for helpfulness and reduced refusal in sensitive applications, while the larger [70b-sft model](https://huggingface.co/Aleph-Alpha/llama-3_1-70b-tfree-hat-sft) is trained on English/German for improved text compression and adaptability.
13