JosephusCheung commited on
Commit
4d49d73
1 Parent(s): 8386742

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +1 -1
README.md CHANGED
@@ -32,7 +32,7 @@ tags:
32
  ---
33
  ![](https://huggingface.co/JosephusCheung/tmp/resolve/main/7.72b.png)
34
 
35
- *Image drawn by GPT-4 DALL路E 3* TL;DR: Perhaps this 7B model, better than all existing models <= 33B, in most quantitative evaluations...
36
 
37
  # CausalLM 7B - Fully Compatible with Meta LLaMA 2
38
  Use the transformers library that does not require remote/external code to load the model, AutoModelForCausalLM and AutoTokenizer (or manually specify LlamaForCausalLM to load LM, GPT2Tokenizer to load Tokenizer), and model quantization is fully compatible with GGUF (llama.cpp), GPTQ, and AWQ.
 
32
  ---
33
  ![](https://huggingface.co/JosephusCheung/tmp/resolve/main/7.72b.png)
34
 
35
+ *Image drawn by GPT-4 DALL路E 3* **TL;DR: Perhaps this 7B model, better than all existing models <= 33B, in most quantitative evaluations...**
36
 
37
  # CausalLM 7B - Fully Compatible with Meta LLaMA 2
38
  Use the transformers library that does not require remote/external code to load the model, AutoModelForCausalLM and AutoTokenizer (or manually specify LlamaForCausalLM to load LM, GPT2Tokenizer to load Tokenizer), and model quantization is fully compatible with GGUF (llama.cpp), GPTQ, and AWQ.