Update app.py
Browse files
app.py
CHANGED
@@ -6,14 +6,12 @@ import gradio as gr
|
|
6 |
nlp = spacy.load('es_core_news_sm')
|
7 |
|
8 |
models = {
|
9 |
-
"stabilityai/stablelm-tuned-alpha-7b": AutoModelForCausalLM.from_pretrained("stabilityai/stablelm-tuned-alpha-7b"),
|
10 |
"CRD716/ggml-LLaMa-65B-quantized": AutoModelForCausalLM.from_pretrained("CRD716/ggml-LLaMa-65B-quantized"),
|
11 |
"RedXeol/bertin-gpt-j-6B-alpaca-4bit-128g": AutoModelForCausalLM.from_pretrained("RedXeol/bertin-gpt-j-6B-alpaca-4bit-128g"),
|
12 |
"bertin-project/bertin-alpaca-lora-7b": AutoModelForCausalLM.from_pretrained("bertin-project/bertin-alpaca-lora-7b")
|
13 |
}
|
14 |
|
15 |
tokenizers = {
|
16 |
-
"stabilityai/stablelm-tuned-alpha-7b": AutoTokenizer.from_pretrained("stabilityai/stablelm-tuned-alpha-7b"),
|
17 |
"CRD716/ggml-LLaMa-65B-quantized": AutoTokenizer.from_pretrained("CRD716/ggml-LLaMa-65B-quantized"),
|
18 |
"RedXeol/bertin-gpt-j-6B-alpaca-4bit-128g": AutoTokenizer.from_pretrained("RedXeol/bertin-gpt-j-6B-alpaca-4bit-128g"),
|
19 |
"bertin-project/bertin-alpaca-lora-7b": AutoTokenizer.from_pretrained("bertin-project/bertin-alpaca-lora-7b")
|
|
|
6 |
nlp = spacy.load('es_core_news_sm')
|
7 |
|
8 |
models = {
|
|
|
9 |
"CRD716/ggml-LLaMa-65B-quantized": AutoModelForCausalLM.from_pretrained("CRD716/ggml-LLaMa-65B-quantized"),
|
10 |
"RedXeol/bertin-gpt-j-6B-alpaca-4bit-128g": AutoModelForCausalLM.from_pretrained("RedXeol/bertin-gpt-j-6B-alpaca-4bit-128g"),
|
11 |
"bertin-project/bertin-alpaca-lora-7b": AutoModelForCausalLM.from_pretrained("bertin-project/bertin-alpaca-lora-7b")
|
12 |
}
|
13 |
|
14 |
tokenizers = {
|
|
|
15 |
"CRD716/ggml-LLaMa-65B-quantized": AutoTokenizer.from_pretrained("CRD716/ggml-LLaMa-65B-quantized"),
|
16 |
"RedXeol/bertin-gpt-j-6B-alpaca-4bit-128g": AutoTokenizer.from_pretrained("RedXeol/bertin-gpt-j-6B-alpaca-4bit-128g"),
|
17 |
"bertin-project/bertin-alpaca-lora-7b": AutoTokenizer.from_pretrained("bertin-project/bertin-alpaca-lora-7b")
|