Spaces:
Paused
Paused
Update app.py
Browse files
app.py
CHANGED
@@ -58,9 +58,12 @@ def group_texts(examples):
|
|
58 |
#base_model = "TheBloke/airoboros-13B-HF" #load_8bit = False (in load_tokenizer_and_model)
|
59 |
base_model = "EleutherAI/gpt-neo-1.3B" #load_8bit = False (in load_tokenizer_and_model)
|
60 |
#base_model = "TheBloke/airoboros-13B-HF" #load_8bit = True
|
|
|
|
|
61 |
tokenizer,model,device = load_tokenizer_and_model(base_model, False)
|
62 |
#tokenizer.add_special_tokens({'pad_token': '[PAD]'}) #not necessary with fast Toekenizers like GPT2
|
63 |
|
|
|
64 |
dataset_neu = daten_laden("alexkueck/tis")
|
65 |
#dataset_neu = daten_laden("EleutherAI/pile")
|
66 |
|
|
|
58 |
#base_model = "TheBloke/airoboros-13B-HF" #load_8bit = False (in load_tokenizer_and_model)
|
59 |
base_model = "EleutherAI/gpt-neo-1.3B" #load_8bit = False (in load_tokenizer_and_model)
|
60 |
#base_model = "TheBloke/airoboros-13B-HF" #load_8bit = True
|
61 |
+
|
62 |
+
#Tokenizer und Model laden
|
63 |
tokenizer,model,device = load_tokenizer_and_model(base_model, False)
|
64 |
#tokenizer.add_special_tokens({'pad_token': '[PAD]'}) #not necessary with fast Toekenizers like GPT2
|
65 |
|
66 |
+
#Datensets für Finetuning laden
|
67 |
dataset_neu = daten_laden("alexkueck/tis")
|
68 |
#dataset_neu = daten_laden("EleutherAI/pile")
|
69 |
|