Update fine_tune_inference_test.py
Browse files
fine_tune_inference_test.py
CHANGED
@@ -4,7 +4,7 @@ import uvicorn
|
|
4 |
from fastapi import FastAPI
|
5 |
from fastapi.responses import HTMLResponse
|
6 |
from pydantic import BaseModel
|
7 |
-
from transformers import AutoTokenizer, AutoModelForCausalLM, pipeline
|
8 |
from datasets import load_dataset
|
9 |
from peft import PeftModel
|
10 |
import torch
|
@@ -128,13 +128,13 @@ def setup_model():
|
|
128 |
log(f"🔍 RAG boyutu: {len(rag)}")
|
129 |
|
130 |
log("🚀 Pipeline oluşturuluyor...")
|
131 |
-
pipe =
|
132 |
-
"text-generation",
|
133 |
model=model,
|
134 |
tokenizer=tokenizer,
|
135 |
torch_dtype=torch.float16 if torch.cuda.is_available() else torch.float32,
|
136 |
device=0 if torch.cuda.is_available() else -1
|
137 |
)
|
|
|
138 |
log("✅ Model ve pipeline başarıyla yüklendi.")
|
139 |
except Exception as e:
|
140 |
log(f"❌ setup_model() sırasında hata oluştu: {e}")
|
|
|
4 |
from fastapi import FastAPI
|
5 |
from fastapi.responses import HTMLResponse
|
6 |
from pydantic import BaseModel
|
7 |
+
from transformers import AutoTokenizer, AutoModelForCausalLM, pipeline, TextGenerationPipeline
|
8 |
from datasets import load_dataset
|
9 |
from peft import PeftModel
|
10 |
import torch
|
|
|
128 |
log(f"🔍 RAG boyutu: {len(rag)}")
|
129 |
|
130 |
log("🚀 Pipeline oluşturuluyor...")
|
131 |
+
pipe = TextGenerationPipeline(
|
|
|
132 |
model=model,
|
133 |
tokenizer=tokenizer,
|
134 |
torch_dtype=torch.float16 if torch.cuda.is_available() else torch.float32,
|
135 |
device=0 if torch.cuda.is_available() else -1
|
136 |
)
|
137 |
+
|
138 |
log("✅ Model ve pipeline başarıyla yüklendi.")
|
139 |
except Exception as e:
|
140 |
log(f"❌ setup_model() sırasında hata oluştu: {e}")
|