KJ24 commited on
Commit
11c5d58
·
verified ·
1 Parent(s): ecd203a

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +4 -0
app.py CHANGED
@@ -8,6 +8,8 @@ from llama_index.core import Document
8
  from llama_index.llms.llama_cpp import LlamaCPP
9
  from llama_index.core.node_parser import SemanticSplitterNodeParser
10
 
 
 
11
  # ✅ Pour l'embedding LOCAL via transformers
12
  from transformers import AutoTokenizer, AutoModel
13
  import torch
@@ -74,6 +76,8 @@ async def chunk_text(data: ChunkRequest):
74
  def get_text_embedding(self, text: str):
75
  return get_embedding(text)
76
 
 
 
77
  # ✅ Nouvelle configuration (⚠️ ne plus utiliser ServiceContext)
78
  Settings.llm = llm
79
  Settings.embed_model = SimpleEmbedding()
 
8
  from llama_index.llms.llama_cpp import LlamaCPP
9
  from llama_index.core.node_parser import SemanticSplitterNodeParser
10
 
11
+ from llama_index.core.base.llms.base import BaseLLM
12
+
13
  # ✅ Pour l'embedding LOCAL via transformers
14
  from transformers import AutoTokenizer, AutoModel
15
  import torch
 
76
  def get_text_embedding(self, text: str):
77
  return get_embedding(text)
78
 
79
+ assert isinstance(llm, BaseLLM), "❌ Ce LLM n’est pas compatible avec Settings.llm"
80
+
81
  # ✅ Nouvelle configuration (⚠️ ne plus utiliser ServiceContext)
82
  Settings.llm = llm
83
  Settings.embed_model = SimpleEmbedding()