File size: 5,844 Bytes
3ec9224
6b7ae1b
ccd769b
9278ac3
6b7ae1b
9278ac3
6b7ae1b
9278ac3
 
f08873e
ccd769b
7e34d60
55bc620
6b7ae1b
2073925
7e34d60
b9d9831
 
 
 
 
 
 
 
 
 
6b7ae1b
2073925
6b7ae1b
 
2073925
6b7ae1b
9278ac3
6b7ae1b
 
 
 
9278ac3
7e34d60
 
2073925
6b7ae1b
2073925
6b7ae1b
 
2073925
6b7ae1b
 
9278ac3
6b7ae1b
 
2073925
6b7ae1b
2073925
ccd769b
7e34d60
2073925
 
 
7e34d60
 
 
9278ac3
 
7e34d60
 
 
2073925
7e34d60
2073925
 
 
 
6b7ae1b
 
 
 
2073925
 
 
 
 
6b7ae1b
2073925
6b7ae1b
 
2073925
5be8df6
2073925
9278ac3
 
71bcd22
2073925
71bcd22
2073925
 
71bcd22
2073925
 
 
 
327ff33
71bcd22
2073925
 
 
 
327ff33
2073925
 
 
 
b9d9831
2073925
 
9278ac3
 
 
 
 
 
2073925
9278ac3
2073925
327ff33
2073925
 
55bc620
 
9278ac3
55bc620
9278ac3
 
2073925
9278ac3
55bc620
327ff33
2073925
 
 
327ff33
 
b9d9831
2073925
 
b9d9831
2073925
b9d9831
 
327ff33
 
2073925
327ff33
b9d9831
2073925
 
71bcd22
 
9278ac3
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
import gradio as gr
import os

from langchain.document_loaders import PyPDFLoader
from langchain.text_splitter import RecursiveCharacterTextSplitter
from langchain.vectorstores import Chroma
from langchain.chains import ConversationalRetrievalChain
from langchain.embeddings import HuggingFaceEmbeddings
from langchain.llms import HuggingFacePipeline
from langchain.memory import ConversationBufferMemory

from transformers import AutoTokenizer, pipeline
import torch

# Lista de modelos 100% abertos e gratuitos
list_llm = [
    "google/flan-t5-xxl",
    "TinyLlama/TinyLlama-1.1B-Chat-v1.0",
    "microsoft/phi-2",
    "facebook/opt-1.3b",
    "EleutherAI/gpt-neo-1.3B",
    "bigscience/bloom-1b7",
    "RWKV/rwkv-4-169m-pile",
    "gpt2-medium",
    "databricks/dolly-v2-3b",
    "mosaicml/mpt-7b-instruct"
]

list_llm_simple = [os.path.basename(llm) for llm in list_llm]

# Função para carregar documentos PDF
def load_doc(list_file_path, chunk_size, chunk_overlap):
    loaders = [PyPDFLoader(file_path) for file_path in list_file_path]
    pages = []
    for loader in loaders:
        pages.extend(loader.load())
    text_splitter = RecursiveCharacterTextSplitter(
        chunk_size=chunk_size,
        chunk_overlap=chunk_overlap
    )
    return text_splitter.split_documents(pages)

# Função para criar banco de dados vetorial
def create_db(splits, collection_name):
    embedding = HuggingFaceEmbeddings()
    return Chroma.from_documents(
        documents=splits,
        embedding=embedding,
        persist_directory=f"./{collection_name}"
    )

# Função para inicializar o modelo LLM
def initialize_llmchain(llm_model, temperature, max_tokens, top_k, vector_db, progress=gr.Progress()):
    progress(0.1, desc="Carregando tokenizer...")
    
    tokenizer = AutoTokenizer.from_pretrained(llm_model)
    
    progress(0.4, desc="Inicializando pipeline...")
    pipeline_obj = pipeline(
        "text-generation",
        model=llm_model,
        tokenizer=tokenizer,
        torch_dtype=torch.bfloat16 if torch.cuda.is_available() else torch.float32,
        device=0 if torch.cuda.is_available() else -1,
        max_new_tokens=max_tokens,
        do_sample=True,
        top_k=top_k,
        temperature=temperature
    )
    
    llm = HuggingFacePipeline(pipeline=pipeline_obj)
    
    progress(0.7, desc="Configurando memória...")
    memory = ConversationBufferMemory(
        memory_key="chat_history",
        return_messages=True
    )
    
    progress(0.8, desc="Criando cadeia...")
    return ConversationalRetrievalChain.from_llm(
        llm=llm,
        retriever=vector_db.as_retriever(),
        memory=memory,
        return_source_documents=True
    )

# Interface Gradio
def demo():
    with gr.Blocks(theme=gr.themes.Soft()) as demo:
        vector_db = gr.State(None)  # Inicializa com None
        qa_chain = gr.State(None)  # Inicializa com None
        
        gr.Markdown("## 🤖 Chatbot para PDFs com Modelos Gratuitos")
        
        with gr.Tab("📤 Upload PDF"):
            pdf_input = gr.Files(label="Selecione seus PDFs", file_types=[".pdf"])
            
        with gr.Tab("⚙️ Processamento"):
            chunk_size = gr.Slider(100, 1000, value=500, label="Tamanho dos Chunks")
            chunk_overlap = gr.Slider(0, 200, value=50, label="Sobreposição")
            process_btn = gr.Button("Processar PDFs")
            process_status = gr.Textbox(label="Status do Processamento", interactive=False)
            
        with gr.Tab("🧠 Modelo"):
            model_selector = gr.Dropdown(list_llm_simple, label="Selecione o Modelo", value=list_llm_simple[0])
            temperature = gr.Slider(0, 1, value=0.7, label="Criatividade")
            load_model_btn = gr.Button("Carregar Modelo")
            model_status = gr.Textbox(label="Status do Modelo", interactive=False)
            
        with gr.Tab("💬 Chat"):
            chatbot = gr.Chatbot(height=400)
            msg = gr.Textbox(label="Sua mensagem")
            clear_btn = gr.Button("Limpar Chat")
            
        # Eventos
        def process_documents(files, cs, co):
            file_paths = [f.name for f in files]
            splits = load_doc(file_paths, cs, co)
            db = create_db(splits, "docs")
            return db, "Documentos processados!"
        
        process_btn.click(
            process_documents,
            inputs=[pdf_input, chunk_size, chunk_overlap],
            outputs=[vector_db, process_status]
        )
        
        def load_model(model, temp, vector_db_state):
            if vector_db_state is None:
                return None, "Por favor, processe os documentos primeiro."
            qa = initialize_llmchain(list_llm[list_llm_simple.index(model)], temp, 512, 3, vector_db_state)
            return qa, "Modelo carregado!"
        
        load_model_btn.click(
            load_model,
            inputs=[model_selector, temperature, vector_db],
            outputs=[qa_chain, model_status]
        )
        
        def respond(message, chat_history):
            if qa_chain.value is None:
                return "Por favor, carregue um modelo primeiro.", chat_history
            
            result = qa_chain.value({"question": message, "chat_history": chat_history})
            response = result["answer"]
            
            sources = "\n".join([f"📄 Página {doc.metadata['page']+1}: {doc.page_content[:50]}..." 
                                   for doc in result.get("source_documents", [])[:2]])
            
            chat_history.append((message, f"{response}\n\n🔍 Fontes:\n{sources}"))
            return "", chat_history
        
        msg.submit(respond, [msg, chatbot], [msg, chatbot])
        clear_btn.click(lambda: [], outputs=[chatbot])
        
    demo.launch()

if __name__ == "__main__":
    demo()