import gradio as gr import os import torch, time from huggingface_hub import hf_hub_download, snapshot_download, login import sys import gc login(token=os.environ["HF_TOKEN"]) tokenizer = os.environ["TOKENIZER"] repo_id = os.environ["REPO_ID"] torch.set_num_threads(1) if torch.cuda.is_available(): torch.backends.cudnn.benchmark = True try: generate_file = hf_hub_download( repo_id=repo_id, filename="gen.py", token=os.environ["HF_TOKEN"] ) os.system(f"cp {generate_file} ./gen.py") except Exception as e: print(f"Error downloading files: {e}") model_path = None sys.path.append('.') from gen import chat_interface with gr.Blocks(title="Madhuram Model Chat") as demo: with gr.Column(elem_id="main-container"): gr.Markdown("# Madhuram Chat Interface", elem_classes="center-text") chatbot = gr.ChatInterface( fn=chat_interface, type="messages" ) gr.Markdown("*Disclaimer - This is a demo version of Madhuram. It may occasionally generate incorrect or incomplete responses. Please verify important information independently. The complete model will be available through our own playground where the missing features will be incorporated.*", elem_classes="disclaimer") demo.css = """ #main-container { max-width: 800px; margin: 0 auto; padding: 20px; } .center-text { text-align: center; } .disclaimer { text-align: center; color: #666; font-size: 0.9em; margin-bottom: 20px; } """ if __name__ == "__main__": demo.launch()