Spaces:
Runtime error
Runtime error
Commit
·
dc959c4
1
Parent(s):
53626ba
Replace with slightly older app.py
Browse files- App_Function_Libraries/__pycache__/Audio_Files.cpython-310.pyc +0 -0
- App_Function_Libraries/__pycache__/Local_File_Processing_Lib.cpython-310.pyc +0 -0
- App_Function_Libraries/__pycache__/Local_LLM_Inference_Engine_Lib.cpython-310.pyc +0 -0
- App_Function_Libraries/__pycache__/PDF_Ingestion_Lib.cpython-310.pyc +0 -0
- App_Function_Libraries/__pycache__/SQLite_DB.cpython-310.pyc +0 -0
- App_Function_Libraries/__pycache__/System_Checks_Lib.cpython-310.pyc +0 -0
- app.py +6 -3
App_Function_Libraries/__pycache__/Audio_Files.cpython-310.pyc
ADDED
|
Binary file (13.9 kB). View file
|
|
|
App_Function_Libraries/__pycache__/Local_File_Processing_Lib.cpython-310.pyc
ADDED
|
Binary file (2.21 kB). View file
|
|
|
App_Function_Libraries/__pycache__/Local_LLM_Inference_Engine_Lib.cpython-310.pyc
ADDED
|
Binary file (13.1 kB). View file
|
|
|
App_Function_Libraries/__pycache__/PDF_Ingestion_Lib.cpython-310.pyc
ADDED
|
Binary file (3.88 kB). View file
|
|
|
App_Function_Libraries/__pycache__/SQLite_DB.cpython-310.pyc
ADDED
|
Binary file (28.1 kB). View file
|
|
|
App_Function_Libraries/__pycache__/System_Checks_Lib.cpython-310.pyc
ADDED
|
Binary file (5.18 kB). View file
|
|
|
app.py
CHANGED
|
@@ -746,7 +746,8 @@ Sample commands:
|
|
| 746 |
help="Use a local LLM from the script(Downloads llamafile from github and 'mistral-7b-instruct-v0.2.Q8' - 8GB model from Huggingface)")
|
| 747 |
parser.add_argument('--server_mode', action='store_true',
|
| 748 |
help='Run in server mode (This exposes the GUI/Server to the network)')
|
| 749 |
-
parser.add_argument('
|
|
|
|
| 750 |
parser.add_argument('--port', type=int, default=7860, help='Port to run the server on')
|
| 751 |
parser.add_argument('--ingest_text_file', action='store_true',
|
| 752 |
help='Ingest .txt files as content instead of treating them as URL lists')
|
|
@@ -768,11 +769,12 @@ Sample commands:
|
|
| 768 |
set_chunk_txt_by_tokens = False
|
| 769 |
set_max_txt_chunk_tokens = 0
|
| 770 |
|
| 771 |
-
if args.share_public
|
| 772 |
share_public = args.share_public
|
| 773 |
else:
|
| 774 |
share_public = None
|
| 775 |
if args.server_mode:
|
|
|
|
| 776 |
server_mode = args.server_mode
|
| 777 |
else:
|
| 778 |
server_mode = None
|
|
@@ -840,12 +842,13 @@ Sample commands:
|
|
| 840 |
sys.exit(0)
|
| 841 |
|
| 842 |
# Launch the GUI
|
|
|
|
| 843 |
if args.user_interface:
|
| 844 |
if local_llm:
|
| 845 |
local_llm_function()
|
| 846 |
time.sleep(2)
|
| 847 |
webbrowser.open_new_tab('http://127.0.0.1:7860')
|
| 848 |
-
launch_ui(
|
| 849 |
elif not args.input_path:
|
| 850 |
parser.print_help()
|
| 851 |
sys.exit(1)
|
|
|
|
| 746 |
help="Use a local LLM from the script(Downloads llamafile from github and 'mistral-7b-instruct-v0.2.Q8' - 8GB model from Huggingface)")
|
| 747 |
parser.add_argument('--server_mode', action='store_true',
|
| 748 |
help='Run in server mode (This exposes the GUI/Server to the network)')
|
| 749 |
+
parser.add_argument('--share_public', type=int, default=7860,
|
| 750 |
+
help="This will use Gradio's built-in ngrok tunneling to share the server publicly on the internet. Specify the port to use (default: 7860)")
|
| 751 |
parser.add_argument('--port', type=int, default=7860, help='Port to run the server on')
|
| 752 |
parser.add_argument('--ingest_text_file', action='store_true',
|
| 753 |
help='Ingest .txt files as content instead of treating them as URL lists')
|
|
|
|
| 769 |
set_chunk_txt_by_tokens = False
|
| 770 |
set_max_txt_chunk_tokens = 0
|
| 771 |
|
| 772 |
+
if args.share_public:
|
| 773 |
share_public = args.share_public
|
| 774 |
else:
|
| 775 |
share_public = None
|
| 776 |
if args.server_mode:
|
| 777 |
+
|
| 778 |
server_mode = args.server_mode
|
| 779 |
else:
|
| 780 |
server_mode = None
|
|
|
|
| 842 |
sys.exit(0)
|
| 843 |
|
| 844 |
# Launch the GUI
|
| 845 |
+
# This is huggingface so:
|
| 846 |
if args.user_interface:
|
| 847 |
if local_llm:
|
| 848 |
local_llm_function()
|
| 849 |
time.sleep(2)
|
| 850 |
webbrowser.open_new_tab('http://127.0.0.1:7860')
|
| 851 |
+
launch_ui(demo_mode=False)
|
| 852 |
elif not args.input_path:
|
| 853 |
parser.print_help()
|
| 854 |
sys.exit(1)
|