Spaces:
Running
Running
Update app.py
Browse files
app.py
CHANGED
@@ -1,11 +1,3 @@
|
|
1 |
-
!pip install streamlit google-generativeai pymupdf pyngrok transformers spacy python-docx nltk dateparser
|
2 |
-
!python -m spacy download en_core_web_sm
|
3 |
-
!python -m nltk.downloader words
|
4 |
-
|
5 |
-
%%writefile combined_resume_analyzer.py
|
6 |
-
|
7 |
-
|
8 |
-
|
9 |
import google.generativeai as genai
|
10 |
import fitz # PyMuPDF for PDF text extraction
|
11 |
import streamlit as st
|
@@ -16,8 +8,6 @@ import re
|
|
16 |
from nltk.corpus import words
|
17 |
import dateparser
|
18 |
from datetime import datetime
|
19 |
-
from pyngrok import ngrok
|
20 |
-
import os
|
21 |
|
22 |
# Load SpaCy model for dependency parsing
|
23 |
nlp_spacy = spacy.load('en_core_web_sm')
|
@@ -187,35 +177,4 @@ def main():
|
|
187 |
st.error(f"Error during processing: {e}")
|
188 |
|
189 |
if __name__ == "__main__":
|
190 |
-
|
191 |
-
|
192 |
-
# Set your authtoken
|
193 |
-
ngrok.set_auth_token("2keP9BS91BCtRFtnf5Ss4tOpzq4_2c6463MYzXPqFM3a95gUM") # Replace YOUR_AUTHTOKEN
|
194 |
-
|
195 |
-
# Terminate any running ngrok processes (if any)
|
196 |
-
!pkill -f streamlit
|
197 |
-
|
198 |
-
# Run Streamlit in the background
|
199 |
-
# The 'port' option should be passed as a keyword argument to the 'ngrok.connect()' function.
|
200 |
-
public_url = ngrok.connect(8501)
|
201 |
-
print("Public URL:", public_url)
|
202 |
-
|
203 |
-
# Launch Streamlit
|
204 |
-
!streamlit run combined_resume_analyzer.py
|
205 |
-
|
206 |
-
|
207 |
-
from pyngrok import ngrok
|
208 |
-
|
209 |
-
# Set your authtoken
|
210 |
-
ngrok.set_auth_token("2keP9BS91BCtRFtnf5Ss4tOpzq4_2c6463MYzXPqFM3a95gUM") # Replace YOUR_AUTHTOKEN
|
211 |
-
|
212 |
-
# Terminate any running ngrok processes (if any)
|
213 |
-
!pkill -f streamlit
|
214 |
-
|
215 |
-
# Run Streamlit in the background
|
216 |
-
# The 'port' option should be passed as a keyword argument to the 'ngrok.connect()' function.
|
217 |
-
public_url = ngrok.connect(8501)
|
218 |
-
print("Public URL:", public_url)
|
219 |
-
|
220 |
-
# Launch Streamlit
|
221 |
-
!streamlit run combined_resume_analyzer.py
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
import google.generativeai as genai
|
2 |
import fitz # PyMuPDF for PDF text extraction
|
3 |
import streamlit as st
|
|
|
8 |
from nltk.corpus import words
|
9 |
import dateparser
|
10 |
from datetime import datetime
|
|
|
|
|
11 |
|
12 |
# Load SpaCy model for dependency parsing
|
13 |
nlp_spacy = spacy.load('en_core_web_sm')
|
|
|
177 |
st.error(f"Error during processing: {e}")
|
178 |
|
179 |
if __name__ == "__main__":
|
180 |
+
main()
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|