File size: 11,055 Bytes
1f5cba5
 
0e29657
4f25f4e
1f5cba5
4f25f4e
3563dd6
c927679
4f25f4e
266fff4
c927679
0c482eb
 
 
 
0e29657
0c482eb
 
 
0e29657
9afd718
e339dd2
0c482eb
 
 
838224c
0c482eb
 
 
838224c
0c482eb
 
 
 
 
 
 
 
 
4f25f4e
1f5cba5
133d76b
 
 
 
 
4f25f4e
 
1f5cba5
e339dd2
4f25f4e
f74ec57
4f25f4e
133d76b
 
 
0c482eb
 
 
 
133d76b
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
0c482eb
 
133d76b
 
 
 
1f5cba5
133d76b
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1f5cba5
133d76b
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
f74ec57
4f25f4e
 
0e29657
4f25f4e
0e29657
4f25f4e
 
 
 
 
 
0e29657
4f25f4e
 
7fb0070
4f25f4e
 
 
7fb0070
4f25f4e
 
 
 
 
 
 
 
 
 
09b1a3d
 
 
4f25f4e
7fb0070
7c5f7b3
0c482eb
 
 
 
7fb0070
09b1a3d
 
0c482eb
09b1a3d
7c5f7b3
7fb0070
e339dd2
4f25f4e
 
 
7fb0070
7c5f7b3
 
 
4f25f4e
7c5f7b3
 
 
0c482eb
 
4f25f4e
 
0c482eb
7c5f7b3
7fb0070
09b1a3d
 
 
 
0c482eb
abff174
03df343
 
 
 
abff174
 
7fb0070
 
838224c
4f25f4e
a59a680
 
 
 
 
4f25f4e
a59a680
 
 
 
 
 
 
 
 
 
894c1ff
4f25f4e
a59a680
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
c927679
a59a680
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
4f25f4e
 
a59a680
 
4f25f4e
a59a680
4f25f4e
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
274
275
276
277
278
279
280
281
282
283
284
285
286
287
288
289
290
291
292
293
294
295
296
297
298
299
300
301
302
303
304
305
306
307
308
309
310
311
312
313
314
315
316
317
318
319
320
321
322
323
# tools.py

import pandas as pd

from pathlib import Path

import regex as re
import time
import os
from duckduckgo_search import DDGS

DEFAULT_API_URL = "https://agents-course-unit4-scoring.hf.space"


def _download_file_for_task(task_id: str, ext: str) -> str:
    """
    Helper: attempt to GET the remote file for a given task_id.
    Saves under ./hf_files/{task_id}.{ext}. Returns the local path if successful,
    or an empty string if no file / download failed.
    """
    
    print("reached _download_file_for_task")
    os.makedirs("hf_files", exist_ok=True)
    local_path = os.path.join("hf_files", f"{task_id}.{ext}")
    url = f"{DEFAULT_API_URL}/files/{task_id}"
    
    try:
        resp = requests.get(url, timeout=10)
        if resp.status_code == 200 and resp.content:
            print(f"Downloaded file from {url} to {local_path}")
            with open(local_path, "wb") as f:
                f.write(resp.content)
            return local_path
    except Exception:
        pass

    # If we get here, either 404 or download error
    return ""

def ocr_image_tool(args: dict) -> str:
    """
    Expects: state["ocr_path"] is either:
      • a local image path (e.g. "./hf_files/abc.png"), OR
      • a Task ID (e.g. "abc123"), in which case we try downloading
        GET {DEFAULT_API_URL}/files/{task_id} with .png/.jpg/.jpeg extensions.

    Returns: "OCR text + brief caption or an error message"
      
    """
    print("reached ocr_image_tool")
    # path_or_id = state.get("ocr_path", "")
    for ext in ("png", "jpg", "jpeg"):
            candidate = _download_file_for_task(args["task_id"], ext)
            if candidate:
                local_img = candidate
                break

    if not local_img or not os.path.exists(local_img):
        return {
            "ocr_path": None,
            "ocr_result": "Error: No image file found (local nonexistent or download failed)."
        }

    # 2) Read raw bytes
    try:
        with open(local_img, "rb") as f:
            image_bytes = f.read()
    except Exception as e:
        return {
            "ocr_path": None,
            "ocr_result": f"Error reading image file: {e}"
        }

    # 3) Prepare HF Inference headers
    hf_token = os.getenv("HF_TOKEN")
    if not hf_token:
        return {
            "ocr_path": None,
            "ocr_result": "Error: HUGGINGFACE_API_KEY not set in environment."
        }

    headers = {"Authorization": f"Bearer {hf_token}"}

    # 4) Call HF’s vision-ocr to extract text
    ocr_text = ""
    try:
        ocr_resp = requests.post(
            "https://api-inference.huggingface.co/models/google/vit-ocr",
            headers=headers,
            files={"file": image_bytes},
            timeout=30
        )
        ocr_resp.raise_for_status()
        ocr_json = ocr_resp.json()

        # The JSON has “pages” → list of blocks → “lines” → each line has “text”
        lines = []
        for page in ocr_json.get("pages", []):
            for line in page.get("lines", []):
                lines.append(line.get("text", "").strip())
        ocr_text = "\n".join(lines).strip() or "(no visible text)"
    except Exception as e:
        ocr_text = f"Error during HF OCR: {e}"

    # 5) Call HF’s image-captioning to get a brief description
    caption = ""
    try:
        cap_resp = requests.post(
            "https://api-inference.huggingface.co/models/Salesforce/blip-image-captioning-base",
            headers=headers,
            files={"file": image_bytes},
            timeout=30
        )
        cap_resp.raise_for_status()
        cap_json = cap_resp.json()
        # The response looks like: {"generated_text": "...caption..."}
        caption = cap_json.get("generated_text", "").strip()
        if not caption:
            caption = "(no caption returned)"
    except Exception as e:
        caption = f"Error during HF captioning: {e}"

    # 6) Combine OCR + caption
    combined = f"OCR text:\n{ocr_text}\n\nImage caption:\n{caption}"
    print("combined: ")
    return combined


def parse_excel_tool(args: dict) -> str:
    """
    Downloads <task_id>.xlsx (if any) and returns a stringified list of
    records from the specified sheet.  No fallback to user-supplied tables.
    Expected keys in `args`:
        • task_id            – required (used to download the file)
        • excel_sheet_name   – optional sheet to load
    returns: stringified list of records from the specified sheet
    """
    task_id = args.get("task_id", "")
    sheet   = args.get("excel_sheet_name", "")

    local_xlsx = _download_file_for_task(task_id, "xlsx")
    if not local_xlsx or not os.path.exists(local_xlsx):
        return "Error: Excel file not found for this task."

    try:
        xls = pd.ExcelFile(local_xlsx)
        df  = pd.read_excel(
            xls,
            sheet_name=sheet if sheet and sheet in xls.sheet_names else xls.sheet_names[0]
        )
        return str(df.to_dict(orient="records"))
    except Exception as e:
        return f"Error reading Excel file: {e}"
    

import openai

def audio_transcriber_tool(args: dict) -> str:
    """
    LangGraph tool for transcribing audio via OpenAI's Whisper API.
    Expects: state["audio_path"] to be either:
      • A local file path (e.g. "./hf_files/abc.mp3"), OR
      • A Task ID (e.g. "abc123"), in which case we try downloading
        GET {DEFAULT_API_URL}/files/{task_id} with .mp3, .wav, .m4a extensions.
    Returns:
      {
        "audio_path": None,
        "transcript": "<text or error message>"
      }
    Always attempts to download the file for the given path or task ID.
    """
    print("reached audio_transcriber_tool")
    # path_or_id = state.get("audio_path", "")
    # if not path_or_id:
    #     return {}

    # Always attempt to download the file, regardless of local existence
    local_audio = ""
    for ext in ("mp3", "wav", "m4a"):
        candidate = _download_file_for_task(args["task_id"], ext)
        if candidate:
            local_audio = candidate
            break

    if not local_audio or not os.path.exists(local_audio):
        return "Error: No audio file found (download failed)."
        

    # Send to OpenAI Whisper
    try:
        openai.api_key = os.getenv("OPENAI_API_KEY")
        if not openai.api_key:
            raise RuntimeError("OPENAI_API_KEY is not set in environment.")

        with open(local_audio, "rb") as audio_file:
            print("reached openai.audio.transcriptions.create")
            response = openai.audio.transcriptions.create(
                model="whisper-1", 
                file=audio_file,
            )
            print("reached response")
        text = response.text.strip()
    except Exception as e:
        text = f"Error during transcription: {e}"
    print(f"Transcripted as transcript: {text}")
    return text
# tools.py

import re
import requests

def wikipedia_search_tool(args: dict) -> str:
    """
    LangGraph wrapper for searching Wikipedia.
    Expects: state["wiki_query"] to be a non‐empty string.
    Returns:
      {
        "wiki_query": None,
        "wiki_result": "<text summary of first matching page or an error message>"
      }
    If no valid wiki_query is provided, returns {}.
    """
    print("reached wikipedia search tool")
    query = args["wiki_query"]
    if not query:
        return {}

    try:
        # 1) Use the MediaWiki API to search for page titles matching the query
        search_params = {
            "action": "query",
            "list": "search",
            "srsearch": query,
            "format": "json",
            "utf8": 1
        }
        search_resp = requests.get("https://en.wikipedia.org/w/api.php", params=search_params, timeout=10)
        search_resp.raise_for_status()
        search_data = search_resp.json()

        search_results = search_data.get("query", {}).get("search", [])
        # print("wikipedia: search_results",search_results)
        if not search_results:
            return {"wiki_query": None, "wiki_result": f"No Wikipedia page found for '{query}'."}

        # 2) Take the first search result's title
        first_title = search_results[0].get("title", "")
        if not first_title:
            return {"wiki_query": None, "wiki_result": "Unexpected format from Wikipedia search."}

        # 3) Fetch the page summary for that title via the REST summary endpoint
        title_for_url = requests.utils.requote_uri(first_title)
        summary_url = f"https://en.wikipedia.org/api/rest_v1/page/summary/{title_for_url}"
        summary_resp = requests.get(summary_url, timeout=10)
        summary_resp.raise_for_status()
        summary_data = summary_resp.json()

        # 4) Extract either the "extract" field or a fallback message
        summary_text = summary_data.get("extract")
        if not summary_text:
            summary_text = summary_data.get("description", "No summary available.")

        return f"Title: {first_title}\n\n{summary_text}"
        

    except requests.exceptions.RequestException as e:
        return f"Wikipedia search error: {e}"
    except Exception as e:
        return f"Unexpected error in wikipedia_search_tool: {e}"





# def web_search_tool(state: AgentState) -> AgentState:
#     """
#     Expects: state["web_search_query"] is a non‐empty string.
#     Returns: {"web_search_query": None, "web_search_result": <string>}.
#     Retries up to 5 times on either a DuckDuckGo “202 Ratelimit” response or any exception (e.g. timeout).
#     """
#     print("reached web_search_tool")
#     query = state.get("web_search_query", "")
#     if not query:
#         return {}  # nothing to do

#     ddg = DDGS()
#     max_retries = 5
#     result_text = ""

#     for attempt in range(1, max_retries + 1):
#         try:
#             result_text = str(ddg.text(query, max_results=5))
#         except Exception as e:
#             # Network error or timeout—retry up to max_retries
#             if attempt < max_retries:
#                 print(f"web_search_tool: exception '{e}', retrying in 4 seconds ({attempt}/{max_retries})")
#                 time.sleep(4)
#                 continue
#             else:
#                 # Final attempt failed
#                 return {
#                     "web_search_query": None,
#                     "web_search_result": f"Error during DuckDuckGo search: {e}"
#                 }

#         # Check for DuckDuckGo rate‐limit indicator
#         if "202 Ratelimit" in result_text:
#             if attempt < max_retries:
#                 print(f"web_search_tool: received '202 Ratelimit', retrying in 4 seconds ({attempt}/{max_retries})")
#                 time.sleep(4)
#                 continue
#             else:
#                 # Final attempt still rate‐limited
#                 break

#         # Successful response (no exception and no rate‐limit text)
#         break

#     return {
#         "web_search_query": None,
#         "web_search_result": result_text
#     }