Spaces:
Running
Running
Update app.py
Browse files
app.py
CHANGED
@@ -342,14 +342,11 @@ class EnhancedFileProcessor:
|
|
342 |
logger.error(f"Archive processing error: {e}")
|
343 |
return dataset
|
344 |
|
345 |
-
def chunk_data(self, data: Union[Dict, List
|
346 |
"""Enhanced data chunking with sequence metadata"""
|
347 |
try:
|
348 |
-
|
349 |
-
|
350 |
-
json_str = json.dumps(data, ensure_ascii=False)
|
351 |
-
else:
|
352 |
-
json_str = data
|
353 |
total_length = len(json_str)
|
354 |
|
355 |
# Calculate overhead for metadata
|
@@ -460,7 +457,7 @@ def generate_qr_codes(data: Union[str, Dict, List], combined: bool = True) -> Li
|
|
460 |
for i, chunk in enumerate(chunks):
|
461 |
filename = f'combined_qr_{int(time.time())}_{i+1}_of_{len(chunks)}.png'
|
462 |
qr_path = generate_stylish_qr(
|
463 |
-
data=chunk
|
464 |
filename=filename,
|
465 |
fill_color="#1a365d", # Deep blue
|
466 |
back_color="#ffffff"
|
@@ -475,7 +472,7 @@ def generate_qr_codes(data: Union[str, Dict, List], combined: bool = True) -> Li
|
|
475 |
for chunk_idx, chunk in enumerate(chunks):
|
476 |
filename = f'item_{idx+1}_chunk_{chunk_idx+1}_of_{len(chunks)}_{int(time.time())}.png'
|
477 |
qr_path = generate_stylish_qr(
|
478 |
-
data=chunk
|
479 |
filename=filename,
|
480 |
fill_color="#1a365d", # Deep blue
|
481 |
back_color="#ffffff"
|
@@ -487,7 +484,7 @@ def generate_qr_codes(data: Union[str, Dict, List], combined: bool = True) -> Li
|
|
487 |
for i, chunk in enumerate(chunks):
|
488 |
filename = f'single_qr_{i+1}_of_{len(chunks)}_{int(time.time())}.png'
|
489 |
qr_path = generate_stylish_qr(
|
490 |
-
data=chunk
|
491 |
filename=filename,
|
492 |
fill_color="#1a365d", # Deep blue
|
493 |
back_color="#ffffff"
|
@@ -806,19 +803,14 @@ def create_modern_interface():
|
|
806 |
for url in url_list:
|
807 |
validation = url_processor.validate_url(url)
|
808 |
if validation['is_valid']:
|
809 |
-
|
810 |
-
if
|
811 |
-
|
812 |
-
|
813 |
-
|
814 |
-
|
815 |
-
|
816 |
-
|
817 |
-
'chunk_index': i + 1,
|
818 |
-
'total_chunks': len(chunks),
|
819 |
-
'content': chunk['data'], # Store the chunked data
|
820 |
-
'timestamp': datetime.now().isoformat()
|
821 |
-
})
|
822 |
|
823 |
# Process files
|
824 |
if files:
|
|
|
342 |
logger.error(f"Archive processing error: {e}")
|
343 |
return dataset
|
344 |
|
345 |
+
def chunk_data(self, data: Union[Dict, List], max_size: int = 2953) -> List[Dict]:
|
346 |
"""Enhanced data chunking with sequence metadata"""
|
347 |
try:
|
348 |
+
# Convert data to JSON string
|
349 |
+
json_str = json.dumps(data, ensure_ascii=False)
|
|
|
|
|
|
|
350 |
total_length = len(json_str)
|
351 |
|
352 |
# Calculate overhead for metadata
|
|
|
457 |
for i, chunk in enumerate(chunks):
|
458 |
filename = f'combined_qr_{int(time.time())}_{i+1}_of_{len(chunks)}.png'
|
459 |
qr_path = generate_stylish_qr(
|
460 |
+
data=chunk,
|
461 |
filename=filename,
|
462 |
fill_color="#1a365d", # Deep blue
|
463 |
back_color="#ffffff"
|
|
|
472 |
for chunk_idx, chunk in enumerate(chunks):
|
473 |
filename = f'item_{idx+1}_chunk_{chunk_idx+1}_of_{len(chunks)}_{int(time.time())}.png'
|
474 |
qr_path = generate_stylish_qr(
|
475 |
+
data=chunk,
|
476 |
filename=filename,
|
477 |
fill_color="#1a365d", # Deep blue
|
478 |
back_color="#ffffff"
|
|
|
484 |
for i, chunk in enumerate(chunks):
|
485 |
filename = f'single_qr_{i+1}_of_{len(chunks)}_{int(time.time())}.png'
|
486 |
qr_path = generate_stylish_qr(
|
487 |
+
data=chunk,
|
488 |
filename=filename,
|
489 |
fill_color="#1a365d", # Deep blue
|
490 |
back_color="#ffffff"
|
|
|
803 |
for url in url_list:
|
804 |
validation = url_processor.validate_url(url)
|
805 |
if validation['is_valid']:
|
806 |
+
content = url_processor.fetch_content(url)
|
807 |
+
if content:
|
808 |
+
results.append({
|
809 |
+
'source': 'url',
|
810 |
+
'url': url,
|
811 |
+
'content': content,
|
812 |
+
'timestamp': datetime.now().isoformat()
|
813 |
+
})
|
|
|
|
|
|
|
|
|
|
|
814 |
|
815 |
# Process files
|
816 |
if files:
|