acecalisto3 commited on
Commit
545818e
·
verified ·
1 Parent(s): f09ec18

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +14 -22
app.py CHANGED
@@ -342,14 +342,11 @@ class EnhancedFileProcessor:
342
  logger.error(f"Archive processing error: {e}")
343
  return dataset
344
 
345
- def chunk_data(self, data: Union[Dict, List, str], max_size: int = 2953) -> List[Dict]:
346
  """Enhanced data chunking with sequence metadata"""
347
  try:
348
- if not isinstance(data, str):
349
- # Convert data to JSON string
350
- json_str = json.dumps(data, ensure_ascii=False)
351
- else:
352
- json_str = data
353
  total_length = len(json_str)
354
 
355
  # Calculate overhead for metadata
@@ -460,7 +457,7 @@ def generate_qr_codes(data: Union[str, Dict, List], combined: bool = True) -> Li
460
  for i, chunk in enumerate(chunks):
461
  filename = f'combined_qr_{int(time.time())}_{i+1}_of_{len(chunks)}.png'
462
  qr_path = generate_stylish_qr(
463
- data=chunk['data'], # Use the 'data' part of the chunk
464
  filename=filename,
465
  fill_color="#1a365d", # Deep blue
466
  back_color="#ffffff"
@@ -475,7 +472,7 @@ def generate_qr_codes(data: Union[str, Dict, List], combined: bool = True) -> Li
475
  for chunk_idx, chunk in enumerate(chunks):
476
  filename = f'item_{idx+1}_chunk_{chunk_idx+1}_of_{len(chunks)}_{int(time.time())}.png'
477
  qr_path = generate_stylish_qr(
478
- data=chunk['data'], # Use the 'data' part of the chunk
479
  filename=filename,
480
  fill_color="#1a365d", # Deep blue
481
  back_color="#ffffff"
@@ -487,7 +484,7 @@ def generate_qr_codes(data: Union[str, Dict, List], combined: bool = True) -> Li
487
  for i, chunk in enumerate(chunks):
488
  filename = f'single_qr_{i+1}_of_{len(chunks)}_{int(time.time())}.png'
489
  qr_path = generate_stylish_qr(
490
- data=chunk['data'], # Use the 'data' part of the chunk
491
  filename=filename,
492
  fill_color="#1a365d", # Deep blue
493
  back_color="#ffffff"
@@ -806,19 +803,14 @@ def create_modern_interface():
806
  for url in url_list:
807
  validation = url_processor.validate_url(url)
808
  if validation['is_valid']:
809
- content_data = url_processor.fetch_content(url)
810
- if content_data and 'content' in content_data:
811
- # Chunk the content of each URL
812
- chunks = file_processor.chunk_data(content_data['content'])
813
- for i, chunk in enumerate(chunks):
814
- results.append({
815
- 'source': 'url',
816
- 'url': url,
817
- 'chunk_index': i + 1,
818
- 'total_chunks': len(chunks),
819
- 'content': chunk['data'], # Store the chunked data
820
- 'timestamp': datetime.now().isoformat()
821
- })
822
 
823
  # Process files
824
  if files:
 
342
  logger.error(f"Archive processing error: {e}")
343
  return dataset
344
 
345
+ def chunk_data(self, data: Union[Dict, List], max_size: int = 2953) -> List[Dict]:
346
  """Enhanced data chunking with sequence metadata"""
347
  try:
348
+ # Convert data to JSON string
349
+ json_str = json.dumps(data, ensure_ascii=False)
 
 
 
350
  total_length = len(json_str)
351
 
352
  # Calculate overhead for metadata
 
457
  for i, chunk in enumerate(chunks):
458
  filename = f'combined_qr_{int(time.time())}_{i+1}_of_{len(chunks)}.png'
459
  qr_path = generate_stylish_qr(
460
+ data=chunk,
461
  filename=filename,
462
  fill_color="#1a365d", # Deep blue
463
  back_color="#ffffff"
 
472
  for chunk_idx, chunk in enumerate(chunks):
473
  filename = f'item_{idx+1}_chunk_{chunk_idx+1}_of_{len(chunks)}_{int(time.time())}.png'
474
  qr_path = generate_stylish_qr(
475
+ data=chunk,
476
  filename=filename,
477
  fill_color="#1a365d", # Deep blue
478
  back_color="#ffffff"
 
484
  for i, chunk in enumerate(chunks):
485
  filename = f'single_qr_{i+1}_of_{len(chunks)}_{int(time.time())}.png'
486
  qr_path = generate_stylish_qr(
487
+ data=chunk,
488
  filename=filename,
489
  fill_color="#1a365d", # Deep blue
490
  back_color="#ffffff"
 
803
  for url in url_list:
804
  validation = url_processor.validate_url(url)
805
  if validation['is_valid']:
806
+ content = url_processor.fetch_content(url)
807
+ if content:
808
+ results.append({
809
+ 'source': 'url',
810
+ 'url': url,
811
+ 'content': content,
812
+ 'timestamp': datetime.now().isoformat()
813
+ })
 
 
 
 
 
814
 
815
  # Process files
816
  if files: