fsb / Unused Codes /downloader.py
privateone's picture
Code Updates & Optimisations :Server Upgraded-Minor Chnages Due to Download problem
d0b286c
import os
import time
import math
import logging
import asyncio
import traceback
from aiohttp import web
from pyrogram import raw
from aiohttp.http_exceptions import BadStatusLine
#---------------------Local Upload---------------------#
from FileStream.config import Telegram
from FileStream.bot import req_client, FileStream
from FileStream import utils, StartTime, __version__
from FileStream.Tools import mime_identifier, Time_ISTKolNow
from FileStream.bot import MULTI_CLIENTS, WORK_LOADS, ACTIVE_CLIENTS
from FileStream.Exceptions import FIleNotFound, InvalidHash
from FileStream.utils.FileProcessors.custom_ul import TeleUploader
async def media_streamer(request: web.Request, db_id: str, speed: str):
# Get the Range header from the request, default to 0 if not present
range_header = request.headers.get("Range", 0)
client = await req_client()
# Log client info if multi-client mode
if Telegram.MULTI_CLIENT:
logging.info(f"Client {client['index']} is now serving {request.headers.get('X-FORWARDED-FOR', request.remote)}")
# Use an existing ByteStreamer or create a new one
tg_connect = ACTIVE_CLIENTS.get(client['client'], None)
if tg_connect is None:
logging.debug(f"Creating new ByteStreamer object for client {client['index']}")
tg_connect = utils.ByteStreamer(client['client'])
ACTIVE_CLIENTS[client['client']] = tg_connect
else:
tg_connect.update_last_activity()
logging.debug(f"Using cached ByteStreamer object for client {client['index']}")
try:
# Fetch file properties once and use it throughout
logging.debug("Fetching file properties")
file_id = await tg_connect.get_file_properties(db_id, MULTI_CLIENTS)
file_size = file_id.file_size
# Parse range header efficiently
from_bytes, until_bytes = parse_range(range_header, file_size)
# If range is invalid, return a 416 error
if from_bytes is None or until_bytes is None:
return web.Response(
status=416,
body="416: Range not satisfiable",
headers={"Content-Range": f"bytes */{file_size}"},
)
# Set chunk size based on speed
chunk_size = 1024 * 1024 if speed == "FAST" else 512 * 1024
# Ensure we don't go past the file size
until_bytes = min(until_bytes, file_size - 1)
# Compute offset and range parts
offset, first_part_cut, last_part_cut, part_count = compute_offsets(from_bytes, until_bytes, chunk_size)
# Request the file chunks
body = tg_connect.yield_file(
file_id, client['index'], offset, first_part_cut, last_part_cut, part_count, chunk_size
)
# Determine MIME type and filename
mime_type = file_id.mime_type or mimetypes.guess_type(file_id.file_name)[0] or "application/octet-stream"
file_name = utils.get_name(file_id)
disposition = "attachment"
# Return the response with proper headers and status
req_length = until_bytes - from_bytes + 1
return web.Response(
status=206 if range_header else 200,
body=body,
headers={
"Content-Type": mime_type,
"Content-Range": f"bytes {from_bytes}-{until_bytes}/{file_size}",
"Content-Length": str(req_length),
"Content-Disposition": f'{disposition}; filename="{file_name}"',
"Accept-Ranges": "bytes",
},
)
except Exception as e:
logging.error(f"Error in media_streamer: {traceback.format_exc()}")
raise web.HTTPInternalServerError() # Re-raise the exception as a server error
def parse_range(range_header: str, file_size: int):
"""Helper function to parse the range header."""
try:
range_str = range_header.replace("bytes=", "")
from_bytes, until_bytes = range_str.split("-")
from_bytes = int(from_bytes)
until_bytes = int(until_bytes) if until_bytes else file_size - 1
return from_bytes, until_bytes
except ValueError:
return None, None
def compute_offsets(from_bytes: int, until_bytes: int, chunk_size: int):
"""Compute the offsets, cuts, and part counts for file chunking."""
offset = from_bytes - (from_bytes % chunk_size)
first_part_cut = from_bytes - offset
last_part_cut = until_bytes % chunk_size + 1
part_count = math.ceil(until_bytes / chunk_size) - math.floor(offset / chunk_size)
return offset, first_part_cut, last_part_cut, part_count