import json import os import random from abc import ABC, abstractmethod from pathlib import Path from typing import Dict, Optional def default_cache_dir(): return os.path.join(Path.home(), ".triton", "cache") def default_override_dir(): return os.path.join(Path.home(), ".triton", "override") def default_dump_dir(): return os.path.join(Path.home(), ".triton", "dump") class CacheManager(ABC): def __init__(self, key): pass @abstractmethod def get_file(self, filename) -> Optional[str]: pass @abstractmethod def has_file(self, filename) -> bool: pass @abstractmethod def put(self, data, filename, binary=True) -> str: pass @abstractmethod def get_group(self, filename: str) -> Optional[Dict[str, str]]: pass @abstractmethod def put_group(self, filename: str, group: Dict[str, str]): pass class FileCacheManager(CacheManager): def __init__(self, key, override=False, dump=False): self.key = key self.lock_path = None if dump: self.cache_dir = default_dump_dir() self.cache_dir = os.path.join(self.cache_dir, self.key) self.lock_path = os.path.join(self.cache_dir, "lock") os.makedirs(self.cache_dir, exist_ok=True) elif override: self.cache_dir = default_override_dir() self.cache_dir = os.path.join(self.cache_dir, self.key) else: # create cache directory if it doesn't exist self.cache_dir = os.getenv("TRITON_CACHE_DIR", "").strip() or default_cache_dir() if self.cache_dir: self.cache_dir = os.path.join(self.cache_dir, self.key) self.lock_path = os.path.join(self.cache_dir, "lock") os.makedirs(self.cache_dir, exist_ok=True) else: raise RuntimeError("Could not create or locate cache dir") def _make_path(self, filename) -> str: return os.path.join(self.cache_dir, filename) def has_file(self, filename) -> bool: if not self.cache_dir: raise RuntimeError("Could not create or locate cache dir") return os.path.exists(self._make_path(filename)) def get_file(self, filename) -> Optional[str]: if self.has_file(filename): return self._make_path(filename) else: return None def get_group(self, filename: str) -> Optional[Dict[str, str]]: grp_filename = f"__grp__{filename}" if not self.has_file(grp_filename): return None grp_filepath = self._make_path(grp_filename) with open(grp_filepath) as f: grp_data = json.load(f) child_paths = grp_data.get("child_paths", None) # Invalid group data. if child_paths is None: return None result = {} for c in child_paths: p = self._make_path(c) if os.path.exists(p): result[c] = p return result # Note a group of pushed files as being part of a group def put_group(self, filename: str, group: Dict[str, str]) -> str: if not self.cache_dir: raise RuntimeError("Could not create or locate cache dir") grp_contents = json.dumps({"child_paths": sorted(list(group.keys()))}) grp_filename = f"__grp__{filename}" return self.put(grp_contents, grp_filename, binary=False) def put(self, data, filename, binary=True) -> str: if not self.cache_dir: raise RuntimeError("Could not create or locate cache dir") binary = isinstance(data, bytes) if not binary: data = str(data) assert self.lock_path is not None filepath = self._make_path(filename) # Random ID to avoid any collisions rnd_id = random.randint(0, 1000000) # we use the PID incase a bunch of these around so we can see what PID made it pid = os.getpid() # use tempfile to be robust against program interruptions temp_path = f"{filepath}.tmp.pid_{pid}_{rnd_id}" mode = "wb" if binary else "w" with open(temp_path, mode) as f: f.write(data) # Replace is guaranteed to be atomic on POSIX systems if it succeeds # so filepath cannot see a partial write os.replace(temp_path, filepath) return filepath __cache_cls = FileCacheManager __cache_cls_nme = "DEFAULT" def get_cache_manager(key) -> CacheManager: import os user_cache_manager = os.environ.get("TRITON_CACHE_MANAGER", None) global __cache_cls global __cache_cls_nme if user_cache_manager is not None and user_cache_manager != __cache_cls_nme: import importlib module_path, clz_nme = user_cache_manager.split(":") module = importlib.import_module(module_path) __cache_cls = getattr(module, clz_nme) __cache_cls_nme = user_cache_manager return __cache_cls(key) def get_override_manager(key) -> CacheManager: return __cache_cls(key, override=True) def get_dump_manager(key) -> CacheManager: return __cache_cls(key, dump=True)