# Description: This is a Stash plugin which updates Stash if any changes occurs in the Stash library paths. # By David Maisonave (aka Axter) Jul-2024 (https://www.axter.com/) # Get the latest developers version from following link: https://github.com/David-Maisonave/Axter-Stash/tree/main/plugins/ChangeFileMonitor import os import sys import shutil import hashlib import json from pathlib import Path import requests import logging from logging.handlers import RotatingFileHandler import stashapi.log as log # Importing stashapi.log as log for critical events ONLY from stashapi.stashapp import StashInterface from changefilemonitor_settings import config # Import settings from changefilemonitor_settings.py from watchdog.observers import Observer # This is also needed for event attributes import watchdog # pip install watchdog # https://pythonhosted.org/watchdog/ from threading import Lock, Condition from multiprocessing import shared_memory # ********************************************************************** # Constant global variables -------------------------------------------- LOG_FILE_PATH = log_file_path = f"{Path(__file__).resolve().parent}\\{Path(__file__).stem}.log" FORMAT = "[%(asctime)s - LN:%(lineno)s] %(message)s" DEFAULT_ENDPOINT = "http://localhost:9999/graphql" # Default GraphQL endpoint PLUGIN_ARGS = False PLUGIN_ARGS_MODE = False # GraphQL query to fetch all scenes QUERY_ALL_SCENES = """ query AllScenes { allScenes { id updated_at } } """ RFH = RotatingFileHandler( filename=LOG_FILE_PATH, mode='a', maxBytes=2*1024*1024, # Configure logging for this script with max log file size of 2000K backupCount=2, encoding=None, delay=0 ) TIMEOUT = 5 CONTINUE_RUNNING_SIG = 99 # ********************************************************************** # Global variables -------------------------------------------- exitMsg = "Change success!!" mutex = Lock() signal = Condition(mutex) shouldUpdate = False TargetPaths = [] # Configure local log file for plugin within plugin folder having a limited max log file size logging.basicConfig(level=logging.INFO, format=FORMAT, datefmt="%y%m%d %H:%M:%S", handlers=[RFH]) logger = logging.getLogger(Path(__file__).stem) # ********************************************************************** # ---------------------------------------------------------------------- # Code section to fetch variables from Plugin UI and from changefilemonitor_settings.py json_input = json.loads(sys.stdin.read()) FRAGMENT_SERVER = json_input["server_connection"] stash = StashInterface(FRAGMENT_SERVER) PLUGINCONFIGURATION = stash.get_configuration()["plugins"] STASHCONFIGURATION = stash.get_configuration()["general"] STASHPATHSCONFIG = STASHCONFIGURATION['stashes'] stashPaths = [] settings = { "scanModified": False, "recursiveDisabled": False, "zgraphqlEndpoint": DEFAULT_ENDPOINT, "zzdebugTracing": False, "zzdryRun": False, } PLUGIN_ID = "changefilemonitor" if PLUGIN_ID in PLUGINCONFIGURATION: settings.update(PLUGINCONFIGURATION[PLUGIN_ID]) # ---------------------------------------------------------------------- debugTracing = settings["zzdebugTracing"] RECURSIVE = settings["recursiveDisabled"] == False SCAN_MODIFIED = settings["scanModified"] for item in STASHPATHSCONFIG: stashPaths.append(item["path"]) # Extract dry_run setting from settings dry_run = settings["zzdryRun"] dry_run_prefix = '' try: PLUGIN_ARGS = json_input['args'] PLUGIN_ARGS_MODE = json_input['args']["mode"] except: pass logger.info(f"\nStarting (debugTracing={debugTracing}) (dry_run={dry_run}) (PLUGIN_ARGS_MODE={PLUGIN_ARGS_MODE}) (PLUGIN_ARGS={PLUGIN_ARGS})************************************************") if debugTracing: logger.info(f"Debug Tracing (stash.get_configuration()={stash.get_configuration()})................") if debugTracing: logger.info("settings: %s " % (settings,)) if debugTracing: logger.info(f"Debug Tracing (STASHCONFIGURATION={STASHCONFIGURATION})................") if debugTracing: logger.info(f"Debug Tracing (stashPaths={stashPaths})................") if dry_run: logger.info("Dry run mode is enabled.") dry_run_prefix = "Would've " if debugTracing: logger.info("Debug Tracing................") # ToDo: Add split logic here to slpit possible string array into an array endpoint = settings["zgraphqlEndpoint"] # GraphQL endpoint if not endpoint or endpoint == "": endpoint = DEFAULT_ENDPOINT if debugTracing: logger.info(f"Debug Tracing (endpoint={endpoint})................") # ---------------------------------------------------------------------- # ********************************************************************** if debugTracing: logger.info(f"Debug Tracing (SCAN_MODIFIED={SCAN_MODIFIED}) (RECURSIVE={RECURSIVE})................") def start_library_monitor(): global shouldUpdate global TargetPaths try: # Create shared memory buffer which can be used as singleton logic or to get a signal to quit task from external script shm_a = shared_memory.SharedMemory(name="DavidMaisonaveAxter_ChangeFileMonitor", create=True, size=4) except: pass logger.info("Could not open shared memory map. Change File Monitor must be running. Can not run multiple instance of Change File Monitor.") return type(shm_a.buf) shm_buffer = shm_a.buf len(shm_buffer) shm_buffer[0] = CONTINUE_RUNNING_SIG if debugTracing: logger.info(f"Shared memory map opended, and flag set to {shm_buffer[0]}") event_handler = watchdog.events.FileSystemEventHandler() def on_created(event): global shouldUpdate global TargetPaths TargetPaths.append(event.src_path) logger.info(f"CREATE *** '{event.src_path}'") with mutex: shouldUpdate = True signal.notify() def on_deleted(event): global shouldUpdate global TargetPaths TargetPaths.append(event.src_path) logger.info(f"DELETE *** '{event.src_path}'") with mutex: shouldUpdate = True signal.notify() def on_modified(event): global shouldUpdate global TargetPaths if SCAN_MODIFIED: TargetPaths.append(event.src_path) logger.info(f"MODIFIED *** '{event.src_path}'") with mutex: shouldUpdate = True signal.notify() else: if debugTracing: logger.info(f"Ignoring modifications due to plugin UI setting. path='{event.src_path}'") def on_moved(event): global shouldUpdate global TargetPaths TargetPaths.append(event.src_path) TargetPaths.append(event.dest_path) logger.info(f"MOVE *** from '{event.src_path}' to '{event.dest_path}'") with mutex: shouldUpdate = True signal.notify() event_handler.on_created = on_created event_handler.on_deleted = on_deleted event_handler.on_modified = on_modified event_handler.on_moved = on_moved observer = Observer() # Iterate through stashPaths for path in stashPaths: observer.schedule(event_handler, path, recursive=RECURSIVE) if debugTracing: logger.info(f"Observing {path}") observer.start() if debugTracing: logger.info("Starting loop................") try: while True: TmpTargetPaths = [] with mutex: while not shouldUpdate: if debugTracing: logger.info("Wait start................") signal.wait() if debugTracing: logger.info("Wait end................") shouldUpdate = False TmpTargetPaths = [] for TargetPath in TargetPaths: TmpTargetPaths.append(os.path.dirname(TargetPath)) TargetPaths = [] TmpTargetPaths = list(set(TmpTargetPaths)) if TmpTargetPaths != []: logger.info(f"Triggering stash scan for path(s) {TmpTargetPaths}") if not dry_run: stash.metadata_scan(paths=TmpTargetPaths) stash.run_plugin_task(plugin_id=PLUGIN_ID, task_name="Start Library Monitor") if debugTracing: logger.info("Exiting plugin so that metadata_scan task can run.") return else: if debugTracing: logger.info("Nothing to scan.") if shm_buffer[0] != CONTINUE_RUNNING_SIG: logger.info(f"Exiting Change File Monitor. (shm_buffer[0]={shm_buffer[0]})") shm_a.close() shm_a.unlink() # Call unlink only once to release the shared memory time.sleep(1) break except KeyboardInterrupt: observer.stop() if debugTracing: logger.info("Stopping observer................") observer.join() if debugTracing: logger.info("Exiting function................") # stop_library_monitor does not work because only one task can run at a time. # def stop_library_monitor(): # if debugTracing: logger.info("Opening shared memory map.") # try: # shm_a = shared_memory.SharedMemory(name="DavidMaisonaveAxter_ChangeFileMonitor", create=False, size=4) # except: # pass # logger.info("Could not open shared memory map. Change File Monitor must not be running.") # return # type(shm_a.buf) # shm_buffer = shm_a.buf # len(shm_buffer) # shm_buffer[0] = 123 # if debugTracing: logger.info(f"Shared memory map opended, and flag set to {shm_buffer[0]}") # shm_a.close() # shm_a.unlink() # Call unlink only once to release the shared memory # time.sleep(1) # return if PLUGIN_ARGS_MODE == "start_library_monitor": start_library_monitor() if debugTracing: logger.info(f"start_library_monitor EXIT................") # elif PLUGIN_ARGS_MODE == "stop_library_monitor": # stop_library_monitor() # if debugTracing: logger.info(f"stop_library_monitor EXIT................") else: logger.info(f"Nothing to do!!! (PLUGIN_ARGS_MODE={PLUGIN_ARGS_MODE})") if debugTracing: logger.info("\n*********************************\nEXITING ***********************\n*********************************")