214 lines
7.6 KiB
Python
214 lines
7.6 KiB
Python
import os
|
|
import json
|
|
import signal
|
|
import subprocess
|
|
import logging
|
|
import time
|
|
import atexit
|
|
from pathlib import Path
|
|
import threading
|
|
import queue
|
|
import sys
|
|
|
|
# Configure logging
|
|
logger = logging.getLogger(__name__)
|
|
|
|
# Configuration
|
|
CONFIG_PATH = './config/main.json'
|
|
CELERY_APP = 'routes.utils.celery_tasks.celery_app'
|
|
CELERY_PROCESS = None
|
|
CONFIG_CHECK_INTERVAL = 30 # seconds
|
|
|
|
class CeleryManager:
|
|
"""
|
|
Manages Celery workers dynamically based on configuration changes.
|
|
"""
|
|
|
|
def __init__(self):
|
|
self.celery_process = None
|
|
self.current_worker_count = 0
|
|
self.monitoring_thread = None
|
|
self.running = False
|
|
self.log_queue = queue.Queue()
|
|
self.output_threads = []
|
|
|
|
def start(self):
|
|
"""Start the Celery manager and initial workers"""
|
|
if self.running:
|
|
return
|
|
|
|
self.running = True
|
|
|
|
# Start initial workers
|
|
self._update_workers()
|
|
|
|
# Start monitoring thread for config changes
|
|
self.monitoring_thread = threading.Thread(target=self._monitor_config, daemon=True)
|
|
self.monitoring_thread.start()
|
|
|
|
# Register shutdown handler
|
|
atexit.register(self.stop)
|
|
|
|
def stop(self):
|
|
"""Stop the Celery manager and all workers"""
|
|
self.running = False
|
|
|
|
# Stop all running threads
|
|
for thread in self.output_threads:
|
|
if thread.is_alive():
|
|
# We can't really stop the threads, but they'll exit on their own
|
|
# when the process is terminated since they're daemon threads
|
|
pass
|
|
|
|
if self.celery_process:
|
|
logger.info("Stopping Celery workers...")
|
|
try:
|
|
# Send SIGTERM to process group
|
|
os.killpg(os.getpgid(self.celery_process.pid), signal.SIGTERM)
|
|
self.celery_process.wait(timeout=5)
|
|
except (subprocess.TimeoutExpired, ProcessLookupError):
|
|
# Force kill if not terminated
|
|
try:
|
|
os.killpg(os.getpgid(self.celery_process.pid), signal.SIGKILL)
|
|
except ProcessLookupError:
|
|
pass
|
|
|
|
self.celery_process = None
|
|
self.current_worker_count = 0
|
|
|
|
def _get_worker_count(self):
|
|
"""Get the configured worker count from config file"""
|
|
try:
|
|
if not Path(CONFIG_PATH).exists():
|
|
return 3 # Default
|
|
|
|
with open(CONFIG_PATH, 'r') as f:
|
|
config = json.load(f)
|
|
|
|
return int(config.get('maxConcurrentDownloads', 3))
|
|
except Exception as e:
|
|
logger.error(f"Error reading worker count from config: {e}")
|
|
return 3 # Default on error
|
|
|
|
def _update_workers(self):
|
|
"""Update workers if needed based on configuration"""
|
|
new_worker_count = self._get_worker_count()
|
|
|
|
if new_worker_count == self.current_worker_count and self.celery_process and self.celery_process.poll() is None:
|
|
return # No change and process is running
|
|
|
|
logger.info(f"Updating Celery workers from {self.current_worker_count} to {new_worker_count}")
|
|
|
|
# Stop existing workers if running
|
|
if self.celery_process:
|
|
try:
|
|
os.killpg(os.getpgid(self.celery_process.pid), signal.SIGTERM)
|
|
self.celery_process.wait(timeout=5)
|
|
except (subprocess.TimeoutExpired, ProcessLookupError):
|
|
try:
|
|
os.killpg(os.getpgid(self.celery_process.pid), signal.SIGKILL)
|
|
except ProcessLookupError:
|
|
pass
|
|
|
|
# Clear output threads list
|
|
self.output_threads = []
|
|
|
|
# Start new workers with updated concurrency
|
|
try:
|
|
# Set environment variables to configure Celery logging
|
|
env = os.environ.copy()
|
|
env['PYTHONUNBUFFERED'] = '1' # Ensure Python output is unbuffered
|
|
|
|
# Construct command with extra logging options
|
|
cmd = [
|
|
'celery',
|
|
'-A', CELERY_APP,
|
|
'worker',
|
|
'--loglevel=info',
|
|
f'--concurrency={new_worker_count}',
|
|
'-Q', 'downloads',
|
|
# Add timestamp to Celery logs
|
|
'--logfile=-', # Output logs to stdout
|
|
'--without-heartbeat', # Reduce log noise
|
|
'--without-gossip', # Reduce log noise
|
|
'--without-mingle' # Reduce log noise
|
|
]
|
|
|
|
self.celery_process = subprocess.Popen(
|
|
cmd,
|
|
stdout=subprocess.PIPE,
|
|
stderr=subprocess.PIPE,
|
|
env=env,
|
|
preexec_fn=os.setsid, # New process group for clean termination
|
|
universal_newlines=True,
|
|
bufsize=1 # Line buffered
|
|
)
|
|
|
|
self.current_worker_count = new_worker_count
|
|
logger.info(f"Started Celery workers with concurrency {new_worker_count}")
|
|
|
|
# Start non-blocking output reader threads for both stdout and stderr
|
|
stdout_thread = threading.Thread(
|
|
target=self._process_output_reader,
|
|
args=(self.celery_process.stdout, "STDOUT"),
|
|
daemon=True
|
|
)
|
|
stdout_thread.start()
|
|
self.output_threads.append(stdout_thread)
|
|
|
|
stderr_thread = threading.Thread(
|
|
target=self._process_output_reader,
|
|
args=(self.celery_process.stderr, "STDERR"),
|
|
daemon=True
|
|
)
|
|
stderr_thread.start()
|
|
self.output_threads.append(stderr_thread)
|
|
|
|
except Exception as e:
|
|
logger.error(f"Error starting Celery workers: {e}")
|
|
|
|
def _process_output_reader(self, pipe, stream_name):
|
|
"""Read and log output from the process"""
|
|
try:
|
|
for line in iter(pipe.readline, ''):
|
|
if not line:
|
|
break
|
|
|
|
line = line.strip()
|
|
if not line:
|
|
continue
|
|
|
|
# Format the message to identify it's from Celery
|
|
if "ERROR" in line or "CRITICAL" in line:
|
|
logger.error(f"Celery[{stream_name}]: {line}")
|
|
elif "WARNING" in line:
|
|
logger.warning(f"Celery[{stream_name}]: {line}")
|
|
elif "DEBUG" in line:
|
|
logger.debug(f"Celery[{stream_name}]: {line}")
|
|
else:
|
|
logger.info(f"Celery[{stream_name}]: {line}")
|
|
|
|
except Exception as e:
|
|
logger.error(f"Error processing Celery output: {e}")
|
|
finally:
|
|
pipe.close()
|
|
|
|
def _monitor_config(self):
|
|
"""Monitor configuration file for changes"""
|
|
logger.info("Starting config monitoring thread")
|
|
last_check_time = 0
|
|
|
|
while self.running:
|
|
try:
|
|
# Check for changes
|
|
if time.time() - last_check_time >= CONFIG_CHECK_INTERVAL:
|
|
self._update_workers()
|
|
last_check_time = time.time()
|
|
|
|
time.sleep(1)
|
|
except Exception as e:
|
|
logger.error(f"Error in config monitoring thread: {e}")
|
|
time.sleep(5) # Wait before retrying
|
|
|
|
# Create single instance
|
|
celery_manager = CeleryManager() |