#!/usr/bin/python3 """ SkywarnPlus.py v0.8.0 by Mason Nelson =============================================================================== SkywarnPlus is a utility that retrieves severe weather alerts from the National Weather Service and integrates these alerts with an Asterisk/app_rpt based radio repeater controller. This utility is designed to be highly configurable, allowing users to specify particular counties for which to check for alerts, the types of alerts to include or block, and how these alerts are integrated into their radio repeater system. This includes features such as automatic voice alerts and a tail message feature for constant updates. All alerts are sorted by severity and cover a broad range of weather conditions such as hurricane warnings, thunderstorms, heat waves, etc. Configurable through a .yaml file, SkywarnPlus serves as a comprehensive and flexible tool for those who need to stay informed about weather conditions and disseminate this information through their radio repeater system. This file is part of SkywarnPlus. SkywarnPlus is free software: you can redistribute it and/or modify it under the terms of the GNU General Public License as published by the Free Software Foundation, either version 3 of the License, or (at your option) any later version. SkywarnPlus is distributed in the hope that it will be useful, but WITHOUT ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License for more details. You should have received a copy of the GNU General Public License along with SkywarnPlus. If not, see . """ import os import json import logging import requests import shutil import fnmatch import subprocess import time import wave import contextlib import math import sys import itertools from datetime import datetime, timezone, timedelta from dateutil import parser from pydub import AudioSegment from ruamel.yaml import YAML from collections import OrderedDict # Use ruamel.yaml instead of PyYAML yaml = YAML() # Directories and Paths BASE_DIR = os.path.dirname(os.path.realpath(__file__)) CONFIG_PATH = os.path.join(BASE_DIR, "config.yaml") COUNTY_CODES_PATH = os.path.join(BASE_DIR, "CountyCodes.md") # Open and read configuration file with open(CONFIG_PATH, "r") as config_file: config = yaml.load(config_file) config = json.loads(json.dumps(config)) # Convert config to a normal dictionary # Define whether SkywarnPlus is enabled in config.yaml MASTER_ENABLE = config.get("SKYWARNPLUS", {}).get("Enable", False) # Define tmp_dir and sounds_path TMP_DIR = config.get("DEV", {}).get("TmpDir", "/tmp/SkywarnPlus") SOUNDS_PATH = config.get("Alerting", {}).get( "SoundsPath", os.path.join(BASE_DIR, "SOUNDS") ) # Create tmp_dir if it doesn't exist if TMP_DIR: os.makedirs(TMP_DIR, exist_ok=True) else: print("Error: tmp_dir is not set.") # Define Blocked events GLOBAL_BLOCKED_EVENTS = config.get("Blocking", {}).get("GlobalBlockedEvents", []) if GLOBAL_BLOCKED_EVENTS is None: GLOBAL_BLOCKED_EVENTS = [] SAYALERT_BLOCKED_EVENTS = config.get("Blocking", {}).get("SayAlertBlockedEvents", []) if SAYALERT_BLOCKED_EVENTS is None: SAYALERT_BLOCKED_EVENTS = [] TAILMESSAGE_BLOCKED_EVENTS = config.get("Blocking", {}).get( "TailmessageBlockedEvents", [] ) if TAILMESSAGE_BLOCKED_EVENTS is None: TAILMESSAGE_BLOCKED_EVENTS = [] # Define Max Alerts MAX_ALERTS = config.get("Alerting", {}).get("MaxAlerts", 99) # Define audio_delay AUDIO_DELAY = config.get("Asterisk", {}).get("AudioDelay", 0) # Define Tailmessage configuration TAILMESSAGE_CONFIG = config.get("Tailmessage", {}) ENABLE_TAILMESSAGE = TAILMESSAGE_CONFIG.get("Enable", False) TAILMESSAGE_FILE = TAILMESSAGE_CONFIG.get( "TailmessagePath", os.path.join(TMP_DIR, "wx-tail.wav") ) # Define IDChange configuration IDCHANGE_CONFIG = config.get("IDChange", {}) ENABLE_IDCHANGE = IDCHANGE_CONFIG.get("Enable", False) # Data file path DATA_FILE = os.path.join(TMP_DIR, "data.json") # Tones directory TONE_DIR = config["CourtesyTones"].get("ToneDir", os.path.join(SOUNDS_PATH, "TONES")) # Define possible alert strings ALERT_STRINGS = [ "911 Telephone Outage Emergency", "Administrative Message", "Air Quality Alert", "Air Stagnation Advisory", "Arroyo And Small Stream Flood Advisory", "Ashfall Advisory", "Ashfall Warning", "Avalanche Advisory", "Avalanche Warning", "Avalanche Watch", "Beach Hazards Statement", "Blizzard Warning", "Blizzard Watch", "Blowing Dust Advisory", "Blowing Dust Warning", "Brisk Wind Advisory", "Child Abduction Emergency", "Civil Danger Warning", "Civil Emergency Message", "Coastal Flood Advisory", "Coastal Flood Statement", "Coastal Flood Warning", "Coastal Flood Watch", "Dense Fog Advisory", "Dense Smoke Advisory", "Dust Advisory", "Dust Storm Warning", "Earthquake Warning", "Evacuation - Immediate", "Excessive Heat Warning", "Excessive Heat Watch", "Extreme Cold Warning", "Extreme Cold Watch", "Extreme Fire Danger", "Extreme Wind Warning", "Fire Warning", "Fire Weather Watch", "Flash Flood Statement", "Flash Flood Warning", "Flash Flood Watch", "Flood Advisory", "Flood Statement", "Flood Warning", "Flood Watch", "Freeze Warning", "Freeze Watch", "Freezing Fog Advisory", "Freezing Rain Advisory", "Freezing Spray Advisory", "Frost Advisory", "Gale Warning", "Gale Watch", "Hard Freeze Warning", "Hard Freeze Watch", "Hazardous Materials Warning", "Hazardous Seas Warning", "Hazardous Seas Watch", "Hazardous Weather Outlook", "Heat Advisory", "Heavy Freezing Spray Warning", "Heavy Freezing Spray Watch", "High Surf Advisory", "High Surf Warning", "High Wind Warning", "High Wind Watch", "Hurricane Force Wind Warning", "Hurricane Force Wind Watch", "Hurricane Local Statement", "Hurricane Warning", "Hurricane Watch", "Hydrologic Advisory", "Hydrologic Outlook", "Ice Storm Warning", "Lake Effect Snow Advisory", "Lake Effect Snow Warning", "Lake Effect Snow Watch", "Lake Wind Advisory", "Lakeshore Flood Advisory", "Lakeshore Flood Statement", "Lakeshore Flood Warning", "Lakeshore Flood Watch", "Law Enforcement Warning", "Local Area Emergency", "Low Water Advisory", "Marine Weather Statement", "Nuclear Power Plant Warning", "Radiological Hazard Warning", "Red Flag Warning", "Rip Current Statement", "Severe Thunderstorm Warning", "Severe Thunderstorm Watch", "Severe Weather Statement", "Shelter In Place Warning", "Short Term Forecast", "Small Craft Advisory", "Small Craft Advisory For Hazardous Seas", "Small Craft Advisory For Rough Bar", "Small Craft Advisory For Winds", "Small Stream Flood Advisory", "Snow Squall Warning", "Special Marine Warning", "Special Weather Statement", "Storm Surge Warning", "Storm Surge Watch", "Storm Warning", "Storm Watch", "Test", "Tornado Warning", "Tornado Watch", "Tropical Depression Local Statement", "Tropical Storm Local Statement", "Tropical Storm Warning", "Tropical Storm Watch", "Tsunami Advisory", "Tsunami Warning", "Tsunami Watch", "Typhoon Local Statement", "Typhoon Warning", "Typhoon Watch", "Urban And Small Stream Flood Advisory", "Volcano Warning", "Wind Advisory", "Wind Chill Advisory", "Wind Chill Warning", "Wind Chill Watch", "Winter Storm Warning", "Winter Storm Watch", "Winter Weather Advisory", ] # Generate the WA list based on the length of WS ALERT_INDEXES = [str(i + 1) for i in range(len(ALERT_STRINGS))] # Test if the script needs to start from a clean slate CLEANSLATE = config.get("DEV", {}).get("CLEANSLATE", False) if CLEANSLATE: shutil.rmtree(TMP_DIR) os.mkdir(TMP_DIR) # Logging setup LOG_CONFIG = config.get("Logging", {}) ENABLE_DEBUG = LOG_CONFIG.get("Debug", False) LOG_FILE = LOG_CONFIG.get("LogPath", os.path.join(TMP_DIR, "SkywarnPlus.log")) # Set up logging LOGGER = logging.getLogger(__name__) LOGGER.setLevel(logging.DEBUG if ENABLE_DEBUG else logging.INFO) # Set up log message formatting LOG_FORMATTER = logging.Formatter("%(asctime)s %(levelname)s %(message)s") # Set up console log handler C_HANDLER = logging.StreamHandler() C_HANDLER.setFormatter(LOG_FORMATTER) LOGGER.addHandler(C_HANDLER) # Set up file log handler F_HANDLER = logging.FileHandler(LOG_FILE) F_HANDLER.setFormatter(LOG_FORMATTER) LOGGER.addHandler(F_HANDLER) # Get the "CountyCodes" from the config COUNTY_CODES_CONFIG = config.get("Alerting", {}).get("CountyCodes", []) # Initialize COUNTY_CODES and COUNTY_WAVS COUNTY_CODES = [] COUNTY_WAVS = [] # Check the type of "CountyCodes" config to handle both list and dictionary if isinstance(COUNTY_CODES_CONFIG, list): # If it's a list, check if it's a list of strings or dictionaries if all(isinstance(i, str) for i in COUNTY_CODES_CONFIG): # It's the old format and we can use it directly COUNTY_CODES = COUNTY_CODES_CONFIG elif all(isinstance(i, dict) for i in COUNTY_CODES_CONFIG): # It's a list of dictionaries with WAV files # We need to separate the county codes and the WAVs for dic in COUNTY_CODES_CONFIG: for key, value in dic.items(): COUNTY_CODES.append(key) COUNTY_WAVS.append(value) elif isinstance(COUNTY_CODES_CONFIG, dict): # If it's a dictionary, it's got WAV files # We need to separate the county codes and the WAVs COUNTY_CODES = list(COUNTY_CODES_CONFIG.keys()) COUNTY_WAVS = list(COUNTY_CODES_CONFIG.values()) else: # Invalid format, set it to an empty list COUNTY_CODES = [] COUNTY_WAVS = [] # Log some debugging information LOGGER.debug("Base directory: %s", BASE_DIR) LOGGER.debug("Temporary directory: %s", TMP_DIR) LOGGER.debug("Sounds path: %s", SOUNDS_PATH) LOGGER.debug("Tailmessage path: %s", TAILMESSAGE_FILE) LOGGER.debug("Global Blocked events: %s", GLOBAL_BLOCKED_EVENTS) LOGGER.debug("SayAlert Blocked events: %s", SAYALERT_BLOCKED_EVENTS) LOGGER.debug("Tailmessage Blocked events: %s", TAILMESSAGE_BLOCKED_EVENTS) def load_state(): """ Load the state from the state file if it exists, else return an initial state. The state file is expected to be a JSON file. If certain keys are missing in the loaded state, this function will provide default values for those keys. """ # Check if the state data file exists if os.path.exists(DATA_FILE): with open(DATA_FILE, "r") as file: state = json.load(file) # Ensure 'alertscript_alerts' key is present in the state, default to an empty list state["alertscript_alerts"] = state.get("alertscript_alerts", []) # Process 'last_alerts' key to maintain the order of alerts using OrderedDict # This step is necessary because JSON does not preserve order by default last_alerts = state.get("last_alerts", []) state["last_alerts"] = OrderedDict((x[0], x[1]) for x in last_alerts) # Ensure 'last_sayalert' and 'active_alerts' keys are present in the state state["last_sayalert"] = state.get("last_sayalert", []) state["active_alerts"] = state.get("active_alerts", []) return state # If the state data file does not exist, return a default state else: return { "ct": None, "id": None, "alertscript_alerts": [], "last_alerts": OrderedDict(), "last_sayalert": [], "active_alerts": [], } def save_state(state): """ Save the state to the state file. The state is saved as a JSON file. The function ensures certain keys in the state are converted to lists before saving, ensuring consistency and ease of processing when the state is later loaded. """ # Convert 'alertscript_alerts', 'last_sayalert', and 'active_alerts' keys to lists # This ensures consistency in data format, especially useful when loading the state later state["alertscript_alerts"] = list(state["alertscript_alerts"]) state["last_sayalert"] = list(state["last_sayalert"]) state["active_alerts"] = list(state["active_alerts"]) # Convert 'last_alerts' from OrderedDict to list of items # This step is necessary because JSON does not natively support OrderedDict state["last_alerts"] = list(state["last_alerts"].items()) # Save the state to the data file in a formatted manner with open(DATA_FILE, "w") as file: json.dump(state, file, ensure_ascii=False, indent=4) def get_alerts(countyCodes): """ Retrieves severe weather alerts for specified county codes and processes them. """ # Define mappings to convert severity levels from various terminologies to a numeric scale severity_mapping_api = { "Extreme": 4, "Severe": 3, "Moderate": 2, "Minor": 1, "Unknown": 0, } severity_mapping_words = {"Warning": 4, "Watch": 3, "Advisory": 2, "Statement": 1} # Initialize storage for the alerts and a set to keep track of processed alerts alerts = OrderedDict() seen_alerts = set() # Log current time for reference current_time = datetime.now(timezone.utc) LOGGER.debug("getAlerts: Current time: %s", current_time) # Handle alert injection for development/testing purposes if config.get("DEV", {}).get("INJECT", False): LOGGER.debug("getAlerts: DEV Alert Injection Enabled") injected_alerts = config["DEV"].get("INJECTALERTS", []) LOGGER.debug("getAlerts: Injecting alerts: %s", injected_alerts) max_counties = len(countyCodes) # Assuming countyCodes is a list of counties county_codes_cycle = itertools.cycle(countyCodes) county_assignment_counter = 1 for alert_info in injected_alerts: if isinstance(alert_info, dict): alert_title = alert_info.get("Title", "") specified_counties = alert_info.get("CountyCodes", []) else: continue # Ignore if not a dictionary last_word = alert_title.split()[-1] severity = severity_mapping_words.get(last_word, 0) description = "This alert was manually injected as a test." end_time_str = alert_info.get("EndTime") county_data = [] # If no counties are specified, use the ones provided to the function in an increasing manner if not specified_counties: # Limit the number of counties assigned to not exceed max_counties counties_to_assign = min(county_assignment_counter, max_counties) specified_counties = [ next(county_codes_cycle) for _ in range(counties_to_assign) ] county_assignment_counter += 1 # Increment for the next injected alert for county in specified_counties: if county not in countyCodes: LOGGER.error( "Specified county code '%s' is not defined in the config. Using next available county code from the config.", county, ) county = next(county_codes_cycle) end_time = ( datetime.strptime(end_time_str, "%Y-%m-%dT%H:%M:%SZ") if end_time_str else current_time + timedelta(hours=1) ) county_data.append( { "county_code": county, "severity": severity, "description": description, "end_time_utc": end_time.strftime("%Y-%m-%dT%H:%M:%S.%fZ"), } ) alerts[alert_title] = ( county_data # Add the list of dictionaries to the alert ) # If injected alerts are used, we return them here and don't proceed with the function. return sort_alerts(alerts) # Configuration specifies whether to use 'effective' or 'onset' time for alerts. # Depending on the configuration, we set the appropriate keys for start and end time. timetype_mode = config.get("Alerting", {}).get("TimeType", "onset") if timetype_mode == "effective": LOGGER.debug("getAlerts: Using effective time for alerting") time_type_start = "effective" time_type_end = "expires" elif timetype_mode == "onset": LOGGER.debug("getAlerts: Using onset time for alerting") time_type_start = "onset" time_type_end = "ends" else: LOGGER.error( "getAlerts: Invalid TimeType specified in config.yaml. Using onset time instead." ) time_type_start = "onset" time_type_end = "ends" # Loop over each county code and retrieve alerts from the API. for countyCode in countyCodes: url = "https://api.weather.gov/alerts/active?zone={}".format(countyCode) # # WARNING: ONLY USE THIS FOR DEVELOPMENT PURPOSES # THIS URL WILL RETURN ALL ACTIVE ALERTS IN THE UNITED STATES # url = "https://api.weather.gov/alerts/active" try: # If we can get a successful response from the API, we process the alerts from the response. response = requests.get(url) response.raise_for_status() LOGGER.debug( "getAlerts: Checking for alerts in %s at URL: %s", countyCode, url ) alert_data = response.json() for feature in alert_data["features"]: # Extract start and end times. If end time is missing, use 'expires' time. start = feature["properties"].get(time_type_start) end = feature["properties"].get(time_type_end) if not end: end = feature["properties"].get("expires") LOGGER.debug( 'getAlerts: %s has no "%s" time, using "expires" time instead: %s', feature["properties"]["event"], time_type_end, end, ) if start and end: # If both start and end times are available, convert them to datetime objects. start_time = parser.isoparse(start) end_time = parser.isoparse(end) # Convert alert times to UTC. start_time_utc = start_time.astimezone(timezone.utc) end_time_utc = end_time.astimezone(timezone.utc) event = feature["properties"]["event"] # If the current time is within the alert's active period, we process it further. if start_time_utc <= current_time < end_time_utc: description = feature["properties"].get("description", "") severity = feature["properties"].get("severity", None) # Check if the event is globally blocked as per the configuration. If it is, skip this event. is_blocked = False for global_blocked_event in GLOBAL_BLOCKED_EVENTS: if fnmatch.fnmatch(event, global_blocked_event): LOGGER.debug( "getAlerts: Globally Blocking %s as per configuration", event, ) is_blocked = True break # If the event is globally blocked, we skip the remaining code in this loop iteration and move to the next one. if is_blocked: continue # Determine severity from event name or API's severity value. if severity is None: last_word = event.split()[-1] severity = severity_mapping_words.get(last_word, 0) else: severity = severity_mapping_api.get(severity, 0) # Log the alerts and their severity level for debugging purposes. LOGGER.debug( "getAlerts: %s - %s - Severity: %s", countyCode, event, severity, ) # Check if the event has already been processed (seen). # If it has been seen, we add a new dictionary to its list of alerts. This dictionary contains details about the alert. if event in seen_alerts: alerts[event].append( { "county_code": countyCode, # the county code the alert is for "severity": severity, # the severity level of the alert "description": description, # a description of the alert "end_time_utc": end_time_utc.strftime( "%Y-%m-%dT%H:%M:%S.%fZ" ), # the time the alert ends in UTC } ) # If the event hasn't been seen before, we create a new list entry in the 'alerts' dictionary for this event. else: alerts[event] = [ { "county_code": countyCode, # the county code the alert is for "severity": severity, # the severity level of the alert "description": description, # a description of the alert "end_time_utc": end_time_utc.strftime( "%Y-%m-%dT%H:%M:%S.%fZ" ), # the time the alert ends in UTC } ] # Add the event to the set of seen alerts. seen_alerts.add(event) # If the current time is not within the alert's active period, we skip it. else: time_difference = time_until(start_time_utc, current_time) LOGGER.debug( "getAlerts: Skipping %s, not active for another %s.", event, time_difference, ) LOGGER.debug( "Current time: %s | Alert start: %s | Alert end %s", current_time, start_time_utc, end_time_utc, ) else: LOGGER.debug( "getAlerts: Skipping %s, missing start or end time.", feature["properties"]["event"], ) except requests.exceptions.RequestException as e: LOGGER.debug("Failed to retrieve alerts for %s. Reason: %s", countyCode, e) LOGGER.debug("API unreachable. Using stored data instead.") # Load alerts from data.json if os.path.isfile(DATA_FILE): with open(DATA_FILE) as f: data = json.load(f) stored_alerts = data.get("last_alerts", []) # Filter alerts by end_time_utc current_time_str = datetime.now(timezone.utc).strftime( "%Y-%m-%dT%H:%M:%S.%fZ" ) LOGGER.debug("Current time: %s", current_time_str) alerts = {} for stored_alert in stored_alerts: event = stored_alert[0] alert_list = stored_alert[1] alerts[event] = [] for alert in alert_list: end_time_str = alert["end_time_utc"] if parser.parse(end_time_str) >= parser.parse( current_time_str ): LOGGER.debug( "getAlerts: Keeping %s because it does not expire until %s", event, end_time_str, ) alerts[event].append(alert) else: LOGGER.debug( "getAlerts: Removing %s because it expired at %s", event, end_time_str, ) else: LOGGER.error("No stored data available.") break alerts = OrderedDict( sorted( alerts.items(), key=lambda item: ( max([x["severity"] for x in item[1]]), # Max Severity severity_mapping_words.get(item[0].split()[-1], 0), # Words severity ), reverse=True, ) ) # If the number of alerts exceeds the maximum defined constant, we truncate the alerts. alerts = OrderedDict(list(alerts.items())[:MAX_ALERTS]) return alerts def sort_alerts(alerts): """ Sorts and limits the alerts based on their severity and word severity. """ # Define mapping for converting common alert terminologies to a numeric severity scale severity_mapping_words = {"Warning": 4, "Watch": 3, "Advisory": 2, "Statement": 1} # Sort the alerts first by their maximum severity, and then by their word severity sorted_alerts = OrderedDict( sorted( alerts.items(), key=lambda item: ( max([x["severity"] for x in item[1]]), # Max Severity for the alert severity_mapping_words.get( item[0].split()[-1], 0 ), # Severity based on last word in the alert title ), reverse=True, # Sort in descending order ) ) # Truncate the list of alerts to the maximum allowed number (MAX_ALERTS) limited_alerts = OrderedDict(list(sorted_alerts.items())[:MAX_ALERTS]) return limited_alerts def time_until(start_time_utc, current_time): """ Calculate the time difference between two datetime objects and returns it as a formatted string. """ # Calculate the time difference between the two datetime objects delta = start_time_utc - current_time # Determine the sign (used for formatting) sign = "-" if delta < timedelta(0) else "" # Decompose the time difference into days, hours, and minutes days, remainder = divmod(abs(delta.total_seconds()), 86400) hours, remainder = divmod(remainder, 3600) minutes, _ = divmod(remainder, 60) # Return the time difference as a formatted string return "{}{} days, {} hours, {} minutes".format( sign, int(days), int(hours), int(minutes) ) def say_alerts(alerts): """ Generate and broadcast severe weather alert sounds on Asterisk. """ # Load the current state state = load_state() # Extract only the alert names and associated counties alert_names_and_counties = { alert: [county["county_code"] for county in counties] for alert, counties in alerts.items() } # Filter out alerts that are blocked based on configuration filtered_alerts_and_counties = {} for alert, county_codes in alert_names_and_counties.items(): if any( fnmatch.fnmatch(alert, blocked_event) for blocked_event in SAYALERT_BLOCKED_EVENTS ): LOGGER.debug("sayAlert: blocking %s as per configuration", alert) continue filtered_alerts_and_counties[alert] = county_codes # Check if the filtered alerts are the same as the last run if filtered_alerts_and_counties == state.get("last_sayalert", {}): LOGGER.debug( "sayAlert: alerts and counties are the same as the last broadcast - skipping." ) return # Update the state with the current alerts state["last_sayalert"] = filtered_alerts_and_counties save_state(state) # Initialize the audio segments and paths alert_file = "{}/alert.wav".format(TMP_DIR) word_space = AudioSegment.silent(duration=600) sound_effect = AudioSegment.from_wav( os.path.join( SOUNDS_PATH, "ALERTS", "EFFECTS", config.get("Alerting", {}).get("AlertSeperator", "Woodblock.wav"), ) ) intro_effect = AudioSegment.from_wav( os.path.join( SOUNDS_PATH, "ALERTS", "EFFECTS", config.get("Alerting", {}).get("AlertSound", "Duncecap.wav.wav"), ) ) combined_sound = ( intro_effect + word_space + AudioSegment.from_wav(os.path.join(SOUNDS_PATH, "ALERTS", "SWP_148.wav")) ) # Build the combined sound with alerts and county names alert_count = 0 for alert, counties in alerts.items(): if alert in filtered_alerts_and_counties: try: descriptions = [county["description"] for county in counties] end_times = [county["end_time_utc"] for county in counties] index = ALERT_STRINGS.index(alert) audio_file = AudioSegment.from_wav( os.path.join( SOUNDS_PATH, "ALERTS", "SWP_{}.wav".format(ALERT_INDEXES[index]) ) ) combined_sound += sound_effect + audio_file LOGGER.debug( "sayAlert: Added %s (SWP_%s.wav) to alert sound", alert, ALERT_INDEXES[index], ) if config["Alerting"]["WithMultiples"]: if len(set(descriptions)) > 1 or len(set(end_times)) > 1: LOGGER.debug( "sayAlert: Found multiple unique instances of the alert %s", alert, ) multiples_sound = AudioSegment.from_wav( os.path.join(SOUNDS_PATH, "ALERTS", "SWP_149.wav") ) combined_sound += ( AudioSegment.silent(duration=200) + multiples_sound ) alert_count += 1 added_county_codes = set() for county in counties: if counties.index(county) == 0: word_space = AudioSegment.silent(duration=600) else: word_space = AudioSegment.silent(duration=400) county_code = county["county_code"] if ( COUNTY_WAVS and county_code in COUNTY_CODES and county_code not in added_county_codes ): index = COUNTY_CODES.index(county_code) county_name_file = COUNTY_WAVS[index] LOGGER.debug( "sayAlert: Adding %s ID %s to %s", county_code, county_name_file, alert, ) try: combined_sound += word_space + AudioSegment.from_wav( os.path.join(SOUNDS_PATH, county_name_file) ) except FileNotFoundError: LOGGER.error( "sayAlert: County audio file not found: %s", os.path.join(SOUNDS_PATH, county_name_file), ) added_county_codes.add(county_code) if counties.index(county) == len(counties) - 1: combined_sound += AudioSegment.silent(duration=600) except ValueError: LOGGER.error("sayAlert: Alert not found: %s", alert) except FileNotFoundError: LOGGER.error( "sayAlert: Alert audio file not found: %s/ALERTS/SWP_%s.wav", SOUNDS_PATH, ALERT_INDEXES[index], ) if alert_count == 0: LOGGER.debug("sayAlert: All alerts were blocked, not broadcasting any alerts.") return alert_suffix = config.get("Alerting", {}).get("SayAlertSuffix", None) if alert_suffix is not None: suffix_silence = AudioSegment.silent(duration=600) LOGGER.debug("sayAlert: Adding alert suffix %s", alert_suffix) suffix_file = os.path.join(SOUNDS_PATH, alert_suffix) suffix_sound = AudioSegment.from_wav(suffix_file) combined_sound += suffix_silence + suffix_sound if AUDIO_DELAY > 0: LOGGER.debug("sayAlert: Prepending audio with %sms of silence", AUDIO_DELAY) silence = AudioSegment.silent(duration=AUDIO_DELAY) combined_sound = silence + combined_sound LOGGER.debug("sayAlert: Exporting alert sound to %s", alert_file) converted_combined_sound = convert_audio(combined_sound) converted_combined_sound.export(alert_file, format="wav") LOGGER.debug("sayAlert: Replacing tailmessage with silence") silence = AudioSegment.silent(duration=100) converted_silence = convert_audio(silence) converted_silence.export(TAILMESSAGE_FILE, format="wav") node_numbers = config.get("Asterisk", {}).get("Nodes", []) for node_number in node_numbers: LOGGER.info("Broadcasting alert on node %s", node_number) command = '/usr/sbin/asterisk -rx "rpt localplay {} {}"'.format( node_number, os.path.splitext(os.path.abspath(alert_file))[0] ) subprocess.run(command, shell=True) # Get the duration of the alert_file with contextlib.closing(wave.open(alert_file, "r")) as f: frames = f.getnframes() rate = f.getframerate() duration = math.ceil(frames / float(rate)) wait_time = duration + 10 LOGGER.debug( "Waiting %s seconds for Asterisk to make announcement to avoid doubling alerts with tailmessage...", wait_time, ) time.sleep(wait_time) def say_allclear(): """ Generate and broadcast 'all clear' message on Asterisk. """ # Load current state and clear the last_sayalert list state = load_state() state["last_sayalert"] = [] save_state(state) # Define file paths for the sounds all_clear_sound_file = os.path.join( config.get("Alerting", {}).get("SoundsPath"), "ALERTS", "EFFECTS", config.get("Alerting", {}).get("AllClearSound"), ) swp_147_file = os.path.join(SOUNDS_PATH, "ALERTS", "SWP_147.wav") # Load sound files into AudioSegment objects all_clear_sound = AudioSegment.from_wav(all_clear_sound_file) swp_147_sound = AudioSegment.from_wav(swp_147_file) # Generate silence for spacing between sounds silence = AudioSegment.silent(duration=600) # 600 ms of silence # Combine the "all clear" sound and SWP_147 sound with the configured silence between them combined_sound = all_clear_sound + silence + swp_147_sound # Add a delay before the sound if configured if AUDIO_DELAY > 0: LOGGER.debug("sayAllClear: Prepending audio with %sms of silence", AUDIO_DELAY) delay_silence = AudioSegment.silent(duration=AUDIO_DELAY) combined_sound = delay_silence + combined_sound # Append a suffix to the sound if configured if config.get("Alerting", {}).get("SayAllClearSuffix", None) is not None: suffix_silence = AudioSegment.silent( duration=600 ) # 600ms silence before the suffix suffix_file = os.path.join( SOUNDS_PATH, config.get("Alerting", {}).get("SayAllClearSuffix") ) LOGGER.debug("sayAllClear: Adding all clear suffix %s", suffix_file) suffix_sound = AudioSegment.from_wav(suffix_file) combined_sound += ( suffix_silence + suffix_sound ) # Append the silence and then the suffix to the combined sound # Now, convert the final combined sound converted_combined_sound = convert_audio(combined_sound) # Export the final converted sound to a file all_clear_file = os.path.join(TMP_DIR, "allclear.wav") converted_combined_sound.export(all_clear_file, format="wav") # Play the "all clear" sound on the configured Asterisk nodes node_numbers = config.get("Asterisk", {}).get("Nodes", []) for node_number in node_numbers: LOGGER.info("Broadcasting all clear message on node %s", node_number) command = '/usr/sbin/asterisk -rx "rpt localplay {} {}"'.format( node_number, os.path.splitext(os.path.abspath(all_clear_file))[0] ) subprocess.run(command, shell=True) def build_tailmessage(alerts): """ Build a tailmessage, which is a short message appended to the end of a transmission to update on the weather conditions. """ # Determine whether the user has enabled county identifiers county_identifiers = config.get("Tailmessage", {}).get("TailmessageCounties", False) # Extract only the alert names from the OrderedDict keys alert_names = [alert for alert in alerts.keys()] # Get the suffix config tailmessage_suffix = config.get("Tailmessage", {}).get("TailmessageSuffix", None) # If alerts is empty if not alerts: LOGGER.debug("buildTailMessage: No alerts, creating silent tailmessage") silence = AudioSegment.silent(duration=100) converted_silence = convert_audio(silence) converted_silence.export(TAILMESSAGE_FILE, format="wav") return combined_sound = AudioSegment.empty() sound_effect = AudioSegment.from_wav( os.path.join( SOUNDS_PATH, "ALERTS", "EFFECTS", config.get("Alerting", {}).get("AlertSeperator", "Woodblock.wav"), ) ) for ( alert, counties, ) in alerts.items(): # Now we loop over both alert name and its associated counties added_counties = set() if any( fnmatch.fnmatch(alert, blocked_event) for blocked_event in TAILMESSAGE_BLOCKED_EVENTS ): LOGGER.debug( "buildTailMessage: Alert blocked by TailmessageBlockedEvents: %s", alert ) continue try: index = ALERT_STRINGS.index(alert) audio_file = AudioSegment.from_wav( os.path.join( SOUNDS_PATH, "ALERTS", "SWP_{}.wav".format(ALERT_INDEXES[index]) ) ) combined_sound += sound_effect + audio_file LOGGER.debug( "buildTailMessage: Added %s (SWP_%s.wav) to tailmessage", alert, ALERT_INDEXES[index], ) descriptions = [county["description"] for county in counties] end_times = [county["end_time_utc"] for county in counties] if config["Alerting"]["WithMultiples"]: if len(set(descriptions)) > 1 or len(set(end_times)) > 1: LOGGER.debug( "buildTailMessage: Found multiple unique instances of the alert %s", alert, ) multiples_sound = AudioSegment.from_wav( os.path.join(SOUNDS_PATH, "ALERTS", "SWP_149.wav") ) combined_sound += ( AudioSegment.silent(duration=200) + multiples_sound ) # Add county names if they exist if county_identifiers: for county in counties: # if its the first county, word_space is 600ms of silence. else it is 400ms if counties.index(county) == 0: word_space = AudioSegment.silent(duration=600) else: word_space = AudioSegment.silent(duration=400) county_code = county["county_code"] if ( COUNTY_WAVS and county_code in COUNTY_CODES and county_code not in added_counties ): index = COUNTY_CODES.index(county_code) county_name_file = COUNTY_WAVS[index] LOGGER.debug( "buildTailMessage: Adding %s ID %s to %s", county_code, county_name_file, alert, ) try: combined_sound += word_space + AudioSegment.from_wav( os.path.join(SOUNDS_PATH, county_name_file) ) # if this is the last county name, add 600ms of silence after the county name if counties.index(county) == len(counties) - 1: combined_sound += AudioSegment.silent(duration=600) added_counties.add(county_code) except FileNotFoundError: LOGGER.error( "buildTailMessage: Audio file not found: %s", os.path.join(SOUNDS_PATH, county_name_file), ) except ValueError: LOGGER.error("Alert not found: %s", alert) except FileNotFoundError: LOGGER.error( "buildTailMessage: Audio file not found: %s/ALERTS/SWP_%s.wav", SOUNDS_PATH, ALERT_INDEXES[index], ) if combined_sound.empty(): LOGGER.debug( "buildTailMessage: All alerts were blocked, creating silent tailmessage" ) combined_sound = AudioSegment.silent(duration=100) elif tailmessage_suffix is not None: suffix_silence = AudioSegment.silent(duration=1000) LOGGER.debug( "buildTailMessage: Adding tailmessage suffix %s", tailmessage_suffix ) suffix_file = os.path.join(SOUNDS_PATH, tailmessage_suffix) suffix_sound = AudioSegment.from_wav(suffix_file) combined_sound += suffix_silence + suffix_sound if AUDIO_DELAY > 0: LOGGER.debug( "buildTailMessage: Prepending audio with %sms of silence", AUDIO_DELAY ) silence = AudioSegment.silent(duration=AUDIO_DELAY) combined_sound = silence + combined_sound converted_combined_sound = convert_audio(combined_sound) LOGGER.info("Built new tailmessage") LOGGER.debug("buildTailMessage: Exporting tailmessage to %s", TAILMESSAGE_FILE) converted_combined_sound.export(TAILMESSAGE_FILE, format="wav") def alert_script(alerts): """ This function reads a list of alerts, then performs actions based on the alert triggers defined in the global configuration file. It supports wildcard matching for triggers and includes functionality to execute commands specifically when transitioning between zero and non-zero active alerts. """ LOGGER.debug("Starting alert_script with alerts: %s", alerts) # Load the saved state state = load_state() LOGGER.debug("Loaded state: %s", state) # Determine the previous and current count of active alerts previous_active_count = len(state.get("active_alerts", [])) LOGGER.debug("Previous active alerts count: %s", previous_active_count) processed_alerts = set( state["alertscript_alerts"] ) # Convert to a set for easier processing active_alerts = set( state.get("active_alerts", []) ) # Load active alerts from state, also as a set LOGGER.debug("Processed alerts from state: %s", processed_alerts) LOGGER.debug("Active alerts from state: %s", active_alerts) # Extract only the alert names from the OrderedDict keys alert_names = set([alert for alert in alerts.keys()]) LOGGER.debug("Extracted alert names: %s", alert_names) # Identify new alerts and cleared alerts new_alerts = alert_names - active_alerts cleared_alerts = active_alerts - alert_names LOGGER.debug("New alerts: %s", new_alerts) LOGGER.debug("Cleared alerts: %s", cleared_alerts) # Update the active alerts in the state state["active_alerts"] = list( alert_names ) # Convert back to list for JSON serialization LOGGER.debug("Updated active alerts in state: %s", state["active_alerts"]) # Fetch AlertScript configuration from global_config alertScript_config = config.get("AlertScript", {}) LOGGER.debug("AlertScript configuration: %s", alertScript_config) # Determine the current count of active alerts after update current_active_count = len(state["active_alerts"]) LOGGER.debug("Current active alerts count: %s", current_active_count) # Check for transition from zero to non-zero active alerts and execute ActiveCommands if previous_active_count == 0 and current_active_count > 0: active_commands = alertScript_config.get("ActiveCommands", []) if active_commands: for command in active_commands: if command["Type"].upper() == "BASH": for cmd in command["Commands"]: LOGGER.info("Executing Active BASH Command: %s", cmd) subprocess.run(cmd, shell=True) elif command["Type"].upper() == "DTMF": for node in command["Nodes"]: for cmd in command["Commands"]: dtmf_cmd = 'asterisk -rx "rpt fun {} {}"'.format(node, cmd) LOGGER.info("Executing Active DTMF Command: %s", dtmf_cmd) subprocess.run(dtmf_cmd, shell=True) # Check for transition from non-zero to zero active alerts and execute InactiveCommands if previous_active_count > 0 and current_active_count == 0: inactive_commands = alertScript_config.get("InactiveCommands", []) if inactive_commands: for command in inactive_commands: if command["Type"].upper() == "BASH": for cmd in command["Commands"]: LOGGER.info("Executing Inactive BASH Command: %s", cmd) subprocess.run(cmd, shell=True) elif command["Type"].upper() == "DTMF": for node in command["Nodes"]: for cmd in command["Commands"]: dtmf_cmd = 'asterisk -rx "rpt fun {} {}"'.format(node, cmd) LOGGER.info("Executing Inactive DTMF Command: %s", dtmf_cmd) subprocess.run(dtmf_cmd, shell=True) # Fetch Mappings from AlertScript configuration mappings = alertScript_config.get("Mappings", []) if mappings is None: mappings = [] LOGGER.debug("Mappings: %s", mappings) # Process each mapping for new alerts and issue a warning for wildcard clear commands for mapping in mappings: if "*" in mapping.get("Triggers", []) and mapping.get("ClearCommands"): LOGGER.warning( "Using ClearCommands with wildcard-based mappings ('*') might not behave as expected for all alert clearances." ) LOGGER.debug("Processing mapping: %s", mapping) triggers = mapping.get("Triggers", []) commands = mapping.get("Commands", []) nodes = mapping.get("Nodes", []) match_type = mapping.get("Match", "ANY").upper() matched_alerts = [ alert for alert in new_alerts if any(fnmatch.fnmatch(alert, trigger) for trigger in triggers) ] LOGGER.debug("Matched alerts for mapping: %s", matched_alerts) # Check if new alerts matched the triggers as per the match type if (match_type == "ANY" and matched_alerts) or ( match_type == "ALL" and len(matched_alerts) == len(triggers) ): for alert in matched_alerts: processed_alerts.add(alert) LOGGER.debug("Processing alert: %s", alert) if mapping.get("Type") == "BASH": for cmd in commands: cmd = cmd.format( alert_title=alert ) # Replace placeholder with alert title LOGGER.info("AlertScript: Executing BASH command: %s", cmd) subprocess.run(cmd, shell=True) elif mapping.get("Type") == "DTMF": for node in nodes: for cmd in commands: dtmf_cmd = 'asterisk -rx "rpt fun {} {}"'.format(node, cmd) LOGGER.info( "AlertScript: Executing DTMF command: %s", dtmf_cmd ) subprocess.run(dtmf_cmd, shell=True) # Process each mapping for cleared alerts for mapping in mappings: LOGGER.debug("Processing clear commands for mapping: %s", mapping) clear_commands = mapping.get("ClearCommands", []) triggers = mapping.get("Triggers", []) match_type = mapping.get("Match", "ANY").upper() matched_cleared_alerts = [ alert for alert in cleared_alerts if any(fnmatch.fnmatch(alert, trigger) for trigger in triggers) ] LOGGER.debug("Matched cleared alerts for mapping: %s", matched_cleared_alerts) # Check if cleared alerts matched the triggers as per the match type if (match_type == "ANY" and matched_cleared_alerts) or ( match_type == "ALL" and len(matched_cleared_alerts) == len(triggers) ): for cmd in clear_commands: LOGGER.debug("Executing clear command: %s", cmd) if mapping.get("Type") == "BASH": LOGGER.info("AlertScript: Executing BASH ClearCommand: %s", cmd) subprocess.run(cmd, shell=True) elif mapping.get("Type") == "DTMF": for node in mapping.get("Nodes", []): dtmf_cmd = 'asterisk -rx "rpt fun {} {}"'.format(node, cmd) LOGGER.info( "AlertScript: Executing DTMF ClearCommand: %s", dtmf_cmd ) subprocess.run(dtmf_cmd, shell=True) # Update the state with the alerts processed in this run state["alertscript_alerts"] = list( processed_alerts ) # Convert back to list for JSON serialization LOGGER.debug("Saving state with processed alerts: %s", state["alertscript_alerts"]) save_state(state) LOGGER.debug("Alert script execution completed.") def send_pushover(message, title=None, priority=0): """ Send a push notification via the Pushover service. This function constructs the payload for the request, including the user key, API token, message, title, and priority. The payload is then sent to the Pushover API endpoint. If the request fails, an error message is logged. """ pushover_config = config["Pushover"] user_key = pushover_config.get("UserKey") token = pushover_config.get("APIToken") # Remove newline from the end of the message message = message.rstrip("\n") url = "https://api.pushover.net/1/messages.json" payload = { "token": token, "user": user_key, "message": message, "title": title, "priority": priority, } response = requests.post(url, data=payload) if response.status_code != 200: LOGGER.error("Failed to send Pushover notification: %s", response.text) def convert_audio(audio): """ Convert audio file to 8000Hz mono for compatibility with Asterisk. """ return audio.set_frame_rate(8000).set_channels(1) def change_ct_id_helper( alerts, specified_alerts, auto_change_enabled, alert_type, pushover_debug, pushover_message, ): """ Check whether the CT or ID needs to be changed, performs the change, and logs the process. """ if auto_change_enabled: LOGGER.debug( "%s auto change is enabled, alerts that require a %s change: %s", alert_type, alert_type, specified_alerts, ) # Extract only the alert names from the OrderedDict keys alert_names = [alert for alert in alerts.keys()] # Check if any alert matches specified_alerts # Here we replace set intersection with a list comprehension intersecting_alerts = [ alert for alert in alert_names if alert in specified_alerts ] if intersecting_alerts: for alert in intersecting_alerts: LOGGER.debug("Alert %s requires a %s change", alert, alert_type) if ( change_ct("WX") if alert_type == "CT" else change_id("WX") ): # If the CT/ID was actually changed if pushover_debug: pushover_message += "Changed {} to WX\n".format(alert_type) break else: # No alerts require a CT/ID change, revert back to normal LOGGER.debug( "No alerts require a %s change, reverting to normal.", alert_type ) if ( change_ct("NORMAL") if alert_type == "CT" else change_id("NORMAL") ): # If the CT/ID was actually changed if pushover_debug: pushover_message += "Changed {} to NORMAL\n".format(alert_type) else: LOGGER.debug("%s auto change is not enabled", alert_type) def change_ct(mode): """ Dynamically changes courtesy tones based on the specified operational mode ('NORMAL' or 'WX') and the detailed configuration provided. This function extends flexibility by supporting multiple courtesy tones, each with configurations for normal and wx modes. :param mode: The operational mode, either 'NORMAL' or 'WX'. :return: True if any changes were made, False otherwise. """ mode = mode.lower() # Normalize the mode to lowercase state = load_state() # Load the current state tone_dir = config["CourtesyTones"]["ToneDir"] tones_config = config["CourtesyTones"]["Tones"] LOGGER.debug("change_ct: Starting courtesy tone change to mode: %s", mode) LOGGER.debug("change_ct: Courtesy tone directory: %s", tone_dir) changed = False for ct_key, tone_settings in tones_config.items(): # Normalize keys in tone_settings to lowercase for case-insensitive comparison tone_settings_lower = {k.lower(): v for k, v in tone_settings.items()} target_tone = tone_settings_lower.get(mode) # Access using normalized mode if not target_tone: LOGGER.error( "change_ct: No target tone specified for %s in %s mode", ct_key, mode ) continue src_file = os.path.join(tone_dir, target_tone) dest_file = os.path.join(tone_dir, "{}.ulaw".format(ct_key)) if not os.path.exists(src_file): LOGGER.error("change_ct: Source tone file does not exist: %s", src_file) continue try: shutil.copyfile(src_file, dest_file) LOGGER.info( "ChangeCT: Updated %s to %s mode with tone %s", ct_key, mode, target_tone, ) changed = True except Exception as e: LOGGER.error("ChangeCT: Failed to update %s: %s", ct_key, str(e)) if changed: LOGGER.debug("ChangeCT: Changes made, updating state to %s mode", mode) state["ct"] = mode # Update the state with the new mode save_state(state) else: LOGGER.debug( "ChangeCT: No changes made, courtesy tones already set for %s mode", mode ) return changed def change_id(id): """ Change the current Identifier (ID) to the one specified. This function first checks if the specified ID is already in use. If so, it does not make any changes. If the ID needs to be changed, it replaces the current ID files with the new ones and updates the state file. """ try: state = load_state() except Exception as e: LOGGER.error("changeID: Failed to load state: %s", e) return False if not state or "id" not in state: LOGGER.error("changeID: State is invalid or missing 'id'") return False current_id = state["id"] id_dir = config["IDChange"].get("IDDir", os.path.join(SOUNDS_PATH, "ID")) normal_id = config["IDChange"]["IDs"]["NormalID"] wx_id = config["IDChange"]["IDs"]["WXID"] rpt_id = config["IDChange"]["IDs"]["RptID"] LOGGER.debug("changeID: ID directory: %s", id_dir) LOGGER.debug("changeID: ID argument: %s", id) if not id: LOGGER.error("changeID: called with no ID specified") return False LOGGER.debug("changeID: Current ID - %s", current_id) if id == current_id: LOGGER.debug("changeID: ID is already %s, no changes made.", id) return False src_file = "" if id == "NORMAL": src_file = os.path.join(id_dir, normal_id) else: src_file = os.path.join(id_dir, wx_id) dest_file = os.path.join(id_dir, rpt_id) if not os.path.exists(src_file): LOGGER.error("changeID: Source file does not exist: %s", src_file) return False try: LOGGER.info("Changing to %s ID", id) LOGGER.debug("changeID: Copying %s to %s", src_file, dest_file) shutil.copyfile(src_file, dest_file) except Exception as e: LOGGER.error( "changeID: Failed to copy file from %s to %s: %s", src_file, dest_file, e ) return False try: state["id"] = id save_state(state) except Exception as e: LOGGER.error("changeID: Failed to save state: %s", e) return False return True def generate_title_string(alerts, county_data): """ Generate a string of alert titles with county names for use in various displays and messages. """ alert_titles_with_counties = [ "{} [{}]".format( alert, ", ".join( sorted( set( replace_with_county_name(x["county_code"], county_data) for x in alerts[alert] ) ) ), ) for alert in alerts ] return alert_titles_with_counties def supermon_back_compat(alerts, county_data): """ Write alerts to a file for backwards compatibility with Supermon. Will NOT work on newer Debian systems (ASL3) where SkywarnPlus is not run as the root user. """ # Exit with a debug statement if we are not the root user if os.getuid() != 0: LOGGER.debug("supermon_back_compat: Not running as root, exiting function") return try: # Ensure the target directory exists for /tmp/AUTOSKY os.makedirs("/tmp/AUTOSKY", exist_ok=True) # Construct alert titles with county names using generate_title_string function alert_titles_with_counties = generate_title_string(alerts, county_data) # Check write permissions before writing to the file if os.access("/tmp/AUTOSKY", os.W_OK): with open("/tmp/AUTOSKY/warnings.txt", "w") as file: file.write("
".join(alert_titles_with_counties)) LOGGER.debug("Successfully wrote alerts to /tmp/AUTOSKY/warnings.txt") else: LOGGER.error("No write permission for /tmp/AUTOSKY") except Exception as e: LOGGER.error("An error occurred while writing to /tmp/AUTOSKY: %s", str(e)) try: # Ensure the target directory exists for /var/www/html/AUTOSKY os.makedirs("/var/www/html/AUTOSKY", exist_ok=True) # Check write permissions before writing to the file if os.access("/var/www/html/AUTOSKY", os.W_OK): with open("/var/www/html/AUTOSKY/warnings.txt", "w") as file: file.write("
".join(alert_titles_with_counties)) LOGGER.debug( "Successfully wrote alerts to /var/www/html/AUTOSKY/warnings.txt" ) else: LOGGER.error("No write permission for /var/www/html/AUTOSKY") except Exception as e: LOGGER.error( "An error occurred while writing to /var/www/html/AUTOSKY: %s", str(e) ) def ast_var_update(): """ Function to mimic the behavior of the ast_var_update.sh script from Supermon2 in HamVoIP. Updated Asterisk channel variables for nodes defined in node_info.ini, including CPU load, temperature, and more. Supermon2 will display these variables in the Node Information section. """ LOGGER.debug("ast_var_update: Starting function") # Exit if Supermon2 directory does not exist if not os.path.exists("/srv/http/supermon2"): LOGGER.debug( "ast_var_update: Supermon2 directory does not exist, exiting function" ) return node_info_path = "/usr/local/sbin/supermon/node_info.ini" allstar_env_path = "/usr/local/etc/allstar.env" WX_CODE = "" WX_LOCATION = "" NODE = "" # Read allstar.env file and extract environment variables env_vars = {} try: LOGGER.debug( "ast_var_update: Reading environment variables from %s", allstar_env_path ) with open(allstar_env_path, "r") as file: for line in file: if line.startswith("export "): key, value = line.split("=", 1) key = key.split()[1].strip() value = value.strip().strip('"') env_vars[key] = value LOGGER.debug( "ast_var_update: Found environment variable %s = %s", key, value ) except Exception as e: LOGGER.error("ast_var_update: Error reading %s: %s", allstar_env_path, e) return # Read node_info.ini file and process lines try: LOGGER.debug("ast_var_update: Reading node information from %s", node_info_path) with open(node_info_path, "r") as file: for line in file: if line.startswith("NODE="): node_value = line.split("=", 1)[1].strip().strip('"') if node_value.startswith("$"): env_var_name = node_value[1:] NODE = env_vars.get(env_var_name, "") else: NODE = node_value LOGGER.debug("ast_var_update: NODE set to %s", NODE) elif line.startswith("WX_CODE="): WX_CODE = line.split("=", 1)[1].strip().strip('"') LOGGER.debug("ast_var_update: WX_CODE set to %s", WX_CODE) elif line.startswith("WX_LOCATION="): WX_LOCATION = line.split("=", 1)[1].strip().strip('"') LOGGER.debug("ast_var_update: WX_LOCATION set to %s", WX_LOCATION) except Exception as e: LOGGER.error("ast_var_update: Error reading %s: %s", node_info_path, e) return if not NODE: LOGGER.debug("ast_var_update: No NODE defined, exiting function") return try: LOGGER.debug("ast_var_update: Retrieving Asterisk registrations") registrations = ( subprocess.check_output(["/bin/asterisk", "-rx", "iax2 show registry"]) .decode("utf-8") .splitlines()[1:] ) except subprocess.CalledProcessError as e: LOGGER.error("ast_var_update: Error getting Asterisk registrations: %s", e) registrations = "No Registrations on this server" if registrations == "No Registrations on this server" or not registrations: LOGGER.debug("ast_var_update: No registrations found") registrations = "No Registrations on this server" else: nodes = {} for reg in registrations: parts = reg.split() node_number = parts[2].split("#")[0] server_ip = parts[0].split(":")[0] try: server_domain = ( subprocess.check_output(["dig", "+short", "-x", server_ip]) .decode("utf-8") .strip() ) LOGGER.debug( "ast_var_update: Resolved %s to %s", server_ip, server_domain ) except subprocess.CalledProcessError as e: LOGGER.error( "ast_var_update: Error resolving domain for IP %s: %s", server_ip, e ) server_domain = "Unknown" node_value = "Node - {} is {}".format(parts[2], parts[5]) if "Registered" in node_value: if "hamvoip" in server_domain: nodes[node_number] = nodes.get(node_number, "") + "Hamvoip " else: nodes[node_number] = nodes.get(node_number, "") + "Allstar " registered = "" for key in nodes: nodes[key] = nodes[key].replace(" ", ",") registered += "{} Registered {} ".format(key, nodes[key]) registered = registered.strip().rstrip(",") registrations = registered LOGGER.debug("ast_var_update: Processed registrations: %s", registrations) try: LOGGER.debug("ast_var_update: Retrieving system uptime") cpu_up = "Up since {}".format( subprocess.check_output(["uptime", "-s"]).decode("utf-8").strip() ) LOGGER.debug("ast_var_update: Retrieving CPU load") cpu_load = ( subprocess.check_output(["uptime"]) .decode("utf-8") .strip() .split("load ")[1] ) LOGGER.debug("ast_var_update: Retrieving CPU temperature") cpu_temp = ( subprocess.check_output(["/opt/vc/bin/vcgencmd", "measure_temp"]) .decode("utf-8") .strip() .split("=")[1][:-2] ) cpu_temp = float(cpu_temp) except subprocess.CalledProcessError as e: LOGGER.error("ast_var_update: Error retrieving system info: %s", e) return cputemp_disp = "{}C".format( "lightgreen" if cpu_temp <= 50 else "yellow" if cpu_temp <= 60 else "#fa4c2d", int(cpu_temp), ) LOGGER.debug("ast_var_update: CPU temperature display: %s", cputemp_disp) try: LOGGER.debug("ast_var_update: Retrieving log size info") logsz = ( subprocess.check_output(["df", "-h", "/var/log"]) .decode("utf-8") .splitlines()[1] .split() ) logs = "Logs - {} {} used, {} remains".format(logsz[2], logsz[4], logsz[3]) except subprocess.CalledProcessError as e: LOGGER.error("ast_var_update: Error retrieving log size info: %s", e) return wx = "" if WX_CODE and WX_LOCATION: try: LOGGER.debug("ast_var_update: Retrieving weather info") wx_info = ( subprocess.check_output(["/usr/local/sbin/weather.sh", WX_CODE, "v"]) .decode("utf-8") .strip() ) wx = "{}   ({})".format(WX_LOCATION, wx_info) except subprocess.CalledProcessError as e: LOGGER.error("ast_var_update: Error retrieving weather info: %s", e) wx = "{}   (Weather info not available)".format(WX_LOCATION) # Filter out the \xb0 character from the string wx = wx.replace("\xb0", "") LOGGER.debug("ast_var_update: Weather info display: %s", wx) try: LOGGER.debug("ast_var_update: Reading alert content from warnings.txt") with open("/tmp/AUTOSKY/warnings.txt") as f: alert_content = f.read().strip() except Exception as e: LOGGER.error("ast_var_update: Error reading warnings.txt: %s", e) alert_content = "" if not MASTER_ENABLE: alert = "SkywarnPlus Disabled" elif not alert_content: alert = "SkywarnPlus Enabled
No Alerts
" else: # Adjusted to remove both '[' and ']' correctly alert_content_cleaned = alert_content.replace("[", "").replace("]", "") alert = "SkywarnPlus Enabled
{}
".format( alert_content ) LOGGER.debug("ast_var_update: Alert display: %s", alert) for ni in NODE.split(): if ni: grep_cmd = "grep -q '[[:blank:]]*\\[{}\\]' /etc/asterisk/rpt.conf".format( ni ) if subprocess.call(grep_cmd, shell=True) == 0: main_cmd = ( 'asterisk -rx "rpt setvar {} cpu_up=\\"{}\\" cpu_load=\\"{}\\" cpu_temp=\\"{}\\" WX=\\"{}\\" LOGS=\\"{}\\" REGISTRATIONS=\\"{}\\""' ).format(ni, cpu_up, cpu_load, cputemp_disp, wx, logs, registrations) alert_cmd = ('asterisk -rx "rpt setvar {} ALERT=\\"{}\\""').format( ni, alert ) main_cmd = main_cmd.encode("ascii", "ignore").decode("ascii") alert_cmd = alert_cmd.encode("ascii", "ignore").decode("ascii") LOGGER.debug("ast_var_update: Running main command: %s", main_cmd) try: subprocess.call(main_cmd, shell=True) except subprocess.CalledProcessError as e: LOGGER.error( "ast_var_update: Error running main command for node %s: %s", ni, e, ) LOGGER.debug("ast_var_update: Running alert command: %s", alert_cmd) try: subprocess.call(alert_cmd, shell=True) except subprocess.CalledProcessError as e: LOGGER.error( "ast_var_update: Error running alert command for node %s: %s", ni, e, ) LOGGER.debug("ast_var_update: Function completed") def detect_county_changes(old_alerts, new_alerts): """ Detect if any counties have been added to or removed from an alert and return the alerts with added or removed counties. """ alerts_with_changed_counties = OrderedDict() changes_detected = {} for alert_name, alert_info in new_alerts.items(): if alert_name not in old_alerts: continue old_alert_info = old_alerts.get(alert_name, []) old_county_codes = {info["county_code"] for info in old_alert_info} new_county_codes = {info["county_code"] for info in alert_info} added_counties = new_county_codes - old_county_codes removed_counties = old_county_codes - new_county_codes added_counties = { code.replace("{", "").replace("}", "").replace('"', "") for code in added_counties } removed_counties = { code.replace("{", "").replace("}", "").replace('"', "") for code in removed_counties } if added_counties or removed_counties: alerts_with_changed_counties[alert_name] = new_alerts[alert_name] changes_detected[alert_name] = { "old": old_county_codes, "added": added_counties, "removed": removed_counties, } return alerts_with_changed_counties, changes_detected def load_county_names(md_file): """ Load county names from separate markdown tables so that county codes can be replaced with county names. """ with open(md_file, "r") as f: lines = f.readlines() county_data = {} in_table = False for line in lines: if line.startswith("| County |"): in_table = True continue # Skip the header elif not in_table or line.strip() == "" or line.startswith("##"): continue else: name, code = [s.strip() for s in line.split("|")[1:-1]] county_data[code] = name return county_data def replace_with_county_name(county_code, county_data): """ Translate county code to county name. """ return county_data.get(county_code, county_code) def main(): """ The main function that orchestrates the entire process of fetching and processing severe weather alerts, then integrating these alerts into an Asterisk/app_rpt based radio repeater system. """ # Fetch configurations say_alert_enabled = config["Alerting"].get("SayAlert", False) say_alert_all = config["Alerting"].get("SayAlertAll", False) say_all_clear_enabled = config["Alerting"].get("SayAllClear", False) alertscript_enabled = config["AlertScript"].get("Enable", False) ct_alerts = config["CourtesyTones"].get("CTAlerts", []) enable_ct_auto_change = config["CourtesyTones"].get("Enable", False) id_alerts = config["IDChange"].get("IDAlerts", []) enable_id_auto_change = config["IDChange"].get("Enable", False) pushover_enabled = config["Pushover"].get("Enable", False) pushover_debug = config["Pushover"].get("Debug", False) supermon_compat_enabled = config["DEV"].get("SupermonCompat", True) say_alerts_changed = config["Alerting"].get("SayAlertsChanged", True) # Check if SkywarnPlus is enabled if not MASTER_ENABLE: print("SkywarnPlus is disabled in config.yaml, exiting...") if supermon_compat_enabled: ast_var_update() exit() # Load previous alert data to compare changes state = load_state() last_alerts = state["last_alerts"] # Load county names from YAML file so that county codes can be replaced with county names in messages county_data = load_county_names(COUNTY_CODES_PATH) # If data file does not exist, assume this is the first run and initialize data file and CT/ID/Tailmessage files if enabled if not os.path.isfile(DATA_FILE): LOGGER.info("Data file does not exist, assuming first run.") LOGGER.info("Initializing data file") save_state(state) if enable_ct_auto_change: LOGGER.info("Initializing CT files") change_ct("NORMAL") if enable_id_auto_change: LOGGER.info("Initializing ID files") change_id("NORMAL") if ENABLE_TAILMESSAGE: LOGGER.info("Initializing Tailmessage file") empty_alerts = OrderedDict() build_tailmessage(empty_alerts) if supermon_compat_enabled: supermon_back_compat(last_alerts, county_data) # Fetch new alert data alerts = get_alerts(COUNTY_CODES) # Placeholder for constructing a pushover message pushover_message = "" # Update HamVoIP Asterisk channel variables if supermon_compat_enabled: supermon_back_compat(alerts, county_data) ast_var_update() # Determine which alerts have been added since the last check added_alerts = [alert for alert in alerts if alert not in last_alerts] for alert in added_alerts: counties = sorted( set( replace_with_county_name(x["county_code"], county_data) for x in alerts[alert] ) ) counties_str = "[" + ", ".join(counties) + "]" LOGGER.info("Added: {} for {}".format(alert, counties_str)) pushover_message += "Added: {} for {}\n".format(alert, counties_str) # Determine which alerts have been removed since the last check removed_alerts = [alert for alert in last_alerts if alert not in alerts] for alert in removed_alerts: counties = sorted( set( replace_with_county_name(x["county_code"], county_data) for x in last_alerts[alert] ) ) counties_str = "[" + ", ".join(counties) + "]" LOGGER.info("Removed: {} for {}".format(alert, counties_str)) pushover_message += "Removed: {} for {}\n".format(alert, counties_str) # Placeholder for storing alerts with changed county codes changed_alerts = {} # If the list of alerts is not empty if alerts: # Compare old and new alerts to detect changes in affected counties changed_alerts, changes_details = detect_county_changes(last_alerts, alerts) for alert, details in changes_details.items(): old_counties = sorted( set( replace_with_county_name(county, county_data) for county in details["old"] ) ) old_counties_str = "[" + ", ".join(old_counties) + "]" added_msg = "" if details["added"]: added_counties = sorted( set( replace_with_county_name(county, county_data) for county in details["added"] ) ) added_counties_str = "[" + ", ".join(added_counties) + "]" added_msg = "is now also affecting {}".format(added_counties_str) removed_msg = "" if details["removed"]: removed_counties = sorted( set( replace_with_county_name(county, county_data) for county in details["removed"] ) ) removed_counties_str = "[" + ", ".join(removed_counties) + "]" removed_msg = "is no longer affecting {}".format(removed_counties_str) # Combining the log messages combined_msg_parts = ["Changed: {} for {}".format(alert, old_counties_str)] if added_msg: combined_msg_parts.append(added_msg) if removed_msg: if ( added_msg ): # if there's an 'added' message, then use 'and' to combine with 'removed' message combined_msg_parts.append("and") combined_msg_parts.append(removed_msg) log_msg = " ".join(combined_msg_parts) LOGGER.info(log_msg) pushover_message += log_msg + "\n" # Process changes in alerts if added_alerts or removed_alerts or changed_alerts: # Save the data state["last_alerts"] = alerts save_state(state) # Send "all clear" messages if alerts have changed but are empty if not alerts: LOGGER.info("Alerts cleared") # Call say_allclear if enabled if say_all_clear_enabled: say_allclear() # Add "Alerts Cleared" to pushover message pushover_message += "Alerts Cleared\n" # If alerts have been added, removed if added_alerts or removed_alerts: # Push alert titles to Supermon if enabled # if supermon_compat_enabled: # supermon_back_compat(alerts) # Change CT/ID if necessary and enabled change_ct_id_helper( alerts, ct_alerts, enable_ct_auto_change, "CT", pushover_debug, pushover_message, ) change_ct_id_helper( alerts, id_alerts, enable_id_auto_change, "ID", pushover_debug, pushover_message, ) # Call alert_script if enabled if alertscript_enabled: alert_script(alerts) # Say alerts if enabled if say_alert_enabled: # If say_alert_all is enabled, say all currently active alerts if say_alert_all: alerts_to_say = alerts # Otherwise, only say newly added alerts else: alerts_to_say = {alert: alerts[alert] for alert in added_alerts} # If County IDs have been set up and there is more than one county code, then also say alerts with county changes # Only if enabled if ( changed_alerts and say_alerts_changed and COUNTY_WAVS and len(COUNTY_CODES) > 1 ): alerts_to_say.update(changed_alerts) # Sort alerts based on severity alerts_to_say = sort_alerts(alerts_to_say) # Say the alerts say_alerts(alerts_to_say) # If alerts have changed, but none added or removed elif changed_alerts: # Say changed alerts only if enabled, County IDs have been set up, and there is more than one county code if say_alerts_changed and COUNTY_WAVS and len(COUNTY_CODES) > 1: # Sort alerts based on severity changed_alerts = sort_alerts(changed_alerts) # Say the alerts say_alerts(changed_alerts) # Alerts have changed, update tailmessage if enabled if ENABLE_TAILMESSAGE: build_tailmessage(alerts) if pushover_debug: pushover_message += ( "WX tailmessage removed\n" if not alerts else "Built WX tailmessage\n" ) # Send pushover message if enabled if pushover_enabled: pushover_message = pushover_message.rstrip("\n") LOGGER.debug("Sending Pushover message: %s", pushover_message) send_pushover(pushover_message, title="SkywarnPlus") # If no changes detected in alerts else: # If this is being run interactively, inform the user that nothing has changed if sys.stdin.isatty(): # Log list of current alerts, unless there aren't any, then current_alerts = "None" if len(alerts) == 0: current_alerts = "None" else: alert_details = [] for alert, counties in alerts.items(): counties_str = ", ".join( sorted( set( replace_with_county_name( county["county_code"], county_data ) for county in counties ) ) ) alert_details.append("{} ({})".format(alert, counties_str)) current_alerts = "; ".join(alert_details) LOGGER.info("No change in alerts.") LOGGER.info("Current alerts: %s.", current_alerts) # If this is being run non-interactively, only log if debug is enabled else: LOGGER.debug("No change in alerts.") if __name__ == "__main__": main()