Update pre-commit versions
This commit is contained in:
parent
66232c8260
commit
3f666e8251
|
@ -294,11 +294,10 @@ class ElasticsearchBackend(StorageBackend):
|
||||||
self.log.error(f"Indexing failed: {result}")
|
self.log.error(f"Indexing failed: {result}")
|
||||||
self.log.debug(f"Indexed {len(matches)} messages in ES")
|
self.log.debug(f"Indexed {len(matches)} messages in ES")
|
||||||
|
|
||||||
async def schedule_query_results(self, rule_object):
|
def prepare_schedule_query(self, rule_object):
|
||||||
"""
|
"""
|
||||||
Helper to run a scheduled query with reduced functionality and async.
|
Helper to run a scheduled query with reduced functionality.
|
||||||
"""
|
"""
|
||||||
|
|
||||||
data = rule_object.parsed
|
data = rule_object.parsed
|
||||||
|
|
||||||
if "tags" in data:
|
if "tags" in data:
|
||||||
|
@ -310,8 +309,6 @@ class ElasticsearchBackend(StorageBackend):
|
||||||
query = data["query"][0]
|
query = data["query"][0]
|
||||||
data["query"] = query
|
data["query"] = query
|
||||||
|
|
||||||
result_map = {}
|
|
||||||
|
|
||||||
add_bool = []
|
add_bool = []
|
||||||
add_top = []
|
add_top = []
|
||||||
if "source" in data:
|
if "source" in data:
|
||||||
|
@ -352,34 +349,13 @@ class ElasticsearchBackend(StorageBackend):
|
||||||
"avg": {"field": "sentiment"},
|
"avg": {"field": "sentiment"},
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
for index in data["index"]:
|
|
||||||
|
|
||||||
if "message" in search_query:
|
return search_query
|
||||||
self.log.error(f"Error parsing query: {search_query['message']}")
|
|
||||||
continue
|
|
||||||
response = await self.async_run_query(
|
|
||||||
rule_object.user,
|
|
||||||
search_query,
|
|
||||||
index=index,
|
|
||||||
)
|
|
||||||
self.log.debug(f"Running scheduled query on {index}: {search_query}")
|
|
||||||
# self.log.debug(f"Response from scheduled query: {response}")
|
|
||||||
if isinstance(response, Exception):
|
|
||||||
error = response.info["error"]["root_cause"][0]["reason"]
|
|
||||||
self.log.error(f"Error running scheduled search: {error}")
|
|
||||||
raise QueryError(error)
|
|
||||||
if len(response["hits"]["hits"]) == 0:
|
|
||||||
# No results, skip
|
|
||||||
continue
|
|
||||||
meta, response = self.parse(response, meta=True)
|
|
||||||
# print("Parsed response", response)
|
|
||||||
if "message" in response:
|
|
||||||
self.log.error(f"Error running scheduled search: {response['message']}")
|
|
||||||
continue
|
|
||||||
result_map[index] = (meta, response)
|
|
||||||
|
|
||||||
# Average aggregation check
|
def schedule_check_aggregations(self, rule_object, result_map):
|
||||||
# Could probably do this in elasticsearch
|
"""
|
||||||
|
Check the results of a scheduled query for aggregations.
|
||||||
|
"""
|
||||||
for index, (meta, result) in result_map.items():
|
for index, (meta, result) in result_map.items():
|
||||||
# Default to true, if no aggs are found, we still want to match
|
# Default to true, if no aggs are found, we still want to match
|
||||||
match = True
|
match = True
|
||||||
|
@ -412,6 +388,71 @@ class ElasticsearchBackend(StorageBackend):
|
||||||
|
|
||||||
return result_map
|
return result_map
|
||||||
|
|
||||||
|
def schedule_query_results_test_sync(self, rule_object):
|
||||||
|
"""
|
||||||
|
Helper to run a scheduled query test with reduced functionality.
|
||||||
|
Sync version for running from Django forms.
|
||||||
|
Does not return results.
|
||||||
|
"""
|
||||||
|
data = rule_object.parsed
|
||||||
|
|
||||||
|
search_query = self.prepare_schedule_query(rule_object)
|
||||||
|
for index in data["index"]:
|
||||||
|
if "message" in search_query:
|
||||||
|
self.log.error(f"Error parsing test query: {search_query['message']}")
|
||||||
|
continue
|
||||||
|
response = self.run_query(
|
||||||
|
rule_object.user,
|
||||||
|
search_query,
|
||||||
|
index=index,
|
||||||
|
)
|
||||||
|
self.log.debug(f"Running scheduled test query on {index}: {search_query}")
|
||||||
|
# self.log.debug(f"Response from scheduled query: {response}")
|
||||||
|
if isinstance(response, Exception):
|
||||||
|
error = response.info["error"]["root_cause"][0]["reason"]
|
||||||
|
self.log.error(f"Error running test scheduled search: {error}")
|
||||||
|
raise QueryError(error)
|
||||||
|
|
||||||
|
async def schedule_query_results(self, rule_object):
|
||||||
|
"""
|
||||||
|
Helper to run a scheduled query with reduced functionality and async.
|
||||||
|
"""
|
||||||
|
result_map = {}
|
||||||
|
data = rule_object.parsed
|
||||||
|
|
||||||
|
search_query = self.prepare_schedule_query(rule_object)
|
||||||
|
|
||||||
|
for index in data["index"]:
|
||||||
|
if "message" in search_query:
|
||||||
|
self.log.error(f"Error parsing query: {search_query['message']}")
|
||||||
|
continue
|
||||||
|
response = await self.async_run_query(
|
||||||
|
rule_object.user,
|
||||||
|
search_query,
|
||||||
|
index=index,
|
||||||
|
)
|
||||||
|
self.log.debug(f"Running scheduled query on {index}: {search_query}")
|
||||||
|
# self.log.debug(f"Response from scheduled query: {response}")
|
||||||
|
if isinstance(response, Exception):
|
||||||
|
error = response.info["error"]["root_cause"][0]["reason"]
|
||||||
|
self.log.error(f"Error running scheduled search: {error}")
|
||||||
|
raise QueryError(error)
|
||||||
|
if len(response["hits"]["hits"]) == 0:
|
||||||
|
# No results, skip
|
||||||
|
continue
|
||||||
|
meta, response = self.parse(response, meta=True)
|
||||||
|
# print("Parsed response", response)
|
||||||
|
if "message" in response:
|
||||||
|
self.log.error(f"Error running scheduled search: {response['message']}")
|
||||||
|
continue
|
||||||
|
result_map[index] = (meta, response)
|
||||||
|
|
||||||
|
# Average aggregation check
|
||||||
|
# Could probably do this in elasticsearch
|
||||||
|
result_map = self.schedule_check_aggregations(rule_object, result_map)
|
||||||
|
|
||||||
|
return result_map
|
||||||
|
|
||||||
def query_results(
|
def query_results(
|
||||||
self,
|
self,
|
||||||
request,
|
request,
|
||||||
|
@ -424,7 +465,6 @@ class ElasticsearchBackend(StorageBackend):
|
||||||
dedup_fields=None,
|
dedup_fields=None,
|
||||||
tags=None,
|
tags=None,
|
||||||
):
|
):
|
||||||
|
|
||||||
add_bool = []
|
add_bool = []
|
||||||
add_top = []
|
add_top = []
|
||||||
add_top_negative = []
|
add_top_negative = []
|
||||||
|
|
|
@ -116,7 +116,7 @@ class NotificationRuleForm(RestrictedFormMixin, ModelForm):
|
||||||
"topic",
|
"topic",
|
||||||
"url",
|
"url",
|
||||||
"service",
|
"service",
|
||||||
"send_empty",
|
"policy",
|
||||||
"enabled",
|
"enabled",
|
||||||
)
|
)
|
||||||
help_texts = {
|
help_texts = {
|
||||||
|
@ -130,7 +130,7 @@ class NotificationRuleForm(RestrictedFormMixin, ModelForm):
|
||||||
"interval": "How often to run the search. On demand evaluates messages as they are received, without running a scheduled search. The remaining options schedule a search of the database with the window below.",
|
"interval": "How often to run the search. On demand evaluates messages as they are received, without running a scheduled search. The remaining options schedule a search of the database with the window below.",
|
||||||
"window": "Time window to search: 1d, 1h, 1m, 1s, etc.",
|
"window": "Time window to search: 1d, 1h, 1m, 1s, etc.",
|
||||||
"amount": "Amount of matches to be returned for scheduled queries. Cannot be used with on-demand queries.",
|
"amount": "Amount of matches to be returned for scheduled queries. Cannot be used with on-demand queries.",
|
||||||
"send_empty": "Send a notification if no matches are found.",
|
"policy": "When to trigger this policy.",
|
||||||
}
|
}
|
||||||
|
|
||||||
def clean(self):
|
def clean(self):
|
||||||
|
|
|
@ -11,7 +11,6 @@ except ImportError:
|
||||||
from datetime import datetime
|
from datetime import datetime
|
||||||
|
|
||||||
import orjson
|
import orjson
|
||||||
from asgiref.sync import async_to_sync
|
|
||||||
from siphashc import siphash
|
from siphashc import siphash
|
||||||
|
|
||||||
from core.lib.notify import sendmsg
|
from core.lib.notify import sendmsg
|
||||||
|
@ -182,6 +181,7 @@ class NotificationRuleData(object):
|
||||||
self.db = db
|
self.db = db
|
||||||
self.data = self.cleaned_data.get("data")
|
self.data = self.cleaned_data.get("data")
|
||||||
self.window = self.cleaned_data.get("window")
|
self.window = self.cleaned_data.get("window")
|
||||||
|
self.policy = self.cleaned_data.get("policy")
|
||||||
self.parsed = None
|
self.parsed = None
|
||||||
self.aggs = {}
|
self.aggs = {}
|
||||||
|
|
||||||
|
@ -323,6 +323,9 @@ class NotificationRuleData(object):
|
||||||
"""
|
"""
|
||||||
current_match = self.get_match(index, message)
|
current_match = self.get_match(index, message)
|
||||||
log.debug(f"Rule matched: {index} - current match: {current_match}")
|
log.debug(f"Rule matched: {index} - current match: {current_match}")
|
||||||
|
|
||||||
|
# Default policy: Trigger only when results change
|
||||||
|
|
||||||
if current_match is False:
|
if current_match is False:
|
||||||
# Matched now, but not before
|
# Matched now, but not before
|
||||||
if "matched" not in meta:
|
if "matched" not in meta:
|
||||||
|
@ -342,6 +345,8 @@ class NotificationRuleData(object):
|
||||||
"""
|
"""
|
||||||
current_match = self.get_match(index, message)
|
current_match = self.get_match(index, message)
|
||||||
log.debug(f"Rule matched: {index} - current match: {current_match}")
|
log.debug(f"Rule matched: {index} - current match: {current_match}")
|
||||||
|
|
||||||
|
# Default policy: Trigger only when results change
|
||||||
if current_match is False:
|
if current_match is False:
|
||||||
# Matched now, but not before
|
# Matched now, but not before
|
||||||
if "matched" not in meta:
|
if "matched" not in meta:
|
||||||
|
@ -361,6 +366,8 @@ class NotificationRuleData(object):
|
||||||
"""
|
"""
|
||||||
current_match = self.get_match(index)
|
current_match = self.get_match(index)
|
||||||
log.debug(f"Rule not matched: {index} - current match: {current_match}")
|
log.debug(f"Rule not matched: {index} - current match: {current_match}")
|
||||||
|
|
||||||
|
# Change policy: When there are no results following a successful run
|
||||||
if current_match is True:
|
if current_match is True:
|
||||||
# Matched before, but not now
|
# Matched before, but not now
|
||||||
if self.object.send_empty:
|
if self.object.send_empty:
|
||||||
|
@ -407,13 +414,10 @@ class NotificationRuleData(object):
|
||||||
def test_schedule(self):
|
def test_schedule(self):
|
||||||
"""
|
"""
|
||||||
Test the schedule query to ensure it is valid.
|
Test the schedule query to ensure it is valid.
|
||||||
Run the query with the async_to_sync helper so we can call it from
|
|
||||||
a form.
|
|
||||||
Raises an exception if the query is invalid.
|
Raises an exception if the query is invalid.
|
||||||
"""
|
"""
|
||||||
if self.db:
|
if self.db:
|
||||||
sync_schedule = async_to_sync(self.db.schedule_query_results)
|
self.db.schedule_query_results_test_sync(self)
|
||||||
sync_schedule(self)
|
|
||||||
|
|
||||||
def validate_schedule_fields(self):
|
def validate_schedule_fields(self):
|
||||||
"""
|
"""
|
||||||
|
@ -476,9 +480,10 @@ class NotificationRuleData(object):
|
||||||
raise RuleParseError(
|
raise RuleParseError(
|
||||||
"Field tags cannot be used with on-demand rules", "data"
|
"Field tags cannot be used with on-demand rules", "data"
|
||||||
)
|
)
|
||||||
if self.cleaned_data["send_empty"]:
|
if self.policy != "default":
|
||||||
raise RuleParseError(
|
raise RuleParseError(
|
||||||
"Field cannot be used with on-demand rules", "send_empty"
|
f"Cannot use {self.cleaned_data['policy']} policy with on-demand rules",
|
||||||
|
"policy",
|
||||||
)
|
)
|
||||||
|
|
||||||
@property
|
@property
|
||||||
|
|
|
@ -0,0 +1,28 @@
|
||||||
|
# Generated by Django 4.1.5 on 2023-02-09 14:38
|
||||||
|
|
||||||
|
from django.db import migrations, models
|
||||||
|
|
||||||
|
|
||||||
|
class Migration(migrations.Migration):
|
||||||
|
|
||||||
|
dependencies = [
|
||||||
|
('core', '0025_alter_notificationrule_id'),
|
||||||
|
]
|
||||||
|
|
||||||
|
operations = [
|
||||||
|
migrations.AddField(
|
||||||
|
model_name='notificationrule',
|
||||||
|
name='policy',
|
||||||
|
field=models.CharField(choices=[('default', 'Only trigger for matched events'), ('change', 'Trigger only if no results found when they were last run'), ('always', 'Always trigger regardless of whether results are found')], default='default', max_length=255),
|
||||||
|
),
|
||||||
|
migrations.AlterField(
|
||||||
|
model_name='notificationrule',
|
||||||
|
name='service',
|
||||||
|
field=models.CharField(choices=[('ntfy', 'NTFY'), ('webhook', 'Custom webhook'), ('none', 'Disabled')], default='ntfy', max_length=255),
|
||||||
|
),
|
||||||
|
migrations.AlterField(
|
||||||
|
model_name='notificationsettings',
|
||||||
|
name='service',
|
||||||
|
field=models.CharField(choices=[('ntfy', 'NTFY'), ('webhook', 'Custom webhook'), ('none', 'Disabled')], default='ntfy', max_length=255),
|
||||||
|
),
|
||||||
|
]
|
|
@ -42,6 +42,15 @@ SERVICE_CHOICES = (
|
||||||
("none", "Disabled"),
|
("none", "Disabled"),
|
||||||
)
|
)
|
||||||
|
|
||||||
|
POLICY_CHOICES = (
|
||||||
|
("default", "Default: Trigger only when there were no results last time"),
|
||||||
|
(
|
||||||
|
"change",
|
||||||
|
"Change: Default + trigger when there are no results (if there were before)",
|
||||||
|
),
|
||||||
|
("always", "Always: Trigger on every run (not recommended for low intervals)"),
|
||||||
|
)
|
||||||
|
|
||||||
|
|
||||||
class Plan(models.Model):
|
class Plan(models.Model):
|
||||||
name = models.CharField(max_length=255, unique=True)
|
name = models.CharField(max_length=255, unique=True)
|
||||||
|
@ -193,6 +202,7 @@ class NotificationRule(models.Model):
|
||||||
match = models.JSONField(null=True, blank=True)
|
match = models.JSONField(null=True, blank=True)
|
||||||
service = models.CharField(choices=SERVICE_CHOICES, max_length=255, default="ntfy")
|
service = models.CharField(choices=SERVICE_CHOICES, max_length=255, default="ntfy")
|
||||||
send_empty = models.BooleanField(default=False)
|
send_empty = models.BooleanField(default=False)
|
||||||
|
policy = models.CharField(choices=POLICY_CHOICES, max_length=255, default="default")
|
||||||
|
|
||||||
def __str__(self):
|
def __str__(self):
|
||||||
return f"{self.user} - {self.name}"
|
return f"{self.user} - {self.name}"
|
||||||
|
|
|
@ -43,7 +43,6 @@ class ColoredFormatter(logging.Formatter):
|
||||||
|
|
||||||
|
|
||||||
def get_logger(name):
|
def get_logger(name):
|
||||||
|
|
||||||
# Define the logging format
|
# Define the logging format
|
||||||
FORMAT = "%(asctime)s %(levelname)18s $BOLD%(name)13s$RESET - %(message)s"
|
FORMAT = "%(asctime)s %(levelname)18s $BOLD%(name)13s$RESET - %(message)s"
|
||||||
COLOR_FORMAT = formatter_message(FORMAT, True)
|
COLOR_FORMAT = formatter_message(FORMAT, True)
|
||||||
|
|
|
@ -121,7 +121,6 @@ class ThresholdIRCNetworkRelayDel(SuperUserRequiredMixin, APIView):
|
||||||
"""
|
"""
|
||||||
deleted = threshold.del_relay(net, num)
|
deleted = threshold.del_relay(net, num)
|
||||||
if deleted["success"]:
|
if deleted["success"]:
|
||||||
|
|
||||||
message = f"Deleted relay {num}"
|
message = f"Deleted relay {num}"
|
||||||
message_class = "success"
|
message_class = "success"
|
||||||
else:
|
else:
|
||||||
|
@ -150,7 +149,6 @@ class ThresholdIRCNetworkRelayProvision(SuperUserRequiredMixin, APIView):
|
||||||
"""
|
"""
|
||||||
provisioned = threshold.irc_provision_relay(net, num)
|
provisioned = threshold.irc_provision_relay(net, num)
|
||||||
if provisioned["success"]:
|
if provisioned["success"]:
|
||||||
|
|
||||||
message = f"Provisioned relay {num}"
|
message = f"Provisioned relay {num}"
|
||||||
message_class = "success"
|
message_class = "success"
|
||||||
else:
|
else:
|
||||||
|
@ -179,7 +177,6 @@ class ThresholdIRCNetworkRelayAuth(SuperUserRequiredMixin, APIView):
|
||||||
"""
|
"""
|
||||||
provisioned = threshold.irc_enable_auth(net, num)
|
provisioned = threshold.irc_enable_auth(net, num)
|
||||||
if provisioned["success"]:
|
if provisioned["success"]:
|
||||||
|
|
||||||
message = f"Enabled authentication on relay {num}"
|
message = f"Enabled authentication on relay {num}"
|
||||||
message_class = "success"
|
message_class = "success"
|
||||||
else:
|
else:
|
||||||
|
|
Loading…
Reference in New Issue