mirror of
https://github.com/apache/superset.git
synced 2026-05-02 06:24:37 +00:00
Compare commits
6 Commits
fix-webpac
...
v2021.8.2
| Author | SHA1 | Date | |
|---|---|---|---|
|
|
a636b3584f | ||
|
|
8ab3da3121 | ||
|
|
5146f39b58 | ||
|
|
59ad4d20bf | ||
|
|
2d13f628b4 | ||
|
|
6c760cd4c7 |
1
setup.py
1
setup.py
@@ -129,6 +129,7 @@ setup(
|
|||||||
"elasticsearch": ["elasticsearch-dbapi>=0.2.0, <0.3.0"],
|
"elasticsearch": ["elasticsearch-dbapi>=0.2.0, <0.3.0"],
|
||||||
"exasol": ["sqlalchemy-exasol>=2.1.0, <2.2"],
|
"exasol": ["sqlalchemy-exasol>=2.1.0, <2.2"],
|
||||||
"excel": ["xlrd>=1.2.0, <1.3"],
|
"excel": ["xlrd>=1.2.0, <1.3"],
|
||||||
|
"firebird": ["sqlalchemy-firebird>=0.7.0, <0.8"],
|
||||||
"gsheets": ["shillelagh>=0.2, <0.3"],
|
"gsheets": ["shillelagh>=0.2, <0.3"],
|
||||||
"hana": ["hdbcli==2.4.162", "sqlalchemy_hana==0.4.0"],
|
"hana": ["hdbcli==2.4.162", "sqlalchemy_hana==0.4.0"],
|
||||||
"hive": ["pyhive[hive]>=0.6.1", "tableschema", "thrift>=0.11.0, <1.0.0"],
|
"hive": ["pyhive[hive]>=0.6.1", "tableschema", "thrift>=0.11.0, <1.0.0"],
|
||||||
|
|||||||
@@ -1,23 +0,0 @@
|
|||||||
/**
|
|
||||||
* Licensed to the Apache Software Foundation (ASF) under one
|
|
||||||
* or more contributor license agreements. See the NOTICE file
|
|
||||||
* distributed with this work for additional information
|
|
||||||
* regarding copyright ownership. The ASF licenses this file
|
|
||||||
* to you under the Apache License, Version 2.0 (the
|
|
||||||
* "License"); you may not use this file except in compliance
|
|
||||||
* with the License. You may obtain a copy of the License at
|
|
||||||
*
|
|
||||||
* http://www.apache.org/licenses/LICENSE-2.0
|
|
||||||
*
|
|
||||||
* Unless required by applicable law or agreed to in writing,
|
|
||||||
* software distributed under the License is distributed on an
|
|
||||||
* "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY
|
|
||||||
* KIND, either express or implied. See the License for the
|
|
||||||
* specific language governing permissions and limitations
|
|
||||||
* under the License.
|
|
||||||
*/
|
|
||||||
{
|
|
||||||
"rules": {
|
|
||||||
"no-restricted-imports": 0
|
|
||||||
}
|
|
||||||
}
|
|
||||||
@@ -456,7 +456,7 @@ class BaseEngineSpec: # pylint: disable=too-many-public-methods
|
|||||||
)
|
)
|
||||||
return database.compile_sqla_query(qry)
|
return database.compile_sqla_query(qry)
|
||||||
|
|
||||||
if LimitMethod.FORCE_LIMIT:
|
if cls.limit_method == LimitMethod.FORCE_LIMIT:
|
||||||
parsed_query = sql_parse.ParsedQuery(sql)
|
parsed_query = sql_parse.ParsedQuery(sql)
|
||||||
sql = parsed_query.set_or_update_query_limit(limit)
|
sql = parsed_query.set_or_update_query_limit(limit)
|
||||||
|
|
||||||
|
|||||||
83
superset/db_engine_specs/firebird.py
Normal file
83
superset/db_engine_specs/firebird.py
Normal file
@@ -0,0 +1,83 @@
|
|||||||
|
# Licensed to the Apache Software Foundation (ASF) under one
|
||||||
|
# or more contributor license agreements. See the NOTICE file
|
||||||
|
# distributed with this work for additional information
|
||||||
|
# regarding copyright ownership. The ASF licenses this file
|
||||||
|
# to you under the Apache License, Version 2.0 (the
|
||||||
|
# "License"); you may not use this file except in compliance
|
||||||
|
# with the License. You may obtain a copy of the License at
|
||||||
|
#
|
||||||
|
# http://www.apache.org/licenses/LICENSE-2.0
|
||||||
|
#
|
||||||
|
# Unless required by applicable law or agreed to in writing,
|
||||||
|
# software distributed under the License is distributed on an
|
||||||
|
# "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY
|
||||||
|
# KIND, either express or implied. See the License for the
|
||||||
|
# specific language governing permissions and limitations
|
||||||
|
# under the License.
|
||||||
|
from datetime import datetime
|
||||||
|
from typing import Optional
|
||||||
|
|
||||||
|
from superset.db_engine_specs.base import BaseEngineSpec, LimitMethod
|
||||||
|
from superset.utils import core as utils
|
||||||
|
|
||||||
|
|
||||||
|
class FirebirdEngineSpec(BaseEngineSpec):
|
||||||
|
"""Engine for Firebird"""
|
||||||
|
|
||||||
|
engine = "firebird"
|
||||||
|
engine_name = "Firebird"
|
||||||
|
|
||||||
|
# Firebird uses FIRST to limit: `SELECT FIRST 10 * FROM table`
|
||||||
|
limit_method = LimitMethod.FETCH_MANY
|
||||||
|
|
||||||
|
_time_grain_expressions = {
|
||||||
|
None: "{col}",
|
||||||
|
"PT1S": (
|
||||||
|
"CAST(CAST({col} AS DATE) "
|
||||||
|
"|| ' ' "
|
||||||
|
"|| EXTRACT(HOUR FROM {col}) "
|
||||||
|
"|| ':' "
|
||||||
|
"|| EXTRACT(MINUTE FROM {col}) "
|
||||||
|
"|| ':' "
|
||||||
|
"|| FLOOR(EXTRACT(SECOND FROM {col})) AS TIMESTAMP)"
|
||||||
|
),
|
||||||
|
"PT1M": (
|
||||||
|
"CAST(CAST({col} AS DATE) "
|
||||||
|
"|| ' ' "
|
||||||
|
"|| EXTRACT(HOUR FROM {col}) "
|
||||||
|
"|| ':' "
|
||||||
|
"|| EXTRACT(MINUTE FROM {col}) "
|
||||||
|
"|| ':00' AS TIMESTAMP)"
|
||||||
|
),
|
||||||
|
"PT1H": (
|
||||||
|
"CAST(CAST({col} AS DATE) "
|
||||||
|
"|| ' ' "
|
||||||
|
"|| EXTRACT(HOUR FROM {col}) "
|
||||||
|
"|| ':00:00' AS TIMESTAMP)"
|
||||||
|
),
|
||||||
|
"P1D": "CAST({col} AS DATE)",
|
||||||
|
"P1M": (
|
||||||
|
"CAST(EXTRACT(YEAR FROM {col}) "
|
||||||
|
"|| '-' "
|
||||||
|
"|| EXTRACT(MONTH FROM {col}) "
|
||||||
|
"|| '-01' AS DATE)"
|
||||||
|
),
|
||||||
|
"P1Y": "CAST(EXTRACT(YEAR FROM {col}) || '-01-01' AS DATE)",
|
||||||
|
}
|
||||||
|
|
||||||
|
@classmethod
|
||||||
|
def epoch_to_dttm(cls) -> str:
|
||||||
|
return "DATEADD(second, {col}, CAST('00:00:00' AS TIMESTAMP))"
|
||||||
|
|
||||||
|
@classmethod
|
||||||
|
def convert_dttm(cls, target_type: str, dttm: datetime) -> Optional[str]:
|
||||||
|
tt = target_type.upper()
|
||||||
|
if tt == utils.TemporalType.TIMESTAMP:
|
||||||
|
dttm_formatted = dttm.isoformat(sep=" ")
|
||||||
|
dttm_valid_precision = dttm_formatted[: len("YYYY-MM-DD HH:MM:SS.MMMM")]
|
||||||
|
return f"CAST('{dttm_valid_precision}' AS TIMESTAMP)"
|
||||||
|
if tt == utils.TemporalType.DATE:
|
||||||
|
return f"CAST('{dttm.date().isoformat()}' AS DATE)"
|
||||||
|
if tt == utils.TemporalType.TIME:
|
||||||
|
return f"CAST('{dttm.time().isoformat()}' AS TIME)"
|
||||||
|
return None
|
||||||
@@ -44,7 +44,10 @@ from superset.reports.commands.exceptions import (
|
|||||||
ReportScheduleUnexpectedError,
|
ReportScheduleUnexpectedError,
|
||||||
ReportScheduleWorkingTimeoutError,
|
ReportScheduleWorkingTimeoutError,
|
||||||
)
|
)
|
||||||
from superset.reports.dao import ReportScheduleDAO
|
from superset.reports.dao import (
|
||||||
|
REPORT_SCHEDULE_ERROR_NOTIFICATION_MARKER,
|
||||||
|
ReportScheduleDAO,
|
||||||
|
)
|
||||||
from superset.reports.notifications import create_notification
|
from superset.reports.notifications import create_notification
|
||||||
from superset.reports.notifications.base import NotificationContent, ScreenshotData
|
from superset.reports.notifications.base import NotificationContent, ScreenshotData
|
||||||
from superset.reports.notifications.exceptions import NotificationError
|
from superset.reports.notifications.exceptions import NotificationError
|
||||||
@@ -147,6 +150,7 @@ class BaseReportState:
|
|||||||
def _get_screenshot(self) -> ScreenshotData:
|
def _get_screenshot(self) -> ScreenshotData:
|
||||||
"""
|
"""
|
||||||
Get a chart or dashboard screenshot
|
Get a chart or dashboard screenshot
|
||||||
|
|
||||||
:raises: ReportScheduleScreenshotFailedError
|
:raises: ReportScheduleScreenshotFailedError
|
||||||
"""
|
"""
|
||||||
screenshot: Optional[BaseScreenshot] = None
|
screenshot: Optional[BaseScreenshot] = None
|
||||||
@@ -170,6 +174,7 @@ class BaseReportState:
|
|||||||
def _get_notification_content(self) -> NotificationContent:
|
def _get_notification_content(self) -> NotificationContent:
|
||||||
"""
|
"""
|
||||||
Gets a notification content, this is composed by a title and a screenshot
|
Gets a notification content, this is composed by a title and a screenshot
|
||||||
|
|
||||||
:raises: ReportScheduleScreenshotFailedError
|
:raises: ReportScheduleScreenshotFailedError
|
||||||
"""
|
"""
|
||||||
screenshot_data = self._get_screenshot()
|
screenshot_data = self._get_screenshot()
|
||||||
@@ -185,14 +190,13 @@ class BaseReportState:
|
|||||||
)
|
)
|
||||||
return NotificationContent(name=name, screenshot=screenshot_data)
|
return NotificationContent(name=name, screenshot=screenshot_data)
|
||||||
|
|
||||||
def send(self) -> None:
|
def _send(self, notification_content: NotificationContent) -> None:
|
||||||
"""
|
"""
|
||||||
Creates the notification content and sends them to all recipients
|
Sends a notification to all recipients
|
||||||
|
|
||||||
:raises: ReportScheduleNotificationError
|
:raises: ReportScheduleNotificationError
|
||||||
"""
|
"""
|
||||||
notification_errors = []
|
notification_errors = []
|
||||||
notification_content = self._get_notification_content()
|
|
||||||
for recipient in self._report_schedule.recipients:
|
for recipient in self._report_schedule.recipients:
|
||||||
notification = create_notification(recipient, notification_content)
|
notification = create_notification(recipient, notification_content)
|
||||||
try:
|
try:
|
||||||
@@ -203,6 +207,24 @@ class BaseReportState:
|
|||||||
if notification_errors:
|
if notification_errors:
|
||||||
raise ReportScheduleNotificationError(";".join(notification_errors))
|
raise ReportScheduleNotificationError(";".join(notification_errors))
|
||||||
|
|
||||||
|
def send(self) -> None:
|
||||||
|
"""
|
||||||
|
Creates the notification content and sends them to all recipients
|
||||||
|
|
||||||
|
:raises: ReportScheduleNotificationError
|
||||||
|
"""
|
||||||
|
notification_content = self._get_notification_content()
|
||||||
|
self._send(notification_content)
|
||||||
|
|
||||||
|
def send_error(self, name: str, message: str) -> None:
|
||||||
|
"""
|
||||||
|
Creates and sends a notification for an error, to all recipients
|
||||||
|
|
||||||
|
:raises: ReportScheduleNotificationError
|
||||||
|
"""
|
||||||
|
notification_content = NotificationContent(name=name, text=message)
|
||||||
|
self._send(notification_content)
|
||||||
|
|
||||||
def is_in_grace_period(self) -> bool:
|
def is_in_grace_period(self) -> bool:
|
||||||
"""
|
"""
|
||||||
Checks if an alert is on it's grace period
|
Checks if an alert is on it's grace period
|
||||||
@@ -218,6 +240,23 @@ class BaseReportState:
|
|||||||
< last_success.end_dttm
|
< last_success.end_dttm
|
||||||
)
|
)
|
||||||
|
|
||||||
|
def is_in_error_grace_period(self) -> bool:
|
||||||
|
"""
|
||||||
|
Checks if an alert/report on error is on it's notification grace period
|
||||||
|
"""
|
||||||
|
last_success = ReportScheduleDAO.find_last_error_notification(
|
||||||
|
self._report_schedule, session=self._session
|
||||||
|
)
|
||||||
|
if not last_success:
|
||||||
|
return False
|
||||||
|
return (
|
||||||
|
last_success is not None
|
||||||
|
and self._report_schedule.grace_period
|
||||||
|
and datetime.utcnow()
|
||||||
|
- timedelta(seconds=self._report_schedule.grace_period)
|
||||||
|
< last_success.end_dttm
|
||||||
|
)
|
||||||
|
|
||||||
def is_on_working_timeout(self) -> bool:
|
def is_on_working_timeout(self) -> bool:
|
||||||
"""
|
"""
|
||||||
Checks if an alert is on a working timeout
|
Checks if an alert is on a working timeout
|
||||||
@@ -256,9 +295,25 @@ class ReportNotTriggeredErrorState(BaseReportState):
|
|||||||
return
|
return
|
||||||
self.send()
|
self.send()
|
||||||
self.set_state_and_log(ReportState.SUCCESS)
|
self.set_state_and_log(ReportState.SUCCESS)
|
||||||
except CommandException as ex:
|
except CommandException as first_ex:
|
||||||
self.set_state_and_log(ReportState.ERROR, error_message=str(ex))
|
self.set_state_and_log(ReportState.ERROR, error_message=str(first_ex))
|
||||||
raise ex
|
# TODO (dpgaspar) convert this logic to a new state eg: ERROR_ON_GRACE
|
||||||
|
if not self.is_in_error_grace_period():
|
||||||
|
try:
|
||||||
|
self.send_error(
|
||||||
|
f"Error occurred for {self._report_schedule.type}:"
|
||||||
|
f" {self._report_schedule.name}",
|
||||||
|
str(first_ex),
|
||||||
|
)
|
||||||
|
self.set_state_and_log(
|
||||||
|
ReportState.ERROR,
|
||||||
|
error_message=REPORT_SCHEDULE_ERROR_NOTIFICATION_MARKER,
|
||||||
|
)
|
||||||
|
except CommandException as second_ex:
|
||||||
|
self.set_state_and_log(
|
||||||
|
ReportState.ERROR, error_message=str(second_ex)
|
||||||
|
)
|
||||||
|
raise first_ex
|
||||||
|
|
||||||
|
|
||||||
class ReportWorkingState(BaseReportState):
|
class ReportWorkingState(BaseReportState):
|
||||||
|
|||||||
@@ -50,9 +50,9 @@ class AsyncPruneReportScheduleLogCommand(BaseCommand):
|
|||||||
report_schedule, from_date, session=session, commit=False
|
report_schedule, from_date, session=session, commit=False
|
||||||
)
|
)
|
||||||
logger.info(
|
logger.info(
|
||||||
"Deleted %s logs for %s",
|
"Deleted %s logs for report schedule id: %s",
|
||||||
str(row_count),
|
str(row_count),
|
||||||
ReportSchedule.name,
|
str(report_schedule.id),
|
||||||
)
|
)
|
||||||
except DAODeleteFailedError as ex:
|
except DAODeleteFailedError as ex:
|
||||||
prune_errors.append(str(ex))
|
prune_errors.append(str(ex))
|
||||||
|
|||||||
@@ -35,6 +35,9 @@ from superset.models.reports import (
|
|||||||
logger = logging.getLogger(__name__)
|
logger = logging.getLogger(__name__)
|
||||||
|
|
||||||
|
|
||||||
|
REPORT_SCHEDULE_ERROR_NOTIFICATION_MARKER = "Notification sent with error"
|
||||||
|
|
||||||
|
|
||||||
class ReportScheduleDAO(BaseDAO):
|
class ReportScheduleDAO(BaseDAO):
|
||||||
model_cls = ReportSchedule
|
model_cls = ReportSchedule
|
||||||
|
|
||||||
@@ -223,6 +226,41 @@ class ReportScheduleDAO(BaseDAO):
|
|||||||
.first()
|
.first()
|
||||||
)
|
)
|
||||||
|
|
||||||
|
@staticmethod
|
||||||
|
def find_last_error_notification(
|
||||||
|
report_schedule: ReportSchedule, session: Optional[Session] = None,
|
||||||
|
) -> Optional[ReportExecutionLog]:
|
||||||
|
"""
|
||||||
|
Finds last error email sent
|
||||||
|
"""
|
||||||
|
session = session or db.session
|
||||||
|
last_error_email_log = (
|
||||||
|
session.query(ReportExecutionLog)
|
||||||
|
.filter(
|
||||||
|
ReportExecutionLog.error_message
|
||||||
|
== REPORT_SCHEDULE_ERROR_NOTIFICATION_MARKER,
|
||||||
|
ReportExecutionLog.report_schedule == report_schedule,
|
||||||
|
)
|
||||||
|
.order_by(ReportExecutionLog.end_dttm.desc())
|
||||||
|
.first()
|
||||||
|
)
|
||||||
|
if not last_error_email_log:
|
||||||
|
return None
|
||||||
|
# Checks that only errors have occurred since the last email
|
||||||
|
report_from_last_email = (
|
||||||
|
session.query(ReportExecutionLog)
|
||||||
|
.filter(
|
||||||
|
ReportExecutionLog.state.notin_(
|
||||||
|
[ReportState.ERROR, ReportState.WORKING]
|
||||||
|
),
|
||||||
|
ReportExecutionLog.report_schedule == report_schedule,
|
||||||
|
ReportExecutionLog.end_dttm < last_error_email_log.end_dttm,
|
||||||
|
)
|
||||||
|
.order_by(ReportExecutionLog.end_dttm.desc())
|
||||||
|
.first()
|
||||||
|
)
|
||||||
|
return last_error_email_log if not report_from_last_email else None
|
||||||
|
|
||||||
@staticmethod
|
@staticmethod
|
||||||
def bulk_delete_logs(
|
def bulk_delete_logs(
|
||||||
model: ReportSchedule,
|
model: ReportSchedule,
|
||||||
|
|||||||
@@ -30,7 +30,8 @@ class ScreenshotData:
|
|||||||
@dataclass
|
@dataclass
|
||||||
class NotificationContent:
|
class NotificationContent:
|
||||||
name: str
|
name: str
|
||||||
screenshot: ScreenshotData
|
screenshot: Optional[ScreenshotData] = None
|
||||||
|
text: Optional[str] = None
|
||||||
|
|
||||||
|
|
||||||
class BaseNotification: # pylint: disable=too-few-public-methods
|
class BaseNotification: # pylint: disable=too-few-public-methods
|
||||||
|
|||||||
@@ -19,7 +19,7 @@ import json
|
|||||||
import logging
|
import logging
|
||||||
from dataclasses import dataclass
|
from dataclasses import dataclass
|
||||||
from email.utils import make_msgid, parseaddr
|
from email.utils import make_msgid, parseaddr
|
||||||
from typing import Dict
|
from typing import Dict, Optional
|
||||||
|
|
||||||
from flask_babel import gettext as __
|
from flask_babel import gettext as __
|
||||||
|
|
||||||
@@ -35,7 +35,7 @@ logger = logging.getLogger(__name__)
|
|||||||
@dataclass
|
@dataclass
|
||||||
class EmailContent:
|
class EmailContent:
|
||||||
body: str
|
body: str
|
||||||
images: Dict[str, bytes]
|
images: Optional[Dict[str, bytes]] = None
|
||||||
|
|
||||||
|
|
||||||
class EmailNotification(BaseNotification): # pylint: disable=too-few-public-methods
|
class EmailNotification(BaseNotification): # pylint: disable=too-few-public-methods
|
||||||
@@ -49,9 +49,21 @@ class EmailNotification(BaseNotification): # pylint: disable=too-few-public-met
|
|||||||
def _get_smtp_domain() -> str:
|
def _get_smtp_domain() -> str:
|
||||||
return parseaddr(app.config["SMTP_MAIL_FROM"])[1].split("@")[1]
|
return parseaddr(app.config["SMTP_MAIL_FROM"])[1].split("@")[1]
|
||||||
|
|
||||||
|
@staticmethod
|
||||||
|
def _error_template(text: str) -> str:
|
||||||
|
return __(
|
||||||
|
"""
|
||||||
|
Error: %(text)s
|
||||||
|
""",
|
||||||
|
text=text,
|
||||||
|
)
|
||||||
|
|
||||||
def _get_content(self) -> EmailContent:
|
def _get_content(self) -> EmailContent:
|
||||||
|
if self._content.text:
|
||||||
|
return EmailContent(body=self._error_template(self._content.text))
|
||||||
# Get the domain from the 'From' address ..
|
# Get the domain from the 'From' address ..
|
||||||
# and make a message id without the < > in the ends
|
# and make a message id without the < > in the end
|
||||||
|
if self._content.screenshot:
|
||||||
domain = self._get_smtp_domain()
|
domain = self._get_smtp_domain()
|
||||||
msgid = make_msgid(domain)[1:-1]
|
msgid = make_msgid(domain)[1:-1]
|
||||||
|
|
||||||
@@ -65,6 +77,7 @@ class EmailNotification(BaseNotification): # pylint: disable=too-few-public-met
|
|||||||
msgid=msgid,
|
msgid=msgid,
|
||||||
)
|
)
|
||||||
return EmailContent(body=body, images=image)
|
return EmailContent(body=body, images=image)
|
||||||
|
return EmailContent(body=self._error_template("Unexpected missing screenshot"))
|
||||||
|
|
||||||
def _get_subject(self) -> str:
|
def _get_subject(self) -> str:
|
||||||
return __(
|
return __(
|
||||||
|
|||||||
@@ -18,13 +18,12 @@
|
|||||||
import json
|
import json
|
||||||
import logging
|
import logging
|
||||||
from io import IOBase
|
from io import IOBase
|
||||||
from typing import cast, Optional, Union
|
from typing import Optional, Union
|
||||||
|
|
||||||
from flask_babel import gettext as __
|
from flask_babel import gettext as __
|
||||||
from retry.api import retry
|
from retry.api import retry
|
||||||
from slack import WebClient
|
from slack import WebClient
|
||||||
from slack.errors import SlackApiError, SlackClientError
|
from slack.errors import SlackApiError, SlackClientError
|
||||||
from slack.web.slack_response import SlackResponse
|
|
||||||
|
|
||||||
from superset import app
|
from superset import app
|
||||||
from superset.models.reports import ReportRecipientType
|
from superset.models.reports import ReportRecipientType
|
||||||
@@ -44,7 +43,21 @@ class SlackNotification(BaseNotification): # pylint: disable=too-few-public-met
|
|||||||
def _get_channel(self) -> str:
|
def _get_channel(self) -> str:
|
||||||
return json.loads(self._recipient.recipient_config_json)["target"]
|
return json.loads(self._recipient.recipient_config_json)["target"]
|
||||||
|
|
||||||
|
@staticmethod
|
||||||
|
def _error_template(name: str, text: str) -> str:
|
||||||
|
return __(
|
||||||
|
"""
|
||||||
|
*%(name)s*\n
|
||||||
|
Error: %(text)s
|
||||||
|
""",
|
||||||
|
name=name,
|
||||||
|
text=text,
|
||||||
|
)
|
||||||
|
|
||||||
def _get_body(self) -> str:
|
def _get_body(self) -> str:
|
||||||
|
if self._content.text:
|
||||||
|
return self._error_template(self._content.name, self._content.text)
|
||||||
|
if self._content.screenshot:
|
||||||
return __(
|
return __(
|
||||||
"""
|
"""
|
||||||
*%(name)s*\n
|
*%(name)s*\n
|
||||||
@@ -53,37 +66,29 @@ class SlackNotification(BaseNotification): # pylint: disable=too-few-public-met
|
|||||||
name=self._content.name,
|
name=self._content.name,
|
||||||
url=self._content.screenshot.url,
|
url=self._content.screenshot.url,
|
||||||
)
|
)
|
||||||
|
return self._error_template(self._content.name, "Unexpected missing screenshot")
|
||||||
|
|
||||||
def _get_inline_screenshot(self) -> Optional[Union[str, IOBase, bytes]]:
|
def _get_inline_screenshot(self) -> Optional[Union[str, IOBase, bytes]]:
|
||||||
|
if self._content.screenshot:
|
||||||
return self._content.screenshot.image
|
return self._content.screenshot.image
|
||||||
|
return None
|
||||||
|
|
||||||
@retry(SlackApiError, delay=10, backoff=2, tries=5)
|
@retry(SlackApiError, delay=10, backoff=2, tries=5)
|
||||||
def send(self) -> None:
|
def send(self) -> None:
|
||||||
file = self._get_inline_screenshot()
|
file = self._get_inline_screenshot()
|
||||||
channel = self._get_channel()
|
channel = self._get_channel()
|
||||||
body = self._get_body()
|
body = self._get_body()
|
||||||
|
|
||||||
try:
|
try:
|
||||||
client = WebClient(
|
client = WebClient(
|
||||||
token=app.config["SLACK_API_TOKEN"], proxy=app.config["SLACK_PROXY"]
|
token=app.config["SLACK_API_TOKEN"], proxy=app.config["SLACK_PROXY"]
|
||||||
)
|
)
|
||||||
# files_upload returns SlackResponse as we run it in sync mode.
|
# files_upload returns SlackResponse as we run it in sync mode.
|
||||||
if file:
|
if file:
|
||||||
response = cast(
|
|
||||||
SlackResponse,
|
|
||||||
client.files_upload(
|
client.files_upload(
|
||||||
channels=channel,
|
channels=channel, file=file, initial_comment=body, title="subject",
|
||||||
file=file,
|
|
||||||
initial_comment=body,
|
|
||||||
title="subject",
|
|
||||||
),
|
|
||||||
)
|
)
|
||||||
assert response["file"], str(response) # the uploaded file
|
|
||||||
else:
|
else:
|
||||||
response = cast(
|
client.chat_postMessage(channel=channel, text=body)
|
||||||
SlackResponse, client.chat_postMessage(channel=channel, text=body),
|
|
||||||
)
|
|
||||||
assert response["message"]["text"], str(response)
|
|
||||||
logger.info("Report sent to slack")
|
logger.info("Report sent to slack")
|
||||||
except SlackClientError as ex:
|
except SlackClientError as ex:
|
||||||
raise NotificationError(ex)
|
raise NotificationError(ex)
|
||||||
|
|||||||
@@ -958,7 +958,7 @@ def send_mime_email(
|
|||||||
smtp.starttls()
|
smtp.starttls()
|
||||||
if smtp_user and smtp_password:
|
if smtp_user and smtp_password:
|
||||||
smtp.login(smtp_user, smtp_password)
|
smtp.login(smtp_user, smtp_password)
|
||||||
logger.info("Sent an email to %s", str(e_to))
|
logger.debug("Sent an email to %s", str(e_to))
|
||||||
smtp.sendmail(e_from, e_to, mime_msg.as_string())
|
smtp.sendmail(e_from, e_to, mime_msg.as_string())
|
||||||
smtp.quit()
|
smtp.quit()
|
||||||
else:
|
else:
|
||||||
|
|||||||
@@ -14,6 +14,9 @@
|
|||||||
# KIND, either express or implied. See the License for the
|
# KIND, either express or implied. See the License for the
|
||||||
# specific language governing permissions and limitations
|
# specific language governing permissions and limitations
|
||||||
# under the License.
|
# under the License.
|
||||||
|
import logging
|
||||||
|
from typing import Optional
|
||||||
|
|
||||||
from flask import flash, request, Response
|
from flask import flash, request, Response
|
||||||
from flask_appbuilder import expose
|
from flask_appbuilder import expose
|
||||||
from flask_appbuilder.security.decorators import has_access_api
|
from flask_appbuilder.security.decorators import has_access_api
|
||||||
@@ -24,11 +27,22 @@ from superset.models import core as models
|
|||||||
from superset.typing import FlaskResponse
|
from superset.typing import FlaskResponse
|
||||||
from superset.views.base import BaseSupersetView
|
from superset.views.base import BaseSupersetView
|
||||||
|
|
||||||
|
logger = logging.getLogger(__name__)
|
||||||
|
|
||||||
|
|
||||||
class R(BaseSupersetView): # pylint: disable=invalid-name
|
class R(BaseSupersetView): # pylint: disable=invalid-name
|
||||||
|
|
||||||
"""used for short urls"""
|
"""used for short urls"""
|
||||||
|
|
||||||
|
@staticmethod
|
||||||
|
def _validate_url(url: Optional[str] = None) -> bool:
|
||||||
|
if url and (
|
||||||
|
url.startswith("//superset/dashboard/")
|
||||||
|
or url.startswith("//superset/explore/")
|
||||||
|
):
|
||||||
|
return True
|
||||||
|
return False
|
||||||
|
|
||||||
@event_logger.log_this
|
@event_logger.log_this
|
||||||
@expose("/<int:url_id>")
|
@expose("/<int:url_id>")
|
||||||
def index(self, url_id: int) -> FlaskResponse: # pylint: disable=no-self-use
|
def index(self, url_id: int) -> FlaskResponse: # pylint: disable=no-self-use
|
||||||
@@ -38,8 +52,9 @@ class R(BaseSupersetView): # pylint: disable=invalid-name
|
|||||||
if url.url.startswith(explore_url):
|
if url.url.startswith(explore_url):
|
||||||
explore_url += f"r={url_id}"
|
explore_url += f"r={url_id}"
|
||||||
return redirect(explore_url[1:])
|
return redirect(explore_url[1:])
|
||||||
|
if self._validate_url(url.url):
|
||||||
return redirect(url.url[1:])
|
return redirect(url.url[1:])
|
||||||
|
return redirect("/")
|
||||||
|
|
||||||
flash("URL to nowhere...", "danger")
|
flash("URL to nowhere...", "danger")
|
||||||
return redirect("/")
|
return redirect("/")
|
||||||
@@ -49,6 +64,9 @@ class R(BaseSupersetView): # pylint: disable=invalid-name
|
|||||||
@expose("/shortner/", methods=["POST"])
|
@expose("/shortner/", methods=["POST"])
|
||||||
def shortner(self) -> FlaskResponse: # pylint: disable=no-self-use
|
def shortner(self) -> FlaskResponse: # pylint: disable=no-self-use
|
||||||
url = request.form.get("data")
|
url = request.form.get("data")
|
||||||
|
if not self._validate_url(url):
|
||||||
|
logger.warning("Invalid URL: %s", url)
|
||||||
|
return Response(f"Invalid URL: {url}", 400)
|
||||||
obj = models.Url(url=url)
|
obj = models.Url(url=url)
|
||||||
db.session.add(obj)
|
db.session.add(obj)
|
||||||
db.session.commit()
|
db.session.commit()
|
||||||
|
|||||||
@@ -634,6 +634,28 @@ class TestCore(SupersetTestCase):
|
|||||||
resp = self.client.post("/r/shortner/", data=dict(data=data))
|
resp = self.client.post("/r/shortner/", data=dict(data=data))
|
||||||
assert re.search(r"\/r\/[0-9]+", resp.data.decode("utf-8"))
|
assert re.search(r"\/r\/[0-9]+", resp.data.decode("utf-8"))
|
||||||
|
|
||||||
|
def test_shortner_invalid(self):
|
||||||
|
self.login(username="admin")
|
||||||
|
invalid_urls = [
|
||||||
|
"hhttp://invalid.com",
|
||||||
|
"hhttps://invalid.com",
|
||||||
|
"www.invalid.com",
|
||||||
|
]
|
||||||
|
for invalid_url in invalid_urls:
|
||||||
|
resp = self.client.post("/r/shortner/", data=dict(data=invalid_url))
|
||||||
|
assert resp.status_code == 400
|
||||||
|
|
||||||
|
def test_redirect_invalid(self):
|
||||||
|
model_url = models.Url(url="hhttp://invalid.com")
|
||||||
|
db.session.add(model_url)
|
||||||
|
db.session.commit()
|
||||||
|
|
||||||
|
self.login(username="admin")
|
||||||
|
response = self.client.get(f"/r/{model_url.id}")
|
||||||
|
assert response.headers["Location"] == "http://localhost/"
|
||||||
|
db.session.delete(model_url)
|
||||||
|
db.session.commit()
|
||||||
|
|
||||||
@skipUnless(
|
@skipUnless(
|
||||||
(is_feature_enabled("KV_STORE")), "skipping as /kv/ endpoints are not enabled"
|
(is_feature_enabled("KV_STORE")), "skipping as /kv/ endpoints are not enabled"
|
||||||
)
|
)
|
||||||
|
|||||||
81
tests/db_engine_specs/firebird_tests.py
Normal file
81
tests/db_engine_specs/firebird_tests.py
Normal file
@@ -0,0 +1,81 @@
|
|||||||
|
# Licensed to the Apache Software Foundation (ASF) under one
|
||||||
|
# or more contributor license agreements. See the NOTICE file
|
||||||
|
# distributed with this work for additional information
|
||||||
|
# regarding copyright ownership. The ASF licenses this file
|
||||||
|
# to you under the Apache License, Version 2.0 (the
|
||||||
|
# "License"); you may not use this file except in compliance
|
||||||
|
# with the License. You may obtain a copy of the License at
|
||||||
|
#
|
||||||
|
# http://www.apache.org/licenses/LICENSE-2.0
|
||||||
|
#
|
||||||
|
# Unless required by applicable law or agreed to in writing,
|
||||||
|
# software distributed under the License is distributed on an
|
||||||
|
# "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY
|
||||||
|
# KIND, either express or implied. See the License for the
|
||||||
|
# specific language governing permissions and limitations
|
||||||
|
# under the License.
|
||||||
|
from datetime import datetime
|
||||||
|
from unittest import mock
|
||||||
|
|
||||||
|
import pytest
|
||||||
|
|
||||||
|
from superset.db_engine_specs.firebird import FirebirdEngineSpec
|
||||||
|
|
||||||
|
grain_expressions = {
|
||||||
|
None: "timestamp_column",
|
||||||
|
"PT1S": (
|
||||||
|
"CAST(CAST(timestamp_column AS DATE) "
|
||||||
|
"|| ' ' "
|
||||||
|
"|| EXTRACT(HOUR FROM timestamp_column) "
|
||||||
|
"|| ':' "
|
||||||
|
"|| EXTRACT(MINUTE FROM timestamp_column) "
|
||||||
|
"|| ':' "
|
||||||
|
"|| FLOOR(EXTRACT(SECOND FROM timestamp_column)) AS TIMESTAMP)"
|
||||||
|
),
|
||||||
|
"PT1M": (
|
||||||
|
"CAST(CAST(timestamp_column AS DATE) "
|
||||||
|
"|| ' ' "
|
||||||
|
"|| EXTRACT(HOUR FROM timestamp_column) "
|
||||||
|
"|| ':' "
|
||||||
|
"|| EXTRACT(MINUTE FROM timestamp_column) "
|
||||||
|
"|| ':00' AS TIMESTAMP)"
|
||||||
|
),
|
||||||
|
"P1D": "CAST(timestamp_column AS DATE)",
|
||||||
|
"P1M": (
|
||||||
|
"CAST(EXTRACT(YEAR FROM timestamp_column) "
|
||||||
|
"|| '-' "
|
||||||
|
"|| EXTRACT(MONTH FROM timestamp_column) "
|
||||||
|
"|| '-01' AS DATE)"
|
||||||
|
),
|
||||||
|
"P1Y": "CAST(EXTRACT(YEAR FROM timestamp_column) || '-01-01' AS DATE)",
|
||||||
|
}
|
||||||
|
|
||||||
|
|
||||||
|
@pytest.mark.parametrize("grain,expected", grain_expressions.items())
|
||||||
|
def test_time_grain_expressions(grain, expected):
|
||||||
|
assert (
|
||||||
|
FirebirdEngineSpec._time_grain_expressions[grain].format(col="timestamp_column")
|
||||||
|
== expected
|
||||||
|
)
|
||||||
|
|
||||||
|
|
||||||
|
def test_epoch_to_dttm():
|
||||||
|
assert (
|
||||||
|
FirebirdEngineSpec.epoch_to_dttm().format(col="timestamp_column")
|
||||||
|
== "DATEADD(second, timestamp_column, CAST('00:00:00' AS TIMESTAMP))"
|
||||||
|
)
|
||||||
|
|
||||||
|
|
||||||
|
def test_convert_dttm():
|
||||||
|
dttm = datetime(2021, 1, 1)
|
||||||
|
assert (
|
||||||
|
FirebirdEngineSpec.convert_dttm("timestamp", dttm)
|
||||||
|
== "CAST('2021-01-01 00:00:00' AS TIMESTAMP)"
|
||||||
|
)
|
||||||
|
assert (
|
||||||
|
FirebirdEngineSpec.convert_dttm("TIMESTAMP", dttm)
|
||||||
|
== "CAST('2021-01-01 00:00:00' AS TIMESTAMP)"
|
||||||
|
)
|
||||||
|
assert FirebirdEngineSpec.convert_dttm("TIME", dttm) == "CAST('00:00:00' AS TIME)"
|
||||||
|
assert FirebirdEngineSpec.convert_dttm("DATE", dttm) == "CAST('2021-01-01' AS DATE)"
|
||||||
|
assert FirebirdEngineSpec.convert_dttm("STRING", dttm) is None
|
||||||
@@ -21,6 +21,7 @@ from unittest.mock import Mock, patch
|
|||||||
|
|
||||||
import pytest
|
import pytest
|
||||||
from contextlib2 import contextmanager
|
from contextlib2 import contextmanager
|
||||||
|
from flask_sqlalchemy import BaseQuery
|
||||||
from freezegun import freeze_time
|
from freezegun import freeze_time
|
||||||
from sqlalchemy.sql import func
|
from sqlalchemy.sql import func
|
||||||
|
|
||||||
@@ -62,13 +63,34 @@ pytestmark = pytest.mark.usefixtures(
|
|||||||
)
|
)
|
||||||
|
|
||||||
|
|
||||||
def get_target_from_report_schedule(report_schedule) -> List[str]:
|
def get_target_from_report_schedule(report_schedule: ReportSchedule) -> List[str]:
|
||||||
return [
|
return [
|
||||||
json.loads(recipient.recipient_config_json)["target"]
|
json.loads(recipient.recipient_config_json)["target"]
|
||||||
for recipient in report_schedule.recipients
|
for recipient in report_schedule.recipients
|
||||||
]
|
]
|
||||||
|
|
||||||
|
|
||||||
|
def get_error_logs_query(report_schedule: ReportSchedule) -> BaseQuery:
|
||||||
|
return (
|
||||||
|
db.session.query(ReportExecutionLog)
|
||||||
|
.filter(
|
||||||
|
ReportExecutionLog.report_schedule == report_schedule,
|
||||||
|
ReportExecutionLog.state == ReportState.ERROR,
|
||||||
|
)
|
||||||
|
.order_by(ReportExecutionLog.end_dttm.desc())
|
||||||
|
)
|
||||||
|
|
||||||
|
|
||||||
|
def get_notification_error_sent_count(report_schedule: ReportSchedule) -> int:
|
||||||
|
logs = get_error_logs_query(report_schedule).all()
|
||||||
|
notification_sent_logs = [
|
||||||
|
log.error_message
|
||||||
|
for log in logs
|
||||||
|
if log.error_message == "Notification sent with error"
|
||||||
|
]
|
||||||
|
return len(notification_sent_logs)
|
||||||
|
|
||||||
|
|
||||||
def assert_log(state: str, error_message: Optional[str] = None):
|
def assert_log(state: str, error_message: Optional[str] = None):
|
||||||
db.session.commit()
|
db.session.commit()
|
||||||
logs = db.session.query(ReportExecutionLog).all()
|
logs = db.session.query(ReportExecutionLog).all()
|
||||||
@@ -77,6 +99,10 @@ def assert_log(state: str, error_message: Optional[str] = None):
|
|||||||
assert logs[0].error_message == error_message
|
assert logs[0].error_message == error_message
|
||||||
assert logs[0].state == state
|
assert logs[0].state == state
|
||||||
return
|
return
|
||||||
|
# On error we send an email
|
||||||
|
if state == ReportState.ERROR:
|
||||||
|
assert len(logs) == 3
|
||||||
|
else:
|
||||||
assert len(logs) == 2
|
assert len(logs) == 2
|
||||||
log_states = [log.state for log in logs]
|
log_states = [log.state for log in logs]
|
||||||
assert ReportState.WORKING in log_states
|
assert ReportState.WORKING in log_states
|
||||||
@@ -94,6 +120,7 @@ def create_report_notification(
|
|||||||
report_type: Optional[str] = None,
|
report_type: Optional[str] = None,
|
||||||
validator_type: Optional[str] = None,
|
validator_type: Optional[str] = None,
|
||||||
validator_config_json: Optional[str] = None,
|
validator_config_json: Optional[str] = None,
|
||||||
|
grace_period: Optional[int] = None,
|
||||||
) -> ReportSchedule:
|
) -> ReportSchedule:
|
||||||
report_type = report_type or ReportScheduleType.REPORT
|
report_type = report_type or ReportScheduleType.REPORT
|
||||||
target = email_target or slack_channel
|
target = email_target or slack_channel
|
||||||
@@ -121,6 +148,7 @@ def create_report_notification(
|
|||||||
recipients=[recipient],
|
recipients=[recipient],
|
||||||
validator_type=validator_type,
|
validator_type=validator_type,
|
||||||
validator_config_json=validator_config_json,
|
validator_config_json=validator_config_json,
|
||||||
|
grace_period=grace_period,
|
||||||
)
|
)
|
||||||
return report_schedule
|
return report_schedule
|
||||||
|
|
||||||
@@ -464,6 +492,7 @@ def create_invalid_sql_alert_email_chart(request):
|
|||||||
validator_config_json=param_config[request.param][
|
validator_config_json=param_config[request.param][
|
||||||
"validator_config_json"
|
"validator_config_json"
|
||||||
],
|
],
|
||||||
|
grace_period=60 * 60,
|
||||||
)
|
)
|
||||||
yield report_schedule
|
yield report_schedule
|
||||||
|
|
||||||
@@ -766,7 +795,8 @@ def test_email_mul_alert(create_mul_alert_email_chart):
|
|||||||
|
|
||||||
|
|
||||||
@pytest.mark.usefixtures("create_invalid_sql_alert_email_chart")
|
@pytest.mark.usefixtures("create_invalid_sql_alert_email_chart")
|
||||||
def test_invalid_sql_alert(create_invalid_sql_alert_email_chart):
|
@patch("superset.reports.notifications.email.send_email_smtp")
|
||||||
|
def test_invalid_sql_alert(email_mock, create_invalid_sql_alert_email_chart):
|
||||||
"""
|
"""
|
||||||
ExecuteReport Command: Test alert with invalid SQL statements
|
ExecuteReport Command: Test alert with invalid SQL statements
|
||||||
"""
|
"""
|
||||||
@@ -775,3 +805,120 @@ def test_invalid_sql_alert(create_invalid_sql_alert_email_chart):
|
|||||||
AsyncExecuteReportScheduleCommand(
|
AsyncExecuteReportScheduleCommand(
|
||||||
create_invalid_sql_alert_email_chart.id, datetime.utcnow()
|
create_invalid_sql_alert_email_chart.id, datetime.utcnow()
|
||||||
).run()
|
).run()
|
||||||
|
|
||||||
|
notification_targets = get_target_from_report_schedule(
|
||||||
|
create_invalid_sql_alert_email_chart
|
||||||
|
)
|
||||||
|
# Assert the email smtp address, asserts a notification was sent with the error
|
||||||
|
assert email_mock.call_args[0][0] == notification_targets[0]
|
||||||
|
|
||||||
|
|
||||||
|
@pytest.mark.usefixtures("create_invalid_sql_alert_email_chart")
|
||||||
|
@patch("superset.reports.notifications.email.send_email_smtp")
|
||||||
|
def test_grace_period_error(email_mock, create_invalid_sql_alert_email_chart):
|
||||||
|
"""
|
||||||
|
ExecuteReport Command: Test alert grace period on error
|
||||||
|
"""
|
||||||
|
with freeze_time("2020-01-01T00:00:00Z"):
|
||||||
|
with pytest.raises((AlertQueryError, AlertQueryInvalidTypeError)):
|
||||||
|
AsyncExecuteReportScheduleCommand(
|
||||||
|
create_invalid_sql_alert_email_chart.id, datetime.utcnow()
|
||||||
|
).run()
|
||||||
|
|
||||||
|
# Only needed for MySQL, understand why
|
||||||
|
db.session.commit()
|
||||||
|
notification_targets = get_target_from_report_schedule(
|
||||||
|
create_invalid_sql_alert_email_chart
|
||||||
|
)
|
||||||
|
# Assert the email smtp address, asserts a notification was sent with the error
|
||||||
|
assert email_mock.call_args[0][0] == notification_targets[0]
|
||||||
|
assert (
|
||||||
|
get_notification_error_sent_count(create_invalid_sql_alert_email_chart) == 1
|
||||||
|
)
|
||||||
|
|
||||||
|
with freeze_time("2020-01-01T00:30:00Z"):
|
||||||
|
with pytest.raises((AlertQueryError, AlertQueryInvalidTypeError)):
|
||||||
|
AsyncExecuteReportScheduleCommand(
|
||||||
|
create_invalid_sql_alert_email_chart.id, datetime.utcnow()
|
||||||
|
).run()
|
||||||
|
db.session.commit()
|
||||||
|
assert (
|
||||||
|
get_notification_error_sent_count(create_invalid_sql_alert_email_chart) == 1
|
||||||
|
)
|
||||||
|
|
||||||
|
# Grace period ends, assert a notification was sent
|
||||||
|
with freeze_time("2020-01-01T01:30:00Z"):
|
||||||
|
with pytest.raises((AlertQueryError, AlertQueryInvalidTypeError)):
|
||||||
|
AsyncExecuteReportScheduleCommand(
|
||||||
|
create_invalid_sql_alert_email_chart.id, datetime.utcnow()
|
||||||
|
).run()
|
||||||
|
db.session.commit()
|
||||||
|
assert (
|
||||||
|
get_notification_error_sent_count(create_invalid_sql_alert_email_chart) == 2
|
||||||
|
)
|
||||||
|
|
||||||
|
|
||||||
|
@pytest.mark.usefixtures("create_invalid_sql_alert_email_chart")
|
||||||
|
@patch("superset.reports.notifications.email.send_email_smtp")
|
||||||
|
@patch("superset.utils.screenshots.ChartScreenshot.compute_and_cache")
|
||||||
|
def test_grace_period_error_flap(
|
||||||
|
screenshot_mock, email_mock, create_invalid_sql_alert_email_chart
|
||||||
|
):
|
||||||
|
"""
|
||||||
|
ExecuteReport Command: Test alert grace period on error
|
||||||
|
"""
|
||||||
|
with freeze_time("2020-01-01T00:00:00Z"):
|
||||||
|
with pytest.raises((AlertQueryError, AlertQueryInvalidTypeError)):
|
||||||
|
AsyncExecuteReportScheduleCommand(
|
||||||
|
create_invalid_sql_alert_email_chart.id, datetime.utcnow()
|
||||||
|
).run()
|
||||||
|
db.session.commit()
|
||||||
|
# Assert we have 1 notification sent on the log
|
||||||
|
assert (
|
||||||
|
get_notification_error_sent_count(create_invalid_sql_alert_email_chart) == 1
|
||||||
|
)
|
||||||
|
|
||||||
|
with freeze_time("2020-01-01T00:30:00Z"):
|
||||||
|
with pytest.raises((AlertQueryError, AlertQueryInvalidTypeError)):
|
||||||
|
AsyncExecuteReportScheduleCommand(
|
||||||
|
create_invalid_sql_alert_email_chart.id, datetime.utcnow()
|
||||||
|
).run()
|
||||||
|
db.session.commit()
|
||||||
|
assert (
|
||||||
|
get_notification_error_sent_count(create_invalid_sql_alert_email_chart) == 1
|
||||||
|
)
|
||||||
|
|
||||||
|
# Change report_schedule to valid
|
||||||
|
create_invalid_sql_alert_email_chart.sql = "SELECT 1 AS metric"
|
||||||
|
create_invalid_sql_alert_email_chart.grace_period = 0
|
||||||
|
db.session.merge(create_invalid_sql_alert_email_chart)
|
||||||
|
db.session.commit()
|
||||||
|
|
||||||
|
with freeze_time("2020-01-01T00:31:00Z"):
|
||||||
|
# One success
|
||||||
|
AsyncExecuteReportScheduleCommand(
|
||||||
|
create_invalid_sql_alert_email_chart.id, datetime.utcnow()
|
||||||
|
).run()
|
||||||
|
# Grace period ends
|
||||||
|
AsyncExecuteReportScheduleCommand(
|
||||||
|
create_invalid_sql_alert_email_chart.id, datetime.utcnow()
|
||||||
|
).run()
|
||||||
|
|
||||||
|
db.session.commit()
|
||||||
|
|
||||||
|
create_invalid_sql_alert_email_chart.sql = "SELECT 'first'"
|
||||||
|
create_invalid_sql_alert_email_chart.grace_period = 10
|
||||||
|
db.session.merge(create_invalid_sql_alert_email_chart)
|
||||||
|
db.session.commit()
|
||||||
|
|
||||||
|
# assert that after a success, when back to error we send the error notification
|
||||||
|
# again
|
||||||
|
with freeze_time("2020-01-01T00:32:00Z"):
|
||||||
|
with pytest.raises((AlertQueryError, AlertQueryInvalidTypeError)):
|
||||||
|
AsyncExecuteReportScheduleCommand(
|
||||||
|
create_invalid_sql_alert_email_chart.id, datetime.utcnow()
|
||||||
|
).run()
|
||||||
|
db.session.commit()
|
||||||
|
assert (
|
||||||
|
get_notification_error_sent_count(create_invalid_sql_alert_email_chart) == 2
|
||||||
|
)
|
||||||
|
|||||||
Reference in New Issue
Block a user