mirror of
https://github.com/apache/superset.git
synced 2026-04-10 03:45:22 +00:00
Co-authored-by: GitHub Action <action@github.com> Co-authored-by: Maxime Beauchemin <maximebeauchemin@gmail.com>
183 lines
6.9 KiB
Python
183 lines
6.9 KiB
Python
# Licensed to the Apache Software Foundation (ASF) under one
|
|
# or more contributor license agreements. See the NOTICE file
|
|
# distributed with this work for additional information
|
|
# regarding copyright ownership. The ASF licenses this file
|
|
# to you under the Apache License, Version 2.0 (the
|
|
# "License"); you may not use this file except in compliance
|
|
# with the License. You may obtain a copy of the License at
|
|
#
|
|
# http://www.apache.org/licenses/LICENSE-2.0
|
|
#
|
|
# Unless required by applicable law or agreed to in writing,
|
|
# software distributed under the License is distributed on an
|
|
# "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY
|
|
# KIND, either express or implied. See the License for the
|
|
# specific language governing permissions and limitations
|
|
# under the License.
|
|
from __future__ import annotations
|
|
|
|
import copy
|
|
import logging
|
|
from typing import Any, cast, TYPE_CHECKING
|
|
|
|
from celery.exceptions import SoftTimeLimitExceeded
|
|
from flask import current_app, g
|
|
from flask_appbuilder.security.sqla.models import User
|
|
from marshmallow import ValidationError
|
|
|
|
from superset.charts.schemas import ChartDataQueryContextSchema
|
|
from superset.exceptions import SupersetVizException
|
|
from superset.extensions import (
|
|
async_query_manager,
|
|
cache_manager,
|
|
celery_app,
|
|
security_manager,
|
|
)
|
|
from superset.utils.cache import generate_cache_key, set_and_log_cache
|
|
from superset.utils.core import override_user
|
|
from superset.views.utils import get_datasource_info, get_viz
|
|
|
|
if TYPE_CHECKING:
|
|
from superset.common.query_context import QueryContext
|
|
|
|
logger = logging.getLogger(__name__)
|
|
query_timeout = current_app.config[
|
|
"SQLLAB_ASYNC_TIME_LIMIT_SEC"
|
|
] # TODO: new config key
|
|
|
|
|
|
def set_form_data(form_data: dict[str, Any]) -> None:
|
|
g.form_data = form_data
|
|
|
|
|
|
def _create_query_context_from_form(form_data: dict[str, Any]) -> QueryContext:
|
|
"""
|
|
Create the query context from the form data.
|
|
|
|
:param form_data: The task form data
|
|
:returns: The query context
|
|
:raises ValidationError: If the request is incorrect
|
|
"""
|
|
|
|
try:
|
|
return ChartDataQueryContextSchema().load(form_data)
|
|
except KeyError as ex:
|
|
raise ValidationError("Request is incorrect") from ex
|
|
|
|
|
|
def _load_user_from_job_metadata(job_metadata: dict[str, Any]) -> User:
|
|
if user_id := job_metadata.get("user_id"):
|
|
# logged in user
|
|
user = security_manager.get_user_by_id(user_id)
|
|
elif guest_token := job_metadata.get("guest_token"):
|
|
# embedded guest user
|
|
user = security_manager.get_guest_user_from_token(guest_token)
|
|
del job_metadata["guest_token"]
|
|
else:
|
|
# default to anonymous user if no user is found
|
|
user = security_manager.get_anonymous_user()
|
|
return user
|
|
|
|
|
|
@celery_app.task(name="load_chart_data_into_cache", soft_time_limit=query_timeout)
|
|
def load_chart_data_into_cache(
|
|
job_metadata: dict[str, Any],
|
|
form_data: dict[str, Any],
|
|
) -> None:
|
|
# pylint: disable=import-outside-toplevel
|
|
from superset.commands.chart.data.get_data_command import ChartDataCommand
|
|
|
|
with override_user(_load_user_from_job_metadata(job_metadata), force=False):
|
|
try:
|
|
set_form_data(form_data)
|
|
query_context = _create_query_context_from_form(form_data)
|
|
command = ChartDataCommand(query_context)
|
|
result = command.run(cache=True)
|
|
cache_key = result["cache_key"]
|
|
result_url = f"/api/v1/chart/data/{cache_key}"
|
|
async_query_manager.update_job(
|
|
job_metadata,
|
|
async_query_manager.STATUS_DONE,
|
|
result_url=result_url,
|
|
)
|
|
except SoftTimeLimitExceeded as ex:
|
|
logger.warning("A timeout occurred while loading chart data, error: %s", ex)
|
|
raise
|
|
except Exception as ex:
|
|
# TODO: QueryContext should support SIP-40 style errors
|
|
error = str(ex.message if hasattr(ex, "message") else ex)
|
|
errors = [{"message": error}]
|
|
async_query_manager.update_job(
|
|
job_metadata, async_query_manager.STATUS_ERROR, errors=errors
|
|
)
|
|
raise
|
|
|
|
|
|
@celery_app.task(name="load_explore_json_into_cache", soft_time_limit=query_timeout)
|
|
def load_explore_json_into_cache( # pylint: disable=too-many-locals
|
|
job_metadata: dict[str, Any],
|
|
form_data: dict[str, Any],
|
|
response_type: str | None = None,
|
|
force: bool = False,
|
|
) -> None:
|
|
cache_key_prefix = "ejr-" # ejr: explore_json request
|
|
|
|
with override_user(_load_user_from_job_metadata(job_metadata), force=False):
|
|
try:
|
|
set_form_data(form_data)
|
|
datasource_id, datasource_type = get_datasource_info(None, None, form_data)
|
|
|
|
# Perform a deep copy here so that below we can cache the original
|
|
# value of the form_data object. This is necessary since the viz
|
|
# objects modify the form_data object. If the modified version were
|
|
# to be cached here, it will lead to a cache miss when clients
|
|
# attempt to retrieve the value of the completed async query.
|
|
original_form_data = copy.deepcopy(form_data)
|
|
|
|
viz_obj = get_viz(
|
|
datasource_type=cast(str, datasource_type),
|
|
datasource_id=datasource_id,
|
|
form_data=form_data,
|
|
force=force,
|
|
)
|
|
# run query & cache results
|
|
payload = viz_obj.get_payload()
|
|
if viz_obj.has_error(payload):
|
|
raise SupersetVizException(errors=payload["errors"])
|
|
|
|
# Cache the original form_data value for async retrieval
|
|
cache_value = {
|
|
"form_data": original_form_data,
|
|
"response_type": response_type,
|
|
}
|
|
cache_key = generate_cache_key(cache_value, cache_key_prefix)
|
|
cache_instance = cache_manager.cache
|
|
cache_timeout = (
|
|
cache_instance.cache.default_timeout if cache_instance.cache else None
|
|
)
|
|
set_and_log_cache(
|
|
cache_instance, cache_key, cache_value, cache_timeout=cache_timeout
|
|
)
|
|
result_url = f"/superset/explore_json/data/{cache_key}"
|
|
async_query_manager.update_job(
|
|
job_metadata,
|
|
async_query_manager.STATUS_DONE,
|
|
result_url=result_url,
|
|
)
|
|
except SoftTimeLimitExceeded as ex:
|
|
logger.warning(
|
|
"A timeout occurred while loading explore json, error: %s", ex
|
|
)
|
|
raise
|
|
except Exception as ex:
|
|
if isinstance(ex, SupersetVizException):
|
|
errors = ex.errors
|
|
else:
|
|
error = ex.message if hasattr(ex, "message") else str(ex)
|
|
errors = [error] # type: ignore
|
|
|
|
async_query_manager.update_job(
|
|
job_metadata, async_query_manager.STATUS_ERROR, errors=errors
|
|
)
|
|
raise
|