# Licensed to the Apache Software Foundation (ASF) under one # or more contributor license agreements. See the NOTICE file # distributed with this work for additional information # regarding copyright ownership. The ASF licenses this file # to you under the Apache License, Version 2.0 (the # "License"); you may not use this file except in compliance # with the License. You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, # software distributed under the License is distributed on an # "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY # KIND, either express or implied. See the License for the # specific language governing permissions and limitations # under the License. """A collection of ORM sqlalchemy models for SQL Lab""" import builtins import inspect import logging import re from collections.abc import Hashable from datetime import datetime from typing import Any, Optional, TYPE_CHECKING import sqlalchemy as sqla from flask import current_app as app from flask_appbuilder import Model from flask_appbuilder.models.decorators import renders from flask_babel import gettext as __ from humanize import naturaltime from jinja2.exceptions import TemplateError from markupsafe import Markup from sqlalchemy import ( Boolean, Column, DateTime, Enum, ForeignKey, Integer, Numeric, String, Text, ) from sqlalchemy.engine.url import URL from sqlalchemy.orm import backref, relationship from sqlalchemy.sql.elements import ColumnElement, literal_column from superset_core.api.models import Query as CoreQuery, SavedQuery as CoreSavedQuery from superset import security_manager from superset.exceptions import SupersetParseError, SupersetSecurityException from superset.explorables.base import TimeGrainDict from superset.jinja_context import BaseTemplateProcessor, get_template_processor from superset.models.helpers import ( AuditMixinNullable, ExploreMixin, ExtraJSONMixin, ImportExportMixin, ) from superset.sql.parse import ( CTASMethod, process_jinja_sql, Table, ) from superset.sqllab.limiting_factor import LimitingFactor from superset.superset_typing import ExplorableData, QueryObjectDict from superset.utils import json from superset.utils.core import ( get_column_name, LongText, MediumText, QueryStatus, user_label, ) if TYPE_CHECKING: from superset.connectors.sqla.models import TableColumn from superset.db_engine_specs import BaseEngineSpec logger = logging.getLogger(__name__) class SqlTablesMixin: # pylint: disable=too-few-public-methods @property def sql_tables(self) -> list[Table]: try: return list( process_jinja_sql( self.sql, # type: ignore self.database, # type: ignore ).tables ) except (SupersetSecurityException, SupersetParseError, TemplateError): return [] class Query( CoreQuery, SqlTablesMixin, ExtraJSONMixin, ExploreMixin, ): # pylint: disable=abstract-method,too-many-public-methods """ORM model for SQL query Now that SQL Lab support multi-statement execution, an entry in this table may represent multiple SQL statements executed sequentially""" __tablename__ = "query" type = "query" id = Column(Integer, primary_key=True) client_id = Column(String(11), unique=True, nullable=False) query_language = "sql" database_id = Column(Integer, ForeignKey("dbs.id"), nullable=False) # Store the tmp table into the DB only if the user asks for it. tmp_table_name = Column(String(256)) tmp_schema_name = Column(String(256)) user_id = Column(Integer, ForeignKey("ab_user.id"), nullable=True) status = Column(String(16), default=QueryStatus.PENDING) tab_name = Column(String(256)) sql_editor_id = Column(String(256), index=True) schema = Column(String(256)) catalog = Column(String(256), nullable=True, default=None) sql = Column(LongText()) # Query to retrieve the results, # used only in case of select_as_cta_used is true. select_sql = Column(LongText()) executed_sql = Column(LongText()) # Could be configured in the superset config. limit = Column(Integer) limiting_factor = Column( Enum(LimitingFactor), server_default=LimitingFactor.UNKNOWN ) select_as_cta = Column(Boolean) select_as_cta_used = Column(Boolean, default=False) ctas_method = Column(String(16), default=CTASMethod.TABLE.name) progress = Column(Integer, default=0) # 1..100 # # of rows in the result set or rows modified. rows = Column(Integer) error_message = Column(Text) # key used to store the results in the results backend results_key = Column(String(64), index=True) # Using Numeric in place of DateTime for sub-second precision # stored as seconds since epoch, allowing for milliseconds start_time = Column(Numeric(precision=20, scale=6)) start_running_time = Column(Numeric(precision=20, scale=6)) end_time = Column(Numeric(precision=20, scale=6)) end_result_backend_time = Column(Numeric(precision=20, scale=6)) tracking_url_raw = Column(Text, name="tracking_url") changed_on = Column( DateTime, default=datetime.utcnow, onupdate=datetime.utcnow, nullable=True ) database = relationship( "Database", foreign_keys=[database_id], backref=backref("queries", cascade="all, delete-orphan"), ) user = relationship(security_manager.user_model, foreign_keys=[user_id]) __table_args__ = (sqla.Index("ti_user_id_changed_on", user_id, changed_on),) def get_template_processor(self, **kwargs: Any) -> BaseTemplateProcessor: return get_template_processor(query=self, database=self.database, **kwargs) def to_dict(self) -> dict[str, Any]: return { "changed_on": self.changed_on.isoformat(), "dbId": self.database_id, "db": self.database.database_name if self.database else None, "endDttm": self.end_time, "errorMessage": self.error_message, "executedSql": self.executed_sql, "id": self.client_id, "queryId": self.id, "limit": self.limit, "limitingFactor": self.limiting_factor, "progress": self.progress, "rows": self.rows, "catalog": self.catalog, "schema": self.schema, "ctas": self.select_as_cta, "serverId": self.id, "sql": self.sql, "sqlEditorId": self.sql_editor_id, "startDttm": self.start_time, "state": self.status.lower(), "tab": self.tab_name, "tempSchema": self.tmp_schema_name, "tempTable": self.tmp_table_name, "userId": self.user_id, "user": user_label(self.user), "resultsKey": self.results_key, "trackingUrl": self.tracking_url, "extra": self.extra, } @property def name(self) -> str: """Name property""" ts = datetime.now().isoformat() ts = ts.replace("-", "").replace(":", "").split(".")[0] tab = self.tab_name.replace(" ", "_").lower() if self.tab_name else "notab" tab = re.sub(r"\W+", "", tab) return f"sqllab_{tab}_{ts}" @property def database_name(self) -> str: return self.database.name @property def username(self) -> str: return self.user.username @property def columns(self) -> list["TableColumn"]: from superset.connectors.sqla.models import ( # pylint: disable=import-outside-toplevel TableColumn, ) return [ TableColumn( column_name=col["column_name"], database=self.database, is_dttm=col["is_dttm"], filterable=True, groupby=True, type=col["type"], ) for col in self.extra.get("columns", []) ] @property def db_extra(self) -> Optional[dict[str, Any]]: return None @property def data(self) -> ExplorableData: """Returns query data for the frontend""" order_by_choices = [] for col in self.columns: column_name = str(col.column_name or "") order_by_choices.append( (json.dumps([column_name, True]), f"{column_name} " + __("[asc]")) ) order_by_choices.append( (json.dumps([column_name, False]), f"{column_name} " + __("[desc]")) ) return { "time_grain_sqla": [ (g.duration, g.name) for g in self.database.grains() or [] ], "filter_select": True, "name": self.tab_name, "columns": [o.data for o in self.columns], "metrics": [], "id": self.id, "type": self.type, "sql": self.sql, "owners": self.owners_data, "database": {"id": self.database_id, "backend": self.database.backend}, "order_by_choices": order_by_choices, "catalog": self.catalog, "schema": self.schema, "verbose_map": {}, } def raise_for_access(self) -> None: """ Raise an exception if the user cannot access the resource. :raises SupersetSecurityException: If the user cannot access the resource """ security_manager.raise_for_access(query=self) @property def db_engine_spec( self, ) -> builtins.type["BaseEngineSpec"]: # pylint: disable=unsubscriptable-object return self.database.db_engine_spec @property def owners_data(self) -> list[dict[str, Any]]: return [] @property def uid(self) -> str: return f"{self.id}__{self.type}" @property def is_rls_supported(self) -> bool: return False @property def cache_timeout(self) -> int: return 0 @property def column_names(self) -> list[Any]: return [col.column_name for col in self.columns] @property def offset(self) -> int: return 0 @property def main_dttm_col(self) -> Optional[str]: for col in self.columns: if col.get("is_dttm"): return col.get("column_name") return None @property def dttm_cols(self) -> list[Any]: return [col.column_name for col in self.columns if col.is_dttm] @property def schema_perm(self) -> str: return f"{self.database.database_name}.{self.schema}" @property def perm(self) -> str: return f"[{self.database.database_name}].[{self.tab_name}](id:{self.id})" @property def default_endpoint(self) -> str: return "" def get_extra_cache_keys(self, query_obj: QueryObjectDict) -> list[Hashable]: return [] def get_time_grains(self) -> list[TimeGrainDict]: """ Get available time granularities from the database. Delegates to the database's time grain definitions. """ return [ { "name": grain.name, "function": grain.function, "duration": grain.duration, } for grain in (self.database.grains() or []) ] def has_drill_by_columns(self, column_names: list[str]) -> bool: """ Check if the specified columns support drill-by operations. For Query objects, all columns are considered drillable since they come from ad-hoc SQL queries without predefined metadata. """ if not column_names: return False return set(column_names).issubset(set(self.column_names)) @property def tracking_url(self) -> Optional[str]: """ Transform tracking url at run time because the exact URL may depend on query properties such as execution and finish time. """ transform = app.config.get("TRACKING_URL_TRANSFORMER") url = self.tracking_url_raw if url and transform: sig = inspect.signature(transform) # for backward compatibility, users may define a transformer function # with only one parameter (`url`). args = [url, self][: len(sig.parameters)] url = transform(*args) logger.debug("Transformed tracking url: %s", url) return url @tracking_url.setter def tracking_url(self, value: str) -> None: self.tracking_url_raw = value def get_column(self, column_name: Optional[str]) -> Optional[dict[str, Any]]: if not column_name: return None for col in self.columns: if col.column_name == column_name: return col return None def adhoc_column_to_sqla( self, col: "AdhocColumn", # type: ignore # noqa: F821 force_type_check: bool = False, template_processor: Optional[BaseTemplateProcessor] = None, ) -> ColumnElement: """ Turn an adhoc column into a sqlalchemy column. :param col: Adhoc column definition :param template_processor: template_processor instance :returns: The metric defined as a sqlalchemy column :rtype: sqlalchemy.sql.column """ label = get_column_name(col) expression = self._process_sql_expression( expression=col["sqlExpression"], database_id=self.database_id, engine=self.database.backend, schema=self.schema, template_processor=template_processor, ) sqla_column = literal_column(expression) return self.make_sqla_column_compatible(sqla_column, label) class SavedQuery( CoreSavedQuery, SqlTablesMixin, AuditMixinNullable, ExtraJSONMixin, ImportExportMixin, ): """ORM model for SQL query""" __tablename__ = "saved_query" id = Column(Integer, primary_key=True) user_id = Column(Integer, ForeignKey("ab_user.id"), nullable=True) db_id = Column(Integer, ForeignKey("dbs.id"), nullable=True) schema = Column(String(128)) catalog = Column(String(256), nullable=True, default=None) label = Column(String(256)) description = Column(Text) sql = Column(MediumText()) template_parameters = Column(Text) user = relationship( security_manager.user_model, backref=backref("saved_queries", cascade="all, delete-orphan"), foreign_keys=[user_id], ) database = relationship( "Database", foreign_keys=[db_id], backref=backref("saved_queries", cascade="all, delete-orphan"), ) rows = Column(Integer, nullable=True) last_run = Column(DateTime, nullable=True) tags = relationship( "Tag", secondary="tagged_object", overlaps="objects,tag,tags", primaryjoin="and_(SavedQuery.id == TaggedObject.object_id, " "TaggedObject.object_type == 'query')", secondaryjoin="TaggedObject.tag_id == Tag.id", viewonly=True, # cascading deletion already handled by superset.tags.models.ObjectUpdater.after_delete # noqa: E501 ) export_parent = "database" export_fields = [ "catalog", "schema", "label", "description", "sql", ] def __repr__(self) -> str: return str(self.label) def to_dict(self) -> dict[str, Any]: return { "id": self.id, } @property def pop_tab_link(self) -> Markup: return Markup( f""" """ ) @property def user_email(self) -> str: return self.user.email @property def sqlalchemy_uri(self) -> URL: return self.database.sqlalchemy_uri def url(self) -> str: return f"/sqllab?savedQueryId={self.id}" @property def last_run_humanized(self) -> str: return naturaltime(datetime.now() - self.changed_on) @property def _last_run_delta_humanized(self) -> str: return naturaltime(datetime.now() - self.changed_on) @renders("changed_on") def last_run_delta_humanized(self) -> str: return self._last_run_delta_humanized class TabState(AuditMixinNullable, ExtraJSONMixin, Model): __tablename__ = "tab_state" # basic info id = Column(Integer, primary_key=True, autoincrement=True) user_id = Column(Integer, ForeignKey("ab_user.id")) label = Column(String(256)) active = Column(Boolean, default=False) # selected DB and schema database_id = Column(Integer, ForeignKey("dbs.id", ondelete="CASCADE")) database = relationship("Database", foreign_keys=[database_id]) schema = Column(String(256)) catalog = Column(String(256), nullable=True, default=None) # tables that are open in the schema browser and their data previews table_schemas = relationship( "TableSchema", cascade="all, delete-orphan", backref="tab_state", passive_deletes=True, ) # the query in the textarea, and results (if any) sql = Column(MediumText()) query_limit = Column(Integer) # latest query that was run latest_query_id = Column( Integer, ForeignKey("query.client_id", ondelete="SET NULL") ) latest_query = relationship("Query") # other properties autorun = Column(Boolean, default=False) template_params = Column(Text) hide_left_bar = Column(Boolean, default=False) # any saved queries that are associated with the Tab State saved_query_id = Column( Integer, ForeignKey("saved_query.id", ondelete="SET NULL"), nullable=True ) saved_query = relationship("SavedQuery", foreign_keys=[saved_query_id]) def to_dict(self) -> dict[str, Any]: return { "id": self.id, "user_id": self.user_id, "label": self.label, "active": self.active, "database_id": self.database_id, "catalog": self.catalog, "schema": self.schema, "table_schemas": [ts.to_dict() for ts in self.table_schemas], "sql": self.sql, "query_limit": self.query_limit, "latest_query": self.latest_query.to_dict() if self.latest_query else None, "autorun": self.autorun, "template_params": self.template_params, "hide_left_bar": self.hide_left_bar, "saved_query": self.saved_query.to_dict() if self.saved_query else None, "extra_json": self.extra, } class TableSchema(AuditMixinNullable, ExtraJSONMixin, Model): __tablename__ = "table_schema" id = Column(Integer, primary_key=True, autoincrement=True) tab_state_id = Column(Integer, ForeignKey("tab_state.id", ondelete="CASCADE")) database_id = Column( Integer, ForeignKey("dbs.id", ondelete="CASCADE"), nullable=False ) database = relationship("Database", foreign_keys=[database_id]) schema = Column(String(256)) catalog = Column(String(256), nullable=True, default=None) table = Column(String(256)) # JSON describing the schema, partitions, latest partition, etc. description = Column(Text) expanded = Column(Boolean, default=False) def to_dict(self) -> dict[str, Any]: try: description = json.loads(self.description) except json.JSONDecodeError: description = None return { "id": self.id, "tab_state_id": self.tab_state_id, "database_id": self.database_id, "catalog": self.catalog, "schema": self.schema, "table": self.table, "description": description, "expanded": self.expanded, }