mirror of
https://github.com/apache/superset.git
synced 2026-05-14 04:15:12 +00:00
Compare commits
8 Commits
showtime-m
...
fix-asyncs
| Author | SHA1 | Date | |
|---|---|---|---|
|
|
83fe647a2b | ||
|
|
cfb704dbeb | ||
|
|
e77f6ece92 | ||
|
|
785a08c7d5 | ||
|
|
d90d3a2dea | ||
|
|
6ee4d694bc | ||
|
|
006a1800be | ||
|
|
c81ba84a46 |
10
.github/workflows/superset-docs-deploy.yml
vendored
10
.github/workflows/superset-docs-deploy.yml
vendored
@@ -17,6 +17,16 @@ on:
|
||||
|
||||
workflow_dispatch: {}
|
||||
|
||||
# Serialize deploys: the action pushes to apache/superset-site without
|
||||
# rebasing, so concurrent runs race on the final push and the loser fails
|
||||
# with `! [rejected] asf-site -> asf-site (fetch first)`. Cancel any
|
||||
# in-progress run as soon as a newer one starts — the destination repo
|
||||
# isn't touched until the final push step, so canceling mid-build is safe,
|
||||
# and the freshest content always wins.
|
||||
concurrency:
|
||||
group: docs-deploy-asf-site
|
||||
cancel-in-progress: true
|
||||
|
||||
jobs:
|
||||
config:
|
||||
runs-on: ubuntu-24.04
|
||||
|
||||
@@ -897,6 +897,86 @@ test('fires onChange when pasting a selection', async () => {
|
||||
await waitFor(() => expect(onChange).toHaveBeenCalledTimes(1));
|
||||
});
|
||||
|
||||
test('replaces cached options with search results instead of merging', async () => {
|
||||
const page0Data = Array.from({ length: 10 }, (_, i) => ({
|
||||
label: `Option ${i}`,
|
||||
value: i,
|
||||
}));
|
||||
const searchData = [{ label: 'Search Match', value: 100 }];
|
||||
const loadOptions = jest.fn(async (search: string) => {
|
||||
if (search === '') {
|
||||
return { data: page0Data, totalCount: 100 };
|
||||
}
|
||||
return { data: searchData, totalCount: 1 };
|
||||
});
|
||||
|
||||
render(<AsyncSelect {...defaultProps} options={loadOptions} />);
|
||||
await open();
|
||||
await waitFor(() => expect(loadOptions).toHaveBeenCalledTimes(1));
|
||||
|
||||
let options = await findAllSelectOptions();
|
||||
expect(options).toHaveLength(10);
|
||||
|
||||
await type('search');
|
||||
await waitFor(() => expect(loadOptions).toHaveBeenCalledTimes(2));
|
||||
|
||||
options = await findAllSelectOptions();
|
||||
expect(options).toHaveLength(1);
|
||||
expect(options[0]).toHaveTextContent('Search Match');
|
||||
});
|
||||
|
||||
test('shows all options when filterOption is false', async () => {
|
||||
const loadOptions = jest.fn(async () => ({
|
||||
data: OPTIONS.slice(0, 10),
|
||||
totalCount: 20,
|
||||
}));
|
||||
|
||||
render(
|
||||
<AsyncSelect
|
||||
{...defaultProps}
|
||||
options={loadOptions}
|
||||
filterOption={false}
|
||||
/>,
|
||||
);
|
||||
await open();
|
||||
await waitFor(() => expect(loadOptions).toHaveBeenCalledTimes(1));
|
||||
|
||||
await type('zzz_no_match');
|
||||
await waitFor(() => expect(loadOptions).toHaveBeenCalledTimes(2));
|
||||
|
||||
const options = await findAllSelectOptions();
|
||||
expect(options.length).toBeGreaterThan(0);
|
||||
});
|
||||
|
||||
test('restores base options when search is cleared', async () => {
|
||||
const page0Data = Array.from({ length: 10 }, (_, i) => ({
|
||||
label: `Option ${i}`,
|
||||
value: i,
|
||||
}));
|
||||
const searchData = [{ label: 'Search Match', value: 100 }];
|
||||
const loadOptions = jest.fn(async (search: string) => {
|
||||
if (search === '') {
|
||||
return { data: page0Data, totalCount: 100 };
|
||||
}
|
||||
return { data: searchData, totalCount: 1 };
|
||||
});
|
||||
|
||||
render(<AsyncSelect {...defaultProps} options={loadOptions} />);
|
||||
await open();
|
||||
await waitFor(() => expect(loadOptions).toHaveBeenCalledTimes(1));
|
||||
|
||||
await type('search');
|
||||
await waitFor(() => expect(loadOptions).toHaveBeenCalledTimes(2));
|
||||
let options = await findAllSelectOptions();
|
||||
expect(options).toHaveLength(1);
|
||||
|
||||
await type('{backspace}{backspace}{backspace}{backspace}{backspace}{backspace}');
|
||||
await waitFor(async () => {
|
||||
options = await findAllSelectOptions();
|
||||
expect(options).toHaveLength(10);
|
||||
});
|
||||
});
|
||||
|
||||
test('does not duplicate options when using numeric values', async () => {
|
||||
render(
|
||||
<AsyncSelect
|
||||
|
||||
@@ -160,6 +160,8 @@ const AsyncSelect = forwardRef(
|
||||
const [allValuesLoaded, setAllValuesLoaded] = useState(false);
|
||||
const selectValueRef = useRef(selectValue);
|
||||
const fetchedQueries = useRef(new Map<string, number>());
|
||||
const initialOptionsRef = useRef<SelectOptionsType>(EMPTY_OPTIONS);
|
||||
const wasSearchingRef = useRef(false);
|
||||
const mappedMode = isSingleMode ? undefined : 'multiple';
|
||||
const allowFetch = !fetchOnlyOnSearch || inputValue;
|
||||
const [maxTagCount, setMaxTagCount] = useState(
|
||||
@@ -335,13 +337,22 @@ const AsyncSelect = forwardRef(
|
||||
const fetchOptions = options as SelectOptionsPagePromise;
|
||||
fetchOptions(search, page, pageSize)
|
||||
.then(({ data, totalCount }: SelectOptionsTypePage) => {
|
||||
const mergedData = mergeData(data);
|
||||
let resultData: SelectOptionsType;
|
||||
if (search && page === 0) {
|
||||
resultData = data.slice().sort(sortComparatorForNoSearch);
|
||||
setSelectOptions(resultData);
|
||||
} else {
|
||||
resultData = mergeData(data);
|
||||
if (!search) {
|
||||
initialOptionsRef.current = resultData;
|
||||
}
|
||||
}
|
||||
fetchedQueries.current.set(key, totalCount);
|
||||
setTotalCount(totalCount);
|
||||
if (
|
||||
!fetchOnlyOnSearch &&
|
||||
search === '' &&
|
||||
mergedData.length >= totalCount
|
||||
resultData.length >= totalCount
|
||||
) {
|
||||
setAllValuesLoaded(true);
|
||||
}
|
||||
@@ -358,6 +369,7 @@ const AsyncSelect = forwardRef(
|
||||
internalOnError,
|
||||
options,
|
||||
pageSize,
|
||||
sortComparatorForNoSearch,
|
||||
],
|
||||
);
|
||||
|
||||
@@ -518,12 +530,26 @@ const AsyncSelect = forwardRef(
|
||||
if (loadingEnabled && allowFetch) {
|
||||
// trigger fetch every time inputValue changes
|
||||
if (inputValue) {
|
||||
wasSearchingRef.current = true;
|
||||
debouncedFetchPage(inputValue, 0);
|
||||
} else {
|
||||
if (wasSearchingRef.current && initialOptionsRef.current.length > 0) {
|
||||
setSelectOptions(
|
||||
[...initialOptionsRef.current].sort(sortComparatorForNoSearch),
|
||||
);
|
||||
}
|
||||
wasSearchingRef.current = false;
|
||||
fetchPage('', 0);
|
||||
}
|
||||
}
|
||||
}, [loadingEnabled, fetchPage, allowFetch, inputValue, debouncedFetchPage]);
|
||||
}, [
|
||||
loadingEnabled,
|
||||
fetchPage,
|
||||
allowFetch,
|
||||
inputValue,
|
||||
debouncedFetchPage,
|
||||
sortComparatorForNoSearch,
|
||||
]);
|
||||
|
||||
useEffect(() => {
|
||||
if (loading !== undefined && loading !== isLoading) {
|
||||
|
||||
@@ -211,6 +211,10 @@ export const handleFilterOptionHelper = (
|
||||
return filterOption(search, option);
|
||||
}
|
||||
|
||||
if (filterOption === false) {
|
||||
return true;
|
||||
}
|
||||
|
||||
if (filterOption) {
|
||||
const searchValue = search.trim().toLowerCase();
|
||||
if (optionFilterProps?.length) {
|
||||
|
||||
@@ -17,16 +17,14 @@
|
||||
* under the License.
|
||||
*/
|
||||
|
||||
import { PureComponent } from 'react';
|
||||
import { memo } from 'react';
|
||||
import { TableRenderer } from './TableRenderers';
|
||||
import type { ComponentProps } from 'react';
|
||||
|
||||
type PivotTableProps = ComponentProps<typeof TableRenderer>;
|
||||
|
||||
class PivotTable extends PureComponent<PivotTableProps> {
|
||||
render() {
|
||||
return <TableRenderer {...this.props} />;
|
||||
}
|
||||
function PivotTable(props: PivotTableProps) {
|
||||
return <TableRenderer {...props} />;
|
||||
}
|
||||
|
||||
export default PivotTable;
|
||||
export default memo(PivotTable);
|
||||
|
||||
File diff suppressed because it is too large
Load Diff
File diff suppressed because it is too large
Load Diff
@@ -30,7 +30,7 @@ import fetchMock from 'fetch-mock';
|
||||
import { SaveDatasetModal } from 'src/SqlLab/components/SaveDatasetModal';
|
||||
import { createDatasource } from 'src/SqlLab/actions/sqlLab';
|
||||
import { user, testQuery, mockdatasets } from 'src/SqlLab/fixtures';
|
||||
import { FeatureFlag } from '@superset-ui/core';
|
||||
import { FeatureFlag, SupersetClient } from '@superset-ui/core';
|
||||
|
||||
const mockedProps = {
|
||||
visible: true,
|
||||
@@ -354,6 +354,131 @@ describe('SaveDatasetModal', () => {
|
||||
});
|
||||
});
|
||||
|
||||
const setupOverwriteFlow = async () => {
|
||||
// Select the "Overwrite existing" radio
|
||||
await userEvent.click(
|
||||
screen.getByRole('radio', { name: /overwrite existing/i }),
|
||||
);
|
||||
// Open the select to load existing-dataset options
|
||||
await userEvent.click(
|
||||
screen.getByRole('combobox', { name: /existing dataset/i }),
|
||||
);
|
||||
// Advance timers to flush debounced fetches in AsyncSelect
|
||||
await act(async () => {
|
||||
jest.runAllTimers();
|
||||
});
|
||||
// Wait for the loading indicator to clear
|
||||
await waitFor(() => {
|
||||
const loading = screen.queryByText('Loading...');
|
||||
expect(loading === null || !loading.checkVisibility()).toBe(true);
|
||||
});
|
||||
// Pick an existing dataset (use the listbox item, not the input mirror)
|
||||
const options = await screen.findAllByText('coolest table 0');
|
||||
await userEvent.click(options[1]);
|
||||
// First overwrite click → confirmation screen
|
||||
await userEvent.click(screen.getByRole('button', { name: /overwrite/i }));
|
||||
// Wait for the confirmation screen to render
|
||||
await screen.findByText(/are you sure you want to overwrite this dataset/i);
|
||||
// Second overwrite click → triggers the PUT
|
||||
await userEvent.click(screen.getByRole('button', { name: /overwrite/i }));
|
||||
};
|
||||
|
||||
test('sends template_params when overwriting a dataset with include template parameters checked', async () => {
|
||||
// @ts-expect-error
|
||||
global.featureFlags = {
|
||||
[FeatureFlag.EnableTemplateProcessing]: true,
|
||||
};
|
||||
|
||||
const putSpy = jest
|
||||
.spyOn(SupersetClient, 'put')
|
||||
.mockResolvedValue({ json: { result: { id: 0 } } } as any);
|
||||
|
||||
const dummyDispatch = jest.fn().mockResolvedValue({});
|
||||
useDispatchMock.mockReturnValue(dummyDispatch);
|
||||
useSelectorMock.mockReturnValue({ ...user });
|
||||
|
||||
const propsWithTemplateParam = {
|
||||
...mockedProps,
|
||||
datasource: {
|
||||
...testQuery,
|
||||
templateParams: JSON.stringify({ my_param: 12, _filters: 'foo' }),
|
||||
},
|
||||
};
|
||||
render(<SaveDatasetModal {...propsWithTemplateParam} />, {
|
||||
useRedux: true,
|
||||
});
|
||||
|
||||
// Check the "Include Template Parameters" checkbox
|
||||
await userEvent.click(screen.getByRole('checkbox'));
|
||||
|
||||
await setupOverwriteFlow();
|
||||
|
||||
await waitFor(() => {
|
||||
expect(
|
||||
putSpy.mock.calls.some(([req]) =>
|
||||
req.endpoint?.includes('api/v1/dataset/'),
|
||||
),
|
||||
).toBe(true);
|
||||
});
|
||||
|
||||
const datasetPutCall = putSpy.mock.calls.find(([req]) =>
|
||||
req.endpoint?.includes('api/v1/dataset/'),
|
||||
)!;
|
||||
const [req] = datasetPutCall;
|
||||
expect(req.endpoint).toContain('override_columns=true');
|
||||
const body = JSON.parse(req.body as string);
|
||||
// _filters should be stripped, but my_param should be preserved
|
||||
expect(body.template_params).toEqual(JSON.stringify({ my_param: 12 }));
|
||||
|
||||
putSpy.mockRestore();
|
||||
});
|
||||
|
||||
test('does not send template_params when overwriting a dataset with include template parameters unchecked', async () => {
|
||||
// @ts-expect-error
|
||||
global.featureFlags = {
|
||||
[FeatureFlag.EnableTemplateProcessing]: true,
|
||||
};
|
||||
|
||||
const putSpy = jest
|
||||
.spyOn(SupersetClient, 'put')
|
||||
.mockResolvedValue({ json: { result: { id: 0 } } } as any);
|
||||
|
||||
const dummyDispatch = jest.fn().mockResolvedValue({});
|
||||
useDispatchMock.mockReturnValue(dummyDispatch);
|
||||
useSelectorMock.mockReturnValue({ ...user });
|
||||
|
||||
const propsWithTemplateParam = {
|
||||
...mockedProps,
|
||||
datasource: {
|
||||
...testQuery,
|
||||
templateParams: JSON.stringify({ my_param: 12 }),
|
||||
},
|
||||
};
|
||||
render(<SaveDatasetModal {...propsWithTemplateParam} />, {
|
||||
useRedux: true,
|
||||
});
|
||||
|
||||
// Do NOT check the "Include Template Parameters" checkbox
|
||||
await setupOverwriteFlow();
|
||||
|
||||
await waitFor(() => {
|
||||
expect(
|
||||
putSpy.mock.calls.some(([req]) =>
|
||||
req.endpoint?.includes('api/v1/dataset/'),
|
||||
),
|
||||
).toBe(true);
|
||||
});
|
||||
|
||||
const datasetPutCall = putSpy.mock.calls.find(([req]) =>
|
||||
req.endpoint?.includes('api/v1/dataset/'),
|
||||
)!;
|
||||
const [req] = datasetPutCall;
|
||||
const body = JSON.parse(req.body as string);
|
||||
expect(body.template_params).toBeUndefined();
|
||||
|
||||
putSpy.mockRestore();
|
||||
});
|
||||
|
||||
test('clears dataset cache when creating new dataset', async () => {
|
||||
const clearDatasetCache = jest.spyOn(
|
||||
require('src/utils/cachedSupersetGet'),
|
||||
|
||||
@@ -149,14 +149,25 @@ const Styles = styled.div`
|
||||
}
|
||||
`}
|
||||
`;
|
||||
const updateDataset = async (
|
||||
dbId: number,
|
||||
datasetId: number,
|
||||
sql: string,
|
||||
columns: Array<Record<string, any>>,
|
||||
owners: [number],
|
||||
overrideColumns: boolean,
|
||||
) => {
|
||||
type UpdateDatasetPayload = {
|
||||
dbId: number;
|
||||
datasetId: number;
|
||||
sql: string;
|
||||
columns: Array<Record<string, any>>;
|
||||
owners: number[];
|
||||
overrideColumns: boolean;
|
||||
templateParams?: string;
|
||||
};
|
||||
|
||||
const updateDataset = async ({
|
||||
dbId,
|
||||
datasetId,
|
||||
sql,
|
||||
columns,
|
||||
owners,
|
||||
overrideColumns,
|
||||
templateParams,
|
||||
}: UpdateDatasetPayload) => {
|
||||
const endpoint = `api/v1/dataset/${datasetId}?override_columns=${overrideColumns}`;
|
||||
const headers = { 'Content-Type': 'application/json' };
|
||||
const body = JSON.stringify({
|
||||
@@ -164,6 +175,7 @@ const updateDataset = async (
|
||||
columns,
|
||||
owners,
|
||||
database_id: dbId,
|
||||
...(templateParams !== undefined && { template_params: templateParams }),
|
||||
});
|
||||
|
||||
const data: JsonResponse = await SupersetClient.put({
|
||||
@@ -179,6 +191,26 @@ const updateDataset = async (
|
||||
|
||||
const UNTITLED = t('Untitled Dataset');
|
||||
|
||||
// The filters param is only used to test jinja templates.
|
||||
// Remove the special filters entry from the templateParams
|
||||
// before saving the dataset.
|
||||
const sanitizeTemplateParams = (
|
||||
templateParams: string | object | null | undefined,
|
||||
): string | undefined => {
|
||||
if (typeof templateParams !== 'string') {
|
||||
return undefined;
|
||||
}
|
||||
try {
|
||||
const parsed = JSON.parse(templateParams) as Record<string, unknown>;
|
||||
// Remove the special _filters entry — it is only used to test jinja templates.
|
||||
const { _filters: _ignored, ...clean } = parsed;
|
||||
return JSON.stringify(clean);
|
||||
} catch (e) {
|
||||
// malformed templateParams, do not include it
|
||||
return undefined;
|
||||
}
|
||||
};
|
||||
|
||||
export const SaveDatasetModal = ({
|
||||
visible,
|
||||
onHide,
|
||||
@@ -232,22 +264,27 @@ export const SaveDatasetModal = ({
|
||||
}
|
||||
setLoading(true);
|
||||
|
||||
const templateParams = includeTemplateParameters
|
||||
? sanitizeTemplateParams(datasource?.templateParams)
|
||||
: undefined;
|
||||
|
||||
try {
|
||||
const [, key] = await Promise.all([
|
||||
updateDataset(
|
||||
datasource?.dbId,
|
||||
datasetToOverwrite?.datasetid,
|
||||
datasource?.sql,
|
||||
datasource?.columns?.map(
|
||||
updateDataset({
|
||||
dbId: datasource?.dbId,
|
||||
datasetId: datasetToOverwrite?.datasetid,
|
||||
sql: datasource?.sql,
|
||||
columns: datasource?.columns?.map(
|
||||
(d: { column_name: string; type: string; is_dttm: boolean }) => ({
|
||||
column_name: d.column_name,
|
||||
type: d.type,
|
||||
is_dttm: d.is_dttm,
|
||||
}),
|
||||
),
|
||||
datasetToOverwrite?.owners?.map((o: DatasetOwner) => o.id),
|
||||
true,
|
||||
),
|
||||
owners: datasetToOverwrite?.owners?.map((o: DatasetOwner) => o.id),
|
||||
overrideColumns: true,
|
||||
templateParams,
|
||||
}),
|
||||
postFormData(datasetToOverwrite.datasetid, 'table', {
|
||||
...formDataWithDefaults,
|
||||
datasource: `${datasetToOverwrite.datasetid}__table`,
|
||||
@@ -319,27 +356,9 @@ export const SaveDatasetModal = ({
|
||||
setLoading(true);
|
||||
const selectedColumns = datasource?.columns ?? [];
|
||||
|
||||
// The filters param is only used to test jinja templates.
|
||||
// Remove the special filters entry from the templateParams
|
||||
// before saving the dataset.
|
||||
let templateParams;
|
||||
if (
|
||||
typeof datasource?.templateParams === 'string' &&
|
||||
includeTemplateParameters
|
||||
) {
|
||||
try {
|
||||
const p = JSON.parse(datasource.templateParams);
|
||||
/* eslint-disable-next-line no-underscore-dangle */
|
||||
if (p._filters) {
|
||||
/* eslint-disable-next-line no-underscore-dangle */
|
||||
delete p._filters;
|
||||
}
|
||||
templateParams = JSON.stringify(p);
|
||||
} catch (e) {
|
||||
// malformed templateParams, do not include it
|
||||
templateParams = undefined;
|
||||
}
|
||||
}
|
||||
const templateParams = includeTemplateParameters
|
||||
? sanitizeTemplateParams(datasource?.templateParams)
|
||||
: undefined;
|
||||
|
||||
dispatch(
|
||||
createDatasource({
|
||||
|
||||
@@ -251,7 +251,7 @@ describe('useTables hook', () => {
|
||||
fetchMock.get(`glob:*/api/v1/database/${expectDbId}/schemas/*`, {
|
||||
result: fakeSchemaApiResult,
|
||||
});
|
||||
const { result, waitFor } = renderHook(
|
||||
const { result } = renderHook(
|
||||
() =>
|
||||
useTables({
|
||||
dbId: expectDbId,
|
||||
|
||||
@@ -22,6 +22,11 @@ import {
|
||||
createListenerMiddleware,
|
||||
StoreEnhancer,
|
||||
} from '@reduxjs/toolkit';
|
||||
import {
|
||||
useDispatch,
|
||||
useSelector,
|
||||
type TypedUseSelectorHook,
|
||||
} from 'react-redux';
|
||||
import thunk from 'redux-thunk';
|
||||
import { api } from 'src/hooks/apiResources/queryApi';
|
||||
import messageToastReducer from 'src/components/MessageToasts/reducers';
|
||||
@@ -177,3 +182,12 @@ export function setupStore({
|
||||
|
||||
export const store = setupStore();
|
||||
export type RootState = ReturnType<typeof store.getState>;
|
||||
|
||||
// Typed Redux hooks. Prefer these over the raw `useDispatch` / `useSelector`
|
||||
// from react-redux: `useAppDispatch` understands the store's middleware (so
|
||||
// thunks resolve correctly), and `useAppSelector` infers `RootState` without
|
||||
// callers having to annotate every selector. Required ahead of the
|
||||
// react-redux v8+ bump, which tightens dispatch typing — see #39927.
|
||||
export type AppDispatch = typeof store.dispatch;
|
||||
export const useAppDispatch: () => AppDispatch = useDispatch;
|
||||
export const useAppSelector: TypedUseSelectorHook<RootState> = useSelector;
|
||||
|
||||
@@ -49,8 +49,9 @@ from superset.datasets.schemas import ImportV1DatasetSchema
|
||||
from superset.extensions import feature_flag_manager
|
||||
from superset.migrations.shared.native_filters import migrate_dashboard
|
||||
from superset.models.core import Database
|
||||
from superset.models.dashboard import dashboard_slices
|
||||
from superset.models.dashboard import Dashboard, dashboard_slices
|
||||
from superset.models.slice import Slice
|
||||
from superset.models.sql_lab import SavedQuery
|
||||
from superset.queries.saved_queries.schemas import ImportV1SavedQuerySchema
|
||||
from superset.utils.decorators import on_error, transaction
|
||||
|
||||
@@ -89,6 +90,9 @@ class ImportAssetsCommand(BaseCommand):
|
||||
)
|
||||
self._configs: dict[str, Any] = {}
|
||||
self.sparse = kwargs.get("sparse", False)
|
||||
# Defaults to ``True`` for backwards compatibility: historically this
|
||||
# command always overwrote existing assets.
|
||||
self.overwrite: bool = kwargs.get("overwrite", True)
|
||||
|
||||
# pylint: disable=too-many-locals
|
||||
@staticmethod
|
||||
@@ -96,6 +100,7 @@ class ImportAssetsCommand(BaseCommand):
|
||||
configs: dict[str, Any],
|
||||
sparse: bool = False,
|
||||
contents: Optional[dict[str, Any]] = None,
|
||||
overwrite: bool = True,
|
||||
) -> None:
|
||||
contents = {} if contents is None else contents
|
||||
# import databases first
|
||||
@@ -116,20 +121,20 @@ class ImportAssetsCommand(BaseCommand):
|
||||
|
||||
for file_name, config in configs.items():
|
||||
if file_name.startswith("databases/"):
|
||||
database = import_database(config, overwrite=True)
|
||||
database = import_database(config, overwrite=overwrite)
|
||||
database_ids[str(database.uuid)] = database.id
|
||||
|
||||
# import saved queries
|
||||
for file_name, config in configs.items():
|
||||
if file_name.startswith("queries/"):
|
||||
config["db_id"] = database_ids[config["database_uuid"]]
|
||||
import_saved_query(config, overwrite=True)
|
||||
import_saved_query(config, overwrite=overwrite)
|
||||
|
||||
# import datasets
|
||||
for file_name, config in configs.items():
|
||||
if file_name.startswith("datasets/"):
|
||||
config["database_id"] = database_ids[config["database_uuid"]]
|
||||
dataset = import_dataset(config, overwrite=True)
|
||||
dataset = import_dataset(config, overwrite=overwrite)
|
||||
dataset_info[str(dataset.uuid)] = {
|
||||
"datasource_id": dataset.id,
|
||||
"datasource_type": dataset.datasource_type,
|
||||
@@ -142,7 +147,7 @@ class ImportAssetsCommand(BaseCommand):
|
||||
if file_name.startswith("charts/"):
|
||||
dataset_dict = dataset_info[config["dataset_uuid"]]
|
||||
config = update_chart_config_dataset(config, dataset_dict)
|
||||
chart = import_chart(config, overwrite=True)
|
||||
chart = import_chart(config, overwrite=overwrite)
|
||||
charts.append(chart)
|
||||
chart_ids[str(chart.uuid)] = chart.id
|
||||
|
||||
@@ -157,7 +162,7 @@ class ImportAssetsCommand(BaseCommand):
|
||||
for file_name, config in configs.items():
|
||||
if file_name.startswith("dashboards/"):
|
||||
config = update_id_refs(config, chart_ids, dataset_info)
|
||||
dashboard = import_dashboard(config, overwrite=True)
|
||||
dashboard = import_dashboard(config, overwrite=overwrite)
|
||||
|
||||
# set ref in the dashboard_slices table
|
||||
dashboard_chart_ids: list[dict[str, int]] = []
|
||||
@@ -206,7 +211,73 @@ class ImportAssetsCommand(BaseCommand):
|
||||
)
|
||||
def run(self) -> None:
|
||||
self.validate()
|
||||
self._import(self._configs, self.sparse, self.contents)
|
||||
self._import(self._configs, self.sparse, self.contents, self.overwrite)
|
||||
|
||||
# Maps asset file prefixes to the model class used to look up UUIDs for
|
||||
# the "already exists" validation check when ``overwrite`` is ``False``.
|
||||
_MODEL_BY_PREFIX: dict[str, Any] = {
|
||||
"databases/": Database,
|
||||
"datasets/": SqlaTable,
|
||||
"charts/": Slice,
|
||||
"dashboards/": Dashboard,
|
||||
"queries/": SavedQuery,
|
||||
}
|
||||
|
||||
def _bundle_entries_by_prefix(self) -> dict[str, list[tuple[str, str]]]:
|
||||
"""Group ``(file_name, uuid)`` pairs from the bundle by asset prefix."""
|
||||
bundle_by_prefix: dict[str, list[tuple[str, str]]] = {
|
||||
prefix: [] for prefix in self._MODEL_BY_PREFIX
|
||||
}
|
||||
for file_name, config in self._configs.items():
|
||||
uuid = config.get("uuid")
|
||||
if not uuid:
|
||||
continue
|
||||
for prefix in bundle_by_prefix:
|
||||
if file_name.startswith(prefix):
|
||||
bundle_by_prefix[prefix].append((file_name, str(uuid)))
|
||||
break
|
||||
return bundle_by_prefix
|
||||
|
||||
def _prevent_overwrite_existing_assets(
|
||||
self, exceptions: list[ValidationError]
|
||||
) -> None:
|
||||
"""
|
||||
When ``overwrite`` is ``False``, raise a clear validation error for any
|
||||
asset in the bundle whose UUID already exists in the database.
|
||||
|
||||
Only the UUIDs present in the import bundle are queried (per prefix),
|
||||
so the cost scales with the bundle size rather than with the total
|
||||
number of stored assets.
|
||||
"""
|
||||
if self.overwrite:
|
||||
return
|
||||
|
||||
for prefix, entries in self._bundle_entries_by_prefix().items():
|
||||
if not entries:
|
||||
continue
|
||||
model_cls = self._MODEL_BY_PREFIX[prefix]
|
||||
incoming_uuids = [uuid for _, uuid in entries]
|
||||
existing_uuids = {
|
||||
str(uuid)
|
||||
for (uuid,) in db.session.query(model_cls.uuid)
|
||||
.filter(model_cls.uuid.in_(incoming_uuids))
|
||||
.all()
|
||||
}
|
||||
if not existing_uuids:
|
||||
continue
|
||||
model_name = model_cls.__name__
|
||||
for file_name, uuid in entries:
|
||||
if uuid in existing_uuids:
|
||||
exceptions.append(
|
||||
ValidationError(
|
||||
{
|
||||
file_name: (
|
||||
f"{model_name} already exists "
|
||||
"and `overwrite=true` was not passed"
|
||||
),
|
||||
}
|
||||
)
|
||||
)
|
||||
|
||||
def validate(self) -> None:
|
||||
exceptions: list[ValidationError] = []
|
||||
@@ -229,6 +300,7 @@ class ImportAssetsCommand(BaseCommand):
|
||||
self.ssh_tunnel_priv_key_passwords,
|
||||
self.encrypted_extra_secrets,
|
||||
)
|
||||
self._prevent_overwrite_existing_assets(exceptions)
|
||||
|
||||
if exceptions:
|
||||
raise CommandInvalidError(
|
||||
|
||||
@@ -30,6 +30,7 @@ from superset.commands.importers.v1.assets import ImportAssetsCommand
|
||||
from superset.commands.importers.v1.utils import get_contents_from_bundle
|
||||
from superset.extensions import event_logger
|
||||
from superset.utils import json
|
||||
from superset.utils.core import parse_boolean_string
|
||||
from superset.views.base_api import BaseSupersetApi, requires_form_data, statsd_metrics
|
||||
|
||||
|
||||
@@ -157,6 +158,12 @@ class ImportExportRestApi(BaseSupersetApi):
|
||||
sparse:
|
||||
description: allow sparse update of resources
|
||||
type: boolean
|
||||
overwrite:
|
||||
description: >-
|
||||
overwrite existing assets? Defaults to ``true`` for
|
||||
backwards compatibility. When ``false``, the import
|
||||
fails if any of the assets already exist.
|
||||
type: boolean
|
||||
responses:
|
||||
200:
|
||||
description: Assets import result
|
||||
@@ -188,6 +195,9 @@ class ImportExportRestApi(BaseSupersetApi):
|
||||
if not contents:
|
||||
raise NoValidFilesFoundError()
|
||||
sparse = request.form.get("sparse") == "true"
|
||||
# Defaults to True for backwards compatibility: historically this
|
||||
# endpoint always overwrote existing assets.
|
||||
overwrite = parse_boolean_string(request.form.get("overwrite", "true"))
|
||||
|
||||
passwords = (
|
||||
json.loads(request.form["passwords"])
|
||||
@@ -218,6 +228,7 @@ class ImportExportRestApi(BaseSupersetApi):
|
||||
command = ImportAssetsCommand(
|
||||
contents,
|
||||
sparse=sparse,
|
||||
overwrite=overwrite,
|
||||
passwords=passwords,
|
||||
ssh_tunnel_passwords=ssh_tunnel_passwords,
|
||||
ssh_tunnel_private_keys=ssh_tunnel_private_keys,
|
||||
|
||||
@@ -19,6 +19,7 @@ import copy
|
||||
from typing import Any, cast
|
||||
|
||||
import yaml
|
||||
from marshmallow.exceptions import ValidationError
|
||||
from pytest_mock import MockerFixture
|
||||
from sqlalchemy.orm.session import Session
|
||||
from sqlalchemy.sql import select
|
||||
@@ -32,6 +33,18 @@ from tests.unit_tests.fixtures.assets_configs import (
|
||||
datasets_config,
|
||||
)
|
||||
|
||||
saved_queries_config: dict[str, Any] = {
|
||||
"queries/examples/my_query.yaml": {
|
||||
"schema": "main",
|
||||
"label": "My saved query",
|
||||
"description": None,
|
||||
"sql": "SELECT 1",
|
||||
"uuid": "e3e4f1f0-5c9d-4a4c-a4e4-0000000000aa",
|
||||
"version": "1.0.0",
|
||||
"database_uuid": "a2dc77af-e654-49bb-b321-40f6b559a1ee",
|
||||
},
|
||||
}
|
||||
|
||||
|
||||
def test_import_new_assets(mocker: MockerFixture, session: Session) -> None:
|
||||
"""
|
||||
@@ -227,6 +240,309 @@ def test_import_assets_skips_tags_when_feature_disabled(
|
||||
assert db.session.query(TaggedObject).count() == 0
|
||||
|
||||
|
||||
def test_import_overwrite_defaults_to_true(session: Session) -> None:
|
||||
"""
|
||||
``ImportAssetsCommand.overwrite`` defaults to ``True`` for backwards
|
||||
compatibility — historically the command always overwrote existing assets.
|
||||
"""
|
||||
from superset.commands.importers.v1.assets import ImportAssetsCommand
|
||||
|
||||
command = ImportAssetsCommand({})
|
||||
assert command.overwrite is True
|
||||
|
||||
explicit_false = ImportAssetsCommand({}, overwrite=False)
|
||||
assert explicit_false.overwrite is False
|
||||
|
||||
|
||||
def test_import_threads_overwrite_flag(mocker: MockerFixture, session: Session) -> None:
|
||||
"""
|
||||
``overwrite`` must be threaded through to ``import_database``,
|
||||
``import_saved_query``, ``import_dataset``, ``import_chart`` and
|
||||
``import_dashboard``. Previously these were hard-coded to ``overwrite=True``
|
||||
which caused the API flag to be ignored.
|
||||
"""
|
||||
from superset import security_manager
|
||||
from superset.commands.importers.v1 import assets as assets_module
|
||||
from superset.commands.importers.v1.assets import ImportAssetsCommand
|
||||
|
||||
mocker.patch.object(security_manager, "can_access", return_value=True)
|
||||
|
||||
mocked_db = mocker.patch.object(assets_module, "import_database")
|
||||
mocked_db.return_value.uuid = "a2dc77af-e654-49bb-b321-40f6b559a1ee"
|
||||
mocked_db.return_value.id = 1
|
||||
mocked_ds = mocker.patch.object(assets_module, "import_dataset")
|
||||
mocked_ds.return_value.uuid = "53d47c0c-c03d-47f0-b9ac-81225f808283"
|
||||
mocked_ds.return_value.id = 1
|
||||
mocked_ds.return_value.datasource_type = "table"
|
||||
mocked_ds.return_value.table_name = "video_game_sales"
|
||||
mocked_chart = mocker.patch.object(assets_module, "import_chart")
|
||||
mocked_chart.return_value.viz_type = "table"
|
||||
mocked_dash = mocker.patch.object(assets_module, "import_dashboard")
|
||||
mocker.patch.object(assets_module, "find_chart_uuids", return_value=[])
|
||||
mocker.patch.object(assets_module, "update_id_refs", side_effect=lambda c, *_: c)
|
||||
mocker.patch.object(assets_module, "migrate_dashboard")
|
||||
mocker.patch("superset.db.session.execute")
|
||||
|
||||
configs = {
|
||||
**copy.deepcopy(databases_config),
|
||||
**copy.deepcopy(datasets_config),
|
||||
**copy.deepcopy(charts_config_1),
|
||||
**copy.deepcopy(dashboards_config_1),
|
||||
}
|
||||
|
||||
ImportAssetsCommand._import(configs, overwrite=False)
|
||||
|
||||
assert mocked_db.called
|
||||
for call in mocked_db.call_args_list:
|
||||
assert call.kwargs["overwrite"] is False
|
||||
for call in mocked_ds.call_args_list:
|
||||
assert call.kwargs["overwrite"] is False
|
||||
for call in mocked_chart.call_args_list:
|
||||
assert call.kwargs["overwrite"] is False
|
||||
for call in mocked_dash.call_args_list:
|
||||
assert call.kwargs["overwrite"] is False
|
||||
|
||||
|
||||
def test_prevent_overwrite_flags_existing_assets(
|
||||
mocker: MockerFixture, session: Session
|
||||
) -> None:
|
||||
"""
|
||||
With ``overwrite=False``, ``_prevent_overwrite_existing_assets`` must
|
||||
surface a clear ``ValidationError`` for each asset whose UUID already
|
||||
exists in the database.
|
||||
"""
|
||||
from superset import db, security_manager
|
||||
from superset.commands.importers.v1.assets import ImportAssetsCommand
|
||||
from superset.models.slice import Slice
|
||||
|
||||
mocker.patch.object(security_manager, "can_access", return_value=True)
|
||||
engine = db.session.get_bind()
|
||||
Slice.metadata.create_all(engine) # pylint: disable=no-member
|
||||
|
||||
# seed the database with the fixture assets
|
||||
seed_configs = {
|
||||
**copy.deepcopy(databases_config),
|
||||
**copy.deepcopy(datasets_config),
|
||||
**copy.deepcopy(charts_config_1),
|
||||
**copy.deepcopy(dashboards_config_1),
|
||||
}
|
||||
ImportAssetsCommand._import(seed_configs)
|
||||
|
||||
command = ImportAssetsCommand({}, overwrite=False)
|
||||
command._configs = {
|
||||
**copy.deepcopy(databases_config),
|
||||
**copy.deepcopy(datasets_config),
|
||||
**copy.deepcopy(charts_config_1),
|
||||
**copy.deepcopy(dashboards_config_1),
|
||||
}
|
||||
|
||||
exceptions: list[ValidationError] = []
|
||||
command._prevent_overwrite_existing_assets(exceptions)
|
||||
|
||||
# one exception for each of the seeded assets (db + datasets + charts + dashboards)
|
||||
expected_count = (
|
||||
len(databases_config)
|
||||
+ len(datasets_config)
|
||||
+ len(charts_config_1)
|
||||
+ len(dashboards_config_1)
|
||||
)
|
||||
assert len(exceptions) == expected_count
|
||||
for exc in exceptions:
|
||||
assert isinstance(exc, ValidationError)
|
||||
[(_, message)] = exc.messages.items()
|
||||
assert "already exists" in message
|
||||
assert "`overwrite=true` was not passed" in message
|
||||
|
||||
|
||||
def test_prevent_overwrite_allows_new_assets(
|
||||
mocker: MockerFixture, session: Session
|
||||
) -> None:
|
||||
"""
|
||||
With ``overwrite=False`` and no conflicting UUIDs in the database, the
|
||||
validation step must not raise.
|
||||
"""
|
||||
from superset import db, security_manager
|
||||
from superset.commands.importers.v1.assets import ImportAssetsCommand
|
||||
from superset.models.slice import Slice
|
||||
|
||||
mocker.patch.object(security_manager, "can_access", return_value=True)
|
||||
engine = db.session.get_bind()
|
||||
Slice.metadata.create_all(engine) # pylint: disable=no-member
|
||||
|
||||
command = ImportAssetsCommand({}, overwrite=False)
|
||||
command._configs = {
|
||||
**copy.deepcopy(databases_config),
|
||||
**copy.deepcopy(datasets_config),
|
||||
**copy.deepcopy(charts_config_1),
|
||||
**copy.deepcopy(dashboards_config_1),
|
||||
}
|
||||
|
||||
exceptions: list[ValidationError] = []
|
||||
command._prevent_overwrite_existing_assets(exceptions)
|
||||
|
||||
assert exceptions == []
|
||||
|
||||
|
||||
def test_prevent_overwrite_noop_when_overwrite_true(
|
||||
mocker: MockerFixture, session: Session
|
||||
) -> None:
|
||||
"""
|
||||
With ``overwrite=True`` (the default) the "already exists" validation must
|
||||
be a no-op even when assets exist in the database — this preserves the
|
||||
historical behavior.
|
||||
"""
|
||||
from superset import db, security_manager
|
||||
from superset.commands.importers.v1.assets import ImportAssetsCommand
|
||||
from superset.models.slice import Slice
|
||||
|
||||
mocker.patch.object(security_manager, "can_access", return_value=True)
|
||||
engine = db.session.get_bind()
|
||||
Slice.metadata.create_all(engine) # pylint: disable=no-member
|
||||
|
||||
seed_configs = {
|
||||
**copy.deepcopy(databases_config),
|
||||
**copy.deepcopy(datasets_config),
|
||||
**copy.deepcopy(charts_config_1),
|
||||
**copy.deepcopy(dashboards_config_1),
|
||||
}
|
||||
ImportAssetsCommand._import(seed_configs)
|
||||
|
||||
command = ImportAssetsCommand({}) # overwrite defaults to True
|
||||
command._configs = copy.deepcopy(seed_configs)
|
||||
|
||||
exceptions: list[ValidationError] = []
|
||||
command._prevent_overwrite_existing_assets(exceptions)
|
||||
|
||||
assert exceptions == []
|
||||
|
||||
|
||||
def test_prevent_overwrite_flags_existing_saved_queries(
|
||||
mocker: MockerFixture, session: Session
|
||||
) -> None:
|
||||
"""
|
||||
Saved queries (``queries/`` prefix) must also be covered by the
|
||||
"already exists" validation when ``overwrite=False`` — otherwise
|
||||
``import_saved_query`` silently returns existing rows and the endpoint
|
||||
would appear to succeed despite the conflict.
|
||||
"""
|
||||
from superset import db, security_manager
|
||||
from superset.commands.importers.v1.assets import ImportAssetsCommand
|
||||
from superset.models.slice import Slice
|
||||
from superset.models.sql_lab import SavedQuery
|
||||
|
||||
mocker.patch.object(security_manager, "can_access", return_value=True)
|
||||
engine = db.session.get_bind()
|
||||
Slice.metadata.create_all(engine) # pylint: disable=no-member
|
||||
SavedQuery.metadata.create_all(engine) # pylint: disable=no-member
|
||||
|
||||
# seed a saved query with a UUID that matches the fixture below
|
||||
saved_query_uuid = next(iter(saved_queries_config.values()))["uuid"]
|
||||
db.session.add(SavedQuery(uuid=saved_query_uuid, label="seeded"))
|
||||
db.session.flush()
|
||||
|
||||
command = ImportAssetsCommand({}, overwrite=False)
|
||||
command._configs = copy.deepcopy(saved_queries_config)
|
||||
|
||||
exceptions: list[ValidationError] = []
|
||||
command._prevent_overwrite_existing_assets(exceptions)
|
||||
|
||||
assert len(exceptions) == 1
|
||||
[(file_name, message)] = exceptions[0].messages.items()
|
||||
assert file_name.startswith("queries/")
|
||||
assert "SavedQuery already exists" in message
|
||||
|
||||
|
||||
def test_prevent_overwrite_partial_conflict(
|
||||
mocker: MockerFixture, session: Session
|
||||
) -> None:
|
||||
"""
|
||||
When only some of the incoming assets already exist, validation must flag
|
||||
exactly the conflicting ones and leave brand-new assets untouched.
|
||||
"""
|
||||
from superset import db, security_manager
|
||||
from superset.commands.importers.v1.assets import ImportAssetsCommand
|
||||
from superset.models.slice import Slice
|
||||
|
||||
mocker.patch.object(security_manager, "can_access", return_value=True)
|
||||
engine = db.session.get_bind()
|
||||
Slice.metadata.create_all(engine) # pylint: disable=no-member
|
||||
|
||||
# seed only databases + datasets; charts and dashboards stay new
|
||||
ImportAssetsCommand._import(
|
||||
{
|
||||
**copy.deepcopy(databases_config),
|
||||
**copy.deepcopy(datasets_config),
|
||||
}
|
||||
)
|
||||
|
||||
command = ImportAssetsCommand({}, overwrite=False)
|
||||
command._configs = {
|
||||
**copy.deepcopy(databases_config),
|
||||
**copy.deepcopy(datasets_config),
|
||||
**copy.deepcopy(charts_config_1),
|
||||
**copy.deepcopy(dashboards_config_1),
|
||||
}
|
||||
|
||||
exceptions: list[ValidationError] = []
|
||||
command._prevent_overwrite_existing_assets(exceptions)
|
||||
|
||||
flagged_files = {next(iter(exc.messages)) for exc in exceptions}
|
||||
assert flagged_files == set(databases_config) | set(datasets_config)
|
||||
|
||||
|
||||
def test_prevent_overwrite_queries_only_bundle_uuids(
|
||||
mocker: MockerFixture, session: Session
|
||||
) -> None:
|
||||
"""
|
||||
The validation must scope its UUID lookup to the UUIDs present in the
|
||||
import bundle (one ``WHERE uuid IN (...)`` query per prefix that has
|
||||
incoming entries) and skip prefixes with no entries entirely. Otherwise
|
||||
every import with ``overwrite=false`` would scan all asset tables in
|
||||
full, regardless of bundle size.
|
||||
"""
|
||||
from superset import db, security_manager
|
||||
from superset.commands.importers.v1.assets import ImportAssetsCommand
|
||||
from superset.connectors.sqla.models import SqlaTable
|
||||
from superset.models.core import Database
|
||||
from superset.models.dashboard import Dashboard
|
||||
from superset.models.slice import Slice
|
||||
from superset.models.sql_lab import SavedQuery
|
||||
|
||||
mocker.patch.object(security_manager, "can_access", return_value=True)
|
||||
engine = db.session.get_bind()
|
||||
Slice.metadata.create_all(engine) # pylint: disable=no-member
|
||||
SavedQuery.metadata.create_all(engine) # pylint: disable=no-member
|
||||
|
||||
# bundle only contains a database — no datasets/charts/dashboards/queries
|
||||
bundle = copy.deepcopy(databases_config)
|
||||
|
||||
spy = mocker.spy(db.session, "query")
|
||||
|
||||
command = ImportAssetsCommand({}, overwrite=False)
|
||||
command._configs = bundle
|
||||
exceptions: list[ValidationError] = []
|
||||
command._prevent_overwrite_existing_assets(exceptions)
|
||||
|
||||
# exactly one UUID query — for the only prefix with bundle entries — and
|
||||
# it targets the Database UUID column. Empty-bundle prefixes (datasets/
|
||||
# charts/dashboards/queries) must not be queried at all, otherwise this
|
||||
# validation degrades to a full-table scan per asset type.
|
||||
queried_columns = [
|
||||
call.args[0]
|
||||
for call in spy.call_args_list
|
||||
if call.args and getattr(call.args[0], "key", None) == "uuid"
|
||||
]
|
||||
assert len(queried_columns) == 1
|
||||
assert queried_columns[0].class_ is Database
|
||||
|
||||
queried_models = {col.class_ for col in queried_columns}
|
||||
for model_cls in (SqlaTable, Slice, Dashboard, SavedQuery):
|
||||
assert model_cls not in queried_models
|
||||
|
||||
# no row matches in an empty table, so no validation errors are raised
|
||||
assert exceptions == []
|
||||
|
||||
|
||||
def test_import_removes_dashboard_charts(
|
||||
mocker: MockerFixture, session: Session
|
||||
) -> None:
|
||||
|
||||
@@ -48,7 +48,9 @@ def test_export_assets(
|
||||
mocked_export_result = [
|
||||
(
|
||||
"metadata.yaml",
|
||||
lambda: "version: 1.0.0\ntype: assets\ntimestamp: '2022-01-01T00:00:00+00:00'\n", # noqa: E501
|
||||
lambda: (
|
||||
"version: 1.0.0\ntype: assets\ntimestamp: '2022-01-01T00:00:00+00:00'\n"
|
||||
), # noqa: E501
|
||||
),
|
||||
("databases/example.yaml", lambda: "<DATABASE CONTENTS>"),
|
||||
]
|
||||
@@ -109,6 +111,7 @@ def test_import_assets(
|
||||
ImportAssetsCommand.assert_called_with(
|
||||
mocked_contents,
|
||||
sparse=False,
|
||||
overwrite=True,
|
||||
passwords=passwords,
|
||||
ssh_tunnel_passwords=None,
|
||||
ssh_tunnel_private_keys=None,
|
||||
@@ -160,6 +163,7 @@ def test_import_assets_with_encrypted_extra_secrets(
|
||||
ImportAssetsCommand.assert_called_with(
|
||||
mocked_contents,
|
||||
sparse=False,
|
||||
overwrite=True,
|
||||
passwords=None,
|
||||
ssh_tunnel_passwords=None,
|
||||
ssh_tunnel_private_keys=None,
|
||||
@@ -168,6 +172,54 @@ def test_import_assets_with_encrypted_extra_secrets(
|
||||
)
|
||||
|
||||
|
||||
def test_import_assets_overwrite_false(
|
||||
mocker: MockerFixture,
|
||||
client: Any,
|
||||
full_api_access: None,
|
||||
) -> None:
|
||||
"""
|
||||
Passing ``overwrite=false`` on the form must be forwarded to
|
||||
``ImportAssetsCommand``. Previously the flag was ignored and assets were
|
||||
always overwritten.
|
||||
"""
|
||||
mocked_contents = {
|
||||
"metadata.yaml": (
|
||||
"version: 1.0.0\ntype: assets\ntimestamp: '2022-01-01T00:00:00+00:00'\n"
|
||||
),
|
||||
"databases/example.yaml": "<DATABASE CONTENTS>",
|
||||
}
|
||||
|
||||
ImportAssetsCommand = mocker.patch("superset.importexport.api.ImportAssetsCommand") # noqa: N806
|
||||
|
||||
root = Path("assets_export")
|
||||
buf = BytesIO()
|
||||
with ZipFile(buf, "w") as bundle:
|
||||
for path, contents in mocked_contents.items():
|
||||
with bundle.open(str(root / path), "w") as fp:
|
||||
fp.write(contents.encode())
|
||||
buf.seek(0)
|
||||
|
||||
form_data = {
|
||||
"bundle": (buf, "assets_export.zip"),
|
||||
"overwrite": "false",
|
||||
}
|
||||
response = client.post(
|
||||
"/api/v1/assets/import/", data=form_data, content_type="multipart/form-data"
|
||||
)
|
||||
assert response.status_code == 200
|
||||
|
||||
ImportAssetsCommand.assert_called_with(
|
||||
mocked_contents,
|
||||
sparse=False,
|
||||
overwrite=False,
|
||||
passwords=None,
|
||||
ssh_tunnel_passwords=None,
|
||||
ssh_tunnel_private_keys=None,
|
||||
ssh_tunnel_priv_key_passwords=None,
|
||||
encrypted_extra_secrets=None,
|
||||
)
|
||||
|
||||
|
||||
def test_import_assets_not_zip(
|
||||
mocker: MockerFixture,
|
||||
client: Any,
|
||||
|
||||
Reference in New Issue
Block a user