fix(ci): resolve OOM issues when building docs locally with Docusaurus Faster + sync docs with latest build result (#38486)

Signed-off-by: hainenber <dotronghai96@gmail.com>
Co-authored-by: Evan Rusackas <evan@preset.io>
Co-authored-by: Claude <claude@anthropic.com>
Co-authored-by: Claude Opus 4.7 <noreply@anthropic.com>
This commit is contained in:
Đỗ Trọng Hải
2026-04-27 15:45:53 +07:00
committed by GitHub
parent 7c24214857
commit 2b13e07521
58 changed files with 407 additions and 163 deletions

View File

@@ -1,13 +1,13 @@
{
"generated": "2026-02-24T20:28:17.222Z",
"generated": "2026-04-25T02:18:43.905Z",
"statistics": {
"totalDatabases": 72,
"withDocumentation": 72,
"withConnectionString": 72,
"totalDatabases": 73,
"withDocumentation": 73,
"withConnectionString": 73,
"withDrivers": 36,
"withAuthMethods": 4,
"supportsJoins": 68,
"supportsSubqueries": 69,
"withAuthMethods": 5,
"supportsJoins": 69,
"supportsSubqueries": 70,
"supportsDynamicSchema": 15,
"supportsCatalog": 9,
"averageScore": 31,
@@ -23,6 +23,7 @@
"Amazon Athena",
"Google BigQuery",
"Databend",
"Google Datastore",
"IBM Db2",
"Denodo",
"Dremio",
@@ -177,10 +178,12 @@
],
"Cloud - Google": [
"Google BigQuery",
"Google Datastore",
"Google Sheets"
],
"Search & NoSQL": [
"Couchbase",
"Google Datastore",
"Amazon DynamoDB",
"Elasticsearch",
"MongoDB",
@@ -751,14 +754,14 @@
"OPEN_SOURCE"
],
"pypi_packages": [
"clickhouse-connect>=0.6.8"
"clickhouse-connect>=0.13.0"
],
"connection_string": "clickhousedb://{username}:{password}@{host}:{port}/{database}",
"default_port": 8123,
"drivers": [
{
"name": "clickhouse-connect (Recommended)",
"pypi_package": "clickhouse-connect>=0.6.8",
"pypi_package": "clickhouse-connect>=0.13.0",
"connection_string": "clickhousedb://{username}:{password}@{host}:{port}/{database}",
"is_recommended": true,
"notes": "Official ClickHouse Python driver with native protocol support."
@@ -781,7 +784,7 @@
"connection_string": "clickhousedb://localhost/default"
}
],
"install_instructions": "echo \"clickhouse-connect>=0.6.8\" >> ./docker/requirements-local.txt",
"install_instructions": "echo \"clickhouse-connect>=0.13.0\" >> ./docker/requirements-local.txt",
"compatible_databases": [
{
"name": "ClickHouse Cloud",
@@ -794,7 +797,7 @@
"HOSTED_OPEN_SOURCE"
],
"pypi_packages": [
"clickhouse-connect>=0.6.8"
"clickhouse-connect>=0.13.0"
],
"connection_string": "clickhousedb://{username}:{password}@{host}:8443/{database}?secure=true",
"parameters": {
@@ -816,7 +819,7 @@
"HOSTED_OPEN_SOURCE"
],
"pypi_packages": [
"clickhouse-connect>=0.6.8"
"clickhouse-connect>=0.13.0"
],
"connection_string": "clickhousedb://{username}:{password}@{host}/{database}?secure=true",
"docs_url": "https://docs.altinity.com/"
@@ -1013,7 +1016,7 @@
"documentation": {
"description": "CrateDB is a distributed SQL database for machine data and IoT workloads.",
"logo": "cratedb.svg",
"homepage_url": "https://crate.io/",
"homepage_url": "https://cratedb.com",
"categories": [
"TIME_SERIES",
"OPEN_SOURCE"
@@ -1296,6 +1299,114 @@
"query_cost_estimation": false,
"sql_validation": false
},
"Google Datastore": {
"engine": "google_datastore",
"engine_name": "Google Datastore",
"module": "datastore",
"documentation": {
"description": "Google Cloud Datastore is a highly scalable NoSQL database for your applications.",
"logo": "datastore.png",
"homepage_url": "https://cloud.google.com/datastore/",
"categories": [
"CLOUD_GCP",
"SEARCH_NOSQL",
"PROPRIETARY"
],
"pypi_packages": [
"python-datastore-sqlalchemy"
],
"connection_string": "datastore://{project_id}/?database={database_id}",
"authentication_methods": [
{
"name": "Service Account JSON",
"description": "Upload service account credentials JSON or paste in Secure Extra",
"secure_extra": {
"credentials_info": {
"type": "service_account",
"project_id": "...",
"private_key_id": "...",
"private_key": "...",
"client_email": "...",
"client_id": "...",
"auth_uri": "...",
"token_uri": "..."
}
}
}
],
"notes": "Create a Service Account via GCP console with access to datastore datasets.",
"docs_url": "https://github.com/splasky/Python-datastore-sqlalchemy",
"custom_errors": [
{
"regex_name": "CONNECTION_DATABASE_PERMISSIONS_REGEX",
"message_template": "Unable to connect. Verify that the following roles are set on the service account: \"Cloud Datastore Viewer\", \"Cloud Datastore User\", \"Cloud Datastore Creator\"",
"error_type": "CONNECTION_DATABASE_PERMISSIONS_ERROR",
"category": "Permissions",
"description": "Insufficient permissions",
"issue_codes": [
1017
]
},
{
"regex_name": "TABLE_DOES_NOT_EXIST_REGEX",
"message_template": "The table \"%(table)s\" does not exist. A valid table must be used to run this query.",
"error_type": "TABLE_DOES_NOT_EXIST_ERROR",
"category": "Query",
"description": "Table not found",
"issue_codes": [
1003,
1005
]
},
{
"regex_name": "COLUMN_DOES_NOT_EXIST_REGEX",
"message_template": "We can't seem to resolve column \"%(column)s\" at line %(location)s.",
"error_type": "COLUMN_DOES_NOT_EXIST_ERROR",
"category": "Query",
"description": "Column not found",
"issue_codes": [
1003,
1004
]
},
{
"regex_name": "SCHEMA_DOES_NOT_EXIST_REGEX",
"message_template": "The schema \"%(schema)s\" does not exist. A valid schema must be used to run this query.",
"error_type": "SCHEMA_DOES_NOT_EXIST_ERROR",
"category": "Query",
"description": "Schema not found",
"issue_codes": [
1003,
1016
]
},
{
"regex_name": "SYNTAX_ERROR_REGEX",
"message_template": "Please check your query for syntax errors at or near \"%(syntax_error)s\". Then, try running your query again.",
"error_type": "SYNTAX_ERROR",
"category": "Query",
"description": "SQL syntax error",
"issue_codes": [
1030
]
}
]
},
"time_grains": {},
"score": 0,
"max_score": 0,
"joins": true,
"subqueries": true,
"supports_dynamic_schema": false,
"supports_catalog": false,
"supports_dynamic_catalog": false,
"ssh_tunneling": false,
"query_cancelation": false,
"supports_file_upload": false,
"user_impersonation": false,
"query_cost_estimation": false,
"sql_validation": false
},
"IBM Db2": {
"engine": "ibm_db2",
"engine_name": "IBM Db2",
@@ -4754,9 +4865,9 @@
}
},
{
"name": "IAM Credentials (Serverless)",
"description": "Use IAM-based credentials for Redshift Serverless",
"requirements": "IAM role must have redshift-serverless:GetCredentials and redshift-serverless:GetWorkgroup permissions",
"name": "IAM Role (Serverless)",
"description": "Authenticate using the IAM role attached to the environment (EC2 instance profile, ECS task role, etc.). No credentials needed.",
"requirements": "The attached IAM role must have redshift-serverless:GetCredentials and redshift-serverless:GetWorkgroup permissions.",
"connection_string": "redshift+redshift_connector://",
"engine_parameters": {
"connect_args": {
@@ -4768,6 +4879,26 @@
"user": "IAMR:<superset iam role name>"
}
}
},
{
"name": "IAM Access Key (Serverless)",
"description": "Authenticate using explicit AWS access key and secret. Suitable for local development or CI environments without an attached IAM role.",
"requirements": "The IAM user must have redshift-serverless:GetCredentials and redshift-serverless:GetWorkgroup permissions.",
"connection_string": "redshift+redshift_connector://",
"engine_parameters": {
"connect_args": {
"iam": true,
"is_serverless": true,
"serverless_acct_id": "<aws account number>",
"serverless_work_group": "<redshift work group>",
"database": "<database>",
"host": "<endpoint>",
"port": 5439,
"region": "<aws region>",
"access_key_id": "<aws access key id>",
"secret_access_key": "<aws secret access key>"
}
}
}
],
"custom_errors": [