BigQuery
Ingesting metadata from BigQuery requires using the bigquery module.
Important Capabilities
Capability | Status | Notes |
---|---|---|
Asset Containers | ✅ | Enabled by default |
Classification | ✅ | Optionally enabled via classification.enabled |
Column-level Lineage | ✅ | Optionally enabled via configuration |
Data Profiling | ✅ | Optionally enabled via configuration |
Dataset Usage | ✅ | Enabled by default, can be disabled via configuration include_usage_statistics |
Descriptions | ✅ | Enabled by default |
Detect Deleted Entities | ✅ | Optionally enabled via stateful_ingestion.remove_stale_metadata |
Domains | ✅ | Supported via the domain config field |
Platform Instance | ❌ | Platform instance is pre-set to the BigQuery project id |
Schema Metadata | ✅ | Enabled by default |
Table-Level Lineage | ✅ | Optionally enabled via configuration |
Prerequisites
To understand how BigQuery ingestion needs to be set up, first familiarize yourself with the concepts in the diagram below:
There are two important concepts to understand and identify:
- Extractor Project: This is the project associated with a service-account, whose credentials you will be configuring in the connector. The connector uses this service-account to run jobs (including queries) within the project.
- Bigquery Projects are the projects from which table metadata, lineage, usage, and profiling data need to be collected. By default, the extractor project is included in the list of projects that DataHub collects metadata from, but you can control that by passing in a specific list of project ids that you want to collect metadata from. Read the configuration section below to understand how to limit the list of projects that DataHub extracts metadata from.
Create a datahub profile in GCP
- Create a custom role for datahub as per BigQuery docs.
- Follow the sections below to grant permissions to this role on this project and other projects.
Basic Requirements (needed for metadata ingestion)
- Identify your Extractor Project where the service account will run queries to extract metadata.
permission | Description | Capability |
---|---|---|
bigquery.jobs.create | Run jobs (e.g. queries) within the project. This only needs for the extractor project where the service account belongs | |
bigquery.jobs.list | Manage the queries that the service account has sent. This only needs for the extractor project where the service account belongs | |
bigquery.readsessions.create | Create a session for streaming large results. This only needs for the extractor project where the service account belongs | |
bigquery.readsessions.getData | Get data from the read session. This only needs for the extractor project where the service account belongs |
- Grant the following permissions to the Service Account on every project where you would like to extract metadata from
If you have multiple projects in your BigQuery setup, the role should be granted these permissions in each of the projects.
Permission | Description | Capability | Default GCP Role Which Contains This Permission |
---|---|---|---|
bigquery.datasets.get | Retrieve metadata about a dataset. | Table Metadata Extraction | roles/bigquery.metadataViewer |
bigquery.datasets.getIamPolicy | Read a dataset's IAM permissions. | Table Metadata Extraction | roles/bigquery.metadataViewer |
bigquery.tables.list | List BigQuery tables. | Table Metadata Extraction | roles/bigquery.metadataViewer |
bigquery.tables.get | Retrieve metadata for a table. | Table Metadata Extraction | roles/bigquery.metadataViewer |
bigquery.routines.get | Get Routines. Needs to retrieve metadata for a table from system table. | Table Metadata Extraction | roles/bigquery.metadataViewer |
bigquery.routines.list | List Routines. Needs to retrieve metadata for a table from system table. | Table Metadata Extraction | roles/bigquery.metadataViewer |
resourcemanager.projects.get | Retrieve project names and metadata. | Table Metadata Extraction | roles/bigquery.metadataViewer |
bigquery.jobs.listAll | List all jobs (queries) submitted by any user. Needs for Lineage extraction. | Lineage Extraction/Usage Extraction | roles/bigquery.resourceViewer |
logging.logEntries.list | Fetch log entries for lineage/usage data. Not required if use_exported_bigquery_audit_metadata is enabled. | Lineage Extraction/Usage Extraction | roles/logging.privateLogViewer |
logging.privateLogEntries.list | Fetch log entries for lineage/usage data. Not required if use_exported_bigquery_audit_metadata is enabled. | Lineage Extraction/Usage Extraction | roles/logging.privateLogViewer |
bigquery.tables.getData | Access table data to extract storage size, last updated at, data profiles etc. | Profiling | |
datacatalog.policyTags.get | Optional Get policy tags for columns with associated policy tags. This permission is required only if extract_policy_tags_from_catalog is enabled. | Policy Tag Extraction | roles/datacatalog.viewer |
Create a service account in the Extractor Project
- Setup a ServiceAccount as per BigQuery docs and assign the previously created role to this service account.
- Download a service account JSON keyfile. Example credential file:
{
"type": "service_account",
"project_id": "project-id-1234567",
"private_key_id": "d0121d0000882411234e11166c6aaa23ed5d74e0",
"private_key": "-----BEGIN PRIVATE KEY-----\nMIIyourkey\n-----END PRIVATE KEY-----",
"client_email": "test@suppproject-id-1234567.iam.gserviceaccount.com",
"client_id": "113545814931671546333",
"auth_uri": "https://accounts.google.com/o/oauth2/auth",
"token_uri": "https://oauth2.googleapis.com/token",
"auth_provider_x509_cert_url": "https://www.googleapis.com/oauth2/v1/certs",
"client_x509_cert_url": "https://www.googleapis.com/robot/v1/metadata/x509/test%suppproject-id-1234567.iam.gserviceaccount.com"
}
To provide credentials to the source, you can either:
Set an environment variable:
$ export GOOGLE_APPLICATION_CREDENTIALS="/path/to/keyfile.json"
or
Set credential config in your source based on the credential json file. For example:
credential:
project_id: project-id-1234567
private_key_id: "d0121d0000882411234e11166c6aaa23ed5d74e0"
private_key: "-----BEGIN PRIVATE KEY-----\nMIIyourkey\n-----END PRIVATE KEY-----\n"
client_email: "test@suppproject-id-1234567.iam.gserviceaccount.com"
client_id: "123456678890"
Profiling Requirements
To profile BigQuery external tables backed by Google Drive document, you need to grant document's "Viewer" access to service account's email address (client_email
in credentials json file). To find the Google Drive document linked to BigQuery table, open the BigQuery console, locate the needed table, select "Details" from the drop-down menu in the top-right corner and refer "Source" field . To share access of Google Drive document, open the document, click "Share" in the top-right corner, add the service account's email address that needs "Viewer" access.
Lineage Computation Details
When use_exported_bigquery_audit_metadata
is set to true
, lineage information will be computed using exported bigquery logs. On how to setup exported bigquery audit logs, refer to the following docs on BigQuery audit logs. Note that only protoPayloads with "type.googleapis.com/google.cloud.audit.BigQueryAuditMetadata" are supported by the current ingestion version. The bigquery_audit_metadata_datasets
parameter will be used only if use_exported_bigquery_audit_metadat
is set to true
.
Note: the bigquery_audit_metadata_datasets
parameter receives a list of datasets, in the format $PROJECT.$DATASET. This way queries from a multiple number of projects can be used to compute lineage information.
Note: Since bigquery source also supports dataset level lineage, the auth client will require additional permissions to be able to access the google audit logs. Refer the permissions section in bigquery-usage section below which also accesses the audit logs.
Profiling Details
For performance reasons, we only profile the latest partition for partitioned tables and the latest shard for sharded tables.
You can set partition explicitly with partition.partition_datetime
property if you want, though note that partition config will be applied to all partitioned tables.
Caveats
- For materialized views, lineage is dependent on logs being retained. If your GCP logging is retained for 30 days (default) and 30 days have passed since the creation of the materialized view we won't be able to get lineage for them.
CLI based Ingestion
Install the Plugin
The bigquery
source works out of the box with acryl-datahub
.
Starter Recipe
Check out the following recipe to get started with ingestion! See below for full configuration options.
For general pointers on writing and running a recipe, see our main recipe guide.
source:
type: bigquery
config:
dataset_pattern:
allow:
- finance_bq_dataset
table_pattern:
deny:
# The exact name of the table is revenue_table_name
# The reason we have this `.*` at the beginning is because the current implmenetation of table_pattern is testing
# project_id.dataset_name.table_name
# We will improve this in the future
- .*revenue_table_name
include_table_lineage: true
include_usage_statistics: true
profiling:
enabled: true
profile_table_level_only: true
sink:
# sink configs
Config Details
- Options
- Schema
Note that a .
is used to denote nested fields in the YAML recipe.
Field | Description |
---|---|
bucket_duration Enum | Size of the time window to aggregate usage stats. Default: DAY |
column_limit integer | Maximum number of columns to process in a table. This is a low level config property which should be touched with care. This restriction is needed because excessively wide tables can result in failure to ingest the schema. Default: 300 |
convert_urns_to_lowercase boolean | Whether to convert dataset urns to lowercase. Default: False |
debug_include_full_payloads boolean | Include full payload into events. It is only for debugging and internal use. Default: False |
enable_legacy_sharded_table_support boolean | Use the legacy sharded table urn suffix added. Default: True |
enable_stateful_lineage_ingestion boolean | Enable stateful lineage ingestion. This will store lineage window timestamps after successful lineage ingestion. and will not run lineage ingestion for same timestamps in subsequent run. Default: True |
enable_stateful_profiling boolean | Enable stateful profiling. This will store profiling timestamps per dataset after successful profiling. and will not run profiling again in subsequent run if table has not been updated. Default: True |
enable_stateful_usage_ingestion boolean | Enable stateful lineage ingestion. This will store usage window timestamps after successful usage ingestion. and will not run usage ingestion for same timestamps in subsequent run. Default: True |
end_time string(date-time) | Latest date of lineage/usage to consider. Default: Current time in UTC |
exclude_empty_projects boolean | Option to exclude empty projects from being ingested. Default: False |
extra_client_options object | Additional options to pass to google.cloud.logging_v2.client.Client. Default: {} |
extract_column_lineage boolean | If enabled, generate column level lineage. Requires lineage_use_sql_parser to be enabled. Default: False |
extract_lineage_from_catalog boolean | This flag enables the data lineage extraction from Data Lineage API exposed by Google Data Catalog. NOTE: This extractor can't build views lineage. It's recommended to enable the view's DDL parsing. Read the docs to have more information about: https://cloud.google.com/data-catalog/docs/concepts/about-data-lineage Default: False |
extract_policy_tags_from_catalog boolean | This flag enables the extraction of policy tags from the Google Data Catalog API. When enabled, the extractor will fetch policy tags associated with BigQuery table columns. For more information about policy tags and column-level security, refer to the documentation: https://cloud.google.com/bigquery/docs/column-level-security-intro Default: False |
include_column_lineage_with_gcs boolean | When enabled, column-level lineage will be extracted from the gcs. Default: True |
include_data_platform_instance boolean | Whether to create a DataPlatformInstance aspect, equal to the BigQuery project id. If enabled, will cause redundancy in the browse path for BigQuery entities in the UI, because the project id is represented as the top-level container. Default: False |
include_external_url boolean | Whether to populate BigQuery Console url to Datasets/Tables Default: True |
include_schema_metadata boolean | Whether to ingest the BigQuery schema, i.e. projects, schemas, tables, and views. Default: True |
include_table_constraints boolean | Whether to ingest table constraints. If you know you don't use table constraints, you can disable it to save one extra query per dataset. In general it should be enabled Default: True |
include_table_lineage boolean | Option to enable/disable lineage generation. Is enabled by default. Default: True |
include_table_location_lineage boolean | If the source supports it, include table lineage to the underlying storage location. Default: True |
include_table_snapshots boolean | Whether table snapshots should be ingested. Default: True |
include_tables boolean | Whether tables should be ingested. Default: True |
include_usage_statistics boolean | Generate usage statistic Default: True |
include_views boolean | Whether views should be ingested. Default: True |
incremental_lineage boolean | When enabled, emits lineage as incremental to existing lineage already in DataHub. When disabled, re-states lineage on each run. Default: False |
lineage_parse_view_ddl boolean | Sql parse view ddl to get lineage. Default: True |
lineage_sql_parser_use_raw_names boolean | This parameter ignores the lowercase pattern stipulated in the SQLParser. NOTE: Ignored if lineage_use_sql_parser is False. Default: False |
lineage_use_sql_parser boolean | Use sql parser to resolve view/table lineage. Default: True |
log_page_size integer | The number of log item will be queried per page for lineage collection Default: 1000 |
match_fully_qualified_names boolean | [deprecated] Whether dataset_pattern is matched against fully qualified dataset name <project_id>.<dataset_name> . Default: True |
max_query_duration number(time-delta) | Correction to pad start_time and end_time with. For handling the case where the read happens within our time range but the query completion event is delayed and happens after the configured end time. Default: 900.0 |
max_threads_dataset_parallelism integer | Number of worker threads to use to parallelize BigQuery Dataset Metadata Extraction. Set to 1 to disable. Default: 20 |
number_of_datasets_process_in_batch_if_profiling_enabled integer | Number of partitioned table queried in batch when getting metadata. This is a low level config property which should be touched with care. This restriction is needed because we query partitions system view which throws error if we try to touch too many tables. Default: 1000 |
options object | Any options specified here will be passed to SQLAlchemy.create_engine as kwargs. |
platform_instance string | The instance of the platform that all assets produced by this recipe belong to. This should be unique within the platform. See https://datahubproject.io/docs/platform-instances/ for more details. |
project_on_behalf string | [Advanced] The BigQuery project in which queries are executed. Will be passed when creating a job. If not passed, falls back to the project associated with the service account. |
rate_limit boolean | Should we rate limit requests made to API. Default: False |
requests_per_min integer | Used to control number of API calls made per min. Only used when rate_limit is set to True . Default: 60 |
schema_resolution_batch_size integer | The number of tables to process in a batch when resolving schema from DataHub. Default: 100 |
scheme string | Default: bigquery |
sharded_table_pattern string | The regex pattern to match sharded tables and group as one table. This is a very low level config parameter, only change if you know what you are doing, Default: ((.+\D)[_$]?)?(\d\d\d\d(?:0[1-9]|1[0-2])(?:0[1-9]|... |
start_time string(date-time) | Earliest date of lineage/usage to consider. Default: Last full day in UTC (or hour, depending on bucket_duration ). You can also specify relative time with respect to end_time such as '-7 days' Or '-7d'. |
temp_table_dataset_prefix string | If you are creating temp tables in a dataset with a particular prefix you can use this config to set the prefix for the dataset. This is to support workflows from before bigquery's introduction of temp tables. By default we use _ because of datasets that begin with an underscore are hidden by default https://cloud.google.com/bigquery/docs/datasets#dataset-naming. Default: _ |
upstream_lineage_in_report boolean | Useful for debugging lineage information. Set to True to see the raw lineage created internally. Default: False |
use_date_sharded_audit_log_tables boolean | Whether to read date sharded tables or time partitioned tables when extracting usage from exported audit logs. Default: False |
use_exported_bigquery_audit_metadata boolean | When configured, use BigQueryAuditMetadata in bigquery_audit_metadata_datasets to compute lineage information. Default: False |
use_file_backed_cache boolean | Whether to use a file backed cache for the view definitions. Default: True |
use_queries_v2 boolean | If enabled, uses the new queries extractor to extract queries from bigquery. Default: False |
use_tables_list_query_v2 boolean | List tables using an improved query that extracts partitions and last modified timestamps more accurately. Requires the ability to read table data. Automatically enabled when profiling is enabled. Default: False |
env string | The environment that all assets produced by this connector belong to Default: PROD |
bigquery_audit_metadata_datasets array | A list of datasets that contain a table named cloudaudit_googleapis_com_data_access which contain BigQuery audit logs, specifically, those containing BigQueryAuditMetadata. It is recommended that the project of the dataset is also specified, for example, projectA.datasetB. |
bigquery_audit_metadata_datasets.string string | |
capture_dataset_label_as_tag One of boolean, AllowDenyPattern | Capture BigQuery dataset labels as DataHub tag Default: False |
capture_dataset_label_as_tag.ignoreCase boolean | Whether to ignore case sensitivity during pattern matching. Default: True |
capture_dataset_label_as_tag.allow array | List of regex patterns to include in ingestion Default: ['.*'] |
capture_dataset_label_as_tag.allow.string string | |
capture_dataset_label_as_tag.deny array | List of regex patterns to exclude from ingestion. Default: [] |
capture_dataset_label_as_tag.deny.string string | |
capture_table_label_as_tag One of boolean, AllowDenyPattern | Capture BigQuery table labels as DataHub tag Default: False |
capture_table_label_as_tag.ignoreCase boolean | Whether to ignore case sensitivity during pattern matching. Default: True |
capture_table_label_as_tag.allow array | List of regex patterns to include in ingestion Default: ['.*'] |
capture_table_label_as_tag.allow.string string | |
capture_table_label_as_tag.deny array | List of regex patterns to exclude from ingestion. Default: [] |
capture_table_label_as_tag.deny.string string | |
capture_view_label_as_tag One of boolean, AllowDenyPattern | Capture BigQuery view labels as DataHub tag Default: False |
capture_view_label_as_tag.ignoreCase boolean | Whether to ignore case sensitivity during pattern matching. Default: True |
capture_view_label_as_tag.allow array | List of regex patterns to include in ingestion Default: ['.*'] |
capture_view_label_as_tag.allow.string string | |
capture_view_label_as_tag.deny array | List of regex patterns to exclude from ingestion. Default: [] |
capture_view_label_as_tag.deny.string string | |
classification ClassificationConfig | For details, refer to Classification. Default: {'enabled': False, 'sample_size': 100, 'max_worker... |
classification.enabled boolean | Whether classification should be used to auto-detect glossary terms Default: False |
classification.info_type_to_term map(str,string) | |
classification.max_workers integer | Number of worker processes to use for classification. Set to 1 to disable. Default: 4 |
classification.sample_size integer | Number of sample values used for classification. Default: 100 |
classification.classifiers array | Classifiers to use to auto-detect glossary terms. If more than one classifier, infotype predictions from the classifier defined later in sequence take precedance. Default: [{'type': 'datahub', 'config': None}] |
classification.classifiers.DynamicTypedClassifierConfig DynamicTypedClassifierConfig | |
classification.classifiers.DynamicTypedClassifierConfig.type ❓ string | The type of the classifier to use. For DataHub, use datahub |
classification.classifiers.DynamicTypedClassifierConfig.config object | The configuration required for initializing the classifier. If not specified, uses defaults for classifer type. |
classification.column_pattern AllowDenyPattern | Regex patterns to filter columns for classification. This is used in combination with other patterns in parent config. Specify regex to match the column name in database.schema.table.column format. Default: {'allow': ['.*'], 'deny': [], 'ignoreCase': True} |
classification.column_pattern.ignoreCase boolean | Whether to ignore case sensitivity during pattern matching. Default: True |
classification.column_pattern.allow array | List of regex patterns to include in ingestion Default: ['.*'] |
classification.column_pattern.allow.string string | |
classification.column_pattern.deny array | List of regex patterns to exclude from ingestion. Default: [] |
classification.column_pattern.deny.string string | |
classification.table_pattern AllowDenyPattern | Regex patterns to filter tables for classification. This is used in combination with other patterns in parent config. Specify regex to match the entire table name in database.schema.table format. e.g. to match all tables starting with customer in Customer database and public schema, use the regex 'Customer.public.customer.*' Default: {'allow': ['.*'], 'deny': [], 'ignoreCase': True} |
classification.table_pattern.ignoreCase boolean | Whether to ignore case sensitivity during pattern matching. Default: True |
classification.table_pattern.allow array | List of regex patterns to include in ingestion Default: ['.*'] |
classification.table_pattern.allow.string string | |
classification.table_pattern.deny array | List of regex patterns to exclude from ingestion. Default: [] |
classification.table_pattern.deny.string string | |
credential BigQueryCredential | BigQuery credential informations |
credential.client_email ❓ string | Client email |
credential.client_id ❓ string | Client Id |
credential.private_key ❓ string | Private key in a form of '-----BEGIN PRIVATE KEY-----\nprivate-key\n-----END PRIVATE KEY-----\n' |
credential.private_key_id ❓ string | Private key id |
credential.project_id ❓ string | Project id to set the credentials |
credential.auth_provider_x509_cert_url string | Auth provider x509 certificate url |
credential.auth_uri string | Authentication uri |
credential.client_x509_cert_url string | If not set it will be default to https://www.googleapis.com/robot/v1/metadata/x509/client_email |
credential.token_uri string | Token uri Default: https://oauth2.googleapis.com/token |
credential.type string | Authentication type Default: service_account |
dataset_pattern AllowDenyPattern | Regex patterns for dataset to filter in ingestion. Specify regex to only match the schema name. e.g. to match all tables in schema analytics, use the regex 'analytics' Default: {'allow': ['.*'], 'deny': [], 'ignoreCase': True} |
dataset_pattern.ignoreCase boolean | Whether to ignore case sensitivity during pattern matching. Default: True |
dataset_pattern.allow array | List of regex patterns to include in ingestion Default: ['.*'] |
dataset_pattern.allow.string string | |
dataset_pattern.deny array | List of regex patterns to exclude from ingestion. Default: [] |
dataset_pattern.deny.string string | |
domain map(str,AllowDenyPattern) | A class to store allow deny regexes |
domain. key .allowarray | List of regex patterns to include in ingestion Default: ['.*'] |
domain. key .allow.stringstring | |
domain. key .ignoreCaseboolean | Whether to ignore case sensitivity during pattern matching. Default: True |
domain. key .denyarray | List of regex patterns to exclude from ingestion. Default: [] |
domain. key .deny.stringstring | |
gcs_lineage_config GcsLineageProviderConfig | Common config for gcs lineage generation Default: {'path_specs': [], 'strip_urls': True, 'ignore_non... |
gcs_lineage_config.ignore_non_path_spec_path boolean | Ignore paths that are not match in path_specs. It only applies if path_specs are specified. Default: False |
gcs_lineage_config.strip_urls boolean | Strip filename from gcs url. It only applies if path_specs are not specified. Default: True |
gcs_lineage_config.path_specs array | List of PathSpec. See below the details about PathSpec Default: [] |
gcs_lineage_config.path_specs.PathSpec PathSpec | |
gcs_lineage_config.path_specs.PathSpec.include ❓ string | Path to table. Name variable {table} is used to mark the folder with dataset. In absence of {table} , file level dataset will be created. Check below examples for more details. |
gcs_lineage_config.path_specs.PathSpec.allow_double_stars boolean | Allow double stars in the include path. This can affect performance significantly if enabled Default: False |
gcs_lineage_config.path_specs.PathSpec.autodetect_partitions boolean | Autodetect partition(s) from the path. If set to true, it will autodetect partition key/value if the folder format is {partition_key}={partition_value} for example year=2024 Default: True |
gcs_lineage_config.path_specs.PathSpec.default_extension string | For files without extension it will assume the specified file type. If it is not set the files without extensions will be skipped. |
gcs_lineage_config.path_specs.PathSpec.enable_compression boolean | Enable or disable processing compressed files. Currently .gz and .bz files are supported. Default: True |
gcs_lineage_config.path_specs.PathSpec.include_hidden_folders boolean | Include hidden folders in the traversal (folders starting with . or _ Default: False |
gcs_lineage_config.path_specs.PathSpec.sample_files boolean | Not listing all the files but only taking a handful amount of sample file to infer the schema. File count and file size calculation will be disabled. This can affect performance significantly if enabled Default: True |
gcs_lineage_config.path_specs.PathSpec.table_name string | Display name of the dataset.Combination of named variables from include path and strings |
gcs_lineage_config.path_specs.PathSpec.traversal_method Enum | Method to traverse the folder. ALL: Traverse all the folders, MIN_MAX: Traverse the folders by finding min and max value, MAX: Traverse the folder with max value Default: MAX |
gcs_lineage_config.path_specs.PathSpec.exclude array | list of paths in glob pattern which will be excluded while scanning for the datasets Default: [] |
gcs_lineage_config.path_specs.PathSpec.exclude.string string | |
gcs_lineage_config.path_specs.PathSpec.file_types array | Files with extenstions specified here (subset of default value) only will be scanned to create dataset. Other files will be omitted. Default: ['csv', 'tsv', 'json', 'parquet', 'avro'] |
gcs_lineage_config.path_specs.PathSpec.file_types.string string | |
profile_pattern AllowDenyPattern | Regex patterns to filter tables (or specific columns) for profiling during ingestion. Note that only tables allowed by the table_pattern will be considered. Default: {'allow': ['.*'], 'deny': [], 'ignoreCase': True} |
profile_pattern.ignoreCase boolean | Whether to ignore case sensitivity during pattern matching. Default: True |
profile_pattern.allow array | List of regex patterns to include in ingestion Default: ['.*'] |
profile_pattern.allow.string string | |
profile_pattern.deny array | List of regex patterns to exclude from ingestion. Default: [] |
profile_pattern.deny.string string | |
project_id_pattern AllowDenyPattern | Regex patterns for project_id to filter in ingestion. Default: {'allow': ['.*'], 'deny': [], 'ignoreCase': True} |
project_id_pattern.ignoreCase boolean | Whether to ignore case sensitivity during pattern matching. Default: True |
project_id_pattern.allow array | List of regex patterns to include in ingestion Default: ['.*'] |
project_id_pattern.allow.string string | |
project_id_pattern.deny array | List of regex patterns to exclude from ingestion. Default: [] |
project_id_pattern.deny.string string | |
project_ids array | Ingests specified project_ids. Use this property if you want to specify what projects to ingest or don't want to give project resourcemanager.projects.list to your service account. Overrides project_id_pattern . |
project_ids.string string | |
project_labels array | Ingests projects with the specified labels. Set value in the format of key:value . Use this property to define which projects to ingest basedon project-level labels. If project_ids or project_id is set, this configuration has no effect. The ingestion process filters projects by label first, and then applies the project_id_pattern. |
project_labels.string string | |
region_qualifiers array | BigQuery regions to be scanned for bigquery jobs when using use_queries_v2 . See this for details. Default: ['region-us', 'region-eu'] |
region_qualifiers.string string | |
table_pattern AllowDenyPattern | Regex patterns for tables to filter in ingestion. Specify regex to match the entire table name in database.schema.table format. e.g. to match all tables starting with customer in Customer database and public schema, use the regex 'Customer.public.customer.*' Default: {'allow': ['.*'], 'deny': [], 'ignoreCase': True} |
table_pattern.ignoreCase boolean | Whether to ignore case sensitivity during pattern matching. Default: True |
table_pattern.allow array | List of regex patterns to include in ingestion Default: ['.*'] |
table_pattern.allow.string string | |
table_pattern.deny array | List of regex patterns to exclude from ingestion. Default: [] |
table_pattern.deny.string string | |
table_snapshot_pattern AllowDenyPattern | Regex patterns for table snapshots to filter in ingestion. Specify regex to match the entire snapshot name in database.schema.snapshot format. e.g. to match all snapshots starting with customer in Customer database and public schema, use the regex 'Customer.public.customer.*' Default: {'allow': ['.*'], 'deny': [], 'ignoreCase': True} |
table_snapshot_pattern.ignoreCase boolean | Whether to ignore case sensitivity during pattern matching. Default: True |
table_snapshot_pattern.allow array | List of regex patterns to include in ingestion Default: ['.*'] |
table_snapshot_pattern.allow.string string | |
table_snapshot_pattern.deny array | List of regex patterns to exclude from ingestion. Default: [] |
table_snapshot_pattern.deny.string string | |
usage BigQueryUsageConfig | Usage related configs Default: {'bucket_duration': 'DAY', 'end_time': '2024-11-22... |
usage.apply_view_usage_to_tables boolean | Whether to apply view's usage to its base tables. If set to False, uses sql parser and applies usage to views / tables mentioned in the query. If set to True, usage is applied to base tables only. Default: False |
usage.bucket_duration Enum | Size of the time window to aggregate usage stats. Default: DAY |
usage.end_time string(date-time) | Latest date of lineage/usage to consider. Default: Current time in UTC |
usage.format_sql_queries boolean | Whether to format sql queries Default: False |
usage.include_operational_stats boolean | Whether to display operational stats. Default: True |
usage.include_read_operational_stats boolean | Whether to report read operational stats. Experimental. Default: False |
usage.include_top_n_queries boolean | Whether to ingest the top_n_queries. Default: True |
usage.max_query_duration number(time-delta) | Correction to pad start_time and end_time with. For handling the case where the read happens within our time range but the query completion event is delayed and happens after the configured end time. Default: 900.0 |
usage.start_time string(date-time) | Earliest date of lineage/usage to consider. Default: Last full day in UTC (or hour, depending on bucket_duration ). You can also specify relative time with respect to end_time such as '-7 days' Or '-7d'. |
usage.top_n_queries integer | Number of top queries to save to each table. Default: 10 |
usage.user_email_pattern AllowDenyPattern | regex patterns for user emails to filter in usage. Default: {'allow': ['.*'], 'deny': [], 'ignoreCase': True} |
usage.user_email_pattern.ignoreCase boolean | Whether to ignore case sensitivity during pattern matching. Default: True |
usage.user_email_pattern.allow array | List of regex patterns to include in ingestion Default: ['.*'] |
usage.user_email_pattern.allow.string string | |
usage.user_email_pattern.deny array | List of regex patterns to exclude from ingestion. Default: [] |
usage.user_email_pattern.deny.string string | |
view_pattern AllowDenyPattern | Regex patterns for views to filter in ingestion. Note: Defaults to table_pattern if not specified. Specify regex to match the entire view name in database.schema.view format. e.g. to match all views starting with customer in Customer database and public schema, use the regex 'Customer.public.customer.*' Default: {'allow': ['.*'], 'deny': [], 'ignoreCase': True} |
view_pattern.ignoreCase boolean | Whether to ignore case sensitivity during pattern matching. Default: True |
view_pattern.allow array | List of regex patterns to include in ingestion Default: ['.*'] |
view_pattern.allow.string string | |
view_pattern.deny array | List of regex patterns to exclude from ingestion. Default: [] |
view_pattern.deny.string string | |
profiling GEProfilingConfig | Default: {'enabled': False, 'operation_config': {'lower_fre... |
profiling.catch_exceptions boolean | Default: True |
profiling.enabled boolean | Whether profiling should be done. Default: False |
profiling.field_sample_values_limit integer | Upper limit for number of sample values to collect for all columns. Default: 20 |
profiling.include_field_distinct_count boolean | Whether to profile for the number of distinct values for each column. Default: True |
profiling.include_field_distinct_value_frequencies boolean | Whether to profile for distinct value frequencies. Default: False |
profiling.include_field_histogram boolean | Whether to profile for the histogram for numeric fields. Default: False |
profiling.include_field_max_value boolean | Whether to profile for the max value of numeric columns. Default: True |
profiling.include_field_mean_value boolean | Whether to profile for the mean value of numeric columns. Default: True |
profiling.include_field_median_value boolean | Whether to profile for the median value of numeric columns. Default: True |
profiling.include_field_min_value boolean | Whether to profile for the min value of numeric columns. Default: True |
profiling.include_field_null_count boolean | Whether to profile for the number of nulls for each column. Default: True |
profiling.include_field_quantiles boolean | Whether to profile for the quantiles of numeric columns. Default: False |
profiling.include_field_sample_values boolean | Whether to profile for the sample values for all columns. Default: True |
profiling.include_field_stddev_value boolean | Whether to profile for the standard deviation of numeric columns. Default: True |
profiling.limit integer | Max number of documents to profile. By default, profiles all documents. |
profiling.max_number_of_fields_to_profile integer | A positive integer that specifies the maximum number of columns to profile for any table. None implies all columns. The cost of profiling goes up significantly as the number of columns to profile goes up. |
profiling.max_workers integer | Number of worker threads to use for profiling. Set to 1 to disable. Default: 20 |
profiling.offset integer | Offset in documents to profile. By default, uses no offset. |
profiling.partition_datetime string(date-time) | If specified, profile only the partition which matches this datetime. If not specified, profile the latest partition. Only Bigquery supports this. |
profiling.partition_profiling_enabled boolean | Whether to profile partitioned tables. Only BigQuery and Aws Athena supports this. If enabled, latest partition data is used for profiling. Default: True |
profiling.profile_external_tables boolean | Whether to profile external tables. Only Snowflake and Redshift supports this. Default: False |
profiling.profile_if_updated_since_days number | Profile table only if it has been updated since these many number of days. If set to null , no constraint of last modified time for tables to profile. Supported only in snowflake and BigQuery . |
profiling.profile_table_level_only boolean | Whether to perform profiling at table-level only, or include column-level profiling as well. Default: False |
profiling.profile_table_row_count_estimate_only boolean | Use an approximate query for row count. This will be much faster but slightly less accurate. Only supported for Postgres and MySQL. Default: False |
profiling.profile_table_row_limit integer | Profile tables only if their row count is less then specified count. If set to null , no limit on the row count of tables to profile. Supported only in snowflake and BigQuery Default: 5000000 |
profiling.profile_table_size_limit integer | Profile tables only if their size is less then specified GBs. If set to null , no limit on the size of tables to profile. Supported only in snowflake and BigQuery Default: 5 |
profiling.query_combiner_enabled boolean | This feature is still experimental and can be disabled if it causes issues. Reduces the total number of queries issued and speeds up profiling by dynamically combining SQL queries where possible. Default: True |
profiling.report_dropped_profiles boolean | Whether to report datasets or dataset columns which were not profiled. Set to True for debugging purposes. Default: False |
profiling.sample_size integer | Number of rows to be sampled from table for column level profiling.Applicable only if use_sampling is set to True. Default: 10000 |
profiling.turn_off_expensive_profiling_metrics boolean | Whether to turn off expensive profiling or not. This turns off profiling for quantiles, distinct_value_frequencies, histogram & sample_values. This also limits maximum number of fields being profiled to 10. Default: False |
profiling.use_sampling boolean | Whether to profile column level stats on sample of table. Only BigQuery and Snowflake support this. If enabled, profiling is done on rows sampled from table. Sampling is not done for smaller tables. Default: True |
profiling.operation_config OperationConfig | Experimental feature. To specify operation configs. |
profiling.operation_config.lower_freq_profile_enabled boolean | Whether to do profiling at lower freq or not. This does not do any scheduling just adds additional checks to when not to run profiling. Default: False |
profiling.operation_config.profile_date_of_month integer | Number between 1 to 31 for date of month (both inclusive). If not specified, defaults to Nothing and this field does not take affect. |
profiling.operation_config.profile_day_of_week integer | Number between 0 to 6 for day of week (both inclusive). 0 is Monday and 6 is Sunday. If not specified, defaults to Nothing and this field does not take affect. |
profiling.tags_to_ignore_sampling array | Fixed list of tags to ignore sampling. If not specified, tables will be sampled based on use_sampling . |
profiling.tags_to_ignore_sampling.string string | |
stateful_ingestion StatefulStaleMetadataRemovalConfig | Base specialized config for Stateful Ingestion with stale metadata removal capability. |
stateful_ingestion.enabled boolean | Whether or not to enable stateful ingest. Default: True if a pipeline_name is set and either a datahub-rest sink or datahub_api is specified, otherwise False Default: False |
stateful_ingestion.remove_stale_metadata boolean | Soft-deletes the entities present in the last successful run but missing in the current run with stateful_ingestion enabled. Default: True |
The JSONSchema for this configuration is inlined below.
{
"title": "BigQueryV2Config",
"description": "Any source that produces gcs lineage from/to Datasets should inherit this class.\nThis is needeed to group all lineage related configs under `gcs_lineage_config` config property.",
"type": "object",
"properties": {
"classification": {
"title": "Classification",
"description": "For details, refer to [Classification](../../../../metadata-ingestion/docs/dev_guides/classification.md).",
"default": {
"enabled": false,
"sample_size": 100,
"max_workers": 4,
"table_pattern": {
"allow": [
".*"
],
"deny": [],
"ignoreCase": true
},
"column_pattern": {
"allow": [
".*"
],
"deny": [],
"ignoreCase": true
},
"info_type_to_term": {},
"classifiers": [
{
"type": "datahub",
"config": null
}
]
},
"allOf": [
{
"$ref": "#/definitions/ClassificationConfig"
}
]
},
"enable_stateful_profiling": {
"title": "Enable Stateful Profiling",
"description": "Enable stateful profiling. This will store profiling timestamps per dataset after successful profiling. and will not run profiling again in subsequent run if table has not been updated. ",
"default": true,
"type": "boolean"
},
"enable_stateful_lineage_ingestion": {
"title": "Enable Stateful Lineage Ingestion",
"description": "Enable stateful lineage ingestion. This will store lineage window timestamps after successful lineage ingestion. and will not run lineage ingestion for same timestamps in subsequent run. ",
"default": true,
"type": "boolean"
},
"bucket_duration": {
"description": "Size of the time window to aggregate usage stats.",
"default": "DAY",
"allOf": [
{
"$ref": "#/definitions/BucketDuration"
}
]
},
"end_time": {
"title": "End Time",
"description": "Latest date of lineage/usage to consider. Default: Current time in UTC",
"type": "string",
"format": "date-time"
},
"start_time": {
"title": "Start Time",
"description": "Earliest date of lineage/usage to consider. Default: Last full day in UTC (or hour, depending on `bucket_duration`). You can also specify relative time with respect to end_time such as '-7 days' Or '-7d'.",
"type": "string",
"format": "date-time"
},
"enable_stateful_usage_ingestion": {
"title": "Enable Stateful Usage Ingestion",
"description": "Enable stateful lineage ingestion. This will store usage window timestamps after successful usage ingestion. and will not run usage ingestion for same timestamps in subsequent run. ",
"default": true,
"type": "boolean"
},
"table_pattern": {
"title": "Table Pattern",
"description": "Regex patterns for tables to filter in ingestion. Specify regex to match the entire table name in database.schema.table format. e.g. to match all tables starting with customer in Customer database and public schema, use the regex 'Customer.public.customer.*'",
"default": {
"allow": [
".*"
],
"deny": [],
"ignoreCase": true
},
"allOf": [
{
"$ref": "#/definitions/AllowDenyPattern"
}
]
},
"view_pattern": {
"title": "View Pattern",
"description": "Regex patterns for views to filter in ingestion. Note: Defaults to table_pattern if not specified. Specify regex to match the entire view name in database.schema.view format. e.g. to match all views starting with customer in Customer database and public schema, use the regex 'Customer.public.customer.*'",
"default": {
"allow": [
".*"
],
"deny": [],
"ignoreCase": true
},
"allOf": [
{
"$ref": "#/definitions/AllowDenyPattern"
}
]
},
"incremental_lineage": {
"title": "Incremental Lineage",
"description": "When enabled, emits lineage as incremental to existing lineage already in DataHub. When disabled, re-states lineage on each run.",
"default": false,
"type": "boolean"
},
"convert_urns_to_lowercase": {
"title": "Convert Urns To Lowercase",
"description": "Whether to convert dataset urns to lowercase.",
"default": false,
"type": "boolean"
},
"env": {
"title": "Env",
"description": "The environment that all assets produced by this connector belong to",
"default": "PROD",
"type": "string"
},
"platform_instance": {
"title": "Platform Instance",
"description": "The instance of the platform that all assets produced by this recipe belong to. This should be unique within the platform. See https://datahubproject.io/docs/platform-instances/ for more details.",
"type": "string"
},
"stateful_ingestion": {
"$ref": "#/definitions/StatefulStaleMetadataRemovalConfig"
},
"options": {
"title": "Options",
"description": "Any options specified here will be passed to [SQLAlchemy.create_engine](https://docs.sqlalchemy.org/en/14/core/engines.html#sqlalchemy.create_engine) as kwargs.",
"type": "object"
},
"profile_pattern": {
"title": "Profile Pattern",
"description": "Regex patterns to filter tables (or specific columns) for profiling during ingestion. Note that only tables allowed by the `table_pattern` will be considered.",
"default": {
"allow": [
".*"
],
"deny": [],
"ignoreCase": true
},
"allOf": [
{
"$ref": "#/definitions/AllowDenyPattern"
}
]
},
"domain": {
"title": "Domain",
"description": "Attach domains to databases, schemas or tables during ingestion using regex patterns. Domain key can be a guid like *urn:li:domain:ec428203-ce86-4db3-985d-5a8ee6df32ba* or a string like \"Marketing\".) If you provide strings, then datahub will attempt to resolve this name to a guid, and will error out if this fails. There can be multiple domain keys specified.",
"default": {},
"type": "object",
"additionalProperties": {
"$ref": "#/definitions/AllowDenyPattern"
}
},
"include_views": {
"title": "Include Views",
"description": "Whether views should be ingested.",
"default": true,
"type": "boolean"
},
"include_tables": {
"title": "Include Tables",
"description": "Whether tables should be ingested.",
"default": true,
"type": "boolean"
},
"include_table_location_lineage": {
"title": "Include Table Location Lineage",
"description": "If the source supports it, include table lineage to the underlying storage location.",
"default": true,
"type": "boolean"
},
"use_file_backed_cache": {
"title": "Use File Backed Cache",
"description": "Whether to use a file backed cache for the view definitions.",
"default": true,
"type": "boolean"
},
"profiling": {
"title": "Profiling",
"default": {
"enabled": false,
"operation_config": {
"lower_freq_profile_enabled": false,
"profile_day_of_week": null,
"profile_date_of_month": null
},
"limit": null,
"offset": null,
"profile_table_level_only": false,
"include_field_null_count": true,
"include_field_distinct_count": true,
"include_field_min_value": true,
"include_field_max_value": true,
"include_field_mean_value": true,
"include_field_median_value": true,
"include_field_stddev_value": true,
"include_field_quantiles": false,
"include_field_distinct_value_frequencies": false,
"include_field_histogram": false,
"include_field_sample_values": true,
"max_workers": 20,
"report_dropped_profiles": false,
"turn_off_expensive_profiling_metrics": false,
"field_sample_values_limit": 20,
"max_number_of_fields_to_profile": null,
"profile_if_updated_since_days": null,
"profile_table_size_limit": 5,
"profile_table_row_limit": 5000000,
"profile_table_row_count_estimate_only": false,
"query_combiner_enabled": true,
"catch_exceptions": true,
"partition_profiling_enabled": true,
"partition_datetime": null,
"use_sampling": true,
"sample_size": 10000,
"profile_external_tables": false,
"tags_to_ignore_sampling": null
},
"allOf": [
{
"$ref": "#/definitions/GEProfilingConfig"
}
]
},
"include_data_platform_instance": {
"title": "Include Data Platform Instance",
"description": "Whether to create a DataPlatformInstance aspect, equal to the BigQuery project id. If enabled, will cause redundancy in the browse path for BigQuery entities in the UI, because the project id is represented as the top-level container.",
"default": false,
"type": "boolean"
},
"enable_legacy_sharded_table_support": {
"title": "Enable Legacy Sharded Table Support",
"description": "Use the legacy sharded table urn suffix added.",
"default": true,
"type": "boolean"
},
"project_ids": {
"title": "Project Ids",
"description": "Ingests specified project_ids. Use this property if you want to specify what projects to ingest or don't want to give project resourcemanager.projects.list to your service account. Overrides `project_id_pattern`.",
"type": "array",
"items": {
"type": "string"
}
},
"project_labels": {
"title": "Project Labels",
"description": "Ingests projects with the specified labels. Set value in the format of `key:value`. Use this property to define which projects to ingest basedon project-level labels. If project_ids or project_id is set, this configuration has no effect. The ingestion process filters projects by label first, and then applies the project_id_pattern.",
"type": "array",
"items": {
"type": "string"
}
},
"project_id_pattern": {
"title": "Project Id Pattern",
"description": "Regex patterns for project_id to filter in ingestion.",
"default": {
"allow": [
".*"
],
"deny": [],
"ignoreCase": true
},
"allOf": [
{
"$ref": "#/definitions/AllowDenyPattern"
}
]
},
"dataset_pattern": {
"title": "Dataset Pattern",
"description": "Regex patterns for dataset to filter in ingestion. Specify regex to only match the schema name. e.g. to match all tables in schema analytics, use the regex 'analytics'",
"default": {
"allow": [
".*"
],
"deny": [],
"ignoreCase": true
},
"allOf": [
{
"$ref": "#/definitions/AllowDenyPattern"
}
]
},
"match_fully_qualified_names": {
"title": "Match Fully Qualified Names",
"description": "[deprecated] Whether `dataset_pattern` is matched against fully qualified dataset name `<project_id>.<dataset_name>`.",
"default": true,
"type": "boolean"
},
"table_snapshot_pattern": {
"title": "Table Snapshot Pattern",
"description": "Regex patterns for table snapshots to filter in ingestion. Specify regex to match the entire snapshot name in database.schema.snapshot format. e.g. to match all snapshots starting with customer in Customer database and public schema, use the regex 'Customer.public.customer.*'",
"default": {
"allow": [
".*"
],
"deny": [],
"ignoreCase": true
},
"allOf": [
{
"$ref": "#/definitions/AllowDenyPattern"
}
]
},
"rate_limit": {
"title": "Rate Limit",
"description": "Should we rate limit requests made to API.",
"default": false,
"type": "boolean"
},
"requests_per_min": {
"title": "Requests Per Min",
"description": "Used to control number of API calls made per min. Only used when `rate_limit` is set to `True`.",
"default": 60,
"type": "integer"
},
"temp_table_dataset_prefix": {
"title": "Temp Table Dataset Prefix",
"description": "If you are creating temp tables in a dataset with a particular prefix you can use this config to set the prefix for the dataset. This is to support workflows from before bigquery's introduction of temp tables. By default we use `_` because of datasets that begin with an underscore are hidden by default https://cloud.google.com/bigquery/docs/datasets#dataset-naming.",
"default": "_",
"type": "string"
},
"sharded_table_pattern": {
"title": "Sharded Table Pattern",
"description": "The regex pattern to match sharded tables and group as one table. This is a very low level config parameter, only change if you know what you are doing, ",
"default": "((.+\\D)[_$]?)?(\\d\\d\\d\\d(?:0[1-9]|1[0-2])(?:0[1-9]|[12][0-9]|3[01]))$",
"deprecated": true,
"type": "string"
},
"credential": {
"title": "Credential",
"description": "BigQuery credential informations",
"allOf": [
{
"$ref": "#/definitions/BigQueryCredential"
}
]
},
"extra_client_options": {
"title": "Extra Client Options",
"description": "Additional options to pass to google.cloud.logging_v2.client.Client.",
"default": {},
"type": "object"
},
"project_on_behalf": {
"title": "Project On Behalf",
"description": "[Advanced] The BigQuery project in which queries are executed. Will be passed when creating a job. If not passed, falls back to the project associated with the service account.",
"type": "string"
},
"gcs_lineage_config": {
"title": "Gcs Lineage Config",
"description": "Common config for gcs lineage generation",
"default": {
"path_specs": [],
"strip_urls": true,
"ignore_non_path_spec_path": false
},
"allOf": [
{
"$ref": "#/definitions/GcsLineageProviderConfig"
}
]
},
"include_schema_metadata": {
"title": "Include Schema Metadata",
"description": "Whether to ingest the BigQuery schema, i.e. projects, schemas, tables, and views.",
"default": true,
"type": "boolean"
},
"usage": {
"title": "Usage",
"description": "Usage related configs",
"default": {
"bucket_duration": "DAY",
"end_time": "2024-11-22T22:54:56.042644+00:00",
"start_time": "2024-11-21T00:00:00+00:00",
"queries_character_limit": 24000,
"top_n_queries": 10,
"user_email_pattern": {
"allow": [
".*"
],
"deny": [],
"ignoreCase": true
},
"include_operational_stats": true,
"include_read_operational_stats": false,
"format_sql_queries": false,
"include_top_n_queries": true,
"max_query_duration": 900.0,
"apply_view_usage_to_tables": false
},
"allOf": [
{
"$ref": "#/definitions/BigQueryUsageConfig"
}
]
},
"include_usage_statistics": {
"title": "Include Usage Statistics",
"description": "Generate usage statistic",
"default": true,
"type": "boolean"
},
"capture_table_label_as_tag": {
"title": "Capture Table Label As Tag",
"description": "Capture BigQuery table labels as DataHub tag",
"default": false,
"anyOf": [
{
"type": "boolean"
},
{
"$ref": "#/definitions/AllowDenyPattern"
}
]
},
"capture_view_label_as_tag": {
"title": "Capture View Label As Tag",
"description": "Capture BigQuery view labels as DataHub tag",
"default": false,
"anyOf": [
{
"type": "boolean"
},
{
"$ref": "#/definitions/AllowDenyPattern"
}
]
},
"capture_dataset_label_as_tag": {
"title": "Capture Dataset Label As Tag",
"description": "Capture BigQuery dataset labels as DataHub tag",
"default": false,
"anyOf": [
{
"type": "boolean"
},
{
"$ref": "#/definitions/AllowDenyPattern"
}
]
},
"include_table_constraints": {
"title": "Include Table Constraints",
"description": "Whether to ingest table constraints. If you know you don't use table constraints, you can disable it to save one extra query per dataset. In general it should be enabled",
"default": true,
"type": "boolean"
},
"include_external_url": {
"title": "Include External Url",
"description": "Whether to populate BigQuery Console url to Datasets/Tables",
"default": true,
"type": "boolean"
},
"include_table_snapshots": {
"title": "Include Table Snapshots",
"description": "Whether table snapshots should be ingested.",
"default": true,
"type": "boolean"
},
"debug_include_full_payloads": {
"title": "Debug Include Full Payloads",
"description": "Include full payload into events. It is only for debugging and internal use.",
"default": false,
"type": "boolean"
},
"number_of_datasets_process_in_batch_if_profiling_enabled": {
"title": "Number Of Datasets Process In Batch If Profiling Enabled",
"description": "Number of partitioned table queried in batch when getting metadata. This is a low level config property which should be touched with care. This restriction is needed because we query partitions system view which throws error if we try to touch too many tables.",
"default": 1000,
"type": "integer"
},
"use_tables_list_query_v2": {
"title": "Use Tables List Query V2",
"description": "List tables using an improved query that extracts partitions and last modified timestamps more accurately. Requires the ability to read table data. Automatically enabled when profiling is enabled.",
"default": false,
"type": "boolean"
},
"use_queries_v2": {
"title": "Use Queries V2",
"description": "If enabled, uses the new queries extractor to extract queries from bigquery.",
"default": false,
"type": "boolean"
},
"column_limit": {
"title": "Column Limit",
"description": "Maximum number of columns to process in a table. This is a low level config property which should be touched with care. This restriction is needed because excessively wide tables can result in failure to ingest the schema.",
"default": 300,
"type": "integer"
},
"lineage_use_sql_parser": {
"title": "Lineage Use Sql Parser",
"description": "Use sql parser to resolve view/table lineage.",
"default": true,
"type": "boolean"
},
"lineage_parse_view_ddl": {
"title": "Lineage Parse View Ddl",
"description": "Sql parse view ddl to get lineage.",
"default": true,
"type": "boolean"
},
"lineage_sql_parser_use_raw_names": {
"title": "Lineage Sql Parser Use Raw Names",
"description": "This parameter ignores the lowercase pattern stipulated in the SQLParser. NOTE: Ignored if lineage_use_sql_parser is False.",
"default": false,
"type": "boolean"
},
"extract_column_lineage": {
"title": "Extract Column Lineage",
"description": "If enabled, generate column level lineage. Requires lineage_use_sql_parser to be enabled.",
"default": false,
"type": "boolean"
},
"extract_lineage_from_catalog": {
"title": "Extract Lineage From Catalog",
"description": "This flag enables the data lineage extraction from Data Lineage API exposed by Google Data Catalog. NOTE: This extractor can't build views lineage. It's recommended to enable the view's DDL parsing. Read the docs to have more information about: https://cloud.google.com/data-catalog/docs/concepts/about-data-lineage",
"default": false,
"type": "boolean"
},
"extract_policy_tags_from_catalog": {
"title": "Extract Policy Tags From Catalog",
"description": "This flag enables the extraction of policy tags from the Google Data Catalog API. When enabled, the extractor will fetch policy tags associated with BigQuery table columns. For more information about policy tags and column-level security, refer to the documentation: https://cloud.google.com/bigquery/docs/column-level-security-intro",
"default": false,
"type": "boolean"
},
"scheme": {
"title": "Scheme",
"default": "bigquery",
"type": "string"
},
"log_page_size": {
"title": "Log Page Size",
"description": "The number of log item will be queried per page for lineage collection",
"default": 1000,
"exclusiveMinimum": 0,
"type": "integer"
},
"include_table_lineage": {
"title": "Include Table Lineage",
"description": "Option to enable/disable lineage generation. Is enabled by default.",
"default": true,
"type": "boolean"
},
"include_column_lineage_with_gcs": {
"title": "Include Column Lineage With Gcs",
"description": "When enabled, column-level lineage will be extracted from the gcs.",
"default": true,
"type": "boolean"
},
"max_query_duration": {
"title": "Max Query Duration",
"description": "Correction to pad start_time and end_time with. For handling the case where the read happens within our time range but the query completion event is delayed and happens after the configured end time.",
"default": 900.0,
"type": "number",
"format": "time-delta"
},
"bigquery_audit_metadata_datasets": {
"title": "Bigquery Audit Metadata Datasets",
"description": "A list of datasets that contain a table named cloudaudit_googleapis_com_data_access which contain BigQuery audit logs, specifically, those containing BigQueryAuditMetadata. It is recommended that the project of the dataset is also specified, for example, projectA.datasetB.",
"type": "array",
"items": {
"type": "string"
}
},
"use_exported_bigquery_audit_metadata": {
"title": "Use Exported Bigquery Audit Metadata",
"description": "When configured, use BigQueryAuditMetadata in bigquery_audit_metadata_datasets to compute lineage information.",
"default": false,
"type": "boolean"
},
"use_date_sharded_audit_log_tables": {
"title": "Use Date Sharded Audit Log Tables",
"description": "Whether to read date sharded tables or time partitioned tables when extracting usage from exported audit logs.",
"default": false,
"type": "boolean"
},
"upstream_lineage_in_report": {
"title": "Upstream Lineage In Report",
"description": "Useful for debugging lineage information. Set to True to see the raw lineage created internally.",
"default": false,
"type": "boolean"
},
"exclude_empty_projects": {
"title": "Exclude Empty Projects",
"description": "Option to exclude empty projects from being ingested.",
"default": false,
"type": "boolean"
},
"schema_resolution_batch_size": {
"title": "Schema Resolution Batch Size",
"description": "The number of tables to process in a batch when resolving schema from DataHub.",
"default": 100,
"hidden_from_schema": true,
"type": "integer"
},
"max_threads_dataset_parallelism": {
"title": "Max Threads Dataset Parallelism",
"description": "Number of worker threads to use to parallelize BigQuery Dataset Metadata Extraction. Set to 1 to disable.",
"default": 20,
"type": "integer"
},
"region_qualifiers": {
"title": "Region Qualifiers",
"description": "BigQuery regions to be scanned for bigquery jobs when using `use_queries_v2`. See [this](https://cloud.google.com/bigquery/docs/information-schema-jobs#scope_and_syntax) for details.",
"default": [
"region-us",
"region-eu"
],
"type": "array",
"items": {
"type": "string"
}
}
},
"additionalProperties": false,
"definitions": {
"AllowDenyPattern": {
"title": "AllowDenyPattern",
"description": "A class to store allow deny regexes",
"type": "object",
"properties": {
"allow": {
"title": "Allow",
"description": "List of regex patterns to include in ingestion",
"default": [
".*"
],
"type": "array",
"items": {
"type": "string"
}
},
"deny": {
"title": "Deny",
"description": "List of regex patterns to exclude from ingestion.",
"default": [],
"type": "array",
"items": {
"type": "string"
}
},
"ignoreCase": {
"title": "Ignorecase",
"description": "Whether to ignore case sensitivity during pattern matching.",
"default": true,
"type": "boolean"
}
},
"additionalProperties": false
},
"DynamicTypedClassifierConfig": {
"title": "DynamicTypedClassifierConfig",
"type": "object",
"properties": {
"type": {
"title": "Type",
"description": "The type of the classifier to use. For DataHub, use `datahub`",
"type": "string"
},
"config": {
"title": "Config",
"description": "The configuration required for initializing the classifier. If not specified, uses defaults for classifer type."
}
},
"required": [
"type"
],
"additionalProperties": false
},
"ClassificationConfig": {
"title": "ClassificationConfig",
"type": "object",
"properties": {
"enabled": {
"title": "Enabled",
"description": "Whether classification should be used to auto-detect glossary terms",
"default": false,
"type": "boolean"
},
"sample_size": {
"title": "Sample Size",
"description": "Number of sample values used for classification.",
"default": 100,
"type": "integer"
},
"max_workers": {
"title": "Max Workers",
"description": "Number of worker processes to use for classification. Set to 1 to disable.",
"default": 4,
"type": "integer"
},
"table_pattern": {
"title": "Table Pattern",
"description": "Regex patterns to filter tables for classification. This is used in combination with other patterns in parent config. Specify regex to match the entire table name in `database.schema.table` format. e.g. to match all tables starting with customer in Customer database and public schema, use the regex 'Customer.public.customer.*'",
"default": {
"allow": [
".*"
],
"deny": [],
"ignoreCase": true
},
"allOf": [
{
"$ref": "#/definitions/AllowDenyPattern"
}
]
},
"column_pattern": {
"title": "Column Pattern",
"description": "Regex patterns to filter columns for classification. This is used in combination with other patterns in parent config. Specify regex to match the column name in `database.schema.table.column` format.",
"default": {
"allow": [
".*"
],
"deny": [],
"ignoreCase": true
},
"allOf": [
{
"$ref": "#/definitions/AllowDenyPattern"
}
]
},
"info_type_to_term": {
"title": "Info Type To Term",
"description": "Optional mapping to provide glossary term identifier for info type",
"default": {},
"type": "object",
"additionalProperties": {
"type": "string"
}
},
"classifiers": {
"title": "Classifiers",
"description": "Classifiers to use to auto-detect glossary terms. If more than one classifier, infotype predictions from the classifier defined later in sequence take precedance.",
"default": [
{
"type": "datahub",
"config": null
}
],
"type": "array",
"items": {
"$ref": "#/definitions/DynamicTypedClassifierConfig"
}
}
},
"additionalProperties": false
},
"BucketDuration": {
"title": "BucketDuration",
"description": "An enumeration.",
"enum": [
"DAY",
"HOUR"
],
"type": "string"
},
"DynamicTypedStateProviderConfig": {
"title": "DynamicTypedStateProviderConfig",
"type": "object",
"properties": {
"type": {
"title": "Type",
"description": "The type of the state provider to use. For DataHub use `datahub`",
"type": "string"
},
"config": {
"title": "Config",
"description": "The configuration required for initializing the state provider. Default: The datahub_api config if set at pipeline level. Otherwise, the default DatahubClientConfig. See the defaults (https://github.com/datahub-project/datahub/blob/master/metadata-ingestion/src/datahub/ingestion/graph/client.py#L19).",
"default": {},
"type": "object"
}
},
"required": [
"type"
],
"additionalProperties": false
},
"StatefulStaleMetadataRemovalConfig": {
"title": "StatefulStaleMetadataRemovalConfig",
"description": "Base specialized config for Stateful Ingestion with stale metadata removal capability.",
"type": "object",
"properties": {
"enabled": {
"title": "Enabled",
"description": "Whether or not to enable stateful ingest. Default: True if a pipeline_name is set and either a datahub-rest sink or `datahub_api` is specified, otherwise False",
"default": false,
"type": "boolean"
},
"remove_stale_metadata": {
"title": "Remove Stale Metadata",
"description": "Soft-deletes the entities present in the last successful run but missing in the current run with stateful_ingestion enabled.",
"default": true,
"type": "boolean"
}
},
"additionalProperties": false
},
"OperationConfig": {
"title": "OperationConfig",
"type": "object",
"properties": {
"lower_freq_profile_enabled": {
"title": "Lower Freq Profile Enabled",
"description": "Whether to do profiling at lower freq or not. This does not do any scheduling just adds additional checks to when not to run profiling.",
"default": false,
"type": "boolean"
},
"profile_day_of_week": {
"title": "Profile Day Of Week",
"description": "Number between 0 to 6 for day of week (both inclusive). 0 is Monday and 6 is Sunday. If not specified, defaults to Nothing and this field does not take affect.",
"type": "integer"
},
"profile_date_of_month": {
"title": "Profile Date Of Month",
"description": "Number between 1 to 31 for date of month (both inclusive). If not specified, defaults to Nothing and this field does not take affect.",
"type": "integer"
}
},
"additionalProperties": false
},
"GEProfilingConfig": {
"title": "GEProfilingConfig",
"type": "object",
"properties": {
"enabled": {
"title": "Enabled",
"description": "Whether profiling should be done.",
"default": false,
"type": "boolean"
},
"operation_config": {
"title": "Operation Config",
"description": "Experimental feature. To specify operation configs.",
"allOf": [
{
"$ref": "#/definitions/OperationConfig"
}
]
},
"limit": {
"title": "Limit",
"description": "Max number of documents to profile. By default, profiles all documents.",
"type": "integer"
},
"offset": {
"title": "Offset",
"description": "Offset in documents to profile. By default, uses no offset.",
"type": "integer"
},
"profile_table_level_only": {
"title": "Profile Table Level Only",
"description": "Whether to perform profiling at table-level only, or include column-level profiling as well.",
"default": false,
"type": "boolean"
},
"include_field_null_count": {
"title": "Include Field Null Count",
"description": "Whether to profile for the number of nulls for each column.",
"default": true,
"type": "boolean"
},
"include_field_distinct_count": {
"title": "Include Field Distinct Count",
"description": "Whether to profile for the number of distinct values for each column.",
"default": true,
"type": "boolean"
},
"include_field_min_value": {
"title": "Include Field Min Value",
"description": "Whether to profile for the min value of numeric columns.",
"default": true,
"type": "boolean"
},
"include_field_max_value": {
"title": "Include Field Max Value",
"description": "Whether to profile for the max value of numeric columns.",
"default": true,
"type": "boolean"
},
"include_field_mean_value": {
"title": "Include Field Mean Value",
"description": "Whether to profile for the mean value of numeric columns.",
"default": true,
"type": "boolean"
},
"include_field_median_value": {
"title": "Include Field Median Value",
"description": "Whether to profile for the median value of numeric columns.",
"default": true,
"type": "boolean"
},
"include_field_stddev_value": {
"title": "Include Field Stddev Value",
"description": "Whether to profile for the standard deviation of numeric columns.",
"default": true,
"type": "boolean"
},
"include_field_quantiles": {
"title": "Include Field Quantiles",
"description": "Whether to profile for the quantiles of numeric columns.",
"default": false,
"type": "boolean"
},
"include_field_distinct_value_frequencies": {
"title": "Include Field Distinct Value Frequencies",
"description": "Whether to profile for distinct value frequencies.",
"default": false,
"type": "boolean"
},
"include_field_histogram": {
"title": "Include Field Histogram",
"description": "Whether to profile for the histogram for numeric fields.",
"default": false,
"type": "boolean"
},
"include_field_sample_values": {
"title": "Include Field Sample Values",
"description": "Whether to profile for the sample values for all columns.",
"default": true,
"type": "boolean"
},
"max_workers": {
"title": "Max Workers",
"description": "Number of worker threads to use for profiling. Set to 1 to disable.",
"default": 20,
"type": "integer"
},
"report_dropped_profiles": {
"title": "Report Dropped Profiles",
"description": "Whether to report datasets or dataset columns which were not profiled. Set to `True` for debugging purposes.",
"default": false,
"type": "boolean"
},
"turn_off_expensive_profiling_metrics": {
"title": "Turn Off Expensive Profiling Metrics",
"description": "Whether to turn off expensive profiling or not. This turns off profiling for quantiles, distinct_value_frequencies, histogram & sample_values. This also limits maximum number of fields being profiled to 10.",
"default": false,
"type": "boolean"
},
"field_sample_values_limit": {
"title": "Field Sample Values Limit",
"description": "Upper limit for number of sample values to collect for all columns.",
"default": 20,
"type": "integer"
},
"max_number_of_fields_to_profile": {
"title": "Max Number Of Fields To Profile",
"description": "A positive integer that specifies the maximum number of columns to profile for any table. `None` implies all columns. The cost of profiling goes up significantly as the number of columns to profile goes up.",
"exclusiveMinimum": 0,
"type": "integer"
},
"profile_if_updated_since_days": {
"title": "Profile If Updated Since Days",
"description": "Profile table only if it has been updated since these many number of days. If set to `null`, no constraint of last modified time for tables to profile. Supported only in `snowflake` and `BigQuery`.",
"exclusiveMinimum": 0,
"type": "number"
},
"profile_table_size_limit": {
"title": "Profile Table Size Limit",
"description": "Profile tables only if their size is less then specified GBs. If set to `null`, no limit on the size of tables to profile. Supported only in `snowflake` and `BigQuery`",
"default": 5,
"type": "integer"
},
"profile_table_row_limit": {
"title": "Profile Table Row Limit",
"description": "Profile tables only if their row count is less then specified count. If set to `null`, no limit on the row count of tables to profile. Supported only in `snowflake` and `BigQuery`",
"default": 5000000,
"type": "integer"
},
"profile_table_row_count_estimate_only": {
"title": "Profile Table Row Count Estimate Only",
"description": "Use an approximate query for row count. This will be much faster but slightly less accurate. Only supported for Postgres and MySQL. ",
"default": false,
"type": "boolean"
},
"query_combiner_enabled": {
"title": "Query Combiner Enabled",
"description": "*This feature is still experimental and can be disabled if it causes issues.* Reduces the total number of queries issued and speeds up profiling by dynamically combining SQL queries where possible.",
"default": true,
"type": "boolean"
},
"catch_exceptions": {
"title": "Catch Exceptions",
"default": true,
"type": "boolean"
},
"partition_profiling_enabled": {
"title": "Partition Profiling Enabled",
"description": "Whether to profile partitioned tables. Only BigQuery and Aws Athena supports this. If enabled, latest partition data is used for profiling.",
"default": true,
"type": "boolean"
},
"partition_datetime": {
"title": "Partition Datetime",
"description": "If specified, profile only the partition which matches this datetime. If not specified, profile the latest partition. Only Bigquery supports this.",
"type": "string",
"format": "date-time"
},
"use_sampling": {
"title": "Use Sampling",
"description": "Whether to profile column level stats on sample of table. Only BigQuery and Snowflake support this. If enabled, profiling is done on rows sampled from table. Sampling is not done for smaller tables. ",
"default": true,
"type": "boolean"
},
"sample_size": {
"title": "Sample Size",
"description": "Number of rows to be sampled from table for column level profiling.Applicable only if `use_sampling` is set to True.",
"default": 10000,
"type": "integer"
},
"profile_external_tables": {
"title": "Profile External Tables",
"description": "Whether to profile external tables. Only Snowflake and Redshift supports this.",
"default": false,
"type": "boolean"
},
"tags_to_ignore_sampling": {
"title": "Tags To Ignore Sampling",
"description": "Fixed list of tags to ignore sampling. If not specified, tables will be sampled based on `use_sampling`.",
"type": "array",
"items": {
"type": "string"
}
}
},
"additionalProperties": false
},
"BigQueryCredential": {
"title": "BigQueryCredential",
"type": "object",
"properties": {
"project_id": {
"title": "Project Id",
"description": "Project id to set the credentials",
"type": "string"
},
"private_key_id": {
"title": "Private Key Id",
"description": "Private key id",
"type": "string"
},
"private_key": {
"title": "Private Key",
"description": "Private key in a form of '-----BEGIN PRIVATE KEY-----\\nprivate-key\\n-----END PRIVATE KEY-----\\n'",
"type": "string"
},
"client_email": {
"title": "Client Email",
"description": "Client email",
"type": "string"
},
"client_id": {
"title": "Client Id",
"description": "Client Id",
"type": "string"
},
"auth_uri": {
"title": "Auth Uri",
"description": "Authentication uri",
"default": "https://accounts.google.com/o/oauth2/auth",
"type": "string"
},
"token_uri": {
"title": "Token Uri",
"description": "Token uri",
"default": "https://oauth2.googleapis.com/token",
"type": "string"
},
"auth_provider_x509_cert_url": {
"title": "Auth Provider X509 Cert Url",
"description": "Auth provider x509 certificate url",
"default": "https://www.googleapis.com/oauth2/v1/certs",
"type": "string"
},
"type": {
"title": "Type",
"description": "Authentication type",
"default": "service_account",
"type": "string"
},
"client_x509_cert_url": {
"title": "Client X509 Cert Url",
"description": "If not set it will be default to https://www.googleapis.com/robot/v1/metadata/x509/client_email",
"type": "string"
}
},
"required": [
"project_id",
"private_key_id",
"private_key",
"client_email",
"client_id"
],
"additionalProperties": false
},
"SortKeyType": {
"title": "SortKeyType",
"description": "An enumeration.",
"enum": [
"STRING",
"INTEGER",
"FLOAT",
"DATETIME",
"DATE"
]
},
"SortKey": {
"title": "SortKey",
"type": "object",
"properties": {
"key": {
"title": "Key",
"description": "The key to sort on. This can be a compound key based on the path_spec variables.",
"type": "string"
},
"type": {
"description": "The date format to use when sorting. This is used to parse the date from the key. The format should follow the java [SimpleDateFormat](https://docs.oracle.com/javase/8/docs/api/java/text/SimpleDateFormat.html) format.",
"default": "STRING",
"allOf": [
{
"$ref": "#/definitions/SortKeyType"
}
]
},
"date_format": {
"title": "Date Format",
"description": "The date format to use when sorting. This is used to parse the date from the key. The format should follow the java [SimpleDateFormat](https://docs.oracle.com/javase/8/docs/api/java/text/SimpleDateFormat.html) format.",
"type": "string"
}
},
"required": [
"key"
],
"additionalProperties": false
},
"FolderTraversalMethod": {
"title": "FolderTraversalMethod",
"description": "An enumeration.",
"enum": [
"ALL",
"MIN_MAX",
"MAX"
]
},
"PathSpec": {
"title": "PathSpec",
"type": "object",
"properties": {
"include": {
"title": "Include",
"description": "Path to table. Name variable `{table}` is used to mark the folder with dataset. In absence of `{table}`, file level dataset will be created. Check below examples for more details.",
"type": "string"
},
"exclude": {
"title": "Exclude",
"description": "list of paths in glob pattern which will be excluded while scanning for the datasets",
"default": [],
"type": "array",
"items": {
"type": "string"
}
},
"file_types": {
"title": "File Types",
"description": "Files with extenstions specified here (subset of default value) only will be scanned to create dataset. Other files will be omitted.",
"default": [
"csv",
"tsv",
"json",
"parquet",
"avro"
],
"type": "array",
"items": {
"type": "string"
}
},
"default_extension": {
"title": "Default Extension",
"description": "For files without extension it will assume the specified file type. If it is not set the files without extensions will be skipped.",
"type": "string"
},
"table_name": {
"title": "Table Name",
"description": "Display name of the dataset.Combination of named variables from include path and strings",
"type": "string"
},
"enable_compression": {
"title": "Enable Compression",
"description": "Enable or disable processing compressed files. Currently .gz and .bz files are supported.",
"default": true,
"type": "boolean"
},
"sample_files": {
"title": "Sample Files",
"description": "Not listing all the files but only taking a handful amount of sample file to infer the schema. File count and file size calculation will be disabled. This can affect performance significantly if enabled",
"default": true,
"type": "boolean"
},
"allow_double_stars": {
"title": "Allow Double Stars",
"description": "Allow double stars in the include path. This can affect performance significantly if enabled",
"default": false,
"type": "boolean"
},
"autodetect_partitions": {
"title": "Autodetect Partitions",
"description": "Autodetect partition(s) from the path. If set to true, it will autodetect partition key/value if the folder format is {partition_key}={partition_value} for example `year=2024`",
"default": true,
"type": "boolean"
},
"traversal_method": {
"description": "Method to traverse the folder. ALL: Traverse all the folders, MIN_MAX: Traverse the folders by finding min and max value, MAX: Traverse the folder with max value",
"default": "MAX",
"allOf": [
{
"$ref": "#/definitions/FolderTraversalMethod"
}
]
},
"include_hidden_folders": {
"title": "Include Hidden Folders",
"description": "Include hidden folders in the traversal (folders starting with . or _",
"default": false,
"type": "boolean"
}
},
"required": [
"include"
],
"additionalProperties": false
},
"GcsLineageProviderConfig": {
"title": "GcsLineageProviderConfig",
"description": "Any source that produces gcs lineage from/to Datasets should inherit this class.",
"type": "object",
"properties": {
"path_specs": {
"title": "Path Specs",
"description": "List of PathSpec. See below the details about PathSpec",
"default": [],
"type": "array",
"items": {
"$ref": "#/definitions/PathSpec"
}
},
"strip_urls": {
"title": "Strip Urls",
"description": "Strip filename from gcs url. It only applies if path_specs are not specified.",
"default": true,
"type": "boolean"
},
"ignore_non_path_spec_path": {
"title": "Ignore Non Path Spec Path",
"description": "Ignore paths that are not match in path_specs. It only applies if path_specs are specified.",
"default": false,
"type": "boolean"
}
},
"additionalProperties": false
},
"BigQueryUsageConfig": {
"title": "BigQueryUsageConfig",
"type": "object",
"properties": {
"bucket_duration": {
"description": "Size of the time window to aggregate usage stats.",
"default": "DAY",
"allOf": [
{
"$ref": "#/definitions/BucketDuration"
}
]
},
"end_time": {
"title": "End Time",
"description": "Latest date of lineage/usage to consider. Default: Current time in UTC",
"type": "string",
"format": "date-time"
},
"start_time": {
"title": "Start Time",
"description": "Earliest date of lineage/usage to consider. Default: Last full day in UTC (or hour, depending on `bucket_duration`). You can also specify relative time with respect to end_time such as '-7 days' Or '-7d'.",
"type": "string",
"format": "date-time"
},
"top_n_queries": {
"title": "Top N Queries",
"description": "Number of top queries to save to each table.",
"default": 10,
"exclusiveMinimum": 0,
"type": "integer"
},
"user_email_pattern": {
"title": "User Email Pattern",
"description": "regex patterns for user emails to filter in usage.",
"default": {
"allow": [
".*"
],
"deny": [],
"ignoreCase": true
},
"allOf": [
{
"$ref": "#/definitions/AllowDenyPattern"
}
]
},
"include_operational_stats": {
"title": "Include Operational Stats",
"description": "Whether to display operational stats.",
"default": true,
"type": "boolean"
},
"include_read_operational_stats": {
"title": "Include Read Operational Stats",
"description": "Whether to report read operational stats. Experimental.",
"default": false,
"type": "boolean"
},
"format_sql_queries": {
"title": "Format Sql Queries",
"description": "Whether to format sql queries",
"default": false,
"type": "boolean"
},
"include_top_n_queries": {
"title": "Include Top N Queries",
"description": "Whether to ingest the top_n_queries.",
"default": true,
"type": "boolean"
},
"max_query_duration": {
"title": "Max Query Duration",
"description": "Correction to pad start_time and end_time with. For handling the case where the read happens within our time range but the query completion event is delayed and happens after the configured end time.",
"default": 900.0,
"type": "number",
"format": "time-delta"
},
"apply_view_usage_to_tables": {
"title": "Apply View Usage To Tables",
"description": "Whether to apply view's usage to its base tables. If set to False, uses sql parser and applies usage to views / tables mentioned in the query. If set to True, usage is applied to base tables only.",
"default": false,
"type": "boolean"
}
},
"additionalProperties": false
}
}
}
Code Coordinates
- Class Name:
datahub.ingestion.source.bigquery_v2.bigquery.BigqueryV2Source
- Browse on GitHub
Questions
If you've got any questions on configuring ingestion for BigQuery, feel free to ping us on our Slack.