...
Property | Macro Enabled? | Version Introduced | Description |
---|---|---|---|
Use Connection | No | 6.7.0/0.20.0 | Optional. Whether to use a connection. If a connection is used, you do not need to provide the credentials. |
Connection | Yes | 6.7.0/0.20.0 | Optional. Name of the connection to use. Project and service account information will be provided by the connection. You can also use the macro function |
Project ID | Yes | Optional. Google Cloud Project ID, which uniquely identifies a project. It can be found on the Dashboard in the Google Cloud Platform Console. This is the project that the BigQuery job will run in. Default is auto-detect. | |
Dataset Project ID | Yes | Optional. Project the dataset belongs to. This is only required if the dataset is not in the same project that the BigQuery job will run in. If no value is given, it will default to the configured Project ID. | |
Service Account Type | Yes | 6.3.0/0.16.0 | Optional. Service account key used for authorization. Select File Path or JSON. If you select File Path, enter the Service Account File Path. If you select JSON, enter the JSON Path. Default is Select File Path. |
Service Account File Path | Yes | Optional. Path on the local file system of the service account key used for authorization. Can be set to 'auto-detect' when running on a Dataproc cluster. When running on other clusters, the file must be present on every node in the cluster. Default is auto-detect. | |
Service Account JSON | Yes | 6.3.0/0.16.0 | Optional. Contents of the service account JSON file. |
Reference Name | No | Required. Name used to uniquely identify this sink for lineage, annotating metadata, etc. | |
Dataset | Yes | Required. Dataset the tables belongs to. A dataset is contained within a specific project. Datasets are top-level containers that are used to organize and control access to tables and views. If dataset does not exist, it will be created. | |
Truncate Table | Yes | Optional. Whether or not to truncate the table before writing to it. Default is False. | |
Temporary Bucket Name | Yes | Optional. Google Cloud Storage bucket to store temporary data in. It will be automatically created if it does not exist, but will not be automatically deleted. Temporary data will be deleted after it is loaded into BigQuery. If it is not provided, a unique bucket will be created and then deleted after the run finishes. Syntax: | |
GCS Upload Request Chunk Size | Yes | Optional. GCS upload request chunk size in bytes. Default is 8388608 bytes. | |
Split Field | Yes | Optional. The name of the field that will be used to determine which table to write to. | |
Allow flexible schemas in Output | Yes | Optional. When enabled, this sink will write out records with arbitrary schemas. Records may not have a well defined schema depending on the source. When disabled, table schemas must be passed in pipeline arguments. | |
Update Table Schema | Yes | Optional. Whether the BigQuery table schema should be modified when it does not match the schema expected by the pipeline.
Compatible changes fall under the following categories:
Incompatible schema changes will result in pipeline failure. Default is False. | |
Location | Yes | Optional. The location where the big query datasets will get created. This value is ignored if the dataset or temporary bucket already exist. Default is US. | |
Encryption Key Name | Yes | 6.5.1/0.18.1 | Optional. The GCP customer managed encryption key (CMEK) used to encrypt data written to any bucket, dataset, or table created by the plugin. If the bucket, dataset, or table already exists, this is ignored. More information can be found here. |
...