Error writing to BigQuery table in different zone

Description

If you write to a BigQuery table that is in a different zone than the dataproc cluster, the pipeline fails with:

The job is not found because it has id
lbg-obd:europe-west2.direct-bigqueryhelper-import-a77c79e4-4e07-4373-b8dd-3a905940876c
and not
lbg-obd:direct-bigqueryhelper-import-a77c79e4-4e07-4373-b8dd-3a905940876c

Moving the datasets and temp storage to the same zone as the dataproc profile causes this issue to go away, but that is clearly not a good solution. I would assume changing the zone for the dataproc profile would also be a workaround.

Release Notes

Fixed a bug in the BigQuery sink that would cause pipelines to fail when writing to a dataset in a different region.

Activity

Show:

Ali Anwar May 9, 2019 at 12:48 PM

Albert Shau May 2, 2019 at 11:06 PM

The data is actually written to the table correctly, it's just that the pipeline fails because it can't lookup the job status correctly.

Albert Shau February 22, 2019 at 4:18 PM

From what I remember, the plugin just uses the Google bigquery outputformat. It may be being configured incorrectly.

Fixed
Pinned fields
Click on the next to a field label to start pinning.

Details

Assignee

Reporter

Labels

Components

Fix versions

Priority

Created February 22, 2019 at 4:17 PM
Updated July 17, 2020 at 12:28 AM
Resolved May 3, 2019 at 12:16 AM