Category "google-bigquery"

How do I create a table with multiple auto incrementing values?

I would like to have an autoincrementing string in my BigQuery table. This seems like a good way of doing it: CREATE TABLE dbo.YourTable ( ID INT IDENTITY(1

Bigquery - which table are affected by another table

I want to delete a table in bigquery. But before that, I want to know which tables are being linked to the table I'm about to delete. Is there a query to extrac

Create/Query different tables based on conditional statement in BQ SQL

To illustrate with an example, say I have two tables in BigQuery called "fruits" and "vegetables" respectively. I want to have a variable at the start that lets

Defining Month Time Partition using Teradata Data Transfer Service Custom Schema

I am looking for a way to define a Teradata Data Transfer custom schema that implements a month based date partition. The documentation only provides a method

Can DBT directly connect to Google sheets?

we currently have a DBT instance that sits over our Google BigQuery data warehouse. Now we've recently been asked to incorporate some data from Google Sheets in

Beam to BigQuery silently failing to create BigQuery table

I am building a data pipeline from PubSub to Beam (Direct/Dataflow Runner) to Big Query. Today we started to run into issues where beam IO BigQuery connector st

how to repeat a block of code in bigquery for multiple columns?

I would like to repeat a block of code with a complex operation (defined in a subquery) for many columns in bigquery. I cannot find a better way than copy and p

How to count elements based on a unique value in BigQuery

I have this table 1 in Bigquery and I need to count the elements in column segments and category that correspond to a single user id. Desired outcome presented

BigQuery Search which results in Error: Scalar subquery produced more than one element

Running this query on BigQuery but resulting in this error: Error: Scalar subquery produced more than one element select * from CRM.organizations where name L

getting error while trying to copy data using google_bigquery_data_transfer_config using terraform

I am trying to setup a bigquery data transfer configuration using terraform. I am using my personal gcp account. I have a setup of terraform on my laptop so tha

Simple way to use bigquery result in python where the query return only one result

I'm simply trying to get one data from bigquery from python. Below is the query select max(date) from `x.x.table_name` all the requirements to access bigquery

Apache beam dataflow Big query IO without schema

Is there any way to write unstructured data to a big query table using apache beam dataflow big query io API (i.e without providing schema upfront)

BigQuery - Scheduled Query Error "PermissionDenied: 403 The caller does not have permission"

Having an issue with Scheduled Queries, using a python script, I get the error: "google.api_core.exceptions.PermissionDenied: 403 The caller does not have permi

Bigquery Regexp pull from 2 different patterns

I'm trying to pull product search colors from urls in my bigquery data, but the urls format changed at one point so I'm trying to pull them from two different f

Grant user permission to bigquery.datasets.create in schedules notebook in Vertex AI

I have a notebook in which I access data through APIs, work with the data and send the results to a BigQuery table as well as to a GCS bucket. Everything works

BigQuery: Set expirationTime on Snapshot

I'm trying to create a BigQuery snapshot of a table with an expiration date 60 minutes in the future. It looks like I can't define a expiration time when creati

Bigquery data transfers from S3 and efficiency of filtering data by last modification datetime

I'm planning a pipeline from an S3 bucket to BigQuery. I have set up daily bigquery data transfers from S3 as described here. So, I have a data transfer configu

I have a Date in a String format and I can't convert it to date in BigQuery

I started with a date in a string format from a JSON extraction using this: json_value(answer, '$.date_created') and got an output 2020-01-02T10:26:47.056-04:00

Unnesting a big quantity of columns in BigQuery and BigTable

I have a table in BigTable, with a single column family, containing some lead data. I was following the Google Cloud guide to querying BigTable data from BigTab

Load local CSV file into BigQuery table with Terraform?

I'm new to terraform. Is it possible to load the content of a CSV file into a BigQuery table without uploading it to GCS? I've studied the document below, but t