I am new to gcp big query. I am trying to model scd2 implementation on both outer and nested table on GCP big query. I would like to know your suggestions wethe
I'm trying to set variable in order to make further operation. But when I set my first variable on line 2 ("SET(new_user)=( "), I get an error Query error: Can
I have some complex BigQuery views that nest each other. I would like to map them in an ER diagram so they are easier to visualize. I can do t
Have ~50k compressed (gzip) json files daily that need to be uploaded to BQ with some transformation, no API calls. The size of the files may be up to 1Gb. What
I understand the BigQuery provides 7 days time travel. I'm trying to get Max and Min of time travel possible for each table like Min and Max of timestamp. Is th
I am trying to create a table in Google BigQuery Table in the console. I have given the correct GCS bucket address where the CSV resides. The CSV is about 140GB
I'm owner in google cloud console. I want transfer data from gcs to Bigquery, but I get "Forbidden: 403 Access Denied: File gs://data/silver/rates/nbu_rates/nbu
imagine we have a clickstream and we want to have some everyday score that follows next rules: increase by 1 each impression increase by 10 each click decrease
I would like to have an autoincrementing string in my BigQuery table. This seems like a good way of doing it: CREATE TABLE dbo.YourTable ( ID INT IDENTITY(1
I want to delete a table in bigquery. But before that, I want to know which tables are being linked to the table I'm about to delete. Is there a query to extrac
To illustrate with an example, say I have two tables in BigQuery called "fruits" and "vegetables" respectively. I want to have a variable at the start that lets
I am looking for a way to define a Teradata Data Transfer custom schema that implements a month based date partition. The documentation only provides a method
we currently have a DBT instance that sits over our Google BigQuery data warehouse. Now we've recently been asked to incorporate some data from Google Sheets in
I am building a data pipeline from PubSub to Beam (Direct/Dataflow Runner) to Big Query. Today we started to run into issues where beam IO BigQuery connector st
I would like to repeat a block of code with a complex operation (defined in a subquery) for many columns in bigquery. I cannot find a better way than copy and p
I have this table 1 in Bigquery and I need to count the elements in column segments and category that correspond to a single user id. Desired outcome presented
Running this query on BigQuery but resulting in this error: Error: Scalar subquery produced more than one element select * from CRM.organizations where name L
I am trying to setup a bigquery data transfer configuration using terraform. I am using my personal gcp account. I have a setup of terraform on my laptop so tha
I'm simply trying to get one data from bigquery from python. Below is the query select max(date) from `x.x.table_name` all the requirements to access bigquery
Is there any way to write unstructured data to a big query table using apache beam dataflow big query io API (i.e without providing schema upfront)