I'm trying to pull product search colors from urls in my bigquery data, but the urls format changed at one point so I'm trying to pull them from two different f
I have a notebook in which I access data through APIs, work with the data and send the results to a BigQuery table as well as to a GCS bucket. Everything works
I'm trying to create a BigQuery snapshot of a table with an expiration date 60 minutes in the future. It looks like I can't define a expiration time when creati
I'm planning a pipeline from an S3 bucket to BigQuery. I have set up daily bigquery data transfers from S3 as described here. So, I have a data transfer configu
I started with a date in a string format from a JSON extraction using this: json_value(answer, '$.date_created') and got an output 2020-01-02T10:26:47.056-04:00
I have a table in BigTable, with a single column family, containing some lead data. I was following the Google Cloud guide to querying BigTable data from BigTab
I'm new to terraform. Is it possible to load the content of a CSV file into a BigQuery table without uploading it to GCS? I've studied the document below, but t
So there is a requirement to build tables and charts in google sheets with data in bigquery. The data can be brought in to google sheets using connected sheets.
Let's take the following example from BigQuery, but it's the same across all other databases I've seen: with tbl as ( select [1,2,3] as arr, '{"IDS": [1
I integrated my Firestore solution with BigQuery. With every Firestore Insert/Update/Delete operation the Data gets transferred to BigQuery. I am trying to sche
Getting below error while connecting Simba driver with BigQuery on OBIEE UNIX Env , any suggesting's. Do we have some license file associated with the driver. U
I am following this guide in setting up a vertex AI pipeline. I am told to supply the "BigQuery project for exporting data", except that the guide is not clear
I'm trying to get the total duration of membership types (member&casual) another application of the same problem would be to get the total count of the popu
I have two date formats 2020-01-02T10:26:47.056-04:00 and January 2, 2020, 10:10 AM that I need to transform into a comparable date format to make a subtraction
I'm working with Cloud Function. I have the following query working correctly: # this is working q = """ SELECT col1, col2 FROM `my_table` WHER
I initially extracted the weeks from a date as 1, 2, 3 using the extract clause but now, I want to have the year along with it as week1, 2022 or week 40, 2021
We recently re-started our license to GA360, having had it 3 years ago. The old dataset which the old export sent data to is in the project where we created the
I would like to know how to calculate the total bytes which were processed by query jobs in a specified period per reference table to reduce BigQuery costs. I c
I'm very new to Google Cloud CLI, sorry for the dumb question but it's really annoying. I'm trying to execute this command: bq mk --transfer_config --target_dat
I have the total amount expected to be saved, the total amount saved, the principal amount expected to be saved and the principal amount saved, now I'm trying t