Category "google-bigquery"

How do I use the BigQueryDate class in the BigQuery nodejs client library?

I have a BigQuery table with a column of type "DATE". When I query the table with the nodejs client library as per below, the date elements in the data are obje

How is a binary/bytes column sorted in sql?

The following query sorts a binary column in BigQuery: with tbl as ( select B'123' as col union all select B'234' ) select * from tbl order by col; ----------

What is the quota limit of Google Analytics 4 Bigquery link

I activated google analytics 4 bigquery linking, and it works correctly. I wonder what is the limitation of this link. Api metrics shows that link using api met

bigquery sql: converting column values into column names [duplicate]

I would like to convert all the keys into column headers and all the values into respective average values in a row underneath, grouped by dat

How to decrease storage size for PostgreSQL on Google Cloud?

I moved some of my data from PostgreSQL to BigQuery. Before PostgreSQL database was using 130 GB of storage, now I only need 30GB. However, in the Google Cloud

Connecting Tranco Google BigQuery with Metabase

I am trying to connect third party ranking management system (https://tranco-list.eu/) with metabase. Tranco is giving us an option to see the record on Google

BigQuery partition pruning while using analytical function as part of a view, on a DATETIME field with DAY granularity

I am trying to use analytical functions (e.g. FIRST_VALUE), while still benefiting from from partition pruning. This while the table is partitioned on a DATETIM

Avoid formatting comments

Is there anyway to avoid formatting the comments? To reproduce, paste below example SQL code to BQ: --important comment on table info select * from mytable1;

Bigquery An internal error occurred and the request could not be completed. Error: 60909208

I'm trying to run the following query but it keeps throwing this error after it keeps running for around 1 to 2 hours An internal error occurred and the reques

Syntax error: Unexpected keyword UNNEST at [12:8] while using unnest

I want to use unnest in the following function to use INkeyword but it is throwing error unexpected keyword UNNEST while using unnest. CREATE TEMPORARY FUNCT

Array specified for non-repeated field, Can I load JSON Format "{a:[1,2,3]}" into BQ

I want upload JSON data which is in the below form {a:[1,2,3]}, into BigQuery, I am familiar Record Field-Type with repeat Mode. I am getting error Array spe

SAS BigQuery Logistic Regression

Is there a way I can match SAS logistic regression results with BigQuery ML logistic regression results (coefficient / intercept values for same data)?

Firebase - Perform Analytics from database/firestore data

I am using Firebase as my authentication and database platform in my React Native-Expo app. I have not yet decided if I will be using the realtime-database or F

Need help creating schema for loading CSV into BigQuery

I am trying to load some CSV files into BigQuery from Google Cloud Storage and wrestling with schema generation. There is an auto-generate option but it is poor

Data Transfers to Big Query error "Error while reading data, error message: JSON table encountered too many errors, giving up. Rows: 1; errors: 1"

Hello im currently trying to establish daily data transfers from Google Cloud Storage to Big Query tables. Theses tables are just meant to store raw data (JSON

BigQuery - extract the most used tables

I'm using the Google BigQuery and looking on the default audit dataset. I know that this dataset contains various data about the queries the users are running.

Big Query - Table 'Copy' from Console Not working, But the same works fine from bq CLI

Across project level when trying to Copying the tables in big query, it works fine using bq CLI but not from Console. Big Query --> Project:Dataset.tableXYZ

Number of rows affected by BigQuery query

I run daily commands to insert new records into a BigQuery table, and would like to log how many records get inserted each day. I create a QueryJob object that

Is there a way to upload the byte type created with zlib to Google bigquery?

I want to input string data into bigquery by implied by pyhton's zlib library. Here is an example code that uses zlib to generate data: import zlib import p

How do I produce a new column with values based on a Partition of one column while also based on the values of two additional columns (BigQuery)?

I have a table that records all the different statuses for a list of Jobs with timestamps. So the ID column has many Ids that appear several times as their stat