Category "google-bigquery"

How to retrieve Performance MAX campaigns Google Ads into Google Big Query

I am using a third-party tool that use the Adwords Reporting https://developers.google.com/adwords/api/docs/appendix/reports/all-reports to retrieve data. At th

How to explicitly split data for training and evaluation with BigQuery ML?

I understand there's already another post, but it's a bit old and it doesn't really answer the question. I understand that we can use the parameter DATA_SPLIT_

Schedule a bq command line in GCP with Cloud Pub/Sub topic

I have a command line that I execute to select data from view and exporting to a staging table in BigQuery: bq mk --transfer_config --headless=true --project_i

Loading data from BigQuery to Salesforce

I have two datasources - Salesforce(SF) and BigQuery(BQ). All the data from SF is available in BQ. So my question is can we load data from BQ to SF?. Here's w

bq load command to load parquet file from GCS to BigQuery with column name start with number

I am loading parquet file into BigQuery using bq load command, my parquet file contains column name start with number (e.g. 00_abc, 01_xyz). since BigQuery don'

How do I use the BigQueryDate class in the BigQuery nodejs client library?

I have a BigQuery table with a column of type "DATE". When I query the table with the nodejs client library as per below, the date elements in the data are obje

How is a binary/bytes column sorted in sql?

The following query sorts a binary column in BigQuery: with tbl as ( select B'123' as col union all select B'234' ) select * from tbl order by col; ----------

What is the quota limit of Google Analytics 4 Bigquery link

I activated google analytics 4 bigquery linking, and it works correctly. I wonder what is the limitation of this link. Api metrics shows that link using api met

bigquery sql: converting column values into column names [duplicate]

I would like to convert all the keys into column headers and all the values into respective average values in a row underneath, grouped by dat

How to decrease storage size for PostgreSQL on Google Cloud?

I moved some of my data from PostgreSQL to BigQuery. Before PostgreSQL database was using 130 GB of storage, now I only need 30GB. However, in the Google Cloud

Connecting Tranco Google BigQuery with Metabase

I am trying to connect third party ranking management system (https://tranco-list.eu/) with metabase. Tranco is giving us an option to see the record on Google

BigQuery partition pruning while using analytical function as part of a view, on a DATETIME field with DAY granularity

I am trying to use analytical functions (e.g. FIRST_VALUE), while still benefiting from from partition pruning. This while the table is partitioned on a DATETIM

Avoid formatting comments

Is there anyway to avoid formatting the comments? To reproduce, paste below example SQL code to BQ: --important comment on table info select * from mytable1;

Bigquery An internal error occurred and the request could not be completed. Error: 60909208

I'm trying to run the following query but it keeps throwing this error after it keeps running for around 1 to 2 hours An internal error occurred and the reques

Syntax error: Unexpected keyword UNNEST at [12:8] while using unnest

I want to use unnest in the following function to use INkeyword but it is throwing error unexpected keyword UNNEST while using unnest. CREATE TEMPORARY FUNCT

Array specified for non-repeated field, Can I load JSON Format "{a:[1,2,3]}" into BQ

I want upload JSON data which is in the below form {a:[1,2,3]}, into BigQuery, I am familiar Record Field-Type with repeat Mode. I am getting error Array spe

SAS BigQuery Logistic Regression

Is there a way I can match SAS logistic regression results with BigQuery ML logistic regression results (coefficient / intercept values for same data)?

Firebase - Perform Analytics from database/firestore data

I am using Firebase as my authentication and database platform in my React Native-Expo app. I have not yet decided if I will be using the realtime-database or F

Need help creating schema for loading CSV into BigQuery

I am trying to load some CSV files into BigQuery from Google Cloud Storage and wrestling with schema generation. There is an auto-generate option but it is poor

Data Transfers to Big Query error "Error while reading data, error message: JSON table encountered too many errors, giving up. Rows: 1; errors: 1"

Hello im currently trying to establish daily data transfers from Google Cloud Storage to Big Query tables. Theses tables are just meant to store raw data (JSON