Category "google-bigquery"

BigQuery table creation from google cloud storage

I am trying to create a table in Google BigQuery Table in the console. I have given the correct GCS bucket address where the CSV resides. The CSV is about 140GB

get Forbidden: 403 Access Denied when do request to transfer data from google cloud storage to bigquery using python

I'm owner in google cloud console. I want transfer data from gcs to Bigquery, but I get "Forbidden: 403 Access Denied: File gs://data/silver/rates/nbu_rates/nbu

Clickstream performance Scoring formula using SQL

imagine we have a clickstream and we want to have some everyday score that follows next rules: increase by 1 each impression increase by 10 each click decrease

How do I create a table with multiple auto incrementing values?

I would like to have an autoincrementing string in my BigQuery table. This seems like a good way of doing it: CREATE TABLE dbo.YourTable ( ID INT IDENTITY(1

Bigquery - which table are affected by another table

I want to delete a table in bigquery. But before that, I want to know which tables are being linked to the table I'm about to delete. Is there a query to extrac

Create/Query different tables based on conditional statement in BQ SQL

To illustrate with an example, say I have two tables in BigQuery called "fruits" and "vegetables" respectively. I want to have a variable at the start that lets

Defining Month Time Partition using Teradata Data Transfer Service Custom Schema

I am looking for a way to define a Teradata Data Transfer custom schema that implements a month based date partition. The documentation only provides a method

Can DBT directly connect to Google sheets?

we currently have a DBT instance that sits over our Google BigQuery data warehouse. Now we've recently been asked to incorporate some data from Google Sheets in

Beam to BigQuery silently failing to create BigQuery table

I am building a data pipeline from PubSub to Beam (Direct/Dataflow Runner) to Big Query. Today we started to run into issues where beam IO BigQuery connector st

how to repeat a block of code in bigquery for multiple columns?

I would like to repeat a block of code with a complex operation (defined in a subquery) for many columns in bigquery. I cannot find a better way than copy and p

How to count elements based on a unique value in BigQuery

I have this table 1 in Bigquery and I need to count the elements in column segments and category that correspond to a single user id. Desired outcome presented

BigQuery Search which results in Error: Scalar subquery produced more than one element

Running this query on BigQuery but resulting in this error: Error: Scalar subquery produced more than one element select * from CRM.organizations where name L

getting error while trying to copy data using google_bigquery_data_transfer_config using terraform

I am trying to setup a bigquery data transfer configuration using terraform. I am using my personal gcp account. I have a setup of terraform on my laptop so tha

Simple way to use bigquery result in python where the query return only one result

I'm simply trying to get one data from bigquery from python. Below is the query select max(date) from `x.x.table_name` all the requirements to access bigquery

Apache beam dataflow Big query IO without schema

Is there any way to write unstructured data to a big query table using apache beam dataflow big query io API (i.e without providing schema upfront)

BigQuery - Scheduled Query Error "PermissionDenied: 403 The caller does not have permission"

Having an issue with Scheduled Queries, using a python script, I get the error: "google.api_core.exceptions.PermissionDenied: 403 The caller does not have permi

Bigquery Regexp pull from 2 different patterns

I'm trying to pull product search colors from urls in my bigquery data, but the urls format changed at one point so I'm trying to pull them from two different f

Grant user permission to bigquery.datasets.create in schedules notebook in Vertex AI

I have a notebook in which I access data through APIs, work with the data and send the results to a BigQuery table as well as to a GCS bucket. Everything works

BigQuery: Set expirationTime on Snapshot

I'm trying to create a BigQuery snapshot of a table with an expiration date 60 minutes in the future. It looks like I can't define a expiration time when creati

Bigquery data transfers from S3 and efficiency of filtering data by last modification datetime

I'm planning a pipeline from an S3 bucket to BigQuery. I have set up daily bigquery data transfers from S3 as described here. So, I have a data transfer configu