Category "google-cloud-platform"

Bigquery keyword Remote is not supported

We are trying out the REMOTE functions within bigquery as per this guide. We created the CLOUD_RESOURCE using the following command : bq mk --connection --disp

Micronaut GCP global pubsub endpoint

I've searched the Micronaut docs for a solution how to define GCP topic endpoint globally, but with no luck. Right now, I'll have to do the below config in ever

GCP Vertex Pipeline - Why kfp.v2.dsl.Output as function arguments work without being provided?

Why kfp.v2.dsl.Output as function argument works without being provided? I am following Create and run ML pipelines with Vertex Pipelines! Jupyter notebook exa

How to use C# and Google.Cloud.Dialogflow.Cx.V3 to generate valid Google DialogFlow CX webhook response JSON

I've created a webhook using C# and ASP.NET Core in order to try to generate the webhook response to DialogFlow, but I'm really struggling with using Google.Clo

gcloud: The user does not have access to service account "default"

I attempting to use an activated service account scoped to create and delete gcloud container clusters (k8s clusters), using the following commands: gcloud con

Allow Public Read access on a GCS bucket?

I am trying to allow anonymous (or just from my applications domain) read access for files in my bucket. When trying to read the files I get ``` <Error&g

Kubeflow Pipelines error on GCP - Run doesn't end

After submitting the "run" using Jupyter notebook, when I go to the Kubeflow pipeline dashboard, I can see my "run" submitted & running but it doesn't end e

Provision a GCP VM instance with no external IP via Terraform

Trying to create a VM in GCP via terraform with External IP as None. network_interface { network = "projects/other-project-name/global/networks/network-name"

How to add an ssh key to an GCP instance using terraform?

So I have a terraform script that creates instances in Google Cloud Platform, I want to be able to have my terraform script also add my ssh key to the instances

How to give service account only access to one bucket (Google Cloud)?

How to give service account only access to one bucket? If you just in case wanna give another 3rd party service access to your private bucket? The problem is by

Google Cloud Functions sometimes doesn't find @google-cloud/* dependencies although the buildpack builds and runs fine in Docker on local

I am deploying a Google Cloud function and it fails to deploy because it sometimes doesn't find the module @google-cloud/pubsub and when it finds it, it instead

How to fix 404 error when installing npm package from GCP artifact registry with yarn?

I'm having an issue with installing an NPM package from GCP. I was able to upload the package to the artifact registry of GCP by doing the following steps: Logi

Firebase functions: cloud storage triggers not working because of insufficient permissions

I have a cloud function that is triggered by updates to storage. when I deploy any function that utilises cloud storage, I get an error code 7 and an insufficie

Using Python to do basic Google Drive operations without web server authentication

I want to create a Python script that can do some really basic things (add/remove files in a shared google drive) The script will be running on remote PCs so an

How to parse json format output of : kubectl get pods using jsonpath

How to parse the json to retrieve a field from output of kubectl get pods -o json From the command line I need to obtain the system generated container name

IAM permission issue with Batch Translation: PermissionDenied: 403 Cloud IAM permission 'cloudtranslate.generalModels.batchPredict' denied

This is my first time running google batch translation, I usually run the API translation, but this time my file size is too large to be translated that way. I

Is there a way to see all GCP Unlabelled resource?

My goal is to see all the unlabelled inventory so I can label them correctly and show it in cost tracking

Google storage permissions error while generating signed url from cloud function

I'm attempting to use a Firebase Cloud Function to create signed download URLs for files stored in a Storage Bucket. Using the snippet below on my local machine

GKE metrics comparsion

Does anyone know how can I compare my query result to some value? For example, I have such query: fetch k8s_container | { metric 'kubernetes.io/container/cpu/co

Django Static - Google Cloud Storage - CDN

I am still new with serving my Django static/media files to Google cloud storage. It's working now but I am not sure if this is right or is this enough already,