Category "kubernetes"

kubernetes mariadb doesnt startt sql's from /docker-entrypoint-initdb.d

pod logs: 2022-01-04 10:43:13+00:00 [Note] [Entrypoint]: Entrypoint script for MariaDB Server 1:10.7.1+maria~focal started. 2022-01-04 10:43:14+00:00 [Note] [En

preinstall VScode remote extensions in a docker container

I'm currently using the vscode extension to attach to running containers in Kubernetes as described here: https://code.visualstudio.com/docs/remote/attach-conta

Regular expression to collect all IPs from Pod definition

I am trying to collect IP addresses of the pods which has 3 network interfaces attached to. From the output of kubectl get pod .. -o yaml, I would like to coll

Expected behavior for ClusterRole with two rules mentioning the same resource

I'm working with RBAC on an Openshift cluster. I was wondering what is the expected behavior for applying a ClusterRole that has two rules which reference the s

Unable to connect to remote Kubernetes cluster

When I attempt to connect to a remote Kubernetes cluster with kubectl cluster-info I am getting an error as in the image below : Trying to curl the cluster als

unable to connect with pod running in GKE

Created 3 node GKE cluster. From cmd prompt local logged into gcloud. Created a pod with nginx container and exposed port 80 apiVersion: v1 kind: Pod metada

How to debug Unknown field error in Kubernetes?

This might be rookie question. I am not well versed with kubernetes. I added this to my deployment.yaml ad.datadoghq.com/helm-chart.check_names: | ["o

getting error while installing kured using fleet

the error iam getting is: error while running post render on files:invalid cluster scoped object[name=namespaces-kured kind=ClusterRole apiversion:rbac.authoriz

how to connect spark workers to spark driver in kubernetes (standalone cluster)

I created a Dockerfile with just debian and apache spark downloaded from the main website. I then created a kubernetes deployment to have 1 pod running spark dr

Kubernetes new pod created, old deleted

I have a nodejs (nestjs) application running in a pod which downloads large csv files and then uploads these to a database (mysql, typeorm) I was tailing the lo

How should I go about writing code that runs on a schedule that will run on multiple containers without race conditions?

I'm wanting to write a service in nestjs that polls another service on a schedule. The problem is that if I have multiple containers how can I make sure one and

variable substitution in a Kubernetes YAML file

I have a deployments.yaml file and for a hostpath in my spec I am doing something like this: spec: volumes: - name: fe-source hostPa

Metrics server is currently unable to handle the request

I am new to kubernetes and was trying to apply horizontal pod autoscaling to my existing application. and after following other stackoverflow details - got to k

Apply node-specific DaemonSet configuration

I've got 4 nodes in a kubernetes cluster and a metric exporter for prometheus that has to run on each node, each one exposing the nodes individual metrics. This

nginx in Docker - returns "server" directive is not allowed here

I have dockerized an Angular application with the following Dockerfile and nginx.conf. The container works fine in Docker desktop. When deploying in Kubernetes

Deployment gaps at fast pace growing application [closed]

Some context: I have little experience with ci/CD and manage a fast paced growing application since it saw the light of the day for the first

Subdirectory of PV directory isn't mounted to persistent volume

I've got a PV that is backed by do-block-storage, and it's mounted to a stateful set, at /var/lib/postgresql, this is the full stateful set: apiVersion: apps/v1

SSE connection not established

I have implemented Server-Sent events with fastify. The client code: const url = `${server}/sse/events`; const connection = new EventSource(url);

how to use postgres.db.name for multiple databases in kubernetes configMaps

So I want to create multiple postgresql databases in the kubernetes deployment. I tried with the below configMaps configuration but the databases are not being

Kubernetes logs from kubectl logs command is different to /var/log/containers in GKE cluster

I have created a GKE cluster on GCP. Kubernetes logs from kubectl logs command is different to /var/log/containers kubectl {"method":"GET","path":"/healthz","fo