input
stringclasses
1 value
output
stringlengths
30
28.3k
text
stringlengths
151
31.8k
instruction
stringlengths
49
30.3k
Fabric8's Kubernetes Client is using a generated model and DSL that has the exact same structure as as the JSON and YAML configuration. So in order to create a Service instance that looks like: { "kind": "Service", "apiVersion": "v1", "metadata": { "name": "myservice" }, "spec": { "ports":...
###Instruction: I am using fabric8.io to orchestrate application containers in Kubernetes. I am looking to create a Service that manages a pod with a certain label on a certain port. Is there a specific example of the API that does this. I couldnt find it in the examples https://github.com/fabric8io/kubernetes-client/...
I am using fabric8.io to orchestrate application containers in Kubernetes. I am looking to create a Service that manages a pod with a certain label on a certain port. Is there a specific example of the API that does this. I couldnt find it in the examples https://github.com/fabric8io/kubernetes-client/blob/master/kube...
In Kubernetes, each pod has an IP and within a pod there exists a so called infrastructure container, which is the first container that the Kubelet instantiates and it acquires the pod’s IP and sets up the network namespace. All the other containers in the pod then join the infra container’s network and IPC namespace. ...
###Instruction: I have known clearly about the usage of the docker option --net=container:NAME_or_ID, I also have read the source code of kubernetes about how to configure the container to use the network of InfraContainer, so I think the only work the process in container gcr.io/google_containers/pause:0.8.0 does is "...
I have known clearly about the usage of the docker option --net=container:NAME_or_ID, I also have read the source code of kubernetes about how to configure the container to use the network of InfraContainer, so I think the only work the process in container gcr.io/google_containers/pause:0.8.0 does is "pause", it will ...
ClusterIP can mean 2 things: a type of service which is only accessible within a Kubernetes cluster, or the internal ("virtual") IP of components within a Kubernetes cluster. Assuming you're asking about finding the internal IP of a cluster, it can be accessed in 3 ways (using the simple-nginx example): Via command li...
###Instruction: I have created a cluster of three nodes: one master, two minions. How to check the cluster IP in Kubernetes? Is it the IP of the master node? ### Response: ClusterIP can mean 2 things: a type of service which is only accessible within a Kubernetes cluster, or the internal ("virtual") IP of components ...
I have created a cluster of three nodes: one master, two minions. How to check the cluster IP in Kubernetes? Is it the IP of the master node?
Can you downgrade docker version to 1.7.2 first? I did the exact what you did above with docker 1.7.2, and everything works. $ curl 127.0.0.1:8080/ { "paths": [ "/api", "/api/v1", "/api/v1beta3", "/healthz", "/healthz/ping", "/logs/", "/metrics", "/resetMetrics", "/swagger-ui/", ...
###Instruction: I am following the Running Kubernetes locally via Docker guide and I am unable to get the master to start normally. Step One: Run etcd docker run --net=host -d gcr.io/google_containers/etcd:2.0.9 /usr/local/bin/etcd --addr=127.0.0.1:4001 --bind-addr=0.0.0.0:4001 --data-dir=/var/etcd/data The etcd contai...
I am following the Running Kubernetes locally via Docker guide and I am unable to get the master to start normally. Step One: Run etcd docker run --net=host -d gcr.io/google_containers/etcd:2.0.9 /usr/local/bin/etcd --addr=127.0.0.1:4001 --bind-addr=0.0.0.0:4001 --data-dir=/var/etcd/data The etcd container appears to s...
Where are you starting your cluster? I tried to reproduce this on GCE using both the 1.0.7 release and from HEAD and wasn't able to. Using the 1.0.7 release: $ kubectl get pods --namespace=kube-system NAME READY STATUS RESTARTS AGE elasticsearch-logging-v1-6x82b ...
###Instruction: I am trying to setup cluster logging following below link http://kubernetes.io/v1.0/docs/getting-started-guides/logging-elasticsearch.html my config-default.sh # Optional: Enable node logging. ENABLE_NODE_LOGGING=**true** LOGGING_DESTINATION=${LOGGING_DESTINATION:-**elasticsearch**} # Optional: When s...
I am trying to setup cluster logging following below link http://kubernetes.io/v1.0/docs/getting-started-guides/logging-elasticsearch.html my config-default.sh # Optional: Enable node logging. ENABLE_NODE_LOGGING=**true** LOGGING_DESTINATION=${LOGGING_DESTINATION:-**elasticsearch**} # Optional: When set to true, Elas...
As @tim-hockin says, we do plan to have a way to partition the network. But, IMO, for systems with more moving parts, (which is where Kubernetes should really shine), I think it will be better to focus on application security. Taking your three-layer example, the PHP pod should be authorized to talk to the database...
###Instruction: TL;DR Kubernetes allows all containers to access all other containers on the entire cluster, this seems to greatly increase the security risks. How to mitigate? Unlike Docker, where one would usually only allow network connection between containers that need to communicate (via --link), each Pod on Kube...
TL;DR Kubernetes allows all containers to access all other containers on the entire cluster, this seems to greatly increase the security risks. How to mitigate? Unlike Docker, where one would usually only allow network connection between containers that need to communicate (via --link), each Pod on Kubernetes can acces...
The default service account for each namespace is created by the service account controller, which is a loop that is part of the kube-controller-manager binary. So, verify that binary is running, and check its logs for anything that suggests it can't create a service account, make sure you set the "--service-account-p...
###Instruction: I'm trying to get Kubernetes running on some local machines running CoreOS. I'm loosely following this guide. Everything seems to be up and running, and I'm able to connect to the api via kubectl. However, when I try to create a pod, I get this error: Pod "redis-master" is forbidden: Missing service ac...
I'm trying to get Kubernetes running on some local machines running CoreOS. I'm loosely following this guide. Everything seems to be up and running, and I'm able to connect to the api via kubectl. However, when I try to create a pod, I get this error: Pod "redis-master" is forbidden: Missing service account default/de...
Can you plot memory/working_set_bytes_guage instead of memory/usage_bytes_guage? The kube UI might be using working set which correlates with free. memory/usage includes pages that the kernel can reclaim on demand.
###Instruction: Kubernetes UI dashboard shows (this matches the free -m on this minion) Memory: 7.29 GB / 7.84 GB This overall memory usage gradually increasing over time. I am trying to get a view into this memory growth using Kubernetes/Grafana default dashboard for this metric: memory/usage_bytes_gauge. However, I s...
Kubernetes UI dashboard shows (this matches the free -m on this minion) Memory: 7.29 GB / 7.84 GB This overall memory usage gradually increasing over time. I am trying to get a view into this memory growth using Kubernetes/Grafana default dashboard for this metric: memory/usage_bytes_gauge. However, I see the following...
Once you're certain your client properly handles disconnections, you can use the following kube-apiserver flag to control how long apiserver lets the watches stay open: https://github.com/kubernetes/kubernetes/blob/release-1.1/docs/admin/kube-apiserver.md --min-request-timeout=1800: An optional field indicating the min...
###Instruction: After 30-45 minutes, chunked HTTP connection to API server is dropped: Transmission Control Protocol, Src Port: http-alt (8080), Dst Port: 55782 (55782), Seq: 751, Ack: 88, Len: 0 .... 0000 0001 0001 = Flags: 0x011 (FIN, ACK) This happens regardless of the activity level, i.e. it happens for connection...
After 30-45 minutes, chunked HTTP connection to API server is dropped: Transmission Control Protocol, Src Port: http-alt (8080), Dst Port: 55782 (55782), Seq: 751, Ack: 88, Len: 0 .... 0000 0001 0001 = Flags: 0x011 (FIN, ACK) This happens regardless of the activity level, i.e. it happens for connections that were idle...
Part 1: What you said about using kubedns to find heapster and querying that REST API is fine. You could also write a client interface that abstracts the interface to heapster -- that would help with unit testing. Take a look at this metrics client: https://github.com/kubernetes/kubernetes/blob/master/pkg/controller/po...
###Instruction: I would like to expand/shrink the number of kubelets being used by kubernetes cluster based on resource usage. I have been looking at the code and have some idea of how to implement it at a high level. I am stuck on 2 things: What will be a good way for accessing the cluster metrics (via Heapster)? Sho...
I would like to expand/shrink the number of kubelets being used by kubernetes cluster based on resource usage. I have been looking at the code and have some idea of how to implement it at a high level. I am stuck on 2 things: What will be a good way for accessing the cluster metrics (via Heapster)? Should I try to use...
There is no command to do exactly what you asked. Here are some close matches. Be careful before running any of these commands. Make sure you are connected to the right cluster, if you use multiple clusters. Consider running. kubectl config view first. You can delete all the pods in a single namespace with this comma...
###Instruction: Upon looking at the docs, there is an API call to delete a single pod, but is there a way to delete all pods in all namespaces? ### Response: There is no command to do exactly what you asked. Here are some close matches. Be careful before running any of these commands. Make sure you are connected to t...
Upon looking at the docs, there is an API call to delete a single pod, but is there a way to delete all pods in all namespaces?
So it's the kubelet that actually creates the pods and talks to the docker daemon. If you do a docker ps -a on your nodes (as in not master) in your cluster, you'll see the containers in your pod running. So the workflow is run a kubectl command, that goes to the API server, which passes it to the controller, say tha...
###Instruction: I have a problem to understand the kubernetes workflow: So as I understand the flow: You have a master which contains etcd, api-server, controller manager and scheduler. You have nodes which contain pods (wich contain containers), kubelet and a proxy. The proxy is working as a basic proxy to make it pos...
I have a problem to understand the kubernetes workflow: So as I understand the flow: You have a master which contains etcd, api-server, controller manager and scheduler. You have nodes which contain pods (wich contain containers), kubelet and a proxy. The proxy is working as a basic proxy to make it possible for a serv...
I really don't understand your questions, but I can explain how flannel integrates with docker. Flannel is managing this file: # cat /usr/lib/systemd/system/docker.service.d/flannel.conf [Service] EnvironmentFile=-/run/flannel/docker Which is setting the docker service to use the values from /run/flannel/docker as env...
###Instruction: When Installing a Kubernetes Master Node via Docker, docker config bip and mtu for Run flannel. --bip=${FLANNEL_SUBNET} --mtu=${FLANNEL_MTU} What's are FLANNEL_SUBNET and FLANNEL_MTU variables? How to set ${FLANNEL_SUBNET} and ${FLANNEL_MTU}? ### Response: I really don't understand your questions, but ...
When Installing a Kubernetes Master Node via Docker, docker config bip and mtu for Run flannel. --bip=${FLANNEL_SUBNET} --mtu=${FLANNEL_MTU} What's are FLANNEL_SUBNET and FLANNEL_MTU variables? How to set ${FLANNEL_SUBNET} and ${FLANNEL_MTU}?
To have the master node access the cluster network, you can run flanneld and kube-proxy on the master node. This should give you the access you need. However, adding these components in the context of using the kube-up.sh method may be a little involved. Seems like you may have a few options while remaining mostly with...
###Instruction: I am trying to setup a small Kubernetes cluster using a VM (master) and 3 bare metal servers (all running Ubuntu 14.04). I am following the Kubernetes install tutorial for Ubuntu. Everything works fine if I use the 4 nodes (VM + servers) as minions. But when I try to use the VM as just a master, it cann...
I am trying to setup a small Kubernetes cluster using a VM (master) and 3 bare metal servers (all running Ubuntu 14.04). I am following the Kubernetes install tutorial for Ubuntu. Everything works fine if I use the 4 nodes (VM + servers) as minions. But when I try to use the VM as just a master, it cannot access the Fl...
In top command, your are looking at physical memory amount. in stats command, this also include the disk cached ram, so it's always bigger than the physical amount of ram. When you really need more RAM, the disk cached will be released for the application to use. In deed, the memmory usage is pulled via cgroup memory.u...
###Instruction: I'm running my rethinkdb container in Kubernetes cluster. Below is what I notice: Running top in the host which is CoreOS, rethinkdb process takes about 3Gb: $ top PID USER PR NI VIRT RES SHR S %CPU %MEM TIME+ COMMAND 981 root 20 0 53.9m 34.5m 20.9m S 15.6 0.4 115...
I'm running my rethinkdb container in Kubernetes cluster. Below is what I notice: Running top in the host which is CoreOS, rethinkdb process takes about 3Gb: $ top PID USER PR NI VIRT RES SHR S %CPU %MEM TIME+ COMMAND 981 root 20 0 53.9m 34.5m 20.9m S 15.6 0.4 1153:34 hyperkube 5...
It looks like your problem is with making sure the corresponding security group is open to ssh from whichever nodes you'd like to connect from. Make sure it's open to the public IP or the private IP, depending on which you're connecting from. For the right ssh key to use: it'll be whichever one you setup when spinning ...
###Instruction: The quickstart mentions a few times that, "You should be able to ssh into any node in your cluster ..." (e.g., http://kubernetes.io/v1.0/docs/user-guide/connecting-applications.html#environment-variables). I have tried as described below but I am getting timed out. I used export KUBERNETES_PROVIDER=aws...
The quickstart mentions a few times that, "You should be able to ssh into any node in your cluster ..." (e.g., http://kubernetes.io/v1.0/docs/user-guide/connecting-applications.html#environment-variables). I have tried as described below but I am getting timed out. I used export KUBERNETES_PROVIDER=aws; curl -sS https...
The problem was with the networking and kube2sky not accessing the API, so couldn't get the services. Changing the docker run for the master from, --config=/etc/kubernetes/manifests to --config=/etc/kubernetes/manifests-multi Then in the skydns-rc.yaml the for kube2sky as well as setting the domain, set the host IP a...
###Instruction: I am trying to prepare a dev environment for my team, so we can develop, stage and deploy with the same (or near same) environment. Getting a Kubernetes Cluster running locally via http://kubernetes.io/v1.0/docs/getting-started-guides/docker.html was nice and simple. I could then use kubectl to start th...
I am trying to prepare a dev environment for my team, so we can develop, stage and deploy with the same (or near same) environment. Getting a Kubernetes Cluster running locally via http://kubernetes.io/v1.0/docs/getting-started-guides/docker.html was nice and simple. I could then use kubectl to start the pods and servi...
That token has no explicit permissions. If you run with any authorization mode other than AllowAll, you will find that account cannot do anything with the API. If you want to stop injecting API tokens, you can remove the service account admission controller from the list (in apiserver options). If you want to stop ge...
###Instruction: Kubernetes automatically places a token and certificate in /var/run/secrets/kubernetes.io/serviceaccount of each running container in a pod. This token allows access to the the API Server from any container. Is it possible to either prevent this directory from being added to a container or specify a se...
Kubernetes automatically places a token and certificate in /var/run/secrets/kubernetes.io/serviceaccount of each running container in a pod. This token allows access to the the API Server from any container. Is it possible to either prevent this directory from being added to a container or specify a service account th...
Rolling update just scales down one replicationController and scales up another one. Therefore, it deletes the old pods and make new pods, at a controlled rate. So, if the new replication controller json file has different env vars and the same image, then the new pods will have that too. In fact, even if you don't...
###Instruction: Background: We're currently using a continuous delivery pipeline and at the end of the pipeline we deploy the generated Docker image to some server(s) together with the latest application configuration (set as environment variables when starting the Docker container). The continuous delivery build numbe...
Background: We're currently using a continuous delivery pipeline and at the end of the pipeline we deploy the generated Docker image to some server(s) together with the latest application configuration (set as environment variables when starting the Docker container). The continuous delivery build number is used as ver...
With @rwehner's help, I was finally able to figure it out. Checking the kubelet log revealed: Output: mount.nfs: rpc.statd is not running but is required for remote locking. As soon as I got rpcbind running, the mount worked.
###Instruction: I'm trying to mount an external nfs share in a Replication Controller. When I create the replication controller, the pod is pending. Getting the details on the pod, I get these events: Events: FirstSeen LastSeen Count From SubobjectPath Reason Message Thu, ...
I'm trying to mount an external nfs share in a Replication Controller. When I create the replication controller, the pod is pending. Getting the details on the pod, I get these events: Events: FirstSeen LastSeen Count From SubobjectPath Reason Message Thu, 05 Nov 2015 11:2...
Not out of the box (at the time of this writing), but the Kubernetes Ingress api is evolving to support internal loadbalancers. Note the following: Kubernetes Services are round robin loadbalanced by default. You can deploy something like the service loadbalancer [1] and access your services on the ClusterIP of the l...
###Instruction: Currently testing out Kubernetes 1.0.7 on AWS and it creates an external load balancer just fine but I want to know if its possible to create an internal load balancer that is only accessible within the internal subnet. ### Response: Not out of the box (at the time of this writing), but the Kubernete...
Currently testing out Kubernetes 1.0.7 on AWS and it creates an external load balancer just fine but I want to know if its possible to create an internal load balancer that is only accessible within the internal subnet.
The template in the ReplicationControllerSpec defines the template for a Pod that the ReplicationController will create, but since it's an abstract template, it doesn't make sense for it to have a creation time. Once the ReplicationController creates a pod, the CreationTimestamp will be set to the time the Pod was crea...
###Instruction: What does it mean when the creationTimestamp for the template is null? "template": { "metadata": { "creationTimestamp": null, "labels": { "name": "kube-template" } }, ### Response: The template in the ReplicationControllerSpec defin...
What does it mean when the creationTimestamp for the template is null? "template": { "metadata": { "creationTimestamp": null, "labels": { "name": "kube-template" } },
Passing the devices to the container is not currently supported in Kubernetes. This issue is tracked in https://github.com/kubernetes/kubernetes/issues/5607
###Instruction: Or to put it another way, what can I do in kubernetes so that the container is run with the equivalent of --device=/dev/tty10, as an example. Otherwise accessing a device like that gives an error. [root@87eb47e75ed4 /]# echo foo >> /dev/tty10 bash: /dev/tty10: Operation not permitted I haven't found a ...
Or to put it another way, what can I do in kubernetes so that the container is run with the equivalent of --device=/dev/tty10, as an example. Otherwise accessing a device like that gives an error. [root@87eb47e75ed4 /]# echo foo >> /dev/tty10 bash: /dev/tty10: Operation not permitted I haven't found a way currently to...
The DNS name for services is as follow: {service-name}.{namespace}.svc.cluster.local. Assuming you configured kubectl to work with your cluster you should be able to get your service and namespace details by the following the steps below. Get your namespace $ kubectl get namespaces NAME LABELS STATUS defaul...
###Instruction: From the kubernetes docs I see that there is a DNS based service discovery mechanism. Does Google Container Engine support this. If so, what's the format of DNS name to discover a service running inside Container Engine. I couldn't find the relevant information in the Container Engine docs. ### Respons...
From the kubernetes docs I see that there is a DNS based service discovery mechanism. Does Google Container Engine support this. If so, what's the format of DNS name to discover a service running inside Container Engine. I couldn't find the relevant information in the Container Engine docs.
From: https://github.com/kubernetes/kubernetes/pull/12717/files This function func ReadDockerConfigFile() (cfg DockerConfig, err error) is used to parse config which is stored in: GetPreferredDockercfgPath() + "/config.json" workingDirPath + "/config.json" $HOME/.docker/config.json /.docker/config.json GetPreferredDoc...
###Instruction: I am attempting to pull private docker images from Docker Hub. Error: image orgname/imagename:latest not found The info I am seeing on the internet... http://kubernetes.io/v1.0/docs/user-guide/images.html#using-a-private-registry https://github.com/kubernetes/kubernetes/issues/7954 Leads me to beli...
I am attempting to pull private docker images from Docker Hub. Error: image orgname/imagename:latest not found The info I am seeing on the internet... http://kubernetes.io/v1.0/docs/user-guide/images.html#using-a-private-registry https://github.com/kubernetes/kubernetes/issues/7954 Leads me to believe I should be ...
Tl;Dr: Watch this space for Kubernetes 1.2 Till now Kubernetes has only supported L4 loadbalancing. This means the GCE/GKE loadbalancer opens up a tcp connection and just sends traffic to your backend, which is responsible for terminating ssl. As of Kubernetes 1.1, Kubernetes has an "Ingress" resource, but it's current...
###Instruction: Background: I'm pretty new to the Google's Cloud platform so I want to make sure that I'm not is missing anything obvious. We're experimenting with GKE and Kubernetes and we'd like to expose some services over https. I've read the documentation for http(s) load-balancing which seem to suggest that you ...
Background: I'm pretty new to the Google's Cloud platform so I want to make sure that I'm not is missing anything obvious. We're experimenting with GKE and Kubernetes and we'd like to expose some services over https. I've read the documentation for http(s) load-balancing which seem to suggest that you should maintain ...
use the latest kubernetes and pass the params --proxy-mode=iptables to kube-proxy start command, HDFS cluster works now
###Instruction: I deployed kubernetes with flanneld.service enabled in coreos. And then I started hdfs namenode and datanode via kubernetes replication-controller. I also created kubernetes service for namenode. The namenode service ip is 10.100.220.223, while the pod ip of namenode is 10.20.96.4. In my case, one namen...
I deployed kubernetes with flanneld.service enabled in coreos. And then I started hdfs namenode and datanode via kubernetes replication-controller. I also created kubernetes service for namenode. The namenode service ip is 10.100.220.223, while the pod ip of namenode is 10.20.96.4. In my case, one namenode and one data...
Flocker is supported in Kubernetes release 1.1. A Flocker dataset can be referenced from a PersistentVolume or directly from a Pod volume. http://kubernetes.io/v1.1/examples/flocker/ http://kubernetes.io/v1.1/docs/api-reference/v1/definitions.html#_v1_persistentvolume
###Instruction: Kubernetes volume support flocker? If support flocker volume, give an example about using flocker volume? Thanks! ### Response: Flocker is supported in Kubernetes release 1.1. A Flocker dataset can be referenced from a PersistentVolume or directly from a Pod volume. http://kubernetes.io/v1.1/example...
Kubernetes volume support flocker? If support flocker volume, give an example about using flocker volume? Thanks!
"kubectl describe pod busybox" or "kubectl get pod busybox -o yaml" output could be useful. Since you didn't specify, I assume that the busybox pod was created in the default namespace, and that no resource requirements nor nodeSelectors were specified. In many cluster setups, including vagrant, we create a LimitRange...
###Instruction: I have a picture below of my mac. K8S Cluster(on VirtualBox, 1*master, 2*workers) OS Ubuntu 15.04 K8S version 1.1.1 When I try to create a pod "busybox.yaml" it goes to pending status. How can I resolve it? I pasted the online status below for understanding with a picture (kubectl describe node). Sta...
I have a picture below of my mac. K8S Cluster(on VirtualBox, 1*master, 2*workers) OS Ubuntu 15.04 K8S version 1.1.1 When I try to create a pod "busybox.yaml" it goes to pending status. How can I resolve it? I pasted the online status below for understanding with a picture (kubectl describe node). Status kubectl get ...
I noted your use case in the feature request for exposing secrets as environment variables: https://github.com/kubernetes/kubernetes/issues/4710 It's mainly the quoting that makes this tricky in shell. The following worked for me interactively and should work in a script, but additional quoting would be needed if speci...
###Instruction: Update: Kubernetes supports adding secrets directly to environment variables now. See pod example on github Original post: I've been using files created by Kubernetes Secrets to store sensitive configs, but I always end up writing an extra layer into the containers or overriding the CMD to get the cont...
Update: Kubernetes supports adding secrets directly to environment variables now. See pod example on github Original post: I've been using files created by Kubernetes Secrets to store sensitive configs, but I always end up writing an extra layer into the containers or overriding the CMD to get the contents of the secr...
You can do this a few ways, one is keep doing what you're doing in a single file, another is to use labels to specify which environment's config to use, and the other is use namespaces. I personally recommend namespaces, this way you can have separate .yml files for each environment that potentially spins up the same ...
###Instruction: We are moving our ruby microservices to kubernetes and we used to hold environment specific configuration in the config/application.yml. With kubernetes, you can create environment specific files for each service, e.g. config/kubernetes/production.yml etc. While kubernetes pod configuration file is abl...
We are moving our ruby microservices to kubernetes and we used to hold environment specific configuration in the config/application.yml. With kubernetes, you can create environment specific files for each service, e.g. config/kubernetes/production.yml etc. While kubernetes pod configuration file is able to hold enviro...
I finally solved the problem. In the yaml files describing the Replication Controllers, I was using keyring: in the volume section: keyring: "ceph.client.admin.keyring" After I generated a Ceph secret and changed the yaml files to use secretRef: secretRef: name: "ceph-secret" Kubernetes was able to map and mount t...
###Instruction: I am setting up a small Kubernetes cluster using a VM (master) and 3 bare metal servers (all running Ubuntu 14.04). I followed the Kubernetes install tutorial for Ubuntu. Each bare metal server also has 2T of disk space exported using Ceph 0.94.5. Everything is working fine, but when I try to start a Re...
I am setting up a small Kubernetes cluster using a VM (master) and 3 bare metal servers (all running Ubuntu 14.04). I followed the Kubernetes install tutorial for Ubuntu. Each bare metal server also has 2T of disk space exported using Ceph 0.94.5. Everything is working fine, but when I try to start a Replication Contro...
Currently, Google Cloud Platform Networks only support IPv4, so connecting to IPv6 addresses from GKE is not possible.
###Instruction: Background I'd like to connect Wordpress docker container to a Google Could SQL instance. By default Google Cloud SQL only expose an IPv6 address and preferably I'd like to connect Wordpress to this address but I can't find a way to do so (see my other stackoverflow post for details). Question I'd like ...
Background I'd like to connect Wordpress docker container to a Google Could SQL instance. By default Google Cloud SQL only expose an IPv6 address and preferably I'd like to connect Wordpress to this address but I can't find a way to do so (see my other stackoverflow post for details). Question I'd like to know if it's ...
The issue is that your client (kubectl) doesn't support the new ingress resource because it hasn't been updated to 1.1 yet. This is mentioned in the Google Container Engine release notes: The packaged kubectl is version 1.0.7, consequently new Kubernetes 1.1 APIs like autoscaling will not be available via kubectl un...
###Instruction: I'd like to try out the new Ingress resource available in Kubernetes 1.1 in Google Container Engine (GKE). But when I try to create for example the following resource: apiVersion: extensions/v1beta1 kind: Ingress metadata: name: test-ingress spec: backend: serviceName: testsvc servicePort: ...
I'd like to try out the new Ingress resource available in Kubernetes 1.1 in Google Container Engine (GKE). But when I try to create for example the following resource: apiVersion: extensions/v1beta1 kind: Ingress metadata: name: test-ingress spec: backend: serviceName: testsvc servicePort: 80 using: $ ku...
The warning is so that folks don't assume that using emptyDir provides a persistent storage layer. An emptyDir volume will persist as long as the pod is running on the same host. But if the host is replaced or it's disk becomes corrupted, then all data would be lost. Using network mounted storage is one way to work aro...
###Instruction: In the Kubernetes example of Elasticsearch production deployment, there is a warning about using emptyDir, and advises to "be adapted according to your storage needs", which is linked to the documentation of persistent storage on Kubernetes. Is it better to use a persistent storage, which is an external...
In the Kubernetes example of Elasticsearch production deployment, there is a warning about using emptyDir, and advises to "be adapted according to your storage needs", which is linked to the documentation of persistent storage on Kubernetes. Is it better to use a persistent storage, which is an external storage for the...
I started a bash shell inside the container and executed my command instead and that turned out to work: command: ["/bin/bash", "-c", "PID=`pidof java` && kill -SIGTERM $PID && while ps -p $PID > /dev/null; do sleep 1; done;"] Without /bin/bash I couldn't get it working.
###Instruction: I've read the docs on graceful termination of a pod in Kubernetes but I'm not quite sure how to map my specific use case of shutting down a Java process gracefully. What I want to do is to run the following bash command as a part of the termination process: $ kill -SIGTERM `pidof java` I've tried this:...
I've read the docs on graceful termination of a pod in Kubernetes but I'm not quite sure how to map my specific use case of shutting down a Java process gracefully. What I want to do is to run the following bash command as a part of the termination process: $ kill -SIGTERM `pidof java` I've tried this: ... lifecyc...
As of Kubernetes v1.1, you can now implement a Horizontal Pod Autoscaler: http://kubernetes.io/v1.1/docs/user-guide/horizontal-pod-autoscaler.html
###Instruction: I'm developing a Docker-based web service, where each subscriber has private access to their own Docker container running in the cloud, exposing port 443. I've used nginx-proxy/docker-gen successfully to serve multiple Docker containers from the same VM, with just port 443 exposed to the public net. Thi...
I'm developing a Docker-based web service, where each subscriber has private access to their own Docker container running in the cloud, exposing port 443. I've used nginx-proxy/docker-gen successfully to serve multiple Docker containers from the same VM, with just port 443 exposed to the public net. This works fine ......
Firstly, please see https://github.com/kubernetes/kubernetes/issues/11793#issuecomment-150410114 if you haven't already. I guess I can kill and restart pod using a mechanism similar to that used by kubectl delete and create but I am unable to get all the required pod and node lists. I would suggest writing a control ...
###Instruction: I want to implement a rescheduler like functionality which basically kills pods if it decides that the pods could be rescheduled in a better way (based on requiring less number of nodes/fitting etc). Till now I have created a new kubectl command which I want to run whenever I want to reschedule. I have ...
I want to implement a rescheduler like functionality which basically kills pods if it decides that the pods could be rescheduled in a better way (based on requiring less number of nodes/fitting etc). Till now I have created a new kubectl command which I want to run whenever I want to reschedule. I have also looked at t...
It appears that you can't currently set a ulimit but it is an open issue: https://github.com/kubernetes/kubernetes/issues/3595
###Instruction: How do I set ulimit for containers in Kubernetes? (specifically ulimit -u) ### Response: It appears that you can't currently set a ulimit but it is an open issue: https://github.com/kubernetes/kubernetes/issues/3595
How do I set ulimit for containers in Kubernetes? (specifically ulimit -u)
I found the Solution, the reason is the version of kube-apiserver,kube-controller-manager and kube-scheduler does not match with the kubelet. the detail: https://github.com/kubernetes/kubernetes/issues/17154
###Instruction: I create a kubernetes cluster to test. but cannot create rc. I got errorreason: 'failedScheduling' no nodes available to schedule pods: I1112 04:24:34.626614 6 factory.go:214] About to try and schedule pod my-nginx-63t4p I1112 04:24:34.626635 6 scheduler.go:127] Failed to schedule: &{{ } {my...
I create a kubernetes cluster to test. but cannot create rc. I got errorreason: 'failedScheduling' no nodes available to schedule pods: I1112 04:24:34.626614 6 factory.go:214] About to try and schedule pod my-nginx-63t4p I1112 04:24:34.626635 6 scheduler.go:127] Failed to schedule: &{{ } {my-nginx-63t4p my-...
Tim, I did run it again using your steps, no difference, didn't work, however today I switched to the version 1.1 docs here: http://kubernetes.io/v1.1/docs/getting-started-guides/docker.html and also switched container versions, currently using: gcr.io/google_containers/etcd:2.2.1 gcr.io/google_containers/hyperkube:v1....
###Instruction: Followed this guide to starting a local-machine kubernetes cluster: http://kubernetes.io/v1.0/docs/getting-started-guides/docker.html I've created various pods with .yaml files and everything works, I can access nginx and mysql using container IPs (in the 172.17.x.x range, with docker0), however when I ...
Followed this guide to starting a local-machine kubernetes cluster: http://kubernetes.io/v1.0/docs/getting-started-guides/docker.html I've created various pods with .yaml files and everything works, I can access nginx and mysql using container IPs (in the 172.17.x.x range, with docker0), however when I create services,...
Try passing this to the shell: "command": [ "/bin/bash", "-c", "export PGPASSWORD=password; psql -h myhost -U root -d AppPostgresDB < /db-backup/backup.sql" ]
###Instruction: I'm trying to create a pod with Postgres. After initialize, the Pod has to execute the following command: "lifecycle": { "postStart": { "exec": { "command": [ "export", "PGPASSWORD=password;", "psql", "-h", "myhost", "-U", "root", "-d", "AppPostgresDB", "<", "/db-...
I'm trying to create a pod with Postgres. After initialize, the Pod has to execute the following command: "lifecycle": { "postStart": { "exec": { "command": [ "export", "PGPASSWORD=password;", "psql", "-h", "myhost", "-U", "root", "-d", "AppPostgresDB", "<", "/db-backup/backup.sq...
Kubernetes uses the IP-per-pod model. If I understand correctly, you want to create three mongo pods, and write IP addresses of the three pods in /etc/hosts of each container. Modifying the /etc/host files directly might not be a good idea for many reasons (e.g., the pod may die and be replaced). For peer discovery in ...
###Instruction: I hope everyone here is doing good. I am trying to find a way to add entries to the containers /etc/hosts file while spinning up a pod. I was just wondering to know if there is any option/parameter that I could mention in my "pod1.json" which adds the entries to the containers /etc/hosts when its being ...
I hope everyone here is doing good. I am trying to find a way to add entries to the containers /etc/hosts file while spinning up a pod. I was just wondering to know if there is any option/parameter that I could mention in my "pod1.json" which adds the entries to the containers /etc/hosts when its being created. Somethi...
Very interesting question! For that to answer I believe we need to walk the Star Wars walk and Use The Source: The ReplicationManager has some hints concerning expectations Then, there is controller_utils.go with some more indications However, the core of the calculation seems to be in updateReplicaCount UPDATE: My ...
###Instruction: When watching a replication controller, it returns it’s most recent replicas count under ReplicationControllerStatus. I could not find anywhere in the documentation what the status of the pod needs to be, in order for it to be included there. Is it enough for the pod to be scheduled? I’ve noticed a repl...
When watching a replication controller, it returns it’s most recent replicas count under ReplicationControllerStatus. I could not find anywhere in the documentation what the status of the pod needs to be, in order for it to be included there. Is it enough for the pod to be scheduled? I’ve noticed a replication controll...
If you name the ports in the pods, you can specify the target ports by name in the service rather than by number, and then the same service can direct target to pods using different port numbers. Or, as Yu-Ju suggested, you can do a read-modify-write of the live state of the service, such as via kubectl edit. The error...
###Instruction: Background: Let's say I have a replication controller with some pods. When these pods were first deployed they were configured to expose port 8080. A service (of type LoadBalancer) was also create to expose port 8080 publicly. Later we decide that we want to export an additional port from the pods (port...
Background: Let's say I have a replication controller with some pods. When these pods were first deployed they were configured to expose port 8080. A service (of type LoadBalancer) was also create to expose port 8080 publicly. Later we decide that we want to export an additional port from the pods (port 8081). We chang...
In descending order, from most mature and battle-tested at scale to less so: Marathon, a Apache Mesos framework Kubernetes Docker Swarm HashiCorp's Nomad
###Instruction: I've been tasked with evaluating container management solutions. I'm aware there is a large number or options, but we need production ready, on premises solution. What are the options? ### Response: In descending order, from most mature and battle-tested at scale to less so: Marathon, a Apache Mesos f...
I've been tasked with evaluating container management solutions. I'm aware there is a large number or options, but we need production ready, on premises solution. What are the options?
This is a bug and is fixed by https://github.com/kubernetes/kubernetes/pull/16432
###Instruction: On my Kubernetes cluster on GKE, I have the following persistent volume claims (PVCs): kind: PersistentVolumeClaim apiVersion: v1 metadata: name: registry spec: accessModes: - ReadWriteOnce resources: requests: storage: 100Gi and: kind: PersistentVolumeClaim apiVersion: v1 metadata:...
On my Kubernetes cluster on GKE, I have the following persistent volume claims (PVCs): kind: PersistentVolumeClaim apiVersion: v1 metadata: name: registry spec: accessModes: - ReadWriteOnce resources: requests: storage: 100Gi and: kind: PersistentVolumeClaim apiVersion: v1 metadata: name: postgre...
By watching changes to a list of Endpoints: GET /api/v1/watch/namespaces/{namespace}/endpoints
###Instruction: Is there a way to discover all the endpoints of a headless service from outside the cluster? Preferably using DNS or Static IPs ### Response: By watching changes to a list of Endpoints: GET /api/v1/watch/namespaces/{namespace}/endpoints
Is there a way to discover all the endpoints of a headless service from outside the cluster? Preferably using DNS or Static IPs
The issue turned out to be an MTU issue between the node and the master. Once that was fixed the problem was resolved.
###Instruction: I've been working with a 6 node cluster for the last few weeks without issue. Earlier today we ran into an open file issue (https://github.com/kubernetes/kubernetes/pull/12443/files) and I patched and restarted kube-proxy. Since then, all rc deployed pods to ALL BUT node-01 get stuck in pending state a...
I've been working with a 6 node cluster for the last few weeks without issue. Earlier today we ran into an open file issue (https://github.com/kubernetes/kubernetes/pull/12443/files) and I patched and restarted kube-proxy. Since then, all rc deployed pods to ALL BUT node-01 get stuck in pending state and there log mes...
Starting with the available documentation: in your case on logging and volumes. Taking these two sources together we arrive at something like the following: ... containers: - name: syslogtest image: ubuntu:14.04 volumeMounts: - name: logvol mountPath: /dev/log readOnly: false volumes: ...
###Instruction: With docker, I can pass log-driver=syslog command line option to forward container logs to syslog. How do I pass these docker arguments via Kubernetes yaml/json descriptor? ### Response: Starting with the available documentation: in your case on logging and volumes. Taking these two sources together we...
With docker, I can pass log-driver=syslog command line option to forward container logs to syslog. How do I pass these docker arguments via Kubernetes yaml/json descriptor?
You'll get "no volume plugins matched" if the rbd command isn't installed and in the path. As the example specifies, you need to ensure that ceph is installed on your Kubernetes nodes. For instance, in Fedora: $ sudo yum -y install ceph-common I'll file an issue to clarify the error messages.
###Instruction: I follow the example to use rbd in kubernetes, but can not success. who can help me!! the error : Nov 09 17:58:03 core-1-97 kubelet[1254]: E1109 17:58:03.289702 1254 volumes.go:114] Could not create volume builder for pod 5df3610e-86c8-11e5-bc34-002590fdf95c: can't use volume plugins for (volume.Spec...
I follow the example to use rbd in kubernetes, but can not success. who can help me!! the error : Nov 09 17:58:03 core-1-97 kubelet[1254]: E1109 17:58:03.289702 1254 volumes.go:114] Could not create volume builder for pod 5df3610e-86c8-11e5-bc34-002590fdf95c: can't use volume plugins for (volume.Spec){Name:(string)r...
There is a known (benevolent) bug in the ingress controller that is automatically added to your cluster in Kubernetes 1.1.1. If you are not using the controller, you can scale the number of replicas to zero: kubectl scale rc l7-lb-controller --namespace=kube-system --replicas=0 which should make your CPU usage go back...
###Instruction: I'm following the container engine walkthrough and I see that my VM Instance has a min CPU usage of ~80%. However, if I ssh into this box and run 'top' I see a much lower utilization. Can someone explain this to me as I must be missing something simple. Thank you. ### Response: There is a known (benev...
I'm following the container engine walkthrough and I see that my VM Instance has a min CPU usage of ~80%. However, if I ssh into this box and run 'top' I see a much lower utilization. Can someone explain this to me as I must be missing something simple. Thank you.
Not so hard, actually. Check my gists may give you some tips: https://gist.github.com/resouer/378bcdaef1d9601ed6aa See, do not try to consume files from outside, just package them in a docker image, and consume them by sidecar mode.
###Instruction: I have a noob question. If I'm using a docker image that uses a folder located in the host to do something, Where should be located the folder in the kubernetes cluster? I'm ok doing this with docker since I know where is my host filesystem but I get lost when I'm on a kubernetes cluster. Actually, I do...
I have a noob question. If I'm using a docker image that uses a folder located in the host to do something, Where should be located the folder in the kubernetes cluster? I'm ok doing this with docker since I know where is my host filesystem but I get lost when I'm on a kubernetes cluster. Actually, I don't know if this...
Starting last week, new GKE clusters should be on created on 1.1.1 by default, and the nfs-common package is installed on all 1.1.1 clusters. (For existing clusters, you'll need to wait until the hosted master is upgraded, then initiate a node upgrade.) See https://github.com/kubernetes/kubernetes/blob/release-1.1/exam...
###Instruction: After I manually install nfs client package under each node, then it works. But in GKE, slave node can be scale in and out. After create a new slave node, I lose nfs client package again. Is there any way we can install software package when kubernetes spin up a new slave node? ### Response: Starting l...
After I manually install nfs client package under each node, then it works. But in GKE, slave node can be scale in and out. After create a new slave node, I lose nfs client package again. Is there any way we can install software package when kubernetes spin up a new slave node?
It is not recommended to use external container garbage collection scripts. Kubernetes relies on exited containers as tombstones to reconstruct the pod status and/or serve logs. Even if you don't care about container logs, if you remove the exited containers before kubernetes examines them and properly records the stat...
###Instruction: I am now doing experiments on using Kubernetes and Docker to provisioning services such as WordPress, Spark and Storm on 10 physical machines But after lots times of launching and terminating Docker Containers, the used memory would increase even I kill all the Containers via Kubernetes delete or Docker...
I am now doing experiments on using Kubernetes and Docker to provisioning services such as WordPress, Spark and Storm on 10 physical machines But after lots times of launching and terminating Docker Containers, the used memory would increase even I kill all the Containers via Kubernetes delete or Docker kill commands. ...
You probably need to remove the manifest files for influxdb from the /etc/kubernetes/addons/ directory on your "master" host. Many of the kube-up.sh implementations use a service (usually at /etc/kubernetes/kube-master-addons.sh) that runs periodically and makes sure that all the manifests in /etc/kubernetes/addons/ ar...
###Instruction: I have spun up a Kubernetes cluster in AWS using the official "kube-up" mechanism. By default, an addon that monitors the cluster and logs to InfluxDB is created. It has been noted in this post that InfluxDB quickly fills up disk space on nodes, and I am seeing this same issue. The problem is, when I ...
I have spun up a Kubernetes cluster in AWS using the official "kube-up" mechanism. By default, an addon that monitors the cluster and logs to InfluxDB is created. It has been noted in this post that InfluxDB quickly fills up disk space on nodes, and I am seeing this same issue. The problem is, when I try to kill the ...
Please also see https://github.com/kubernetes/kubernetes/issues/16741 where we're discussing nfs and pretty much exactly this problem (amongst others)
###Instruction: After I manually install nfs client package under each node, then it works. But in GKE, slave node can be scale in and out. After create a new slave node, I lose nfs client package again. Is there any way we can install software package when kubernetes spin up a new slave node? ### Response: Please als...
After I manually install nfs client package under each node, then it works. But in GKE, slave node can be scale in and out. After create a new slave node, I lose nfs client package again. Is there any way we can install software package when kubernetes spin up a new slave node?
Unfortunately those URLs are incomplete. Influx's ports are named, so you need to say which port you want. https://MASTER_IP/api/v1/proxy/namespaces/kube-system/services/monitoring-influxdb:http or https://MASTER_IP/api/v1/proxy/namespaces/kube-system/services/monitoring-influxdb:api There's a bug open to give better ...
###Instruction: I have a new Kubernetes cluster on AWS that was built using the kube-up script from v1.1.1. I can successfully access the Elasticsearch/Kibana/KubeUI/Grafana endpoints, but cannot access Heapster/KubeDNS/InfluxDB from my machine, through the API proxy. I have seen some ancillary issues related to this...
I have a new Kubernetes cluster on AWS that was built using the kube-up script from v1.1.1. I can successfully access the Elasticsearch/Kibana/KubeUI/Grafana endpoints, but cannot access Heapster/KubeDNS/InfluxDB from my machine, through the API proxy. I have seen some ancillary issues related to this on the K8S proj...
Then this will not work ( ReferenceError: process is not defined ) , as these angularjs code is executed at the client side. If the client is outside the cluster, the only way it will be able to access the NodeJS API is if you expose it to the client's network, which is probably the public internet. If you're concern...
###Instruction: Using Kubernetes -- Gogle Container Enginer setup , Within the Same google cloud Cluster, I am having the Front end Service -> nginx + Angular JS and REST API service --> NodeJS API. I don't want to expose NodeJS API KubeCTL Service public domain. So, 'ServiceType' is set to only 'ClusterIP' . How do we...
Using Kubernetes -- Gogle Container Enginer setup , Within the Same google cloud Cluster, I am having the Front end Service -> nginx + Angular JS and REST API service --> NodeJS API. I don't want to expose NodeJS API KubeCTL Service public domain. So, 'ServiceType' is set to only 'ClusterIP' . How do we infer this NODE...
An ingress controller can completely bypass kube-proxy. The haproxy controller for example, does this and goes straight to endpoints. However it doesn't use the Ingress in the typical sense. You could do the same with the nginx controller, all you need to lookup endpoints and insert them instead of the DNS name it cur...
###Instruction: I'd like to implement a sticky-session Ingress controller. Cookies or IP hashing would both be fine; I'm happy as long as the same client is generally routed to the same pod. What I'm stuck on: it seems like the Kubernetes service model means my connections are going to be proxied randomly no matter wha...
I'd like to implement a sticky-session Ingress controller. Cookies or IP hashing would both be fine; I'm happy as long as the same client is generally routed to the same pod. What I'm stuck on: it seems like the Kubernetes service model means my connections are going to be proxied randomly no matter what. I can configu...
You need to increase the quota assigned for your project. Please see https://cloud.google.com/compute/docs/resource-quotas for the explanation of resource quotas, and follow the link on that page to check and/or request a quota increase.
###Instruction: I'm using Kubernetes and I'm trying to create an ingress resource. I create it using: $ kubectl create -f my-ingress.yaml I wait a while and a load balancer doesn't seem to be created. Running: $ kubectl describe ing my-ingress returns: Events: FirstSeen LastSeen Count From Subob...
I'm using Kubernetes and I'm trying to create an ingress resource. I create it using: $ kubectl create -f my-ingress.yaml I wait a while and a load balancer doesn't seem to be created. Running: $ kubectl describe ing my-ingress returns: Events: FirstSeen LastSeen Count From SubobjectPath Reaso...
With Windows, you need docker-machine and boot2docker VMs to run anything docker related. There is no (not yet) "docker for Windows". Note that issue 7428 mentioned "Can't run kubernetes within boot2docker". So even when you follow instructions (from a default VM created with docker-machine), you might still get errors...
###Instruction: I'm very new to kubernetes and trying to conceptualize it as well as set it up locally in order to try developing something on it. There's a confound though that I am running on a windows machine. Their "getting started" documentation in github says you have to run Linux to use kubernetes. As docker run...
I'm very new to kubernetes and trying to conceptualize it as well as set it up locally in order to try developing something on it. There's a confound though that I am running on a windows machine. Their "getting started" documentation in github says you have to run Linux to use kubernetes. As docker runs on windows, I ...
Yes, it is possible. The Replication Controller template or Pod spec image isn't special. If you specify image: redis you will get the latest tag of the official Docker Hub library Redis image, just as if you did docker pull redis.
###Instruction: Still new to Containers and Kubernetes here but I am dabbling with deploying a cluster on Google Containers Engine and was wondering if you can use a docker hub hosted image to deploy containers, so in my .yaml configuration file I'd say: ... image: hub.docker.com/r/my-team/my-image:latest ... Is...
Still new to Containers and Kubernetes here but I am dabbling with deploying a cluster on Google Containers Engine and was wondering if you can use a docker hub hosted image to deploy containers, so in my .yaml configuration file I'd say: ... image: hub.docker.com/r/my-team/my-image:latest ... Is this possible? ...
So you basically you need to expose the mysql port to your host: docker run --name mysql -p 3306:3306 -e MYSQL_ROOT_PASSWORD=password -d mysql/mysql-server:latest Then you can access from your host using the mysql command line: mysql -h127.0.0.1 -ppassword -uroot Not sure why you are trying to run another container t...
###Instruction: I have already googled on this subject and found few threads. Based on these threads I have followed the following steps. But I am facing a problem. Basically, I want to create a docker image for mysql and then connect to it from my host machine (Mac OS X). Based on this post , I have to share the mysql...
I have already googled on this subject and found few threads. Based on these threads I have followed the following steps. But I am facing a problem. Basically, I want to create a docker image for mysql and then connect to it from my host machine (Mac OS X). Based on this post , I have to share the mysql unix socket wit...
This is missing functionality in the ABAC authorizer. The fix is in progress: #16148. As for a workaround, from the authorization doc: For miscellaneous endpoints, like /version, the resource is the empty string. So you may be able to solve by defining a policy: {"user":"bob", "readonly": true, "resource": ""} ...
###Instruction: I sent up a 4 node cluster (1 master 3 workers) running Kubernetes on Ubuntu. I turned on --authorization-mode=ABAC and set up a policy file with an entry like the following {"user":"bob", "readonly": true, "namespace": "projectgino"} I want user bob to only be able to look at resources in projectgino...
I sent up a 4 node cluster (1 master 3 workers) running Kubernetes on Ubuntu. I turned on --authorization-mode=ABAC and set up a policy file with an entry like the following {"user":"bob", "readonly": true, "namespace": "projectgino"} I want user bob to only be able to look at resources in projectgino. I'm having pro...
There is currently no supported way to achieve what you're trying to do. But there is a Kubernetes project under way to address it, which goes under the name of Kubernetes Cluster Federation, alternatively known as "Ubernetes". Further details are available here: http://www.slideshare.net/quintonh/federation-of-kuber...
###Instruction: Is it in any way possible to configure a Kubernetes Cluster that utilizes ressources from multiple IaaS providers at the same time e.g. a cluster running partially on GCE and AWS? Or a Kubernetes Cluster running on your bare metal and an IaaS provider? Maybe in combination with some other tools like Mes...
Is it in any way possible to configure a Kubernetes Cluster that utilizes ressources from multiple IaaS providers at the same time e.g. a cluster running partially on GCE and AWS? Or a Kubernetes Cluster running on your bare metal and an IaaS provider? Maybe in combination with some other tools like Mesos? Are there an...
Actually, once you delete the pod, it will be in "terminating" state until it is destroyed (after terminationGracePeriodSeconds) which means it is removed from the service load balancer, but still capable of serving existing requests. We also use "readiness" health checks, and preStop is synchronous, so you could make ...
###Instruction: Standard practice for a rolling update of hosts behind load balancer is to gracefully take the hosts out of rotation. This can be done by marking the host "un-healthy" and ensuring the host is no longer receiving requests from the load balancer. Does Kubernetes do something similar for pods managed by ...
Standard practice for a rolling update of hosts behind load balancer is to gracefully take the hosts out of rotation. This can be done by marking the host "un-healthy" and ensuring the host is no longer receiving requests from the load balancer. Does Kubernetes do something similar for pods managed by a ReplicationCon...
You would need to create the SSL cert like Yu-Ju Hong said, then you would have to tell ruby to use the certificate when connecting something like http://makandracards.com/makandra/1701-use-ssl-for-amazon-rds-mysql-and-your-rails-app The bit about: sslca: /path/to/mysql-ssl-ca-cert.pem
###Instruction: I am using Kubernetes to deploy a Rails application to Google Container Engine. The database is using Google Cloud SQL. I know the database's ip address and set it into my Kubernetes config file: # web-controller.yml apiVersion: v1 kind: ReplicationController metadata: labels: name: web name: we...
I am using Kubernetes to deploy a Rails application to Google Container Engine. The database is using Google Cloud SQL. I know the database's ip address and set it into my Kubernetes config file: # web-controller.yml apiVersion: v1 kind: ReplicationController metadata: labels: name: web name: web-controller spe...
Looking at your question, I believe by account you mean project. The command for pulling an image from the registry is: $ gcloud docker pull gcr.io/your-project-id/example-image This means as long as your account is a member of the project which the image belongs to, you can pull the image from that project to any ot...
###Instruction: We'd like to have a separate test and prod project on the Google Cloud Platform but we want to reuse the same docker images in both environments. Is it possible for the Kubernetes cluster running on the test project to use images pushed to the prod project? If so, how? ### Response: Looking at your que...
We'd like to have a separate test and prod project on the Google Cloud Platform but we want to reuse the same docker images in both environments. Is it possible for the Kubernetes cluster running on the test project to use images pushed to the prod project? If so, how?
TL;DR Google Container Engine running Kubernetes v1.1 supports loadBalancerIP just mark the auto-assigned IP as static first. Kubernetes v1.1 supports externalIPs: apiVersion: v1 kind: Service spec: type: LoadBalancer loadBalancerIP: 10.10.10.10 ... So far there isn't a really good consistent documentation on ho...
###Instruction: I have a Kubernetes cluster running on Google Compute Engine and I would like to assign static IP addresses to my external services (type: LoadBalancer). I am unsure about whether this is possible at the moment or not. I found the following sources on that topic: Kubernetes Service Documentation lets y...
I have a Kubernetes cluster running on Google Compute Engine and I would like to assign static IP addresses to my external services (type: LoadBalancer). I am unsure about whether this is possible at the moment or not. I found the following sources on that topic: Kubernetes Service Documentation lets you define an ext...
PR #12112 changed kubectl get pods to not show terminated pods by default. You can get the old behavior (show all pods) by using kubectl get pods -a
###Instruction: All running computation Monte Carlo jobs on Google Compute Engine. Last time I ran them was September, and things have changed a bit since then. I used to run a lot of jobs with kubectl from some pod.json file, no RC, no restart, fire-and-forget setup. After I started jobs I used to get pods (kubectl ge...
All running computation Monte Carlo jobs on Google Compute Engine. Last time I ran them was September, and things have changed a bit since then. I used to run a lot of jobs with kubectl from some pod.json file, no RC, no restart, fire-and-forget setup. After I started jobs I used to get pods (kubectl get pods) and typi...
You'll need to tell Jenkins how and where to run your builds by adding your Kubernetes cluster as a 'cloud' in the Jenkins configuration. Go to Manage Jenkins -> Configure System -> Cloud -> Add new cloud and select 'Kubernetes'. You'll find the server certificate key, user name and password in your local kubectl confi...
###Instruction: I've been struggling with setting up the Jenkins Kubernetes Plugin on the Google Container Engine. I have the plugin installed but I think all my builds are still running on master. I haven't found any good documentation or guides on configuring this. UPDATE I removed the master executor from my Jenkins...
I've been struggling with setting up the Jenkins Kubernetes Plugin on the Google Container Engine. I have the plugin installed but I think all my builds are still running on master. I haven't found any good documentation or guides on configuring this. UPDATE I removed the master executor from my Jenkins image. So now m...
In a standard Kubernetes installation, all pods (even across namespaces) share a flat IP space and can all communicate with each other. To get isolation, you'll need to customize your install to prevent cross namespace communication. One way to do this is to use OpenContrail. They recently wrote a blog post describing...
###Instruction: We need to know about pods network isolation. Is there a possibility to access one pod from another one in cluster? Maybe by namespace dividing? We also need pod's membership in local networks, which are not accessible from outside. Any plans? Is it will be soon? ### Response: In a standard Kuber...
We need to know about pods network isolation. Is there a possibility to access one pod from another one in cluster? Maybe by namespace dividing? We also need pod's membership in local networks, which are not accessible from outside. Any plans? Is it will be soon?
Yes, it is new behavior in v1.1.1. PR #9165 added graceful deletion of pods, which causes them to appear in the "Terminating" state for a short amount of time. Issue #1535 has some more background discussion.
###Instruction: A new GKE cluster created at v1.1.1 and using latest kubectl (from gcloud components update) when deleting resources (say a pod) sometimes kubectl get pods in a 'Terminating' state and other times they delete (are removed from kubectl get pods output) right away. NAME READY S...
A new GKE cluster created at v1.1.1 and using latest kubectl (from gcloud components update) when deleting resources (say a pod) sometimes kubectl get pods in a 'Terminating' state and other times they delete (are removed from kubectl get pods output) right away. NAME READY STATUS RES...
There are definitely folks running Kubernetes HA masters in production following the instructions for High Availability Kubernetes Clusters. As noted at the beginning of that page, it's an advanced use case and requires in-depth knowledge of how the Kubernetes master components work.
###Instruction: Hi all we are looking for practically and tested guide or reference for kubernetes master high availability or other solution for master node fail over. ### Response: There are definitely folks running Kubernetes HA masters in production following the instructions for High Availability Kubernetes Cl...
Hi all we are looking for practically and tested guide or reference for kubernetes master high availability or other solution for master node fail over.
You do not need to install Powerstrip anymore. (it's been deprecated) Powerstrip was a useful tool early on to prototype docker extensions but we've moved on since Docker has added the docker api via the plugins model. (Powerstrip was essentially a Precurser to docker plugins) docker --volume-driver=flocker docker volu...
###Instruction: I am using flocker volumes. Should I install Powerstrip? I have installed flocker, but not installed Powerstrip. I create flocker pod fail: Unable to mount volumes for pod "flocker-web-3gy69_default": Get https://localhost:4523/v1/configuration/datasets: x509: certificate is valid for control-...
I am using flocker volumes. Should I install Powerstrip? I have installed flocker, but not installed Powerstrip. I create flocker pod fail: Unable to mount volumes for pod "flocker-web-3gy69_default": Get https://localhost:4523/v1/configuration/datasets: x509: certificate is valid for control-service, hostnam...
command: ["/bin/sh","-c"] args: ["command one; command two && command three"] Explanation: The command ["/bin/sh", "-c"] says "run a shell, and execute the following instructions". The args are then passed as commands to the shell. In shell scripting a semicolon separates commands, and && conditionally runs the follow...
###Instruction: In this official document, it can run command in a yaml config file: https://kubernetes.io/docs/tasks/configure-pod-container/ apiVersion: v1 kind: Pod metadata: name: hello-world spec: # specification of the pod’s contents restartPolicy: Never containers: - name: hello image: "ubuntu:14....
In this official document, it can run command in a yaml config file: https://kubernetes.io/docs/tasks/configure-pod-container/ apiVersion: v1 kind: Pod metadata: name: hello-world spec: # specification of the pod’s contents restartPolicy: Never containers: - name: hello image: "ubuntu:14.04" env: ...
POSTing this to the API server will have no effect if you have not configured an Ingress controller. You need to choose the ingress controller implementation that is the best fit for your cluster, or implement one. Examples and instructions can be found here.
###Instruction: I have a setup of kubernetes on a coreos baremetal. For now I did the connection from outside world to service with a nginx reverse-proxy. I'm trying the new Ingress resource. for now I have added a simple ingress: apiVersion: extensions/v1beta1 kind: Ingress metadata: name: kube-ui spec: backend: ...
I have a setup of kubernetes on a coreos baremetal. For now I did the connection from outside world to service with a nginx reverse-proxy. I'm trying the new Ingress resource. for now I have added a simple ingress: apiVersion: extensions/v1beta1 kind: Ingress metadata: name: kube-ui spec: backend: serviceName: ...
Routes are http, HTTPS, or TCP wrapped with TLS. You can use a service with a "node port", which load balances your app instances over TCP or udp at a high port exposed on each node. Routes point to services to get their source data, but since routes expect to be able to identify which backend service to route traff...
###Instruction: I am going through the Openshift V3 documentation and got confused by services and routes details. The description in service says that: Each node also runs a simple network proxy that reflects the services defined in the API on that node. This allows the node to do simple TCP and UDP stream forwarding...
I am going through the Openshift V3 documentation and got confused by services and routes details. The description in service says that: Each node also runs a simple network proxy that reflects the services defined in the API on that node. This allows the node to do simple TCP and UDP stream forwarding across a set of...
You cannot add or remove containers in a running pod. If you are using replication controller, kubectl rolling-update is the easiest solution, but this will require editing the pod template. That said, are you sure you need to add your containers to the existing pod? Unless strictly necessary, it's better to just run t...
###Instruction: I can add a container to a pod by editing the pod template, but I'm looking for something simpler. Is there any way to add a container to a deployed OpenShift pod without editing the pod template? CLI preferable. ### Response: You cannot add or remove containers in a running pod. If you are using repli...
I can add a container to a pod by editing the pod template, but I'm looking for something simpler. Is there any way to add a container to a deployed OpenShift pod without editing the pod template? CLI preferable.
Check out the Downward API, which allows the container to know more about itself.
###Instruction: When we create a yml for the replication controller, we can give labels for the pod that is being created. apiVersion: v1 kind: ReplicationController metadata: name: redis spec: template: metadata: labels: app: redis tier: backend Can the containers that reside in this pod...
When we create a yml for the replication controller, we can give labels for the pod that is being created. apiVersion: v1 kind: ReplicationController metadata: name: redis spec: template: metadata: labels: app: redis tier: backend Can the containers that reside in this pod access those la...
Sorry about this question, I had find it. http://kubernetes.io/v1.1/docs/api-reference/v1/definitions.html
###Instruction: I feel very confused when use the kubernetes! Where can I find about the specify api about the components of kubernetes ? such as pod, service, volumes, and Persistent Volumes and so on, when I create the components use the configure files. Who can help me? ### Response: Sorry about this question, I h...
I feel very confused when use the kubernetes! Where can I find about the specify api about the components of kubernetes ? such as pod, service, volumes, and Persistent Volumes and so on, when I create the components use the configure files. Who can help me?
If you used this command to create your certificate: openssl x509 -req -days 365 -in server.csr -CA ca.pem -CAkey ca-key.pem \ -CAcreateserial -out server-cert.pem Then your issue can be resolved by doing the following as the 'client' cert uses an -extfile extfile.cnf: echo subjectAltName = IP:worker_node_ip > ext...
###Instruction: I'm trying to use kubectl exec to enter one of my containers, but I'm getting stuck on this error. $ kubectl exec -it ubuntu -- bash error: Unable to upgrade connection: { "kind": "Status", "apiVersion": "v1", "metadata": {}, "status": "Failure", "message": "x509: cannot validate cer...
I'm trying to use kubectl exec to enter one of my containers, but I'm getting stuck on this error. $ kubectl exec -it ubuntu -- bash error: Unable to upgrade connection: { "kind": "Status", "apiVersion": "v1", "metadata": {}, "status": "Failure", "message": "x509: cannot validate certificate for <wo...
This is one of the large problems that Kubernetes is trying to solve with Cross-Cluster Service Discovery as a part of the Cluster Federation plans. You can also check out/contribute to the Federation SIG. If you've used one of the hacks solutions described here, you might be able to hack up your /etc/resolve.conf to a...
###Instruction: Relates to How to call a service exposed by a Kubernetes cluster from another Kubernetes cluster in same project. Asking again since Kubernetes has been changes a lot since July. Context: I'm working on an infrastructure with multiple clusters serving different purposes, e.g.: Cluster A runs servic...
Relates to How to call a service exposed by a Kubernetes cluster from another Kubernetes cluster in same project. Asking again since Kubernetes has been changes a lot since July. Context: I'm working on an infrastructure with multiple clusters serving different purposes, e.g.: Cluster A runs services/apps creating...
There is no command today that makes it easy to add a container to the pod template for an RC or deployment. You can use oc new-app to quickly generate deployment configs that have multiple containers with oc new-app php+apache+somethingelse But this won't let you deeply customize those containers. Agree this would b...
###Instruction: I can add a container to a pod by editing the pod template, but I'm looking for something simpler. Is there any way to add a container to a deployed OpenShift pod without editing the pod template? CLI preferable. ### Response: There is no command today that makes it easy to add a container to the pod t...
I can add a container to a pod by editing the pod template, but I'm looking for something simpler. Is there any way to add a container to a deployed OpenShift pod without editing the pod template? CLI preferable.
You can't clone only one file. gitRepo executes git clone which only allows you to clone the entire repository. volumeMounts doesn't support executing command in it.
###Instruction: I know how to mount git repo when I start pod. See: apiVersion: v1 kind: Pod metadata: name: server spec: containers: - image: nginx name: nginx volumeMounts: - mountPath: /mypath name: git-volume volumes: - name: git-volume gitRepo: repository: "git@somewhere:me/m...
I know how to mount git repo when I start pod. See: apiVersion: v1 kind: Pod metadata: name: server spec: containers: - image: nginx name: nginx volumeMounts: - mountPath: /mypath name: git-volume volumes: - name: git-volume gitRepo: repository: "git@somewhere:me/my-git-repository...
A few potential problems here: 1) Ceph needs its nodes to be up and running to be accessible: did you say the nodes were mounting disk from a different Ceph cluster, or is the Ceph cluster running on the same nodes? If the same nodes, then it makes sense that the drive not being accessible paralyzes K8s. 2) There is a ...
###Instruction: I am setting up a small Kubernetes cluster using a VM (master) and 3 bare metal servers (all running Ubuntu 14.04). I followed the Kubernetes install tutorial for Ubuntu. Each bare metal server also has 2T of disk space exported using Ceph 0.94.5. Everything was working fine, but when one node failed to...
I am setting up a small Kubernetes cluster using a VM (master) and 3 bare metal servers (all running Ubuntu 14.04). I followed the Kubernetes install tutorial for Ubuntu. Each bare metal server also has 2T of disk space exported using Ceph 0.94.5. Everything was working fine, but when one node failed to start (it wasn'...
You can't really do this right now. The issue to follow if you're interested in templating is https://github.com/kubernetes/kubernetes/issues/11492 Also, this is currently a topic of discussion in the configuration SIG https://groups.google.com/forum/#!forum/kubernetes-sig-config I'd suggest you register your interest ...
###Instruction: I have two separate Kubernetes clusters that will be used for a staging and a production environment, respectively. I want to have YAML manifests for the Kubernetes API objects I will be submitting to each cluster, but some of the objects will have slightly different configurations between the two envir...
I have two separate Kubernetes clusters that will be used for a staging and a production environment, respectively. I want to have YAML manifests for the Kubernetes API objects I will be submitting to each cluster, but some of the objects will have slightly different configurations between the two environments. As a ma...
The containers in a pod continue to run, yes. If the master components are not available this means no new pods/services can be launched, but existing ones continue to operate. Note that this behaviour is also one of the good practices and lessons learned from Borg.
###Instruction: If kube-master or etcd service down in my kubernetes cluster, can my Pod/Service still work? Also, if the network is still work? ### Response: The containers in a pod continue to run, yes. If the master components are not available this means no new pods/services can be launched, but existing ones cont...
If kube-master or etcd service down in my kubernetes cluster, can my Pod/Service still work? Also, if the network is still work?
Two options: Talk to the Kubernetes API to get the endpoints for the service. (either with kubectl get endpoints SVCNAME or by GETing the /api/v1/namespaces/{namespace}/endpoints/{svcname} path on the apiserver) Less likely to be of use, but if you create a service without a cluster IP, the DNS for that service will r...
###Instruction: I have a proxy service that wraps 3 pods (say pod A, pod B, pod C). Some container inside pod A needs to get virtual IPs of other two pods. How can I do this? ### Response: Two options: Talk to the Kubernetes API to get the endpoints for the service. (either with kubectl get endpoints SVCNAME or by GE...
I have a proxy service that wraps 3 pods (say pod A, pod B, pod C). Some container inside pod A needs to get virtual IPs of other two pods. How can I do this?
docker compose is just a way to declare the container you have to start: it has no notion of node or cluster, unless it launches swarm master and swarm nodes, but that is docker swarm) Update July 2016, 7 months later: docker 1.12 blurs the lines and includes a "swarm mode". It is vastly different from kubernetes, a go...
###Instruction: Both Kubernetes Pods and the results of Docker Compose scripts (henceforth: "Compositions") appear to result in clusters of virtual computers. The computers in the clusters can all be configured to talk to each other so you can write a single script that mirrors your entire end-to-end production config....
Both Kubernetes Pods and the results of Docker Compose scripts (henceforth: "Compositions") appear to result in clusters of virtual computers. The computers in the clusters can all be configured to talk to each other so you can write a single script that mirrors your entire end-to-end production config. A single script...
After a lot of experimentations I believe adding the script to the Dockerfile: ADD prod_start.sh /backend/prod_start.sh And then calling the command like this in the yaml controller file: command: ['/bin/sh', './prod_start.sh'] Fixed it.
###Instruction: I am trying to run a shell script at the start of a docker container running on Google Cloud Containers using Kubernetes. The structure of my app directory is something like this. I'd like to run prod_start.sh script at the start of the container (I don't want to put it as part of the Dockerfile though)...
I am trying to run a shell script at the start of a docker container running on Google Cloud Containers using Kubernetes. The structure of my app directory is something like this. I'd like to run prod_start.sh script at the start of the container (I don't want to put it as part of the Dockerfile though). The current se...
There can only be a single entrypoint in a container... if you want to run multiple commands like that, make bash be the entry point, and make all the other commands be an argument for bash to run: command: ["/bin/bash","-c","touch /foo && echo 'here' && ls /"]
###Instruction: I want send multiple entrypoint commands to a Docker container in the command tag of kubernetes config file. apiVersion: v1 kind: Pod metadata: name: hello-world spec: # specification of the pod’s contents restartPolicy: Never containers: - name: hello image: "ubuntu:14.04" command: ["c...
I want send multiple entrypoint commands to a Docker container in the command tag of kubernetes config file. apiVersion: v1 kind: Pod metadata: name: hello-world spec: # specification of the pod’s contents restartPolicy: Never containers: - name: hello image: "ubuntu:14.04" command: ["command1 arg1 arg...
The doc you link to describes the methodology used (specifically the master VM size). The cluster is created in Google Compute Engine using the default cluster/kube-up.sh script from the repository, with all the default settings implied by that. How large is the master that you're using? If it's really small, it's poss...
###Instruction: From the performance test report we can find that the kubernetes can support 100 nodes. To do the same test, I have set up a 100 nodes kubernetes cluster, but the kube-apiserver became slow when the cluster is set up. That meant when I typed kubectl get nodes, it suspended and can not get any respond. T...
From the performance test report we can find that the kubernetes can support 100 nodes. To do the same test, I have set up a 100 nodes kubernetes cluster, but the kube-apiserver became slow when the cluster is set up. That meant when I typed kubectl get nodes, it suspended and can not get any respond. To find the reaso...
For future, the results of iptables-save are much easier to read (to me anyway). I don't see anything missing here. KUBE-SERVICES traps 10.116.0.2 port 53/UDP and passes it to KUBE-SVC-TCOU7JCQXEZGVUNU KUBE-SVC-TCOU7JCQXEZGVUNU has just one endpoint so jumps to KUBE-SEP-GSM3BZTEXEBWDXPN KUBE-SEP-GSM3BZTEXEBWDXPN DNATs ...
###Instruction: I have Kubernetes: v.1.1.1 iptables v1.4.21 kernel: 4.2.0-18-generic which come with Ubuntu wily Networking is done via L2 VLAN terminated on switch no cloud provider what I do I'm experimenting with iptables mode for kube-proxy. I have enabled it with --proxy_mode=iptables argument. It seems some r...
I have Kubernetes: v.1.1.1 iptables v1.4.21 kernel: 4.2.0-18-generic which come with Ubuntu wily Networking is done via L2 VLAN terminated on switch no cloud provider what I do I'm experimenting with iptables mode for kube-proxy. I have enabled it with --proxy_mode=iptables argument. It seems some rule is missing: ...
If someone will reach this question then I want to let you know that external load balancer support is available in latest kubernetes version. Link to the documentation
###Instruction: I have a kubernetes (0.15) cluster running on CoreOS instances on Amazon EC2 When I create a service that I want to be publicly accessible, I currently add some private IP addresses of the EC2 instances to the service description like so: { "kind": "Service", "apiVersion": "v1beta3", "metadata": {...
I have a kubernetes (0.15) cluster running on CoreOS instances on Amazon EC2 When I create a service that I want to be publicly accessible, I currently add some private IP addresses of the EC2 instances to the service description like so: { "kind": "Service", "apiVersion": "v1beta3", "metadata": { "name": "ap...
I just stumbled across the same thing and found out after some digging, that they actually changed the volume mount syntax. Based on that knowledge I created this PR for documentation update. See https://github.com/kubernetes/kubernetes/pull/17958 for tracking that and more info, follow the link to the bug and the orig...
###Instruction: I created a volume using the following command. aws ec2 create-volume --size 10 --region us-east-1 --availability-zone us-east-1c --volume-type gp2 Then I used the file below to create a pod that uses the volume. But when I login to the pod, I don't see the volume. Is there something that I might be do...
I created a volume using the following command. aws ec2 create-volume --size 10 --region us-east-1 --availability-zone us-east-1c --volume-type gp2 Then I used the file below to create a pod that uses the volume. But when I login to the pod, I don't see the volume. Is there something that I might be doing wrong? Did I...
Kubernetes is a set of daemons/binaries: kube-apiserver (AKA the master), kubelet (start/stop containers, sync conf.), kube-scheduler (resources manager) kube-controller-manager (monitor RC, and maintain the desired state) kube-proxy (expose services on each node) kubectl (CLI) The hyperkube binary is an all in one ...
###Instruction: I am trying to setup kubernetes in aws and following the guides at https://github.com/kubernetes/kubernetes/blob/master/docs/getting-started-guides/docker-multinode I couldn't understand what is meant by hyperkube. Can someone please explain to me what it is and how does it work? And another question I ...
I am trying to setup kubernetes in aws and following the guides at https://github.com/kubernetes/kubernetes/blob/master/docs/getting-started-guides/docker-multinode I couldn't understand what is meant by hyperkube. Can someone please explain to me what it is and how does it work? And another question I have is while ru...
In order to use this UI, go to the saturnism/gcp-live-k8s-visualizer GitHub repo and follow the steps, there.
###Instruction: I was going through a Kubernetes tutorial on Youtube and found the following UI which demonstrates pod and service arrangements of Kubernetes cluster.How can I install this UI in my Kubernetes setup? ### Response: In order to use this UI, go to the saturnism/gcp-live-k8s-visualizer GitHub repo and f...
I was going through a Kubernetes tutorial on Youtube and found the following UI which demonstrates pod and service arrangements of Kubernetes cluster.How can I install this UI in my Kubernetes setup?
You can pass the --cluster= or --context= flags to kubectl to set a single run. For example, if I have two clusters in my ~/.kube/config "foo" and "bar": $ kubectl --cluster=foo get pods NAME READY STATUS RESTARTS AGE foo-ht1qh 1/1 Running 0 3h foo-wf8f4 1/1 Running 0...
###Instruction: Background We're using Jenkins to deploy a new version of a Kubernetes (k8s) replication controller to our test or prod cluster. The test and prod (k8s) clusters are located under different (google cloud platform) projects. We have configured two profiles for our gcloud SDK on Jenkins, one for test (tes...
Background We're using Jenkins to deploy a new version of a Kubernetes (k8s) replication controller to our test or prod cluster. The test and prod (k8s) clusters are located under different (google cloud platform) projects. We have configured two profiles for our gcloud SDK on Jenkins, one for test (test-profile) and o...