/* Copyright 2016 The Kubernetes Authors. Licensed under the Apache License, Version 2.0 (the "License"); you may not use this file except in compliance with the License. You may obtain a copy of the License at http://www.apache.org/licenses/LICENSE-2.0 Unless required by applicable law or agreed to in writing, software distributed under the License is distributed on an "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the License for the specific language governing permissions and limitations under the License. */ package e2e import ( "fmt" "time" . "github.com/onsi/ginkgo" . "github.com/onsi/gomega" metav1 "k8s.io/apimachinery/pkg/apis/meta/v1" "k8s.io/apimachinery/pkg/util/wait" "k8s.io/client-go/kubernetes" "k8s.io/client-go/pkg/api/v1" extensions "k8s.io/client-go/pkg/apis/extensions/v1beta1" policy "k8s.io/client-go/pkg/apis/policy/v1beta1" "k8s.io/client-go/pkg/util/intstr" "k8s.io/kubernetes/test/e2e/framework" ) // schedulingTimeout is longer specifically because sometimes we need to wait // awhile to guarantee that we've been patient waiting for something ordinary // to happen: a pod to get scheduled and move into Ready const ( bigClusterSize = 7 schedulingTimeout = 10 * time.Minute timeout = 60 * time.Second ) var _ = framework.KubeDescribe("DisruptionController", func() { f := framework.NewDefaultFramework("disruption") var ns string var cs *kubernetes.Clientset BeforeEach(func() { cs = f.StagingClient ns = f.Namespace.Name }) It("should create a PodDisruptionBudget", func() { createPodDisruptionBudgetOrDie(cs, ns, intstr.FromString("1%")) }) It("should update PodDisruptionBudget status", func() { createPodDisruptionBudgetOrDie(cs, ns, intstr.FromInt(2)) createPodsOrDie(cs, ns, 3) waitForPodsOrDie(cs, ns, 3) // Since disruptionAllowed starts out 0, if we see it ever become positive, // that means the controller is working. err := wait.PollImmediate(framework.Poll, timeout, func() (bool, error) { pdb, err := cs.Policy().PodDisruptionBudgets(ns).Get("foo", metav1.GetOptions{}) if err != nil { return false, err } return pdb.Status.PodDisruptionsAllowed > 0, nil }) Expect(err).NotTo(HaveOccurred()) }) evictionCases := []struct { description string minAvailable intstr.IntOrString podCount int replicaSetSize int32 shouldDeny bool exclusive bool skipForBigClusters bool }{ { description: "no PDB", minAvailable: intstr.FromString(""), podCount: 1, shouldDeny: false, }, { description: "too few pods, absolute", minAvailable: intstr.FromInt(2), podCount: 2, shouldDeny: true, }, { description: "enough pods, absolute", minAvailable: intstr.FromInt(2), podCount: 3, shouldDeny: false, }, { description: "enough pods, replicaSet, percentage", minAvailable: intstr.FromString("90%"), replicaSetSize: 10, exclusive: false, shouldDeny: false, }, { description: "too few pods, replicaSet, percentage", minAvailable: intstr.FromString("90%"), replicaSetSize: 10, exclusive: true, shouldDeny: true, // This tests assumes that there is less than replicaSetSize nodes in the cluster. skipForBigClusters: true, }, } for i := range evictionCases { c := evictionCases[i] expectation := "should allow an eviction" if c.shouldDeny { expectation = "should not allow an eviction" } It(fmt.Sprintf("evictions: %s => %s", c.description, expectation), func() { if c.skipForBigClusters { framework.SkipUnlessNodeCountIsAtMost(bigClusterSize - 1) } createPodsOrDie(cs, ns, c.podCount) if c.replicaSetSize > 0 { createReplicaSetOrDie(cs, ns, c.replicaSetSize, c.exclusive) } if c.minAvailable.String() != "" { createPodDisruptionBudgetOrDie(cs, ns, c.minAvailable) } // Locate a running pod. var pod v1.Pod err := wait.PollImmediate(framework.Poll, schedulingTimeout, func() (bool, error) { podList, err := cs.Pods(ns).List(metav1.ListOptions{}) if err != nil { return false, err } for i := range podList.Items { if podList.Items[i].Status.Phase == v1.PodRunning { pod = podList.Items[i] return true, nil } } return false, nil }) Expect(err).NotTo(HaveOccurred()) e := &policy.Eviction{ ObjectMeta: metav1.ObjectMeta{ Name: pod.Name, Namespace: ns, }, } if c.shouldDeny { // Since disruptionAllowed starts out false, wait at least 60s hoping that // this gives the controller enough time to have truly set the status. time.Sleep(timeout) err = cs.Pods(ns).Evict(e) Expect(err).Should(MatchError("Cannot evict pod as it would violate the pod's disruption budget.")) } else { // Only wait for running pods in the "allow" case // because one of shouldDeny cases relies on the // replicaSet not fitting on the cluster. waitForPodsOrDie(cs, ns, c.podCount+int(c.replicaSetSize)) // Since disruptionAllowed starts out false, if an eviction is ever allowed, // that means the controller is working. err = wait.PollImmediate(framework.Poll, timeout, func() (bool, error) { err = cs.Pods(ns).Evict(e) if err != nil { return false, nil } else { return true, nil } }) Expect(err).NotTo(HaveOccurred()) } }) } }) func createPodDisruptionBudgetOrDie(cs *kubernetes.Clientset, ns string, minAvailable intstr.IntOrString) { pdb := policy.PodDisruptionBudget{ ObjectMeta: metav1.ObjectMeta{ Name: "foo", Namespace: ns, }, Spec: policy.PodDisruptionBudgetSpec{ Selector: &metav1.LabelSelector{MatchLabels: map[string]string{"foo": "bar"}}, MinAvailable: minAvailable, }, } _, err := cs.Policy().PodDisruptionBudgets(ns).Create(&pdb) Expect(err).NotTo(HaveOccurred()) } func createPodsOrDie(cs *kubernetes.Clientset, ns string, n int) { for i := 0; i < n; i++ { pod := &v1.Pod{ ObjectMeta: metav1.ObjectMeta{ Name: fmt.Sprintf("pod-%d", i), Namespace: ns, Labels: map[string]string{"foo": "bar"}, }, Spec: v1.PodSpec{ Containers: []v1.Container{ { Name: "busybox", Image: "gcr.io/google_containers/echoserver:1.4", }, }, RestartPolicy: v1.RestartPolicyAlways, }, } _, err := cs.Pods(ns).Create(pod) framework.ExpectNoError(err, "Creating pod %q in namespace %q", pod.Name, ns) } } func waitForPodsOrDie(cs *kubernetes.Clientset, ns string, n int) { By("Waiting for all pods to be running") err := wait.PollImmediate(framework.Poll, schedulingTimeout, func() (bool, error) { pods, err := cs.Core().Pods(ns).List(metav1.ListOptions{LabelSelector: "foo=bar"}) if err != nil { return false, err } if pods == nil { return false, fmt.Errorf("pods is nil") } if len(pods.Items) < n { framework.Logf("pods: %v < %v", len(pods.Items), n) return false, nil } ready := 0 for i := 0; i < n; i++ { if pods.Items[i].Status.Phase == v1.PodRunning { ready++ } } if ready < n { framework.Logf("running pods: %v < %v", ready, n) return false, nil } return true, nil }) framework.ExpectNoError(err, "Waiting for pods in namespace %q to be ready", ns) } func createReplicaSetOrDie(cs *kubernetes.Clientset, ns string, size int32, exclusive bool) { container := v1.Container{ Name: "busybox", Image: "gcr.io/google_containers/echoserver:1.4", } if exclusive { container.Ports = []v1.ContainerPort{ {HostPort: 5555, ContainerPort: 5555}, } } rs := &extensions.ReplicaSet{ ObjectMeta: metav1.ObjectMeta{ Name: "rs", Namespace: ns, }, Spec: extensions.ReplicaSetSpec{ Replicas: &size, Selector: &metav1.LabelSelector{ MatchLabels: map[string]string{"foo": "bar"}, }, Template: v1.PodTemplateSpec{ ObjectMeta: metav1.ObjectMeta{ Labels: map[string]string{"foo": "bar"}, }, Spec: v1.PodSpec{ Containers: []v1.Container{container}, }, }, }, } _, err := cs.Extensions().ReplicaSets(ns).Create(rs) framework.ExpectNoError(err, "Creating replica set %q in namespace %q", rs.Name, ns) }