mirror of
				https://github.com/optim-enterprises-bv/kubernetes.git
				synced 2025-11-04 04:08:16 +00:00 
			
		
		
		
	
		
			
				
	
	
		
			859 lines
		
	
	
		
			29 KiB
		
	
	
	
		
			Go
		
	
	
	
	
	
			
		
		
	
	
			859 lines
		
	
	
		
			29 KiB
		
	
	
	
		
			Go
		
	
	
	
	
	
/*
 | 
						|
Copyright 2017 The Kubernetes Authors.
 | 
						|
 | 
						|
Licensed under the Apache License, Version 2.0 (the "License");
 | 
						|
you may not use this file except in compliance with the License.
 | 
						|
You may obtain a copy of the License at
 | 
						|
 | 
						|
    http://www.apache.org/licenses/LICENSE-2.0
 | 
						|
 | 
						|
Unless required by applicable law or agreed to in writing, software
 | 
						|
distributed under the License is distributed on an "AS IS" BASIS,
 | 
						|
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
 | 
						|
See the License for the specific language governing permissions and
 | 
						|
limitations under the License.
 | 
						|
*/
 | 
						|
 | 
						|
package daemon
 | 
						|
 | 
						|
import (
 | 
						|
	"context"
 | 
						|
	"reflect"
 | 
						|
	"testing"
 | 
						|
	"time"
 | 
						|
 | 
						|
	apps "k8s.io/api/apps/v1"
 | 
						|
	v1 "k8s.io/api/core/v1"
 | 
						|
	metav1 "k8s.io/apimachinery/pkg/apis/meta/v1"
 | 
						|
	"k8s.io/apimachinery/pkg/labels"
 | 
						|
	"k8s.io/apimachinery/pkg/util/intstr"
 | 
						|
	"k8s.io/klog/v2/ktesting"
 | 
						|
	podutil "k8s.io/kubernetes/pkg/api/v1/pod"
 | 
						|
	"k8s.io/kubernetes/pkg/controller/daemon/util"
 | 
						|
	testingclock "k8s.io/utils/clock/testing"
 | 
						|
)
 | 
						|
 | 
						|
func TestDaemonSetUpdatesPods(t *testing.T) {
 | 
						|
	_, ctx := ktesting.NewTestContext(t)
 | 
						|
	ds := newDaemonSet("foo")
 | 
						|
	manager, podControl, _, err := newTestController(ctx, ds)
 | 
						|
	if err != nil {
 | 
						|
		t.Fatalf("error creating DaemonSets controller: %v", err)
 | 
						|
	}
 | 
						|
	maxUnavailable := 2
 | 
						|
	addNodes(manager.nodeStore, 0, 5, nil)
 | 
						|
	manager.dsStore.Add(ds)
 | 
						|
	expectSyncDaemonSets(t, manager, ds, podControl, 5, 0, 0)
 | 
						|
	markPodsReady(podControl.podStore)
 | 
						|
 | 
						|
	ds.Spec.Template.Spec.Containers[0].Image = "foo2/bar2"
 | 
						|
	ds.Spec.UpdateStrategy.Type = apps.RollingUpdateDaemonSetStrategyType
 | 
						|
	intStr := intstr.FromInt32(int32(maxUnavailable))
 | 
						|
	ds.Spec.UpdateStrategy.RollingUpdate = &apps.RollingUpdateDaemonSet{MaxUnavailable: &intStr}
 | 
						|
	manager.dsStore.Update(ds)
 | 
						|
 | 
						|
	clearExpectations(t, manager, ds, podControl)
 | 
						|
	expectSyncDaemonSets(t, manager, ds, podControl, 0, maxUnavailable, 0)
 | 
						|
	clearExpectations(t, manager, ds, podControl)
 | 
						|
	expectSyncDaemonSets(t, manager, ds, podControl, maxUnavailable, 0, 0)
 | 
						|
	markPodsReady(podControl.podStore)
 | 
						|
 | 
						|
	clearExpectations(t, manager, ds, podControl)
 | 
						|
	expectSyncDaemonSets(t, manager, ds, podControl, 0, maxUnavailable, 0)
 | 
						|
	clearExpectations(t, manager, ds, podControl)
 | 
						|
	expectSyncDaemonSets(t, manager, ds, podControl, maxUnavailable, 0, 0)
 | 
						|
	markPodsReady(podControl.podStore)
 | 
						|
 | 
						|
	clearExpectations(t, manager, ds, podControl)
 | 
						|
	expectSyncDaemonSets(t, manager, ds, podControl, 0, 1, 0)
 | 
						|
	clearExpectations(t, manager, ds, podControl)
 | 
						|
	expectSyncDaemonSets(t, manager, ds, podControl, 1, 0, 0)
 | 
						|
	markPodsReady(podControl.podStore)
 | 
						|
 | 
						|
	clearExpectations(t, manager, ds, podControl)
 | 
						|
	expectSyncDaemonSets(t, manager, ds, podControl, 0, 0, 0)
 | 
						|
	clearExpectations(t, manager, ds, podControl)
 | 
						|
}
 | 
						|
 | 
						|
func TestDaemonSetUpdatesPodsWithMaxSurge(t *testing.T) {
 | 
						|
	_, ctx := ktesting.NewTestContext(t)
 | 
						|
	ds := newDaemonSet("foo")
 | 
						|
	manager, podControl, _, err := newTestController(ctx, ds)
 | 
						|
	if err != nil {
 | 
						|
		t.Fatalf("error creating DaemonSets controller: %v", err)
 | 
						|
	}
 | 
						|
	addNodes(manager.nodeStore, 0, 5, nil)
 | 
						|
	manager.dsStore.Add(ds)
 | 
						|
	expectSyncDaemonSets(t, manager, ds, podControl, 5, 0, 0)
 | 
						|
	markPodsReady(podControl.podStore)
 | 
						|
 | 
						|
	// surge is thhe controlling amount
 | 
						|
	maxSurge := 2
 | 
						|
	ds.Spec.Template.Spec.Containers[0].Image = "foo2/bar2"
 | 
						|
	ds.Spec.UpdateStrategy = newUpdateSurge(intstr.FromInt32(int32(maxSurge)))
 | 
						|
	manager.dsStore.Update(ds)
 | 
						|
 | 
						|
	clearExpectations(t, manager, ds, podControl)
 | 
						|
	expectSyncDaemonSets(t, manager, ds, podControl, maxSurge, 0, 0)
 | 
						|
	clearExpectations(t, manager, ds, podControl)
 | 
						|
	expectSyncDaemonSets(t, manager, ds, podControl, 0, 0, 0)
 | 
						|
	markPodsReady(podControl.podStore)
 | 
						|
 | 
						|
	clearExpectations(t, manager, ds, podControl)
 | 
						|
	expectSyncDaemonSets(t, manager, ds, podControl, maxSurge, maxSurge, 0)
 | 
						|
	clearExpectations(t, manager, ds, podControl)
 | 
						|
	expectSyncDaemonSets(t, manager, ds, podControl, 0, 0, 0)
 | 
						|
	markPodsReady(podControl.podStore)
 | 
						|
 | 
						|
	clearExpectations(t, manager, ds, podControl)
 | 
						|
	expectSyncDaemonSets(t, manager, ds, podControl, 5%maxSurge, maxSurge, 0)
 | 
						|
	clearExpectations(t, manager, ds, podControl)
 | 
						|
	expectSyncDaemonSets(t, manager, ds, podControl, 0, 0, 0)
 | 
						|
	markPodsReady(podControl.podStore)
 | 
						|
 | 
						|
	clearExpectations(t, manager, ds, podControl)
 | 
						|
	expectSyncDaemonSets(t, manager, ds, podControl, 0, 5%maxSurge, 0)
 | 
						|
	clearExpectations(t, manager, ds, podControl)
 | 
						|
	expectSyncDaemonSets(t, manager, ds, podControl, 0, 0, 0)
 | 
						|
}
 | 
						|
 | 
						|
func TestDaemonSetUpdatesPodsNotMatchTainstWithMaxSurge(t *testing.T) {
 | 
						|
	_, ctx := ktesting.NewTestContext(t)
 | 
						|
 | 
						|
	ds := newDaemonSet("foo")
 | 
						|
	maxSurge := 1
 | 
						|
	ds.Spec.UpdateStrategy = newUpdateSurge(intstr.FromInt(maxSurge))
 | 
						|
	tolerations := []v1.Toleration{
 | 
						|
		{Key: "node-role.kubernetes.io/control-plane", Operator: v1.TolerationOpExists},
 | 
						|
	}
 | 
						|
	setDaemonSetToleration(ds, tolerations)
 | 
						|
	manager, podControl, _, err := newTestController(ctx, ds)
 | 
						|
	if err != nil {
 | 
						|
		t.Fatalf("error creating DaemonSets controller: %v", err)
 | 
						|
	}
 | 
						|
	err = manager.dsStore.Add(ds)
 | 
						|
	if err != nil {
 | 
						|
		t.Fatal(err)
 | 
						|
	}
 | 
						|
 | 
						|
	// Add five nodes and taint to one node
 | 
						|
	addNodes(manager.nodeStore, 0, 5, nil)
 | 
						|
	taints := []v1.Taint{
 | 
						|
		{Key: "node-role.kubernetes.io/control-plane", Effect: v1.TaintEffectNoSchedule},
 | 
						|
	}
 | 
						|
	node := newNode("node-0", nil)
 | 
						|
	setNodeTaint(node, taints)
 | 
						|
	err = manager.nodeStore.Update(node)
 | 
						|
	if err != nil {
 | 
						|
		t.Fatal(err)
 | 
						|
	}
 | 
						|
 | 
						|
	// Create DaemonSet with toleration
 | 
						|
	expectSyncDaemonSets(t, manager, ds, podControl, 5, 0, 0)
 | 
						|
	markPodsReady(podControl.podStore)
 | 
						|
 | 
						|
	// RollingUpdate DaemonSet without toleration
 | 
						|
	ds.Spec.Template.Spec.Tolerations = nil
 | 
						|
	err = manager.dsStore.Update(ds)
 | 
						|
	if err != nil {
 | 
						|
		t.Fatal(err)
 | 
						|
	}
 | 
						|
 | 
						|
	clearExpectations(t, manager, ds, podControl)
 | 
						|
	expectSyncDaemonSets(t, manager, ds, podControl, maxSurge, 1, 0)
 | 
						|
	clearExpectations(t, manager, ds, podControl)
 | 
						|
	expectSyncDaemonSets(t, manager, ds, podControl, 0, 0, 0)
 | 
						|
	markPodsReady(podControl.podStore)
 | 
						|
 | 
						|
	clearExpectations(t, manager, ds, podControl)
 | 
						|
	expectSyncDaemonSets(t, manager, ds, podControl, maxSurge, maxSurge, 0)
 | 
						|
	clearExpectations(t, manager, ds, podControl)
 | 
						|
	expectSyncDaemonSets(t, manager, ds, podControl, 0, 0, 0)
 | 
						|
	markPodsReady(podControl.podStore)
 | 
						|
 | 
						|
	clearExpectations(t, manager, ds, podControl)
 | 
						|
	expectSyncDaemonSets(t, manager, ds, podControl, maxSurge, maxSurge, 0)
 | 
						|
	clearExpectations(t, manager, ds, podControl)
 | 
						|
	expectSyncDaemonSets(t, manager, ds, podControl, 0, 0, 0)
 | 
						|
	markPodsReady(podControl.podStore)
 | 
						|
 | 
						|
	clearExpectations(t, manager, ds, podControl)
 | 
						|
	expectSyncDaemonSets(t, manager, ds, podControl, maxSurge, maxSurge, 0)
 | 
						|
	clearExpectations(t, manager, ds, podControl)
 | 
						|
	expectSyncDaemonSets(t, manager, ds, podControl, 0, 0, 0)
 | 
						|
	markPodsReady(podControl.podStore)
 | 
						|
 | 
						|
	clearExpectations(t, manager, ds, podControl)
 | 
						|
	expectSyncDaemonSets(t, manager, ds, podControl, 0, maxSurge, 0)
 | 
						|
	clearExpectations(t, manager, ds, podControl)
 | 
						|
	expectSyncDaemonSets(t, manager, ds, podControl, 0, 0, 0)
 | 
						|
}
 | 
						|
 | 
						|
func TestDaemonSetUpdatesWhenNewPodIsNotReady(t *testing.T) {
 | 
						|
	_, ctx := ktesting.NewTestContext(t)
 | 
						|
	ds := newDaemonSet("foo")
 | 
						|
	manager, podControl, _, err := newTestController(ctx, ds)
 | 
						|
	if err != nil {
 | 
						|
		t.Fatalf("error creating DaemonSets controller: %v", err)
 | 
						|
	}
 | 
						|
	maxUnavailable := 3
 | 
						|
	addNodes(manager.nodeStore, 0, 5, nil)
 | 
						|
	err = manager.dsStore.Add(ds)
 | 
						|
	if err != nil {
 | 
						|
		t.Fatal(err)
 | 
						|
	}
 | 
						|
	expectSyncDaemonSets(t, manager, ds, podControl, 5, 0, 0)
 | 
						|
	markPodsReady(podControl.podStore)
 | 
						|
 | 
						|
	ds.Spec.Template.Spec.Containers[0].Image = "foo2/bar2"
 | 
						|
	ds.Spec.UpdateStrategy.Type = apps.RollingUpdateDaemonSetStrategyType
 | 
						|
	intStr := intstr.FromInt32(int32(maxUnavailable))
 | 
						|
	ds.Spec.UpdateStrategy.RollingUpdate = &apps.RollingUpdateDaemonSet{MaxUnavailable: &intStr}
 | 
						|
	err = manager.dsStore.Update(ds)
 | 
						|
	if err != nil {
 | 
						|
		t.Fatal(err)
 | 
						|
	}
 | 
						|
 | 
						|
	// new pods are not ready numUnavailable == maxUnavailable
 | 
						|
	clearExpectations(t, manager, ds, podControl)
 | 
						|
	expectSyncDaemonSets(t, manager, ds, podControl, 0, maxUnavailable, 0)
 | 
						|
 | 
						|
	clearExpectations(t, manager, ds, podControl)
 | 
						|
	expectSyncDaemonSets(t, manager, ds, podControl, maxUnavailable, 0, 0)
 | 
						|
 | 
						|
	clearExpectations(t, manager, ds, podControl)
 | 
						|
	expectSyncDaemonSets(t, manager, ds, podControl, 0, 0, 0)
 | 
						|
	clearExpectations(t, manager, ds, podControl)
 | 
						|
}
 | 
						|
 | 
						|
func TestDaemonSetUpdatesSomeOldPodsNotReady(t *testing.T) {
 | 
						|
	_, ctx := ktesting.NewTestContext(t)
 | 
						|
	ds := newDaemonSet("foo")
 | 
						|
	manager, podControl, _, err := newTestController(ctx, ds)
 | 
						|
	if err != nil {
 | 
						|
		t.Fatalf("error creating DaemonSets controller: %v", err)
 | 
						|
	}
 | 
						|
	maxUnavailable := 2
 | 
						|
	addNodes(manager.nodeStore, 0, 5, nil)
 | 
						|
	err = manager.dsStore.Add(ds)
 | 
						|
	if err != nil {
 | 
						|
		t.Fatal(err)
 | 
						|
	}
 | 
						|
	expectSyncDaemonSets(t, manager, ds, podControl, 5, 0, 0)
 | 
						|
	markPodsReady(podControl.podStore)
 | 
						|
 | 
						|
	ds.Spec.Template.Spec.Containers[0].Image = "foo2/bar2"
 | 
						|
	ds.Spec.UpdateStrategy.Type = apps.RollingUpdateDaemonSetStrategyType
 | 
						|
	intStr := intstr.FromInt(maxUnavailable)
 | 
						|
	ds.Spec.UpdateStrategy.RollingUpdate = &apps.RollingUpdateDaemonSet{MaxUnavailable: &intStr}
 | 
						|
	err = manager.dsStore.Update(ds)
 | 
						|
	if err != nil {
 | 
						|
		t.Fatal(err)
 | 
						|
	}
 | 
						|
 | 
						|
	// All old pods are available, should update 2
 | 
						|
	clearExpectations(t, manager, ds, podControl)
 | 
						|
	expectSyncDaemonSets(t, manager, ds, podControl, 0, maxUnavailable, 0)
 | 
						|
 | 
						|
	clearExpectations(t, manager, ds, podControl)
 | 
						|
	expectSyncDaemonSets(t, manager, ds, podControl, maxUnavailable, 0, 0)
 | 
						|
 | 
						|
	clearExpectations(t, manager, ds, podControl)
 | 
						|
	expectSyncDaemonSets(t, manager, ds, podControl, 0, 0, 0)
 | 
						|
	clearExpectations(t, manager, ds, podControl)
 | 
						|
 | 
						|
	// Perform another update, verify we delete and create 2 pods, three available should remain untouched
 | 
						|
 | 
						|
	ds.Spec.Template.Spec.Containers[0].Image = "foo2/bar3"
 | 
						|
	err = manager.dsStore.Update(ds)
 | 
						|
	if err != nil {
 | 
						|
		t.Fatal(err)
 | 
						|
	}
 | 
						|
 | 
						|
	clearExpectations(t, manager, ds, podControl)
 | 
						|
	expectSyncDaemonSets(t, manager, ds, podControl, 0, maxUnavailable, 0)
 | 
						|
 | 
						|
	clearExpectations(t, manager, ds, podControl)
 | 
						|
	expectSyncDaemonSets(t, manager, ds, podControl, maxUnavailable, 0, 0)
 | 
						|
 | 
						|
	clearExpectations(t, manager, ds, podControl)
 | 
						|
	expectSyncDaemonSets(t, manager, ds, podControl, 0, 0, 0)
 | 
						|
	clearExpectations(t, manager, ds, podControl)
 | 
						|
 | 
						|
	readyCount := 0
 | 
						|
	expectedReadyCount := 5 - maxUnavailable
 | 
						|
	for _, obj := range podControl.podStore.List() {
 | 
						|
		pod := obj.(*v1.Pod)
 | 
						|
		n, condition := podutil.GetPodCondition(&pod.Status, v1.PodReady)
 | 
						|
		if n != -1 && condition.Status == v1.ConditionTrue {
 | 
						|
			readyCount++
 | 
						|
		}
 | 
						|
	}
 | 
						|
 | 
						|
	if readyCount != expectedReadyCount {
 | 
						|
		t.Fatalf("Expected %d old ready pods, but found %d", expectedReadyCount, readyCount)
 | 
						|
	}
 | 
						|
}
 | 
						|
 | 
						|
func TestDaemonSetUpdatesAllOldPodsNotReady(t *testing.T) {
 | 
						|
	_, ctx := ktesting.NewTestContext(t)
 | 
						|
	ds := newDaemonSet("foo")
 | 
						|
	manager, podControl, _, err := newTestController(ctx, ds)
 | 
						|
	if err != nil {
 | 
						|
		t.Fatalf("error creating DaemonSets controller: %v", err)
 | 
						|
	}
 | 
						|
	maxUnavailable := 3
 | 
						|
	addNodes(manager.nodeStore, 0, 5, nil)
 | 
						|
	err = manager.dsStore.Add(ds)
 | 
						|
	if err != nil {
 | 
						|
		t.Fatal(err)
 | 
						|
	}
 | 
						|
	expectSyncDaemonSets(t, manager, ds, podControl, 5, 0, 0)
 | 
						|
 | 
						|
	ds.Spec.Template.Spec.Containers[0].Image = "foo2/bar2"
 | 
						|
	ds.Spec.UpdateStrategy.Type = apps.RollingUpdateDaemonSetStrategyType
 | 
						|
	intStr := intstr.FromInt32(int32(maxUnavailable))
 | 
						|
	ds.Spec.UpdateStrategy.RollingUpdate = &apps.RollingUpdateDaemonSet{MaxUnavailable: &intStr}
 | 
						|
	err = manager.dsStore.Update(ds)
 | 
						|
	if err != nil {
 | 
						|
		t.Fatal(err)
 | 
						|
	}
 | 
						|
 | 
						|
	// all old pods are unavailable so should be removed
 | 
						|
	clearExpectations(t, manager, ds, podControl)
 | 
						|
	expectSyncDaemonSets(t, manager, ds, podControl, 0, 5, 0)
 | 
						|
 | 
						|
	clearExpectations(t, manager, ds, podControl)
 | 
						|
	expectSyncDaemonSets(t, manager, ds, podControl, 5, 0, 0)
 | 
						|
 | 
						|
	clearExpectations(t, manager, ds, podControl)
 | 
						|
	expectSyncDaemonSets(t, manager, ds, podControl, 0, 0, 0)
 | 
						|
	clearExpectations(t, manager, ds, podControl)
 | 
						|
}
 | 
						|
 | 
						|
func TestDaemonSetUpdatesAllOldPodsNotReadyMaxSurge(t *testing.T) {
 | 
						|
	_, ctx := ktesting.NewTestContext(t)
 | 
						|
	ds := newDaemonSet("foo")
 | 
						|
	manager, podControl, _, err := newTestController(ctx, ds)
 | 
						|
	if err != nil {
 | 
						|
		t.Fatalf("error creating DaemonSets controller: %v", err)
 | 
						|
	}
 | 
						|
	addNodes(manager.nodeStore, 0, 5, nil)
 | 
						|
	manager.dsStore.Add(ds)
 | 
						|
	expectSyncDaemonSets(t, manager, ds, podControl, 5, 0, 0)
 | 
						|
 | 
						|
	maxSurge := 3
 | 
						|
	ds.Spec.Template.Spec.Containers[0].Image = "foo2/bar2"
 | 
						|
	ds.Spec.UpdateStrategy = newUpdateSurge(intstr.FromInt32(int32(maxSurge)))
 | 
						|
	manager.dsStore.Update(ds)
 | 
						|
 | 
						|
	// all old pods are unavailable so should be surged
 | 
						|
	manager.failedPodsBackoff.Clock = testingclock.NewFakeClock(time.Unix(100, 0))
 | 
						|
	clearExpectations(t, manager, ds, podControl)
 | 
						|
	expectSyncDaemonSets(t, manager, ds, podControl, 5, 0, 0)
 | 
						|
 | 
						|
	// waiting for pods to go ready, old pods are deleted
 | 
						|
	manager.failedPodsBackoff.Clock = testingclock.NewFakeClock(time.Unix(200, 0))
 | 
						|
	clearExpectations(t, manager, ds, podControl)
 | 
						|
	expectSyncDaemonSets(t, manager, ds, podControl, 0, 5, 0)
 | 
						|
 | 
						|
	setPodReadiness(t, manager, true, 5, func(_ *v1.Pod) bool { return true })
 | 
						|
	ds.Spec.MinReadySeconds = 15
 | 
						|
	ds.Spec.Template.Spec.Containers[0].Image = "foo3/bar3"
 | 
						|
	manager.dsStore.Update(ds)
 | 
						|
 | 
						|
	manager.failedPodsBackoff.Clock = testingclock.NewFakeClock(time.Unix(300, 0))
 | 
						|
	clearExpectations(t, manager, ds, podControl)
 | 
						|
	expectSyncDaemonSets(t, manager, ds, podControl, 3, 0, 0)
 | 
						|
 | 
						|
	hash, err := currentDSHash(manager, ds)
 | 
						|
	if err != nil {
 | 
						|
		t.Fatal(err)
 | 
						|
	}
 | 
						|
	currentPods := podsByNodeMatchingHash(manager, hash)
 | 
						|
	// mark two updated pods as ready at time 300
 | 
						|
	setPodReadiness(t, manager, true, 2, func(pod *v1.Pod) bool {
 | 
						|
		return pod.Labels[apps.ControllerRevisionHashLabelKey] == hash
 | 
						|
	})
 | 
						|
	// mark one of the old pods that is on a node without an updated pod as unready
 | 
						|
	setPodReadiness(t, manager, false, 1, func(pod *v1.Pod) bool {
 | 
						|
		nodeName, err := util.GetTargetNodeName(pod)
 | 
						|
		if err != nil {
 | 
						|
			t.Fatal(err)
 | 
						|
		}
 | 
						|
		return pod.Labels[apps.ControllerRevisionHashLabelKey] != hash && len(currentPods[nodeName]) == 0
 | 
						|
	})
 | 
						|
 | 
						|
	// the new pods should still be considered waiting to hit min readiness, so one pod should be created to replace
 | 
						|
	// the deleted old pod
 | 
						|
	manager.failedPodsBackoff.Clock = testingclock.NewFakeClock(time.Unix(310, 0))
 | 
						|
	clearExpectations(t, manager, ds, podControl)
 | 
						|
	expectSyncDaemonSets(t, manager, ds, podControl, 1, 0, 0)
 | 
						|
 | 
						|
	// the new pods are now considered available, so delete the old pods
 | 
						|
	manager.failedPodsBackoff.Clock = testingclock.NewFakeClock(time.Unix(320, 0))
 | 
						|
	clearExpectations(t, manager, ds, podControl)
 | 
						|
	expectSyncDaemonSets(t, manager, ds, podControl, 1, 3, 0)
 | 
						|
 | 
						|
	// mark all updated pods as ready at time 320
 | 
						|
	currentPods = podsByNodeMatchingHash(manager, hash)
 | 
						|
	setPodReadiness(t, manager, true, 3, func(pod *v1.Pod) bool {
 | 
						|
		return pod.Labels[apps.ControllerRevisionHashLabelKey] == hash
 | 
						|
	})
 | 
						|
 | 
						|
	// the new pods are now considered available, so delete the old pods
 | 
						|
	manager.failedPodsBackoff.Clock = testingclock.NewFakeClock(time.Unix(340, 0))
 | 
						|
	clearExpectations(t, manager, ds, podControl)
 | 
						|
	expectSyncDaemonSets(t, manager, ds, podControl, 0, 2, 0)
 | 
						|
 | 
						|
	// controller has completed upgrade
 | 
						|
	manager.failedPodsBackoff.Clock = testingclock.NewFakeClock(time.Unix(350, 0))
 | 
						|
	clearExpectations(t, manager, ds, podControl)
 | 
						|
	expectSyncDaemonSets(t, manager, ds, podControl, 0, 0, 0)
 | 
						|
}
 | 
						|
 | 
						|
func podsByNodeMatchingHash(dsc *daemonSetsController, hash string) map[string][]string {
 | 
						|
	byNode := make(map[string][]string)
 | 
						|
	for _, obj := range dsc.podStore.List() {
 | 
						|
		pod := obj.(*v1.Pod)
 | 
						|
		if pod.Labels[apps.ControllerRevisionHashLabelKey] != hash {
 | 
						|
			continue
 | 
						|
		}
 | 
						|
		nodeName, err := util.GetTargetNodeName(pod)
 | 
						|
		if err != nil {
 | 
						|
			panic(err)
 | 
						|
		}
 | 
						|
		byNode[nodeName] = append(byNode[nodeName], pod.Name)
 | 
						|
	}
 | 
						|
	return byNode
 | 
						|
}
 | 
						|
 | 
						|
func setPodReadiness(t *testing.T, dsc *daemonSetsController, ready bool, count int, fn func(*v1.Pod) bool) {
 | 
						|
	t.Helper()
 | 
						|
	logger, _ := ktesting.NewTestContext(t)
 | 
						|
	for _, obj := range dsc.podStore.List() {
 | 
						|
		if count <= 0 {
 | 
						|
			break
 | 
						|
		}
 | 
						|
		pod := obj.(*v1.Pod)
 | 
						|
		if pod.DeletionTimestamp != nil {
 | 
						|
			continue
 | 
						|
		}
 | 
						|
		if podutil.IsPodReady(pod) == ready {
 | 
						|
			continue
 | 
						|
		}
 | 
						|
		if !fn(pod) {
 | 
						|
			continue
 | 
						|
		}
 | 
						|
		condition := v1.PodCondition{Type: v1.PodReady}
 | 
						|
		if ready {
 | 
						|
			condition.Status = v1.ConditionTrue
 | 
						|
		} else {
 | 
						|
			condition.Status = v1.ConditionFalse
 | 
						|
		}
 | 
						|
		if !podutil.UpdatePodCondition(&pod.Status, &condition) {
 | 
						|
			t.Fatal("failed to update pod")
 | 
						|
		}
 | 
						|
		// TODO: workaround UpdatePodCondition calling time.Now() directly
 | 
						|
		setCondition := podutil.GetPodReadyCondition(pod.Status)
 | 
						|
		setCondition.LastTransitionTime.Time = dsc.failedPodsBackoff.Clock.Now()
 | 
						|
		logger.Info("marked pod ready", "pod", pod.Name, "ready", ready)
 | 
						|
		count--
 | 
						|
	}
 | 
						|
	if count > 0 {
 | 
						|
		t.Fatalf("could not mark %d pods ready=%t", count, ready)
 | 
						|
	}
 | 
						|
}
 | 
						|
 | 
						|
func currentDSHash(dsc *daemonSetsController, ds *apps.DaemonSet) (string, error) {
 | 
						|
	// Construct histories of the DaemonSet, and get the hash of current history
 | 
						|
	cur, _, err := dsc.constructHistory(context.TODO(), ds)
 | 
						|
	if err != nil {
 | 
						|
		return "", err
 | 
						|
	}
 | 
						|
	return cur.Labels[apps.DefaultDaemonSetUniqueLabelKey], nil
 | 
						|
 | 
						|
}
 | 
						|
 | 
						|
func TestDaemonSetUpdatesNoTemplateChanged(t *testing.T) {
 | 
						|
	_, ctx := ktesting.NewTestContext(t)
 | 
						|
	ds := newDaemonSet("foo")
 | 
						|
	manager, podControl, _, err := newTestController(ctx, ds)
 | 
						|
	if err != nil {
 | 
						|
		t.Fatalf("error creating DaemonSets controller: %v", err)
 | 
						|
	}
 | 
						|
	maxUnavailable := 3
 | 
						|
	addNodes(manager.nodeStore, 0, 5, nil)
 | 
						|
	manager.dsStore.Add(ds)
 | 
						|
	expectSyncDaemonSets(t, manager, ds, podControl, 5, 0, 0)
 | 
						|
 | 
						|
	ds.Spec.UpdateStrategy.Type = apps.RollingUpdateDaemonSetStrategyType
 | 
						|
	intStr := intstr.FromInt32(int32(maxUnavailable))
 | 
						|
	ds.Spec.UpdateStrategy.RollingUpdate = &apps.RollingUpdateDaemonSet{MaxUnavailable: &intStr}
 | 
						|
	manager.dsStore.Update(ds)
 | 
						|
 | 
						|
	// template is not changed no pod should be removed
 | 
						|
	clearExpectations(t, manager, ds, podControl)
 | 
						|
	expectSyncDaemonSets(t, manager, ds, podControl, 0, 0, 0)
 | 
						|
	clearExpectations(t, manager, ds, podControl)
 | 
						|
}
 | 
						|
 | 
						|
func newUpdateSurge(value intstr.IntOrString) apps.DaemonSetUpdateStrategy {
 | 
						|
	zero := intstr.FromInt32(0)
 | 
						|
	return apps.DaemonSetUpdateStrategy{
 | 
						|
		Type: apps.RollingUpdateDaemonSetStrategyType,
 | 
						|
		RollingUpdate: &apps.RollingUpdateDaemonSet{
 | 
						|
			MaxUnavailable: &zero,
 | 
						|
			MaxSurge:       &value,
 | 
						|
		},
 | 
						|
	}
 | 
						|
}
 | 
						|
 | 
						|
func newUpdateUnavailable(value intstr.IntOrString) apps.DaemonSetUpdateStrategy {
 | 
						|
	return apps.DaemonSetUpdateStrategy{
 | 
						|
		Type: apps.RollingUpdateDaemonSetStrategyType,
 | 
						|
		RollingUpdate: &apps.RollingUpdateDaemonSet{
 | 
						|
			MaxUnavailable: &value,
 | 
						|
		},
 | 
						|
	}
 | 
						|
}
 | 
						|
 | 
						|
func TestGetUnavailableNumbers(t *testing.T) {
 | 
						|
	cases := []struct {
 | 
						|
		name                   string
 | 
						|
		ManagerFunc            func(ctx context.Context) *daemonSetsController
 | 
						|
		ds                     *apps.DaemonSet
 | 
						|
		nodeToPods             map[string][]*v1.Pod
 | 
						|
		maxSurge               int
 | 
						|
		maxUnavailable         int
 | 
						|
		desiredNumberScheduled int
 | 
						|
		emptyNodes             int
 | 
						|
		Err                    error
 | 
						|
	}{
 | 
						|
		{
 | 
						|
			name: "No nodes",
 | 
						|
			ManagerFunc: func(ctx context.Context) *daemonSetsController {
 | 
						|
				manager, _, _, err := newTestController(ctx)
 | 
						|
				if err != nil {
 | 
						|
					t.Fatalf("error creating DaemonSets controller: %v", err)
 | 
						|
				}
 | 
						|
				return manager
 | 
						|
			},
 | 
						|
			ds: func() *apps.DaemonSet {
 | 
						|
				ds := newDaemonSet("x")
 | 
						|
				ds.Spec.UpdateStrategy = newUpdateUnavailable(intstr.FromInt32(0))
 | 
						|
				return ds
 | 
						|
			}(),
 | 
						|
			nodeToPods:     make(map[string][]*v1.Pod),
 | 
						|
			maxUnavailable: 0,
 | 
						|
			emptyNodes:     0,
 | 
						|
		},
 | 
						|
		{
 | 
						|
			name: "Two nodes with ready pods",
 | 
						|
			ManagerFunc: func(ctx context.Context) *daemonSetsController {
 | 
						|
				manager, _, _, err := newTestController(ctx)
 | 
						|
				if err != nil {
 | 
						|
					t.Fatalf("error creating DaemonSets controller: %v", err)
 | 
						|
				}
 | 
						|
				addNodes(manager.nodeStore, 0, 2, nil)
 | 
						|
				return manager
 | 
						|
			},
 | 
						|
			ds: func() *apps.DaemonSet {
 | 
						|
				ds := newDaemonSet("x")
 | 
						|
				ds.Spec.UpdateStrategy = newUpdateUnavailable(intstr.FromInt32(1))
 | 
						|
				return ds
 | 
						|
			}(),
 | 
						|
			nodeToPods: func() map[string][]*v1.Pod {
 | 
						|
				mapping := make(map[string][]*v1.Pod)
 | 
						|
				pod0 := newPod("pod-0", "node-0", simpleDaemonSetLabel, nil)
 | 
						|
				pod1 := newPod("pod-1", "node-1", simpleDaemonSetLabel, nil)
 | 
						|
				markPodReady(pod0)
 | 
						|
				markPodReady(pod1)
 | 
						|
				mapping["node-0"] = []*v1.Pod{pod0}
 | 
						|
				mapping["node-1"] = []*v1.Pod{pod1}
 | 
						|
				return mapping
 | 
						|
			}(),
 | 
						|
			maxUnavailable:         1,
 | 
						|
			desiredNumberScheduled: 2,
 | 
						|
			emptyNodes:             0,
 | 
						|
		},
 | 
						|
		{
 | 
						|
			name: "Two nodes, one node without pods",
 | 
						|
			ManagerFunc: func(ctx context.Context) *daemonSetsController {
 | 
						|
				manager, _, _, err := newTestController(ctx)
 | 
						|
				if err != nil {
 | 
						|
					t.Fatalf("error creating DaemonSets controller: %v", err)
 | 
						|
				}
 | 
						|
				addNodes(manager.nodeStore, 0, 2, nil)
 | 
						|
				return manager
 | 
						|
			},
 | 
						|
			ds: func() *apps.DaemonSet {
 | 
						|
				ds := newDaemonSet("x")
 | 
						|
				ds.Spec.UpdateStrategy = newUpdateUnavailable(intstr.FromInt32(0))
 | 
						|
				return ds
 | 
						|
			}(),
 | 
						|
			nodeToPods: func() map[string][]*v1.Pod {
 | 
						|
				mapping := make(map[string][]*v1.Pod)
 | 
						|
				pod0 := newPod("pod-0", "node-0", simpleDaemonSetLabel, nil)
 | 
						|
				markPodReady(pod0)
 | 
						|
				mapping["node-0"] = []*v1.Pod{pod0}
 | 
						|
				return mapping
 | 
						|
			}(),
 | 
						|
			maxUnavailable:         1,
 | 
						|
			desiredNumberScheduled: 2,
 | 
						|
			emptyNodes:             1,
 | 
						|
		},
 | 
						|
		{
 | 
						|
			name: "Two nodes, one node without pods, surge",
 | 
						|
			ManagerFunc: func(ctx context.Context) *daemonSetsController {
 | 
						|
				manager, _, _, err := newTestController(ctx)
 | 
						|
				if err != nil {
 | 
						|
					t.Fatalf("error creating DaemonSets controller: %v", err)
 | 
						|
				}
 | 
						|
				addNodes(manager.nodeStore, 0, 2, nil)
 | 
						|
				return manager
 | 
						|
			},
 | 
						|
			ds: func() *apps.DaemonSet {
 | 
						|
				ds := newDaemonSet("x")
 | 
						|
				ds.Spec.UpdateStrategy = newUpdateSurge(intstr.FromInt32(0))
 | 
						|
				return ds
 | 
						|
			}(),
 | 
						|
			nodeToPods: func() map[string][]*v1.Pod {
 | 
						|
				mapping := make(map[string][]*v1.Pod)
 | 
						|
				pod0 := newPod("pod-0", "node-0", simpleDaemonSetLabel, nil)
 | 
						|
				markPodReady(pod0)
 | 
						|
				mapping["node-0"] = []*v1.Pod{pod0}
 | 
						|
				return mapping
 | 
						|
			}(),
 | 
						|
			maxUnavailable:         1,
 | 
						|
			desiredNumberScheduled: 2,
 | 
						|
			emptyNodes:             1,
 | 
						|
		},
 | 
						|
		{
 | 
						|
			name: "Two nodes with pods, MaxUnavailable in percents",
 | 
						|
			ManagerFunc: func(ctx context.Context) *daemonSetsController {
 | 
						|
				manager, _, _, err := newTestController(ctx)
 | 
						|
				if err != nil {
 | 
						|
					t.Fatalf("error creating DaemonSets controller: %v", err)
 | 
						|
				}
 | 
						|
				addNodes(manager.nodeStore, 0, 2, nil)
 | 
						|
				return manager
 | 
						|
			},
 | 
						|
			ds: func() *apps.DaemonSet {
 | 
						|
				ds := newDaemonSet("x")
 | 
						|
				ds.Spec.UpdateStrategy = newUpdateUnavailable(intstr.FromString("50%"))
 | 
						|
				return ds
 | 
						|
			}(),
 | 
						|
			nodeToPods: func() map[string][]*v1.Pod {
 | 
						|
				mapping := make(map[string][]*v1.Pod)
 | 
						|
				pod0 := newPod("pod-0", "node-0", simpleDaemonSetLabel, nil)
 | 
						|
				pod1 := newPod("pod-1", "node-1", simpleDaemonSetLabel, nil)
 | 
						|
				markPodReady(pod0)
 | 
						|
				markPodReady(pod1)
 | 
						|
				mapping["node-0"] = []*v1.Pod{pod0}
 | 
						|
				mapping["node-1"] = []*v1.Pod{pod1}
 | 
						|
				return mapping
 | 
						|
			}(),
 | 
						|
			maxUnavailable:         1,
 | 
						|
			desiredNumberScheduled: 2,
 | 
						|
			emptyNodes:             0,
 | 
						|
		},
 | 
						|
		{
 | 
						|
			name: "Two nodes with pods, MaxUnavailable in percents, surge",
 | 
						|
			ManagerFunc: func(ctx context.Context) *daemonSetsController {
 | 
						|
				manager, _, _, err := newTestController(ctx)
 | 
						|
				if err != nil {
 | 
						|
					t.Fatalf("error creating DaemonSets controller: %v", err)
 | 
						|
				}
 | 
						|
				addNodes(manager.nodeStore, 0, 2, nil)
 | 
						|
				return manager
 | 
						|
			},
 | 
						|
			ds: func() *apps.DaemonSet {
 | 
						|
				ds := newDaemonSet("x")
 | 
						|
				ds.Spec.UpdateStrategy = newUpdateSurge(intstr.FromString("50%"))
 | 
						|
				return ds
 | 
						|
			}(),
 | 
						|
			nodeToPods: func() map[string][]*v1.Pod {
 | 
						|
				mapping := make(map[string][]*v1.Pod)
 | 
						|
				pod0 := newPod("pod-0", "node-0", simpleDaemonSetLabel, nil)
 | 
						|
				pod1 := newPod("pod-1", "node-1", simpleDaemonSetLabel, nil)
 | 
						|
				markPodReady(pod0)
 | 
						|
				markPodReady(pod1)
 | 
						|
				mapping["node-0"] = []*v1.Pod{pod0}
 | 
						|
				mapping["node-1"] = []*v1.Pod{pod1}
 | 
						|
				return mapping
 | 
						|
			}(),
 | 
						|
			maxSurge:               1,
 | 
						|
			maxUnavailable:         0,
 | 
						|
			desiredNumberScheduled: 2,
 | 
						|
			emptyNodes:             0,
 | 
						|
		},
 | 
						|
		{
 | 
						|
			name: "Two nodes with pods, MaxUnavailable is 100%, surge",
 | 
						|
			ManagerFunc: func(ctx context.Context) *daemonSetsController {
 | 
						|
				manager, _, _, err := newTestController(ctx)
 | 
						|
				if err != nil {
 | 
						|
					t.Fatalf("error creating DaemonSets controller: %v", err)
 | 
						|
				}
 | 
						|
				addNodes(manager.nodeStore, 0, 2, nil)
 | 
						|
				return manager
 | 
						|
			},
 | 
						|
			ds: func() *apps.DaemonSet {
 | 
						|
				ds := newDaemonSet("x")
 | 
						|
				ds.Spec.UpdateStrategy = newUpdateSurge(intstr.FromString("100%"))
 | 
						|
				return ds
 | 
						|
			}(),
 | 
						|
			nodeToPods: func() map[string][]*v1.Pod {
 | 
						|
				mapping := make(map[string][]*v1.Pod)
 | 
						|
				pod0 := newPod("pod-0", "node-0", simpleDaemonSetLabel, nil)
 | 
						|
				pod1 := newPod("pod-1", "node-1", simpleDaemonSetLabel, nil)
 | 
						|
				markPodReady(pod0)
 | 
						|
				markPodReady(pod1)
 | 
						|
				mapping["node-0"] = []*v1.Pod{pod0}
 | 
						|
				mapping["node-1"] = []*v1.Pod{pod1}
 | 
						|
				return mapping
 | 
						|
			}(),
 | 
						|
			maxSurge:               2,
 | 
						|
			maxUnavailable:         0,
 | 
						|
			desiredNumberScheduled: 2,
 | 
						|
			emptyNodes:             0,
 | 
						|
		},
 | 
						|
		{
 | 
						|
			name: "Two nodes with pods, MaxUnavailable in percents, pod terminating",
 | 
						|
			ManagerFunc: func(ctx context.Context) *daemonSetsController {
 | 
						|
				manager, _, _, err := newTestController(ctx)
 | 
						|
				if err != nil {
 | 
						|
					t.Fatalf("error creating DaemonSets controller: %v", err)
 | 
						|
				}
 | 
						|
				addNodes(manager.nodeStore, 0, 3, nil)
 | 
						|
				return manager
 | 
						|
			},
 | 
						|
			ds: func() *apps.DaemonSet {
 | 
						|
				ds := newDaemonSet("x")
 | 
						|
				ds.Spec.UpdateStrategy = newUpdateUnavailable(intstr.FromString("50%"))
 | 
						|
				return ds
 | 
						|
			}(),
 | 
						|
			nodeToPods: func() map[string][]*v1.Pod {
 | 
						|
				mapping := make(map[string][]*v1.Pod)
 | 
						|
				pod0 := newPod("pod-0", "node-0", simpleDaemonSetLabel, nil)
 | 
						|
				pod1 := newPod("pod-1", "node-1", simpleDaemonSetLabel, nil)
 | 
						|
				now := metav1.Now()
 | 
						|
				markPodReady(pod0)
 | 
						|
				markPodReady(pod1)
 | 
						|
				pod1.DeletionTimestamp = &now
 | 
						|
				mapping["node-0"] = []*v1.Pod{pod0}
 | 
						|
				mapping["node-1"] = []*v1.Pod{pod1}
 | 
						|
				return mapping
 | 
						|
			}(),
 | 
						|
			maxUnavailable:         2,
 | 
						|
			desiredNumberScheduled: 3,
 | 
						|
			emptyNodes:             1,
 | 
						|
		},
 | 
						|
	}
 | 
						|
 | 
						|
	for _, c := range cases {
 | 
						|
		t.Run(c.name, func(t *testing.T) {
 | 
						|
			_, ctx := ktesting.NewTestContext(t)
 | 
						|
			manager := c.ManagerFunc(ctx)
 | 
						|
			manager.dsStore.Add(c.ds)
 | 
						|
			nodeList, err := manager.nodeLister.List(labels.Everything())
 | 
						|
			if err != nil {
 | 
						|
				t.Fatalf("error listing nodes: %v", err)
 | 
						|
			}
 | 
						|
			maxSurge, maxUnavailable, desiredNumberScheduled, err := manager.updatedDesiredNodeCounts(ctx, c.ds, nodeList, c.nodeToPods)
 | 
						|
			if err != nil && c.Err != nil {
 | 
						|
				if c.Err != err {
 | 
						|
					t.Fatalf("Expected error: %v but got: %v", c.Err, err)
 | 
						|
				}
 | 
						|
			}
 | 
						|
			if err != nil {
 | 
						|
				t.Fatalf("Unexpected error: %v", err)
 | 
						|
			}
 | 
						|
			if maxSurge != c.maxSurge || maxUnavailable != c.maxUnavailable || desiredNumberScheduled != c.desiredNumberScheduled {
 | 
						|
				t.Errorf("Wrong values. maxSurge: %d, expected %d, maxUnavailable: %d, expected: %d, desiredNumberScheduled: %d, expected: %d", maxSurge, c.maxSurge, maxUnavailable, c.maxUnavailable, desiredNumberScheduled, c.desiredNumberScheduled)
 | 
						|
			}
 | 
						|
			var emptyNodes int
 | 
						|
			for _, pods := range c.nodeToPods {
 | 
						|
				if len(pods) == 0 {
 | 
						|
					emptyNodes++
 | 
						|
				}
 | 
						|
			}
 | 
						|
			if emptyNodes != c.emptyNodes {
 | 
						|
				t.Errorf("expected numEmpty to be %d, was %d", c.emptyNodes, emptyNodes)
 | 
						|
			}
 | 
						|
		})
 | 
						|
	}
 | 
						|
}
 | 
						|
 | 
						|
func TestControlledHistories(t *testing.T) {
 | 
						|
	ds1 := newDaemonSet("ds1")
 | 
						|
	crOfDs1 := newControllerRevision(ds1.GetName()+"-x1", ds1.GetNamespace(), ds1.Spec.Template.Labels,
 | 
						|
		[]metav1.OwnerReference{*metav1.NewControllerRef(ds1, controllerKind)})
 | 
						|
	orphanCrInSameNsWithDs1 := newControllerRevision(ds1.GetName()+"-x2", ds1.GetNamespace(), ds1.Spec.Template.Labels, nil)
 | 
						|
	orphanCrNotInSameNsWithDs1 := newControllerRevision(ds1.GetName()+"-x3", ds1.GetNamespace()+"-other", ds1.Spec.Template.Labels, nil)
 | 
						|
	cases := []struct {
 | 
						|
		name                      string
 | 
						|
		managerFunc               func(ctx context.Context) *daemonSetsController
 | 
						|
		historyCRAll              []*apps.ControllerRevision
 | 
						|
		expectControllerRevisions []*apps.ControllerRevision
 | 
						|
	}{
 | 
						|
		{
 | 
						|
			name: "controller revision in the same namespace",
 | 
						|
			managerFunc: func(ctx context.Context) *daemonSetsController {
 | 
						|
				manager, _, _, err := newTestController(ctx, ds1, crOfDs1, orphanCrInSameNsWithDs1)
 | 
						|
				if err != nil {
 | 
						|
					t.Fatalf("error creating DaemonSets controller: %v", err)
 | 
						|
				}
 | 
						|
				manager.dsStore.Add(ds1)
 | 
						|
				return manager
 | 
						|
			},
 | 
						|
			historyCRAll:              []*apps.ControllerRevision{crOfDs1, orphanCrInSameNsWithDs1},
 | 
						|
			expectControllerRevisions: []*apps.ControllerRevision{crOfDs1, orphanCrInSameNsWithDs1},
 | 
						|
		},
 | 
						|
		{
 | 
						|
			name: "Skip adopting the controller revision in namespace other than the one in which DS lives",
 | 
						|
			managerFunc: func(ctx context.Context) *daemonSetsController {
 | 
						|
				manager, _, _, err := newTestController(ctx, ds1, orphanCrNotInSameNsWithDs1)
 | 
						|
				if err != nil {
 | 
						|
					t.Fatalf("error creating DaemonSets controller: %v", err)
 | 
						|
				}
 | 
						|
				manager.dsStore.Add(ds1)
 | 
						|
				return manager
 | 
						|
			},
 | 
						|
			historyCRAll:              []*apps.ControllerRevision{orphanCrNotInSameNsWithDs1},
 | 
						|
			expectControllerRevisions: []*apps.ControllerRevision{},
 | 
						|
		},
 | 
						|
	}
 | 
						|
	for _, c := range cases {
 | 
						|
		_, ctx := ktesting.NewTestContext(t)
 | 
						|
		manager := c.managerFunc(ctx)
 | 
						|
		for _, h := range c.historyCRAll {
 | 
						|
			manager.historyStore.Add(h)
 | 
						|
		}
 | 
						|
		crList, err := manager.controlledHistories(context.TODO(), ds1)
 | 
						|
		if err != nil {
 | 
						|
			t.Fatalf("Test case: %s. Unexpected error: %v", c.name, err)
 | 
						|
		}
 | 
						|
		if len(crList) != len(c.expectControllerRevisions) {
 | 
						|
			t.Errorf("Test case: %s, expect controllerrevision count %d but got:%d",
 | 
						|
				c.name, len(c.expectControllerRevisions), len(crList))
 | 
						|
		} else {
 | 
						|
			// check controller revisions match
 | 
						|
			for _, cr := range crList {
 | 
						|
				found := false
 | 
						|
				for _, expectCr := range c.expectControllerRevisions {
 | 
						|
					if reflect.DeepEqual(cr, expectCr) {
 | 
						|
						found = true
 | 
						|
						break
 | 
						|
					}
 | 
						|
				}
 | 
						|
				if !found {
 | 
						|
					t.Errorf("Test case: %s, controllerrevision %v not expected",
 | 
						|
						c.name, cr)
 | 
						|
				}
 | 
						|
			}
 | 
						|
			t.Logf("Test case: %s done", c.name)
 | 
						|
		}
 | 
						|
	}
 | 
						|
}
 |