mirror of
				https://github.com/optim-enterprises-bv/kubernetes.git
				synced 2025-11-04 12:18:16 +00:00 
			
		
		
		
	Signed-off-by: Ashutosh Kumar <sonasingh46@gmail.com> Co-authored-by: Ashutosh Kumar <sonasingh46@gmail.com> Co-authored-by: xing-yang <xingyang105@gmail.com>
		
			
				
	
	
		
			605 lines
		
	
	
		
			20 KiB
		
	
	
	
		
			Go
		
	
	
	
	
	
			
		
		
	
	
			605 lines
		
	
	
		
			20 KiB
		
	
	
	
		
			Go
		
	
	
	
	
	
/*
 | 
						|
Copyright 2015 The Kubernetes Authors.
 | 
						|
 | 
						|
Licensed under the Apache License, Version 2.0 (the "License");
 | 
						|
you may not use this file except in compliance with the License.
 | 
						|
You may obtain a copy of the License at
 | 
						|
 | 
						|
    http://www.apache.org/licenses/LICENSE-2.0
 | 
						|
 | 
						|
Unless required by applicable law or agreed to in writing, software
 | 
						|
distributed under the License is distributed on an "AS IS" BASIS,
 | 
						|
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
 | 
						|
See the License for the specific language governing permissions and
 | 
						|
limitations under the License.
 | 
						|
*/
 | 
						|
 | 
						|
package podgc
 | 
						|
 | 
						|
import (
 | 
						|
	"context"
 | 
						|
	"sync"
 | 
						|
	"testing"
 | 
						|
	"time"
 | 
						|
 | 
						|
	v1 "k8s.io/api/core/v1"
 | 
						|
	metav1 "k8s.io/apimachinery/pkg/apis/meta/v1"
 | 
						|
	"k8s.io/apimachinery/pkg/labels"
 | 
						|
	"k8s.io/apimachinery/pkg/util/sets"
 | 
						|
	"k8s.io/apimachinery/pkg/util/wait"
 | 
						|
	utilfeature "k8s.io/apiserver/pkg/util/feature"
 | 
						|
	"k8s.io/client-go/informers"
 | 
						|
	coreinformers "k8s.io/client-go/informers/core/v1"
 | 
						|
	clientset "k8s.io/client-go/kubernetes"
 | 
						|
	"k8s.io/client-go/kubernetes/fake"
 | 
						|
	"k8s.io/client-go/util/workqueue"
 | 
						|
	featuregatetesting "k8s.io/component-base/featuregate/testing"
 | 
						|
	"k8s.io/kubernetes/pkg/controller"
 | 
						|
	"k8s.io/kubernetes/pkg/controller/testutil"
 | 
						|
	"k8s.io/kubernetes/pkg/features"
 | 
						|
	testingclock "k8s.io/utils/clock/testing"
 | 
						|
)
 | 
						|
 | 
						|
func alwaysReady() bool { return true }
 | 
						|
 | 
						|
func NewFromClient(kubeClient clientset.Interface, terminatedPodThreshold int) (*PodGCController, coreinformers.PodInformer, coreinformers.NodeInformer) {
 | 
						|
	informerFactory := informers.NewSharedInformerFactory(kubeClient, controller.NoResyncPeriodFunc())
 | 
						|
	podInformer := informerFactory.Core().V1().Pods()
 | 
						|
	nodeInformer := informerFactory.Core().V1().Nodes()
 | 
						|
	controller := NewPodGC(context.TODO(), kubeClient, podInformer, nodeInformer, terminatedPodThreshold)
 | 
						|
	controller.podListerSynced = alwaysReady
 | 
						|
	return controller, podInformer, nodeInformer
 | 
						|
}
 | 
						|
 | 
						|
func compareStringSetToList(set sets.String, list []string) bool {
 | 
						|
	for _, item := range list {
 | 
						|
		if !set.Has(item) {
 | 
						|
			return false
 | 
						|
		}
 | 
						|
	}
 | 
						|
	if len(list) != len(set) {
 | 
						|
		return false
 | 
						|
	}
 | 
						|
	return true
 | 
						|
}
 | 
						|
 | 
						|
func TestGCTerminated(t *testing.T) {
 | 
						|
	type nameToPhase struct {
 | 
						|
		name  string
 | 
						|
		phase v1.PodPhase
 | 
						|
	}
 | 
						|
 | 
						|
	testCases := []struct {
 | 
						|
		name            string
 | 
						|
		pods            []nameToPhase
 | 
						|
		threshold       int
 | 
						|
		deletedPodNames sets.String
 | 
						|
	}{
 | 
						|
		{
 | 
						|
			name: "threshold = 0, disables terminated pod deletion",
 | 
						|
			pods: []nameToPhase{
 | 
						|
				{name: "a", phase: v1.PodFailed},
 | 
						|
				{name: "b", phase: v1.PodSucceeded},
 | 
						|
			},
 | 
						|
			threshold: 0,
 | 
						|
			// threshold = 0 disables terminated pod deletion
 | 
						|
			deletedPodNames: sets.NewString(),
 | 
						|
		},
 | 
						|
		{
 | 
						|
			name: "threshold = 1, delete pod a which is PodFailed and pod b which is PodSucceeded",
 | 
						|
			pods: []nameToPhase{
 | 
						|
				{name: "a", phase: v1.PodFailed},
 | 
						|
				{name: "b", phase: v1.PodSucceeded},
 | 
						|
				{name: "c", phase: v1.PodFailed},
 | 
						|
			},
 | 
						|
			threshold:       1,
 | 
						|
			deletedPodNames: sets.NewString("a", "b"),
 | 
						|
		},
 | 
						|
		{
 | 
						|
			name: "threshold = 1, delete pod b which is PodSucceeded",
 | 
						|
			pods: []nameToPhase{
 | 
						|
				{name: "a", phase: v1.PodRunning},
 | 
						|
				{name: "b", phase: v1.PodSucceeded},
 | 
						|
				{name: "c", phase: v1.PodFailed},
 | 
						|
			},
 | 
						|
			threshold:       1,
 | 
						|
			deletedPodNames: sets.NewString("b"),
 | 
						|
		},
 | 
						|
		{
 | 
						|
			name: "threshold = 1, delete pod a which is PodFailed",
 | 
						|
			pods: []nameToPhase{
 | 
						|
				{name: "a", phase: v1.PodFailed},
 | 
						|
				{name: "b", phase: v1.PodSucceeded},
 | 
						|
			},
 | 
						|
			threshold:       1,
 | 
						|
			deletedPodNames: sets.NewString("a"),
 | 
						|
		},
 | 
						|
		{
 | 
						|
			name: "threshold = 5, don't delete pod",
 | 
						|
			pods: []nameToPhase{
 | 
						|
				{name: "a", phase: v1.PodFailed},
 | 
						|
				{name: "b", phase: v1.PodSucceeded},
 | 
						|
			},
 | 
						|
			threshold:       5,
 | 
						|
			deletedPodNames: sets.NewString(),
 | 
						|
		},
 | 
						|
	}
 | 
						|
 | 
						|
	for i, test := range testCases {
 | 
						|
		t.Run(test.name, func(t *testing.T) {
 | 
						|
			client := fake.NewSimpleClientset(&v1.NodeList{Items: []v1.Node{*testutil.NewNode("node")}})
 | 
						|
			gcc, podInformer, _ := NewFromClient(client, test.threshold)
 | 
						|
			deletedPodNames := make([]string, 0)
 | 
						|
			var lock sync.Mutex
 | 
						|
			gcc.deletePod = func(_, name string) error {
 | 
						|
				lock.Lock()
 | 
						|
				defer lock.Unlock()
 | 
						|
				deletedPodNames = append(deletedPodNames, name)
 | 
						|
				return nil
 | 
						|
			}
 | 
						|
 | 
						|
			creationTime := time.Unix(0, 0)
 | 
						|
			for _, pod := range test.pods {
 | 
						|
				creationTime = creationTime.Add(1 * time.Hour)
 | 
						|
				podInformer.Informer().GetStore().Add(&v1.Pod{
 | 
						|
					ObjectMeta: metav1.ObjectMeta{Name: pod.name, CreationTimestamp: metav1.Time{Time: creationTime}},
 | 
						|
					Status:     v1.PodStatus{Phase: pod.phase},
 | 
						|
					Spec:       v1.PodSpec{NodeName: "node"},
 | 
						|
				})
 | 
						|
			}
 | 
						|
 | 
						|
			gcc.gc(context.TODO())
 | 
						|
 | 
						|
			if pass := compareStringSetToList(test.deletedPodNames, deletedPodNames); !pass {
 | 
						|
				t.Errorf("[%v]pod's deleted expected and actual did not match.\n\texpected: %v\n\tactual: %v",
 | 
						|
					i, test.deletedPodNames.List(), deletedPodNames)
 | 
						|
			}
 | 
						|
		})
 | 
						|
	}
 | 
						|
}
 | 
						|
 | 
						|
func makePod(name string, nodeName string, phase v1.PodPhase) *v1.Pod {
 | 
						|
	return &v1.Pod{
 | 
						|
		ObjectMeta: metav1.ObjectMeta{
 | 
						|
			Name: name,
 | 
						|
		},
 | 
						|
		Spec:   v1.PodSpec{NodeName: nodeName},
 | 
						|
		Status: v1.PodStatus{Phase: phase},
 | 
						|
	}
 | 
						|
}
 | 
						|
 | 
						|
func waitForAdded(q workqueue.DelayingInterface, depth int) error {
 | 
						|
	return wait.Poll(1*time.Millisecond, 10*time.Second, func() (done bool, err error) {
 | 
						|
		if q.Len() == depth {
 | 
						|
			return true, nil
 | 
						|
		}
 | 
						|
 | 
						|
		return false, nil
 | 
						|
	})
 | 
						|
}
 | 
						|
 | 
						|
func TestGCOrphaned(t *testing.T) {
 | 
						|
	testCases := []struct {
 | 
						|
		name                 string
 | 
						|
		initialClientNodes   []*v1.Node
 | 
						|
		initialInformerNodes []*v1.Node
 | 
						|
		delay                time.Duration
 | 
						|
		addedClientNodes     []*v1.Node
 | 
						|
		deletedClientNodes   []*v1.Node
 | 
						|
		addedInformerNodes   []*v1.Node
 | 
						|
		deletedInformerNodes []*v1.Node
 | 
						|
		pods                 []*v1.Pod
 | 
						|
		itemsInQueue         int
 | 
						|
		deletedPodNames      sets.String
 | 
						|
	}{
 | 
						|
		{
 | 
						|
			name: "nodes present in lister",
 | 
						|
			initialInformerNodes: []*v1.Node{
 | 
						|
				testutil.NewNode("existing1"),
 | 
						|
				testutil.NewNode("existing2"),
 | 
						|
			},
 | 
						|
			delay: 2 * quarantineTime,
 | 
						|
			pods: []*v1.Pod{
 | 
						|
				makePod("a", "existing1", v1.PodRunning),
 | 
						|
				makePod("b", "existing2", v1.PodFailed),
 | 
						|
				makePod("c", "existing2", v1.PodSucceeded),
 | 
						|
			},
 | 
						|
			itemsInQueue:    0,
 | 
						|
			deletedPodNames: sets.NewString(),
 | 
						|
		},
 | 
						|
		{
 | 
						|
			name: "nodes present in client",
 | 
						|
			initialClientNodes: []*v1.Node{
 | 
						|
				testutil.NewNode("existing1"),
 | 
						|
				testutil.NewNode("existing2"),
 | 
						|
			},
 | 
						|
			delay: 2 * quarantineTime,
 | 
						|
			pods: []*v1.Pod{
 | 
						|
				makePod("a", "existing1", v1.PodRunning),
 | 
						|
				makePod("b", "existing2", v1.PodFailed),
 | 
						|
				makePod("c", "existing2", v1.PodSucceeded),
 | 
						|
			},
 | 
						|
			itemsInQueue:    2,
 | 
						|
			deletedPodNames: sets.NewString(),
 | 
						|
		},
 | 
						|
		{
 | 
						|
			name:  "no nodes",
 | 
						|
			delay: 2 * quarantineTime,
 | 
						|
			pods: []*v1.Pod{
 | 
						|
				makePod("a", "deleted", v1.PodFailed),
 | 
						|
				makePod("b", "deleted", v1.PodSucceeded),
 | 
						|
			},
 | 
						|
			itemsInQueue:    1,
 | 
						|
			deletedPodNames: sets.NewString("a", "b"),
 | 
						|
		},
 | 
						|
		{
 | 
						|
			name:  "quarantine not finished",
 | 
						|
			delay: quarantineTime / 2,
 | 
						|
			pods: []*v1.Pod{
 | 
						|
				makePod("a", "deleted", v1.PodFailed),
 | 
						|
			},
 | 
						|
			itemsInQueue:    0,
 | 
						|
			deletedPodNames: sets.NewString(),
 | 
						|
		},
 | 
						|
		{
 | 
						|
			name:                 "wrong nodes",
 | 
						|
			initialInformerNodes: []*v1.Node{testutil.NewNode("existing")},
 | 
						|
			delay:                2 * quarantineTime,
 | 
						|
			pods: []*v1.Pod{
 | 
						|
				makePod("a", "deleted", v1.PodRunning),
 | 
						|
			},
 | 
						|
			itemsInQueue:    1,
 | 
						|
			deletedPodNames: sets.NewString("a"),
 | 
						|
		},
 | 
						|
		{
 | 
						|
			name:                 "some nodes missing",
 | 
						|
			initialInformerNodes: []*v1.Node{testutil.NewNode("existing")},
 | 
						|
			delay:                2 * quarantineTime,
 | 
						|
			pods: []*v1.Pod{
 | 
						|
				makePod("a", "deleted", v1.PodFailed),
 | 
						|
				makePod("b", "existing", v1.PodFailed),
 | 
						|
				makePod("c", "deleted", v1.PodSucceeded),
 | 
						|
				makePod("d", "deleted", v1.PodRunning),
 | 
						|
			},
 | 
						|
			itemsInQueue:    1,
 | 
						|
			deletedPodNames: sets.NewString("a", "c", "d"),
 | 
						|
		},
 | 
						|
		{
 | 
						|
			name:             "node added to client after quarantine",
 | 
						|
			delay:            2 * quarantineTime,
 | 
						|
			addedClientNodes: []*v1.Node{testutil.NewNode("node")},
 | 
						|
			pods: []*v1.Pod{
 | 
						|
				makePod("a", "node", v1.PodRunning),
 | 
						|
			},
 | 
						|
			itemsInQueue:    1,
 | 
						|
			deletedPodNames: sets.NewString(),
 | 
						|
		},
 | 
						|
		{
 | 
						|
			name:               "node added to informer after quarantine",
 | 
						|
			delay:              2 * quarantineTime,
 | 
						|
			addedInformerNodes: []*v1.Node{testutil.NewNode("node")},
 | 
						|
			pods: []*v1.Pod{
 | 
						|
				makePod("a", "node", v1.PodFailed),
 | 
						|
			},
 | 
						|
			itemsInQueue:    1,
 | 
						|
			deletedPodNames: sets.NewString(),
 | 
						|
		},
 | 
						|
		{
 | 
						|
			// It shouldn't happen that client will be lagging behind informer.
 | 
						|
			// This test case is more a sanity check.
 | 
						|
			name:               "node deleted from client after quarantine",
 | 
						|
			initialClientNodes: []*v1.Node{testutil.NewNode("node")},
 | 
						|
			delay:              2 * quarantineTime,
 | 
						|
			deletedClientNodes: []*v1.Node{testutil.NewNode("node")},
 | 
						|
			pods: []*v1.Pod{
 | 
						|
				makePod("a", "node", v1.PodFailed),
 | 
						|
			},
 | 
						|
			itemsInQueue:    1,
 | 
						|
			deletedPodNames: sets.NewString("a"),
 | 
						|
		},
 | 
						|
		{
 | 
						|
			name:                 "node deleted from informer after quarantine",
 | 
						|
			initialInformerNodes: []*v1.Node{testutil.NewNode("node")},
 | 
						|
			delay:                2 * quarantineTime,
 | 
						|
			deletedInformerNodes: []*v1.Node{testutil.NewNode("node")},
 | 
						|
			pods: []*v1.Pod{
 | 
						|
				makePod("a", "node", v1.PodSucceeded),
 | 
						|
			},
 | 
						|
			itemsInQueue:    0,
 | 
						|
			deletedPodNames: sets.NewString(),
 | 
						|
		},
 | 
						|
	}
 | 
						|
 | 
						|
	for _, test := range testCases {
 | 
						|
		t.Run(test.name, func(t *testing.T) {
 | 
						|
			nodeList := &v1.NodeList{}
 | 
						|
			for _, node := range test.initialClientNodes {
 | 
						|
				nodeList.Items = append(nodeList.Items, *node)
 | 
						|
			}
 | 
						|
			client := fake.NewSimpleClientset(nodeList)
 | 
						|
			gcc, podInformer, nodeInformer := NewFromClient(client, -1)
 | 
						|
			for _, node := range test.initialInformerNodes {
 | 
						|
				nodeInformer.Informer().GetStore().Add(node)
 | 
						|
			}
 | 
						|
			for _, pod := range test.pods {
 | 
						|
				podInformer.Informer().GetStore().Add(pod)
 | 
						|
			}
 | 
						|
			// Overwrite queue
 | 
						|
			fakeClock := testingclock.NewFakeClock(time.Now())
 | 
						|
			gcc.nodeQueue.ShutDown()
 | 
						|
			gcc.nodeQueue = workqueue.NewDelayingQueueWithCustomClock(fakeClock, "podgc_test_queue")
 | 
						|
 | 
						|
			deletedPodNames := make([]string, 0)
 | 
						|
			var lock sync.Mutex
 | 
						|
			gcc.deletePod = func(_, name string) error {
 | 
						|
				lock.Lock()
 | 
						|
				defer lock.Unlock()
 | 
						|
				deletedPodNames = append(deletedPodNames, name)
 | 
						|
				return nil
 | 
						|
			}
 | 
						|
 | 
						|
			// First GC of orphaned pods
 | 
						|
			gcc.gc(context.TODO())
 | 
						|
			if len(deletedPodNames) > 0 {
 | 
						|
				t.Errorf("no pods should be deleted at this point.\n\tactual: %v", deletedPodNames)
 | 
						|
			}
 | 
						|
 | 
						|
			// Move clock forward
 | 
						|
			fakeClock.Step(test.delay)
 | 
						|
			// Wait for queue goroutine to process items
 | 
						|
			if test.itemsInQueue > 0 {
 | 
						|
				err := waitForAdded(gcc.nodeQueue, test.itemsInQueue)
 | 
						|
				if err != nil {
 | 
						|
					t.Errorf("wrong number of items in the node queue.\n\texpected: %v\n\tactual: %v",
 | 
						|
						test.itemsInQueue, gcc.nodeQueue.Len())
 | 
						|
				}
 | 
						|
			}
 | 
						|
 | 
						|
			// Execute planned nodes changes
 | 
						|
			for _, node := range test.addedClientNodes {
 | 
						|
				client.CoreV1().Nodes().Create(context.TODO(), node, metav1.CreateOptions{})
 | 
						|
			}
 | 
						|
			for _, node := range test.deletedClientNodes {
 | 
						|
				client.CoreV1().Nodes().Delete(context.TODO(), node.Name, metav1.DeleteOptions{})
 | 
						|
			}
 | 
						|
			for _, node := range test.addedInformerNodes {
 | 
						|
				nodeInformer.Informer().GetStore().Add(node)
 | 
						|
			}
 | 
						|
			for _, node := range test.deletedInformerNodes {
 | 
						|
				nodeInformer.Informer().GetStore().Delete(node)
 | 
						|
			}
 | 
						|
 | 
						|
			// Actual pod deletion
 | 
						|
			gcc.gc(context.TODO())
 | 
						|
 | 
						|
			if pass := compareStringSetToList(test.deletedPodNames, deletedPodNames); !pass {
 | 
						|
				t.Errorf("pod's deleted expected and actual did not match.\n\texpected: %v\n\tactual: %v",
 | 
						|
					test.deletedPodNames.List(), deletedPodNames)
 | 
						|
			}
 | 
						|
		})
 | 
						|
	}
 | 
						|
}
 | 
						|
 | 
						|
func TestGCUnscheduledTerminating(t *testing.T) {
 | 
						|
	type nameToPhase struct {
 | 
						|
		name              string
 | 
						|
		phase             v1.PodPhase
 | 
						|
		deletionTimeStamp *metav1.Time
 | 
						|
		nodeName          string
 | 
						|
	}
 | 
						|
 | 
						|
	testCases := []struct {
 | 
						|
		name            string
 | 
						|
		pods            []nameToPhase
 | 
						|
		deletedPodNames sets.String
 | 
						|
	}{
 | 
						|
		{
 | 
						|
			name: "Unscheduled pod in any phase must be deleted",
 | 
						|
			pods: []nameToPhase{
 | 
						|
				{name: "a", phase: v1.PodFailed, deletionTimeStamp: &metav1.Time{}, nodeName: ""},
 | 
						|
				{name: "b", phase: v1.PodSucceeded, deletionTimeStamp: &metav1.Time{}, nodeName: ""},
 | 
						|
				{name: "c", phase: v1.PodRunning, deletionTimeStamp: &metav1.Time{}, nodeName: ""},
 | 
						|
			},
 | 
						|
			deletedPodNames: sets.NewString("a", "b", "c"),
 | 
						|
		},
 | 
						|
		{
 | 
						|
			name: "Scheduled pod in any phase must not be deleted",
 | 
						|
			pods: []nameToPhase{
 | 
						|
				{name: "a", phase: v1.PodFailed, deletionTimeStamp: nil, nodeName: ""},
 | 
						|
				{name: "b", phase: v1.PodSucceeded, deletionTimeStamp: nil, nodeName: "node"},
 | 
						|
				{name: "c", phase: v1.PodRunning, deletionTimeStamp: &metav1.Time{}, nodeName: "node"},
 | 
						|
			},
 | 
						|
			deletedPodNames: sets.NewString(),
 | 
						|
		},
 | 
						|
	}
 | 
						|
 | 
						|
	for i, test := range testCases {
 | 
						|
		t.Run(test.name, func(t *testing.T) {
 | 
						|
			client := fake.NewSimpleClientset()
 | 
						|
			gcc, podInformer, _ := NewFromClient(client, -1)
 | 
						|
			deletedPodNames := make([]string, 0)
 | 
						|
			var lock sync.Mutex
 | 
						|
			gcc.deletePod = func(_, name string) error {
 | 
						|
				lock.Lock()
 | 
						|
				defer lock.Unlock()
 | 
						|
				deletedPodNames = append(deletedPodNames, name)
 | 
						|
				return nil
 | 
						|
			}
 | 
						|
 | 
						|
			creationTime := time.Unix(0, 0)
 | 
						|
			for _, pod := range test.pods {
 | 
						|
				creationTime = creationTime.Add(1 * time.Hour)
 | 
						|
				podInformer.Informer().GetStore().Add(&v1.Pod{
 | 
						|
					ObjectMeta: metav1.ObjectMeta{Name: pod.name, CreationTimestamp: metav1.Time{Time: creationTime},
 | 
						|
						DeletionTimestamp: pod.deletionTimeStamp},
 | 
						|
					Status: v1.PodStatus{Phase: pod.phase},
 | 
						|
					Spec:   v1.PodSpec{NodeName: pod.nodeName},
 | 
						|
				})
 | 
						|
			}
 | 
						|
 | 
						|
			pods, err := podInformer.Lister().List(labels.Everything())
 | 
						|
			if err != nil {
 | 
						|
				t.Errorf("Error while listing all Pods: %v", err)
 | 
						|
				return
 | 
						|
			}
 | 
						|
			gcc.gcUnscheduledTerminating(pods)
 | 
						|
 | 
						|
			if pass := compareStringSetToList(test.deletedPodNames, deletedPodNames); !pass {
 | 
						|
				t.Errorf("[%v]pod's deleted expected and actual did not match.\n\texpected: %v\n\tactual: %v, test: %v",
 | 
						|
					i, test.deletedPodNames.List(), deletedPodNames, test.name)
 | 
						|
			}
 | 
						|
		})
 | 
						|
	}
 | 
						|
}
 | 
						|
 | 
						|
func TestGCTerminating(t *testing.T) {
 | 
						|
	defer featuregatetesting.SetFeatureGateDuringTest(t, utilfeature.DefaultFeatureGate, features.NodeOutOfServiceVolumeDetach, true)()
 | 
						|
	type node struct {
 | 
						|
		name           string
 | 
						|
		readyCondition v1.ConditionStatus
 | 
						|
		taints         []v1.Taint
 | 
						|
	}
 | 
						|
 | 
						|
	type nameToPodConfig struct {
 | 
						|
		name              string
 | 
						|
		phase             v1.PodPhase
 | 
						|
		deletionTimeStamp *metav1.Time
 | 
						|
		nodeName          string
 | 
						|
	}
 | 
						|
 | 
						|
	testCases := []struct {
 | 
						|
		name            string
 | 
						|
		pods            []nameToPodConfig
 | 
						|
		nodes           []node
 | 
						|
		deletedPodNames sets.String
 | 
						|
	}{
 | 
						|
		{
 | 
						|
			name: "pods have deletion timestamp set and the corresponding nodes are not ready",
 | 
						|
			nodes: []node{
 | 
						|
				{name: "worker-0", readyCondition: v1.ConditionFalse},
 | 
						|
				{name: "worker-1", readyCondition: v1.ConditionFalse},
 | 
						|
			},
 | 
						|
			pods: []nameToPodConfig{
 | 
						|
				{name: "a", deletionTimeStamp: &metav1.Time{}, nodeName: "worker-0"},
 | 
						|
				{name: "b", deletionTimeStamp: &metav1.Time{}, nodeName: "worker-1"},
 | 
						|
			},
 | 
						|
			deletedPodNames: sets.NewString(),
 | 
						|
		},
 | 
						|
 | 
						|
		{
 | 
						|
			name: "some pods have deletion timestamp and/or phase set and some of the corresponding nodes have an" +
 | 
						|
				"outofservice taint that are not ready",
 | 
						|
			nodes: []node{
 | 
						|
				// terminated pods on this node should be force deleted
 | 
						|
				{name: "worker-0", readyCondition: v1.ConditionFalse, taints: []v1.Taint{{Key: v1.TaintNodeOutOfService,
 | 
						|
					Effect: v1.TaintEffectNoExecute}}},
 | 
						|
				// terminated pods on this node should not be force deleted
 | 
						|
				{name: "worker-1", readyCondition: v1.ConditionFalse},
 | 
						|
				// terminated pods on this node should not be force deleted
 | 
						|
				{name: "worker-2", readyCondition: v1.ConditionTrue},
 | 
						|
				// terminated pods on this node should be force deleted
 | 
						|
				{name: "worker-3", readyCondition: v1.ConditionFalse, taints: []v1.Taint{{Key: v1.TaintNodeOutOfService,
 | 
						|
					Effect: v1.TaintEffectNoSchedule}}},
 | 
						|
				// terminated pods on this node should be force deleted
 | 
						|
				{name: "worker-4", readyCondition: v1.ConditionFalse, taints: []v1.Taint{{Key: v1.TaintNodeOutOfService,
 | 
						|
					Effect: v1.TaintEffectPreferNoSchedule}}},
 | 
						|
				// terminated pods on this node should be force deleted
 | 
						|
				{name: "worker-5", readyCondition: v1.ConditionFalse, taints: []v1.Taint{{Key: v1.TaintNodeOutOfService,
 | 
						|
					Value: "any-value", Effect: v1.TaintEffectNoExecute}}},
 | 
						|
			},
 | 
						|
			pods: []nameToPodConfig{
 | 
						|
				// pods a1, b1, c1, d1 and e1 are on node worker-0
 | 
						|
				{name: "a1", nodeName: "worker-0"},
 | 
						|
				{name: "b1", deletionTimeStamp: &metav1.Time{}, nodeName: "worker-0"},
 | 
						|
				{name: "c1", phase: v1.PodPending, nodeName: "worker-0"},
 | 
						|
				{name: "d1", phase: v1.PodRunning, nodeName: "worker-0"},
 | 
						|
				{name: "e1", phase: v1.PodUnknown, nodeName: "worker-0"},
 | 
						|
 | 
						|
				// pods a2, b2, c2, d2 and e2 are on node worker-1
 | 
						|
				{name: "a2", nodeName: "worker-1"},
 | 
						|
				{name: "b2", deletionTimeStamp: &metav1.Time{}, nodeName: "worker-1"},
 | 
						|
				{name: "c2", phase: v1.PodPending, nodeName: "worker-1"},
 | 
						|
				{name: "d2", phase: v1.PodRunning, nodeName: "worker-1"},
 | 
						|
				{name: "e2", phase: v1.PodUnknown, nodeName: "worker-1"},
 | 
						|
 | 
						|
				// pods a3, b3, c3, d3 and e3 are on node worker-2
 | 
						|
				{name: "a3", nodeName: "worker-2"},
 | 
						|
				{name: "b3", deletionTimeStamp: &metav1.Time{}, nodeName: "worker-2"},
 | 
						|
				{name: "c3", phase: v1.PodPending, nodeName: "worker-2"},
 | 
						|
				{name: "d3", phase: v1.PodRunning, nodeName: "worker-2"},
 | 
						|
				{name: "e3", phase: v1.PodUnknown, nodeName: "worker-2"},
 | 
						|
 | 
						|
				// pods a4, b4, c4, d4 and e4 are on node worker-3
 | 
						|
				{name: "a4", nodeName: "worker-3"},
 | 
						|
				{name: "b4", deletionTimeStamp: &metav1.Time{}, nodeName: "worker-3"},
 | 
						|
				{name: "c4", phase: v1.PodPending, nodeName: "worker-3"},
 | 
						|
				{name: "d4", phase: v1.PodRunning, nodeName: "worker-3"},
 | 
						|
				{name: "e4", phase: v1.PodUnknown, nodeName: "worker-3"},
 | 
						|
 | 
						|
				// pods a5, b5, c5, d5 and e5 are on node worker-4
 | 
						|
				{name: "a5", nodeName: "worker-3"},
 | 
						|
				{name: "b5", deletionTimeStamp: &metav1.Time{}, nodeName: "worker-4"},
 | 
						|
				{name: "c5", phase: v1.PodPending, nodeName: "worker-4"},
 | 
						|
				{name: "d5", phase: v1.PodRunning, nodeName: "worker-4"},
 | 
						|
				{name: "e5", phase: v1.PodUnknown, nodeName: "worker-4"},
 | 
						|
 | 
						|
				// pods a6, b6, c6, d6 and e6 are on node worker-5
 | 
						|
				{name: "a6", nodeName: "worker-5"},
 | 
						|
				{name: "b6", deletionTimeStamp: &metav1.Time{}, nodeName: "worker-5"},
 | 
						|
				{name: "c6", phase: v1.PodPending, nodeName: "worker-5"},
 | 
						|
				{name: "d6", phase: v1.PodRunning, nodeName: "worker-5"},
 | 
						|
				{name: "e6", phase: v1.PodUnknown, nodeName: "worker-5"},
 | 
						|
			},
 | 
						|
			deletedPodNames: sets.NewString("b1", "b4", "b5", "b6"),
 | 
						|
		},
 | 
						|
	}
 | 
						|
	for i, test := range testCases {
 | 
						|
		t.Run(test.name, func(t *testing.T) {
 | 
						|
			client := fake.NewSimpleClientset(&v1.NodeList{Items: []v1.Node{*testutil.NewNode("node-a")}})
 | 
						|
			gcc, podInformer, nodeInformer := NewFromClient(client, -1)
 | 
						|
			deletedPodNames := make([]string, 0)
 | 
						|
			var lock sync.Mutex
 | 
						|
			gcc.deletePod = func(_, name string) error {
 | 
						|
				lock.Lock()
 | 
						|
				defer lock.Unlock()
 | 
						|
				deletedPodNames = append(deletedPodNames, name)
 | 
						|
				return nil
 | 
						|
			}
 | 
						|
			creationTime := time.Unix(0, 0)
 | 
						|
			for _, node := range test.nodes {
 | 
						|
				creationTime = creationTime.Add(2 * time.Hour)
 | 
						|
				nodeInformer.Informer().GetStore().Add(&v1.Node{
 | 
						|
					ObjectMeta: metav1.ObjectMeta{Name: node.name, CreationTimestamp: metav1.Time{Time: creationTime}},
 | 
						|
					Spec: v1.NodeSpec{
 | 
						|
						Taints: node.taints,
 | 
						|
					},
 | 
						|
					Status: v1.NodeStatus{
 | 
						|
						Conditions: []v1.NodeCondition{
 | 
						|
							{
 | 
						|
								Type:   v1.NodeReady,
 | 
						|
								Status: node.readyCondition,
 | 
						|
							},
 | 
						|
						},
 | 
						|
					},
 | 
						|
				})
 | 
						|
			}
 | 
						|
 | 
						|
			for _, pod := range test.pods {
 | 
						|
				creationTime = creationTime.Add(1 * time.Hour)
 | 
						|
				podInformer.Informer().GetStore().Add(&v1.Pod{
 | 
						|
					ObjectMeta: metav1.ObjectMeta{Name: pod.name, CreationTimestamp: metav1.Time{Time: creationTime},
 | 
						|
						DeletionTimestamp: pod.deletionTimeStamp},
 | 
						|
					Status: v1.PodStatus{Phase: pod.phase},
 | 
						|
					Spec:   v1.PodSpec{NodeName: pod.nodeName},
 | 
						|
				})
 | 
						|
			}
 | 
						|
 | 
						|
			gcc.gc(context.TODO())
 | 
						|
			if pass := compareStringSetToList(test.deletedPodNames, deletedPodNames); !pass {
 | 
						|
				t.Errorf("[%v]pod's deleted expected and actual did not match.\n\texpected: %v\n\tactual: %v",
 | 
						|
					i, test.deletedPodNames.List(), deletedPodNames)
 | 
						|
			}
 | 
						|
		})
 | 
						|
	}
 | 
						|
}
 |