mirror of
				https://github.com/optim-enterprises-bv/kubernetes.git
				synced 2025-11-04 04:08:16 +00:00 
			
		
		
		
	
		
			
				
	
	
		
			387 lines
		
	
	
		
			12 KiB
		
	
	
	
		
			Go
		
	
	
	
	
	
			
		
		
	
	
			387 lines
		
	
	
		
			12 KiB
		
	
	
	
		
			Go
		
	
	
	
	
	
/*
 | 
						|
Copyright 2018 The Kubernetes Authors.
 | 
						|
 | 
						|
Licensed under the Apache License, Version 2.0 (the "License");
 | 
						|
you may not use this file except in compliance with the License.
 | 
						|
You may obtain a copy of the License at
 | 
						|
 | 
						|
    http://www.apache.org/licenses/LICENSE-2.0
 | 
						|
 | 
						|
Unless required by applicable law or agreed to in writing, software
 | 
						|
distributed under the License is distributed on an "AS IS" BASIS,
 | 
						|
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
 | 
						|
See the License for the specific language governing permissions and
 | 
						|
limitations under the License.
 | 
						|
*/
 | 
						|
 | 
						|
package statefulset
 | 
						|
 | 
						|
import (
 | 
						|
	"context"
 | 
						|
	"fmt"
 | 
						|
	"sync"
 | 
						|
	"testing"
 | 
						|
	"time"
 | 
						|
 | 
						|
	"k8s.io/kubernetes/test/utils/ktesting"
 | 
						|
 | 
						|
	appsv1 "k8s.io/api/apps/v1"
 | 
						|
	v1 "k8s.io/api/core/v1"
 | 
						|
	"k8s.io/apimachinery/pkg/api/resource"
 | 
						|
	metav1 "k8s.io/apimachinery/pkg/apis/meta/v1"
 | 
						|
	"k8s.io/apimachinery/pkg/labels"
 | 
						|
	"k8s.io/apimachinery/pkg/util/wait"
 | 
						|
	"k8s.io/apiserver/pkg/admission"
 | 
						|
	"k8s.io/client-go/informers"
 | 
						|
	clientset "k8s.io/client-go/kubernetes"
 | 
						|
	typedappsv1 "k8s.io/client-go/kubernetes/typed/apps/v1"
 | 
						|
	typedv1 "k8s.io/client-go/kubernetes/typed/core/v1"
 | 
						|
	restclient "k8s.io/client-go/rest"
 | 
						|
	"k8s.io/client-go/util/retry"
 | 
						|
	kubeapiservertesting "k8s.io/kubernetes/cmd/kube-apiserver/app/testing"
 | 
						|
	api "k8s.io/kubernetes/pkg/apis/core"
 | 
						|
 | 
						|
	//svc "k8s.io/kubernetes/pkg/api/v1/service"
 | 
						|
	"k8s.io/kubernetes/pkg/controller/statefulset"
 | 
						|
	"k8s.io/kubernetes/test/integration/framework"
 | 
						|
)
 | 
						|
 | 
						|
const (
 | 
						|
	pollInterval = 100 * time.Millisecond
 | 
						|
	pollTimeout  = 60 * time.Second
 | 
						|
)
 | 
						|
 | 
						|
func labelMap() map[string]string {
 | 
						|
	return map[string]string{"foo": "bar"}
 | 
						|
}
 | 
						|
 | 
						|
// newService returns a service with a fake name for StatefulSet to be created soon
 | 
						|
func newHeadlessService(namespace string) *v1.Service {
 | 
						|
	return &v1.Service{
 | 
						|
		TypeMeta: metav1.TypeMeta{
 | 
						|
			Kind:       "Service",
 | 
						|
			APIVersion: "v1",
 | 
						|
		},
 | 
						|
		ObjectMeta: metav1.ObjectMeta{
 | 
						|
			Namespace: namespace,
 | 
						|
			Name:      "fake-service-name",
 | 
						|
		},
 | 
						|
		Spec: v1.ServiceSpec{
 | 
						|
			ClusterIP: "None",
 | 
						|
			Ports: []v1.ServicePort{
 | 
						|
				{Port: 80, Name: "http", Protocol: "TCP"},
 | 
						|
			},
 | 
						|
			Selector: labelMap(),
 | 
						|
		},
 | 
						|
	}
 | 
						|
}
 | 
						|
 | 
						|
// newSTS returns a StatefulSet with a fake container image
 | 
						|
func newSTS(name, namespace string, replicas int) *appsv1.StatefulSet {
 | 
						|
	replicasCopy := int32(replicas)
 | 
						|
	return &appsv1.StatefulSet{
 | 
						|
		TypeMeta: metav1.TypeMeta{
 | 
						|
			Kind:       "StatefulSet",
 | 
						|
			APIVersion: "apps/v1",
 | 
						|
		},
 | 
						|
		ObjectMeta: metav1.ObjectMeta{
 | 
						|
			Namespace: namespace,
 | 
						|
			Name:      name,
 | 
						|
		},
 | 
						|
		Spec: appsv1.StatefulSetSpec{
 | 
						|
			PodManagementPolicy: appsv1.ParallelPodManagement,
 | 
						|
			Replicas:            &replicasCopy,
 | 
						|
			Selector: &metav1.LabelSelector{
 | 
						|
				MatchLabels: labelMap(),
 | 
						|
			},
 | 
						|
			Template: v1.PodTemplateSpec{
 | 
						|
				ObjectMeta: metav1.ObjectMeta{
 | 
						|
					Labels: labelMap(),
 | 
						|
				},
 | 
						|
				Spec: v1.PodSpec{
 | 
						|
					Containers: []v1.Container{
 | 
						|
						{
 | 
						|
							Name:  "fake-name",
 | 
						|
							Image: "fakeimage",
 | 
						|
							VolumeMounts: []v1.VolumeMount{
 | 
						|
								{Name: "datadir", MountPath: "/data/"},
 | 
						|
								{Name: "home", MountPath: "/home"},
 | 
						|
							},
 | 
						|
						},
 | 
						|
					},
 | 
						|
					Volumes: []v1.Volume{
 | 
						|
						{
 | 
						|
							Name: "datadir",
 | 
						|
							VolumeSource: v1.VolumeSource{
 | 
						|
								PersistentVolumeClaim: &v1.PersistentVolumeClaimVolumeSource{
 | 
						|
									ClaimName: "fake-pvc-name",
 | 
						|
								},
 | 
						|
							},
 | 
						|
						},
 | 
						|
						{
 | 
						|
							Name: "home",
 | 
						|
							VolumeSource: v1.VolumeSource{
 | 
						|
								HostPath: &v1.HostPathVolumeSource{
 | 
						|
									Path: fmt.Sprintf("/tmp/%v", "home"),
 | 
						|
								},
 | 
						|
							},
 | 
						|
						},
 | 
						|
					},
 | 
						|
				},
 | 
						|
			},
 | 
						|
			ServiceName: "fake-service-name",
 | 
						|
			UpdateStrategy: appsv1.StatefulSetUpdateStrategy{
 | 
						|
				Type: appsv1.RollingUpdateStatefulSetStrategyType,
 | 
						|
			},
 | 
						|
			VolumeClaimTemplates: []v1.PersistentVolumeClaim{
 | 
						|
				// for volume mount "datadir"
 | 
						|
				newStatefulSetPVC("fake-pvc-name"),
 | 
						|
			},
 | 
						|
		},
 | 
						|
	}
 | 
						|
}
 | 
						|
 | 
						|
func newStatefulSetPVC(name string) v1.PersistentVolumeClaim {
 | 
						|
	return v1.PersistentVolumeClaim{
 | 
						|
		ObjectMeta: metav1.ObjectMeta{
 | 
						|
			Name: name,
 | 
						|
			Annotations: map[string]string{
 | 
						|
				"volume.alpha.kubernetes.io/storage-class": "anything",
 | 
						|
			},
 | 
						|
		},
 | 
						|
		Spec: v1.PersistentVolumeClaimSpec{
 | 
						|
			AccessModes: []v1.PersistentVolumeAccessMode{
 | 
						|
				v1.ReadWriteOnce,
 | 
						|
			},
 | 
						|
			Resources: v1.VolumeResourceRequirements{
 | 
						|
				Requests: v1.ResourceList{
 | 
						|
					v1.ResourceStorage: *resource.NewQuantity(1, resource.BinarySI),
 | 
						|
				},
 | 
						|
			},
 | 
						|
		},
 | 
						|
	}
 | 
						|
}
 | 
						|
 | 
						|
// scSetup sets up necessities for Statefulset integration test, including control plane, apiserver, informers, and clientset
 | 
						|
func scSetup(t *testing.T) (context.Context, kubeapiservertesting.TearDownFunc, *statefulset.StatefulSetController, informers.SharedInformerFactory, clientset.Interface) {
 | 
						|
	tCtx := ktesting.Init(t)
 | 
						|
	// Disable ServiceAccount admission plugin as we don't have serviceaccount controller running.
 | 
						|
	server := kubeapiservertesting.StartTestServerOrDie(t, nil, framework.DefaultTestServerFlags(), framework.SharedEtcd())
 | 
						|
 | 
						|
	config := restclient.CopyConfig(server.ClientConfig)
 | 
						|
	clientSet, err := clientset.NewForConfig(config)
 | 
						|
	if err != nil {
 | 
						|
		t.Fatalf("error in create clientset: %v", err)
 | 
						|
	}
 | 
						|
	resyncPeriod := 12 * time.Hour
 | 
						|
	informers := informers.NewSharedInformerFactory(clientset.NewForConfigOrDie(restclient.AddUserAgent(config, "statefulset-informers")), resyncPeriod)
 | 
						|
 | 
						|
	sc := statefulset.NewStatefulSetController(
 | 
						|
		tCtx,
 | 
						|
		informers.Core().V1().Pods(),
 | 
						|
		informers.Apps().V1().StatefulSets(),
 | 
						|
		informers.Core().V1().PersistentVolumeClaims(),
 | 
						|
		informers.Apps().V1().ControllerRevisions(),
 | 
						|
		clientset.NewForConfigOrDie(restclient.AddUserAgent(config, "statefulset-controller")),
 | 
						|
	)
 | 
						|
 | 
						|
	teardown := func() {
 | 
						|
		tCtx.Cancel("tearing down controller")
 | 
						|
		server.TearDownFn()
 | 
						|
	}
 | 
						|
	return tCtx, teardown, sc, informers, clientSet
 | 
						|
}
 | 
						|
 | 
						|
// Run STS controller and informers
 | 
						|
func runControllerAndInformers(ctx context.Context, sc *statefulset.StatefulSetController, informers informers.SharedInformerFactory) context.CancelFunc {
 | 
						|
	ctx, cancel := context.WithCancel(ctx)
 | 
						|
	informers.Start(ctx.Done())
 | 
						|
	go sc.Run(ctx, 5)
 | 
						|
	return cancel
 | 
						|
}
 | 
						|
 | 
						|
func createHeadlessService(t *testing.T, clientSet clientset.Interface, headlessService *v1.Service) {
 | 
						|
	_, err := clientSet.CoreV1().Services(headlessService.Namespace).Create(context.TODO(), headlessService, metav1.CreateOptions{})
 | 
						|
	if err != nil {
 | 
						|
		t.Fatalf("failed creating headless service: %v", err)
 | 
						|
	}
 | 
						|
}
 | 
						|
 | 
						|
func createSTSs(t *testing.T, clientSet clientset.Interface, stss []*appsv1.StatefulSet) []*appsv1.StatefulSet {
 | 
						|
	var createdSTSs []*appsv1.StatefulSet
 | 
						|
	for _, sts := range stss {
 | 
						|
		createdSTS, err := clientSet.AppsV1().StatefulSets(sts.Namespace).Create(context.TODO(), sts, metav1.CreateOptions{})
 | 
						|
		if err != nil {
 | 
						|
			t.Fatalf("failed to create sts %s: %v", sts.Name, err)
 | 
						|
		}
 | 
						|
		createdSTSs = append(createdSTSs, createdSTS)
 | 
						|
	}
 | 
						|
	return createdSTSs
 | 
						|
}
 | 
						|
 | 
						|
func createPods(t *testing.T, clientSet clientset.Interface, pods []*v1.Pod) []*v1.Pod {
 | 
						|
	var createdPods []*v1.Pod
 | 
						|
	for _, pod := range pods {
 | 
						|
		createdPod, err := clientSet.CoreV1().Pods(pod.Namespace).Create(context.TODO(), pod, metav1.CreateOptions{})
 | 
						|
		if err != nil {
 | 
						|
			t.Fatalf("failed to create pod %s: %v", pod.Name, err)
 | 
						|
		}
 | 
						|
		createdPods = append(createdPods, createdPod)
 | 
						|
	}
 | 
						|
 | 
						|
	return createdPods
 | 
						|
}
 | 
						|
 | 
						|
func createSTSsPods(t *testing.T, clientSet clientset.Interface, stss []*appsv1.StatefulSet, pods []*v1.Pod) ([]*appsv1.StatefulSet, []*v1.Pod) {
 | 
						|
	return createSTSs(t, clientSet, stss), createPods(t, clientSet, pods)
 | 
						|
}
 | 
						|
 | 
						|
// Verify .Status.Replicas is equal to .Spec.Replicas
 | 
						|
func waitSTSStable(t *testing.T, clientSet clientset.Interface, sts *appsv1.StatefulSet) {
 | 
						|
	stsClient := clientSet.AppsV1().StatefulSets(sts.Namespace)
 | 
						|
	desiredGeneration := sts.Generation
 | 
						|
	if err := wait.PollImmediate(pollInterval, pollTimeout, func() (bool, error) {
 | 
						|
		newSTS, err := stsClient.Get(context.TODO(), sts.Name, metav1.GetOptions{})
 | 
						|
		if err != nil {
 | 
						|
			return false, err
 | 
						|
		}
 | 
						|
		return newSTS.Status.Replicas == *newSTS.Spec.Replicas && newSTS.Status.ObservedGeneration >= desiredGeneration, nil
 | 
						|
	}); err != nil {
 | 
						|
		t.Fatalf("failed to verify .Status.Replicas is equal to .Spec.Replicas for sts %s: %v", sts.Name, err)
 | 
						|
	}
 | 
						|
}
 | 
						|
 | 
						|
func updatePod(t *testing.T, podClient typedv1.PodInterface, podName string, updateFunc func(*v1.Pod)) *v1.Pod {
 | 
						|
	var pod *v1.Pod
 | 
						|
	if err := retry.RetryOnConflict(retry.DefaultBackoff, func() error {
 | 
						|
		newPod, err := podClient.Get(context.TODO(), podName, metav1.GetOptions{})
 | 
						|
		if err != nil {
 | 
						|
			return err
 | 
						|
		}
 | 
						|
		updateFunc(newPod)
 | 
						|
		pod, err = podClient.Update(context.TODO(), newPod, metav1.UpdateOptions{})
 | 
						|
		return err
 | 
						|
	}); err != nil {
 | 
						|
		t.Fatalf("failed to update pod %s: %v", podName, err)
 | 
						|
	}
 | 
						|
	return pod
 | 
						|
}
 | 
						|
 | 
						|
func updatePodStatus(t *testing.T, podClient typedv1.PodInterface, podName string, updateStatusFunc func(*v1.Pod)) *v1.Pod {
 | 
						|
	var pod *v1.Pod
 | 
						|
	if err := retry.RetryOnConflict(retry.DefaultBackoff, func() error {
 | 
						|
		newPod, err := podClient.Get(context.TODO(), podName, metav1.GetOptions{})
 | 
						|
		if err != nil {
 | 
						|
			return err
 | 
						|
		}
 | 
						|
		updateStatusFunc(newPod)
 | 
						|
		pod, err = podClient.UpdateStatus(context.TODO(), newPod, metav1.UpdateOptions{})
 | 
						|
		return err
 | 
						|
	}); err != nil {
 | 
						|
		t.Fatalf("failed to update status of pod %s: %v", podName, err)
 | 
						|
	}
 | 
						|
	return pod
 | 
						|
}
 | 
						|
 | 
						|
func getPods(t *testing.T, podClient typedv1.PodInterface, labelMap map[string]string) *v1.PodList {
 | 
						|
	podSelector := labels.Set(labelMap).AsSelector()
 | 
						|
	options := metav1.ListOptions{LabelSelector: podSelector.String()}
 | 
						|
	pods, err := podClient.List(context.TODO(), options)
 | 
						|
	if err != nil {
 | 
						|
		t.Fatalf("failed obtaining a list of pods that match the pod labels %v: %v", labelMap, err)
 | 
						|
	}
 | 
						|
	if pods == nil {
 | 
						|
		t.Fatalf("obtained a nil list of pods")
 | 
						|
	}
 | 
						|
	return pods
 | 
						|
}
 | 
						|
 | 
						|
func getStatefulSetPVCs(t *testing.T, pvcClient typedv1.PersistentVolumeClaimInterface, sts *appsv1.StatefulSet) []*v1.PersistentVolumeClaim {
 | 
						|
	pvcs := []*v1.PersistentVolumeClaim{}
 | 
						|
	for i := int32(0); i < *sts.Spec.Replicas; i++ {
 | 
						|
		pvcName := fmt.Sprintf("%s-%s-%d", sts.Spec.VolumeClaimTemplates[0].Name, sts.Name, i)
 | 
						|
		pvc, err := pvcClient.Get(context.TODO(), pvcName, metav1.GetOptions{})
 | 
						|
		if err != nil {
 | 
						|
			t.Fatalf("failed to get PVC %s: %v", pvcName, err)
 | 
						|
		}
 | 
						|
		pvcs = append(pvcs, pvc)
 | 
						|
	}
 | 
						|
	return pvcs
 | 
						|
}
 | 
						|
 | 
						|
func verifyOwnerRef(t *testing.T, pvc *v1.PersistentVolumeClaim, kind string, expected bool) {
 | 
						|
	found := false
 | 
						|
	for _, ref := range pvc.GetOwnerReferences() {
 | 
						|
		if ref.Kind == kind {
 | 
						|
			if expected {
 | 
						|
				found = true
 | 
						|
			} else {
 | 
						|
				t.Fatalf("Found %s ref but expected none for PVC %s", kind, pvc.Name)
 | 
						|
			}
 | 
						|
		}
 | 
						|
	}
 | 
						|
	if expected && !found {
 | 
						|
		t.Fatalf("Expected %s ref but found none for PVC %s", kind, pvc.Name)
 | 
						|
	}
 | 
						|
}
 | 
						|
 | 
						|
func updateSTS(t *testing.T, stsClient typedappsv1.StatefulSetInterface, stsName string, updateFunc func(*appsv1.StatefulSet)) *appsv1.StatefulSet {
 | 
						|
	var sts *appsv1.StatefulSet
 | 
						|
	if err := retry.RetryOnConflict(retry.DefaultBackoff, func() error {
 | 
						|
		newSTS, err := stsClient.Get(context.TODO(), stsName, metav1.GetOptions{})
 | 
						|
		if err != nil {
 | 
						|
			return err
 | 
						|
		}
 | 
						|
		updateFunc(newSTS)
 | 
						|
		sts, err = stsClient.Update(context.TODO(), newSTS, metav1.UpdateOptions{})
 | 
						|
		return err
 | 
						|
	}); err != nil {
 | 
						|
		t.Fatalf("failed to update sts %s: %v", stsName, err)
 | 
						|
	}
 | 
						|
	return sts
 | 
						|
}
 | 
						|
 | 
						|
// Update .Spec.Replicas to replicas and verify .Status.Replicas is changed accordingly
 | 
						|
func scaleSTS(t *testing.T, c clientset.Interface, sts *appsv1.StatefulSet, replicas int32) {
 | 
						|
	stsClient := c.AppsV1().StatefulSets(sts.Namespace)
 | 
						|
	if err := retry.RetryOnConflict(retry.DefaultBackoff, func() error {
 | 
						|
		newSTS, err := stsClient.Get(context.TODO(), sts.Name, metav1.GetOptions{})
 | 
						|
		if err != nil {
 | 
						|
			return err
 | 
						|
		}
 | 
						|
		*newSTS.Spec.Replicas = replicas
 | 
						|
		sts, err = stsClient.Update(context.TODO(), newSTS, metav1.UpdateOptions{})
 | 
						|
		return err
 | 
						|
	}); err != nil {
 | 
						|
		t.Fatalf("failed to update .Spec.Replicas to %d for sts %s: %v", replicas, sts.Name, err)
 | 
						|
	}
 | 
						|
	waitSTSStable(t, c, sts)
 | 
						|
}
 | 
						|
 | 
						|
var _ admission.ValidationInterface = &fakePodFailAdmission{}
 | 
						|
 | 
						|
type fakePodFailAdmission struct {
 | 
						|
	lock             sync.Mutex
 | 
						|
	limitedPodNumber int
 | 
						|
	succeedPodsCount int
 | 
						|
}
 | 
						|
 | 
						|
func (f *fakePodFailAdmission) Handles(operation admission.Operation) bool {
 | 
						|
	return operation == admission.Create
 | 
						|
}
 | 
						|
 | 
						|
func (f *fakePodFailAdmission) Validate(ctx context.Context, attr admission.Attributes, o admission.ObjectInterfaces) (err error) {
 | 
						|
	if attr.GetKind().GroupKind() != api.Kind("Pod") {
 | 
						|
		return nil
 | 
						|
	}
 | 
						|
 | 
						|
	f.lock.Lock()
 | 
						|
	defer f.lock.Unlock()
 | 
						|
 | 
						|
	if f.succeedPodsCount >= f.limitedPodNumber {
 | 
						|
		return fmt.Errorf("fakePodFailAdmission error")
 | 
						|
	}
 | 
						|
	f.succeedPodsCount++
 | 
						|
	return nil
 | 
						|
}
 |