mirror of
				https://github.com/optim-enterprises-bv/kubernetes.git
				synced 2025-10-30 17:58:14 +00:00 
			
		
		
		
	Merge pull request #129723 from carlory/NFSPersistentVolumes
remove unrunnable tests from NFSPersistentVolumes
This commit is contained in:
		| @@ -19,7 +19,6 @@ package storage | ||||
| import ( | ||||
| 	"context" | ||||
| 	"fmt" | ||||
| 	"net" | ||||
| 	"time" | ||||
|  | ||||
| 	"github.com/onsi/ginkgo/v2" | ||||
| @@ -28,15 +27,12 @@ import ( | ||||
| 	v1 "k8s.io/api/core/v1" | ||||
| 	metav1 "k8s.io/apimachinery/pkg/apis/meta/v1" | ||||
| 	"k8s.io/apimachinery/pkg/labels" | ||||
| 	utilerrors "k8s.io/apimachinery/pkg/util/errors" | ||||
| 	clientset "k8s.io/client-go/kubernetes" | ||||
| 	"k8s.io/kubernetes/test/e2e/framework" | ||||
| 	e2ekubesystem "k8s.io/kubernetes/test/e2e/framework/kubesystem" | ||||
| 	e2enode "k8s.io/kubernetes/test/e2e/framework/node" | ||||
| 	e2epod "k8s.io/kubernetes/test/e2e/framework/pod" | ||||
| 	e2epv "k8s.io/kubernetes/test/e2e/framework/pv" | ||||
| 	e2eskipper "k8s.io/kubernetes/test/e2e/framework/skipper" | ||||
| 	e2essh "k8s.io/kubernetes/test/e2e/framework/ssh" | ||||
| 	e2evolume "k8s.io/kubernetes/test/e2e/framework/volume" | ||||
| 	"k8s.io/kubernetes/test/e2e/storage/utils" | ||||
| 	admissionapi "k8s.io/pod-security-admission/api" | ||||
| @@ -48,34 +44,7 @@ type disruptiveTest struct { | ||||
| 	runTest    testBody | ||||
| } | ||||
|  | ||||
| // checkForControllerManagerHealthy checks that the controller manager does not crash within "duration" | ||||
| func checkForControllerManagerHealthy(ctx context.Context, duration time.Duration) error { | ||||
| 	var PID string | ||||
| 	cmd := "pidof kube-controller-manager" | ||||
| 	for start := time.Now(); time.Since(start) < duration && ctx.Err() == nil; time.Sleep(5 * time.Second) { | ||||
| 		result, err := e2essh.SSH(ctx, cmd, net.JoinHostPort(framework.APIAddress(), e2essh.SSHPort), framework.TestContext.Provider) | ||||
| 		if err != nil { | ||||
| 			// We don't necessarily know that it crashed, pipe could just be broken | ||||
| 			e2essh.LogResult(result) | ||||
| 			return fmt.Errorf("master unreachable after %v", time.Since(start)) | ||||
| 		} else if result.Code != 0 { | ||||
| 			e2essh.LogResult(result) | ||||
| 			return fmt.Errorf("SSH result code not 0. actually: %v after %v", result.Code, time.Since(start)) | ||||
| 		} else if result.Stdout != PID { | ||||
| 			if PID == "" { | ||||
| 				PID = result.Stdout | ||||
| 			} else { | ||||
| 				//its dead | ||||
| 				return fmt.Errorf("controller manager crashed, old PID: %s, new PID: %s", PID, result.Stdout) | ||||
| 			} | ||||
| 		} else { | ||||
| 			framework.Logf("kube-controller-manager still healthy after %v", time.Since(start)) | ||||
| 		} | ||||
| 	} | ||||
| 	return nil | ||||
| } | ||||
|  | ||||
| var _ = utils.SIGDescribe("NFSPersistentVolumes", framework.WithDisruptive(), framework.WithFlaky(), func() { | ||||
| var _ = utils.SIGDescribe("NFSPersistentVolumes", framework.WithDisruptive(), func() { | ||||
|  | ||||
| 	f := framework.NewDefaultFramework("disruptive-pv") | ||||
| 	f.NamespacePodSecurityLevel = admissionapi.LevelPrivileged | ||||
| @@ -136,98 +105,6 @@ var _ = utils.SIGDescribe("NFSPersistentVolumes", framework.WithDisruptive(), fr | ||||
| 		} | ||||
| 	}) | ||||
|  | ||||
| 	ginkgo.Context("when kube-controller-manager restarts", func() { | ||||
| 		var ( | ||||
| 			diskName1, diskName2 string | ||||
| 			err                  error | ||||
| 			pvConfig1, pvConfig2 e2epv.PersistentVolumeConfig | ||||
| 			pv1, pv2             *v1.PersistentVolume | ||||
| 			pvSource1, pvSource2 *v1.PersistentVolumeSource | ||||
| 			pvc1, pvc2           *v1.PersistentVolumeClaim | ||||
| 			clientPod            *v1.Pod | ||||
| 		) | ||||
|  | ||||
| 		ginkgo.BeforeEach(func(ctx context.Context) { | ||||
| 			e2eskipper.SkipUnlessProviderIs("gce") | ||||
| 			e2eskipper.SkipUnlessSSHKeyPresent() | ||||
|  | ||||
| 			ginkgo.By("Initializing first PD with PVPVC binding") | ||||
| 			pvSource1, diskName1 = createGCEVolume(ctx) | ||||
| 			framework.ExpectNoError(err) | ||||
| 			pvConfig1 = e2epv.PersistentVolumeConfig{ | ||||
| 				NamePrefix: "gce-", | ||||
| 				Labels:     volLabel, | ||||
| 				PVSource:   *pvSource1, | ||||
| 				Prebind:    nil, | ||||
| 			} | ||||
| 			pv1, pvc1, err = e2epv.CreatePVPVC(ctx, c, f.Timeouts, pvConfig1, pvcConfig, ns, false) | ||||
| 			framework.ExpectNoError(err) | ||||
| 			framework.ExpectNoError(e2epv.WaitOnPVandPVC(ctx, c, f.Timeouts, ns, pv1, pvc1)) | ||||
|  | ||||
| 			ginkgo.By("Initializing second PD with PVPVC binding") | ||||
| 			pvSource2, diskName2 = createGCEVolume(ctx) | ||||
| 			framework.ExpectNoError(err) | ||||
| 			pvConfig2 = e2epv.PersistentVolumeConfig{ | ||||
| 				NamePrefix: "gce-", | ||||
| 				Labels:     volLabel, | ||||
| 				PVSource:   *pvSource2, | ||||
| 				Prebind:    nil, | ||||
| 			} | ||||
| 			pv2, pvc2, err = e2epv.CreatePVPVC(ctx, c, f.Timeouts, pvConfig2, pvcConfig, ns, false) | ||||
| 			framework.ExpectNoError(err) | ||||
| 			framework.ExpectNoError(e2epv.WaitOnPVandPVC(ctx, c, f.Timeouts, ns, pv2, pvc2)) | ||||
|  | ||||
| 			ginkgo.By("Attaching both PVC's to a single pod") | ||||
| 			clientPod, err = e2epod.CreatePod(ctx, c, ns, nil, []*v1.PersistentVolumeClaim{pvc1, pvc2}, f.NamespacePodSecurityLevel, "") | ||||
| 			framework.ExpectNoError(err) | ||||
| 		}) | ||||
|  | ||||
| 		ginkgo.AfterEach(func(ctx context.Context) { | ||||
| 			// Delete client/user pod first | ||||
| 			framework.ExpectNoError(e2epod.DeletePodWithWait(ctx, c, clientPod)) | ||||
|  | ||||
| 			// Delete PV and PVCs | ||||
| 			if errs := e2epv.PVPVCCleanup(ctx, c, ns, pv1, pvc1); len(errs) > 0 { | ||||
| 				framework.Failf("AfterEach: Failed to delete PVC and/or PV. Errors: %v", utilerrors.NewAggregate(errs)) | ||||
| 			} | ||||
| 			pv1, pvc1 = nil, nil | ||||
| 			if errs := e2epv.PVPVCCleanup(ctx, c, ns, pv2, pvc2); len(errs) > 0 { | ||||
| 				framework.Failf("AfterEach: Failed to delete PVC and/or PV. Errors: %v", utilerrors.NewAggregate(errs)) | ||||
| 			} | ||||
| 			pv2, pvc2 = nil, nil | ||||
|  | ||||
| 			// Delete the actual disks | ||||
| 			if diskName1 != "" { | ||||
| 				framework.ExpectNoError(e2epv.DeletePDWithRetry(ctx, diskName1)) | ||||
| 			} | ||||
| 			if diskName2 != "" { | ||||
| 				framework.ExpectNoError(e2epv.DeletePDWithRetry(ctx, diskName2)) | ||||
| 			} | ||||
| 		}) | ||||
|  | ||||
| 		ginkgo.It("should delete a bound PVC from a clientPod, restart the kube-control-manager, and ensure the kube-controller-manager does not crash", func(ctx context.Context) { | ||||
| 			e2eskipper.SkipUnlessSSHKeyPresent() | ||||
|  | ||||
| 			ginkgo.By("Deleting PVC for volume 2") | ||||
| 			err = e2epv.DeletePersistentVolumeClaim(ctx, c, pvc2.Name, ns) | ||||
| 			framework.ExpectNoError(err) | ||||
| 			pvc2 = nil | ||||
|  | ||||
| 			ginkgo.By("Restarting the kube-controller-manager") | ||||
| 			err = e2ekubesystem.RestartControllerManager(ctx) | ||||
| 			framework.ExpectNoError(err) | ||||
| 			err = e2ekubesystem.WaitForControllerManagerUp(ctx) | ||||
| 			framework.ExpectNoError(err) | ||||
| 			framework.Logf("kube-controller-manager restarted") | ||||
|  | ||||
| 			ginkgo.By("Observing the kube-controller-manager healthy for at least 2 minutes") | ||||
| 			// Continue checking for 2 minutes to make sure kube-controller-manager is healthy | ||||
| 			err = checkForControllerManagerHealthy(ctx, 2*time.Minute) | ||||
| 			framework.ExpectNoError(err) | ||||
| 		}) | ||||
|  | ||||
| 	}) | ||||
|  | ||||
| 	ginkgo.Context("when kubelet restarts", func() { | ||||
| 		var ( | ||||
| 			clientPod *v1.Pod | ||||
| @@ -277,19 +154,6 @@ var _ = utils.SIGDescribe("NFSPersistentVolumes", framework.WithDisruptive(), fr | ||||
| 	}) | ||||
| }) | ||||
|  | ||||
| // createGCEVolume creates PersistentVolumeSource for GCEVolume. | ||||
| func createGCEVolume(ctx context.Context) (*v1.PersistentVolumeSource, string) { | ||||
| 	diskName, err := e2epv.CreatePDWithRetry(ctx) | ||||
| 	framework.ExpectNoError(err) | ||||
| 	return &v1.PersistentVolumeSource{ | ||||
| 		GCEPersistentDisk: &v1.GCEPersistentDiskVolumeSource{ | ||||
| 			PDName:   diskName, | ||||
| 			FSType:   "ext3", | ||||
| 			ReadOnly: false, | ||||
| 		}, | ||||
| 	}, diskName | ||||
| } | ||||
|  | ||||
| // initTestCase initializes spec resources (pv, pvc, and pod) and returns pointers to be consumed | ||||
| // by the test. | ||||
| func initTestCase(ctx context.Context, f *framework.Framework, c clientset.Interface, pvConfig e2epv.PersistentVolumeConfig, pvcConfig e2epv.PersistentVolumeClaimConfig, ns, nodeName string) (*v1.Pod, *v1.PersistentVolume, *v1.PersistentVolumeClaim) { | ||||
|   | ||||
		Reference in New Issue
	
	Block a user
	 Kubernetes Prow Robot
					Kubernetes Prow Robot