mirror of
				https://github.com/optim-enterprises-bv/kubernetes.git
				synced 2025-10-30 17:58:14 +00:00 
			
		
		
		
	 2f6c4f5eab
			
		
	
	2f6c4f5eab
	
	
	
		
			
			All code must use the context from Ginkgo when doing API calls or polling for a change, otherwise the code would not return immediately when the test gets aborted.
		
			
				
	
	
		
			389 lines
		
	
	
		
			16 KiB
		
	
	
	
		
			Go
		
	
	
	
	
	
			
		
		
	
	
			389 lines
		
	
	
		
			16 KiB
		
	
	
	
		
			Go
		
	
	
	
	
	
| /*
 | |
| Copyright 2018 The Kubernetes Authors.
 | |
| 
 | |
| Licensed under the Apache License, Version 2.0 (the "License");
 | |
| you may not use this file except in compliance with the License.
 | |
| You may obtain a copy of the License at
 | |
| 
 | |
|     http://www.apache.org/licenses/LICENSE-2.0
 | |
| 
 | |
| Unless required by applicable law or agreed to in writing, software
 | |
| distributed under the License is distributed on an "AS IS" BASIS,
 | |
| WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
 | |
| See the License for the specific language governing permissions and
 | |
| limitations under the License.
 | |
| */
 | |
| 
 | |
| package node
 | |
| 
 | |
| import (
 | |
| 	"context"
 | |
| 	"fmt"
 | |
| 	"time"
 | |
| 
 | |
| 	v1 "k8s.io/api/core/v1"
 | |
| 	nodev1 "k8s.io/api/node/v1"
 | |
| 	apierrors "k8s.io/apimachinery/pkg/api/errors"
 | |
| 	"k8s.io/apimachinery/pkg/api/resource"
 | |
| 	metav1 "k8s.io/apimachinery/pkg/apis/meta/v1"
 | |
| 	"k8s.io/apimachinery/pkg/fields"
 | |
| 	"k8s.io/apimachinery/pkg/labels"
 | |
| 	types "k8s.io/apimachinery/pkg/types"
 | |
| 	"k8s.io/apimachinery/pkg/util/wait"
 | |
| 	"k8s.io/apimachinery/pkg/watch"
 | |
| 	"k8s.io/kubernetes/pkg/kubelet/events"
 | |
| 	runtimeclasstest "k8s.io/kubernetes/pkg/kubelet/runtimeclass/testing"
 | |
| 	"k8s.io/kubernetes/test/e2e/framework"
 | |
| 	e2eevents "k8s.io/kubernetes/test/e2e/framework/events"
 | |
| 	e2enode "k8s.io/kubernetes/test/e2e/framework/node"
 | |
| 	e2epod "k8s.io/kubernetes/test/e2e/framework/pod"
 | |
| 	e2eskipper "k8s.io/kubernetes/test/e2e/framework/skipper"
 | |
| 	admissionapi "k8s.io/pod-security-admission/api"
 | |
| 
 | |
| 	"github.com/onsi/ginkgo/v2"
 | |
| )
 | |
| 
 | |
| var _ = SIGDescribe("RuntimeClass", func() {
 | |
| 	f := framework.NewDefaultFramework("runtimeclass")
 | |
| 	f.NamespacePodSecurityEnforceLevel = admissionapi.LevelBaseline
 | |
| 
 | |
| 	/*
 | |
| 		Release: v1.20
 | |
| 		Testname: Pod with the non-existing RuntimeClass is rejected.
 | |
| 		Description: The Pod requesting the non-existing RuntimeClass must be rejected.
 | |
| 	*/
 | |
| 	framework.ConformanceIt("should reject a Pod requesting a non-existent RuntimeClass [NodeConformance]", func(ctx context.Context) {
 | |
| 		rcName := f.Namespace.Name + "-nonexistent"
 | |
| 		expectPodRejection(ctx, f, e2enode.NewRuntimeClassPod(rcName))
 | |
| 	})
 | |
| 
 | |
| 	// The test CANNOT be made a Conformance as it depends on a container runtime to have a specific handler not being installed.
 | |
| 	ginkgo.It("should reject a Pod requesting a RuntimeClass with an unconfigured handler [NodeFeature:RuntimeHandler]", func(ctx context.Context) {
 | |
| 		handler := f.Namespace.Name + "-handler"
 | |
| 		rcName := createRuntimeClass(ctx, f, "unconfigured-handler", handler, nil)
 | |
| 		ginkgo.DeferCleanup(deleteRuntimeClass, f, rcName)
 | |
| 		pod := e2epod.NewPodClient(f).Create(ctx, e2enode.NewRuntimeClassPod(rcName))
 | |
| 		eventSelector := fields.Set{
 | |
| 			"involvedObject.kind":      "Pod",
 | |
| 			"involvedObject.name":      pod.Name,
 | |
| 			"involvedObject.namespace": f.Namespace.Name,
 | |
| 			"reason":                   events.FailedCreatePodSandBox,
 | |
| 		}.AsSelector().String()
 | |
| 		// Events are unreliable, don't depend on the event. It's used only to speed up the test.
 | |
| 		err := e2eevents.WaitTimeoutForEvent(ctx, f.ClientSet, f.Namespace.Name, eventSelector, handler, framework.PodEventTimeout)
 | |
| 		if err != nil {
 | |
| 			framework.Logf("Warning: did not get event about FailedCreatePodSandBox. Err: %v", err)
 | |
| 		}
 | |
| 		// Check the pod is still not running
 | |
| 		p, err := f.ClientSet.CoreV1().Pods(f.Namespace.Name).Get(ctx, pod.Name, metav1.GetOptions{})
 | |
| 		framework.ExpectNoError(err, "could not re-read the pod after event (or timeout)")
 | |
| 		framework.ExpectEqual(p.Status.Phase, v1.PodPending, "Pod phase isn't pending")
 | |
| 	})
 | |
| 
 | |
| 	// This test requires that the PreconfiguredRuntimeClassHandler has already been set up on nodes.
 | |
| 	// The test CANNOT be made a Conformance as it depends on a container runtime to have a specific handler installed and working.
 | |
| 	ginkgo.It("should run a Pod requesting a RuntimeClass with a configured handler [NodeFeature:RuntimeHandler]", func(ctx context.Context) {
 | |
| 		// Requires special setup of test-handler which is only done in GCE kube-up environment
 | |
| 		// see https://github.com/kubernetes/kubernetes/blob/eb729620c522753bc7ae61fc2c7b7ea19d4aad2f/cluster/gce/gci/configure-helper.sh#L3069-L3076
 | |
| 		e2eskipper.SkipUnlessProviderIs("gce")
 | |
| 
 | |
| 		rcName := createRuntimeClass(ctx, f, "preconfigured-handler", e2enode.PreconfiguredRuntimeClassHandler, nil)
 | |
| 		ginkgo.DeferCleanup(deleteRuntimeClass, f, rcName)
 | |
| 		pod := e2epod.NewPodClient(f).Create(ctx, e2enode.NewRuntimeClassPod(rcName))
 | |
| 		expectPodSuccess(ctx, f, pod)
 | |
| 	})
 | |
| 
 | |
| 	/*
 | |
| 		Release: v1.20
 | |
| 		Testname: Can schedule a pod requesting existing RuntimeClass.
 | |
| 		Description: The Pod requesting the existing RuntimeClass must be scheduled.
 | |
| 		This test doesn't validate that the Pod will actually start because this functionality
 | |
| 		depends on container runtime and preconfigured handler. Runtime-specific functionality
 | |
| 		is not being tested here.
 | |
| 	*/
 | |
| 	framework.ConformanceIt("should schedule a Pod requesting a RuntimeClass without PodOverhead [NodeConformance]", func(ctx context.Context) {
 | |
| 		rcName := createRuntimeClass(ctx, f, "preconfigured-handler", e2enode.PreconfiguredRuntimeClassHandler, nil)
 | |
| 		ginkgo.DeferCleanup(deleteRuntimeClass, f, rcName)
 | |
| 		pod := e2epod.NewPodClient(f).Create(ctx, e2enode.NewRuntimeClassPod(rcName))
 | |
| 		// there is only one pod in the namespace
 | |
| 		label := labels.SelectorFromSet(labels.Set(map[string]string{}))
 | |
| 		pods, err := e2epod.WaitForPodsWithLabelScheduled(ctx, f.ClientSet, f.Namespace.Name, label)
 | |
| 		framework.ExpectNoError(err, "Failed to schedule Pod with the RuntimeClass")
 | |
| 
 | |
| 		framework.ExpectEqual(len(pods.Items), 1)
 | |
| 		scheduledPod := &pods.Items[0]
 | |
| 		framework.ExpectEqual(scheduledPod.Name, pod.Name)
 | |
| 
 | |
| 		// Overhead should not be set
 | |
| 		framework.ExpectEqual(len(scheduledPod.Spec.Overhead), 0)
 | |
| 	})
 | |
| 
 | |
| 	/*
 | |
| 		Release: v1.24
 | |
| 		Testname: RuntimeClass Overhead field must be respected.
 | |
| 		Description: The Pod requesting the existing RuntimeClass must be scheduled.
 | |
| 		This test doesn't validate that the Pod will actually start because this functionality
 | |
| 		depends on container runtime and preconfigured handler. Runtime-specific functionality
 | |
| 		is not being tested here.
 | |
| 	*/
 | |
| 	framework.ConformanceIt("should schedule a Pod requesting a RuntimeClass and initialize its Overhead [NodeConformance]", func(ctx context.Context) {
 | |
| 		rcName := createRuntimeClass(ctx, f, "preconfigured-handler", e2enode.PreconfiguredRuntimeClassHandler, &nodev1.Overhead{
 | |
| 			PodFixed: v1.ResourceList{
 | |
| 				v1.ResourceName(v1.ResourceCPU):    resource.MustParse("10m"),
 | |
| 				v1.ResourceName(v1.ResourceMemory): resource.MustParse("1Mi"),
 | |
| 			},
 | |
| 		})
 | |
| 		ginkgo.DeferCleanup(deleteRuntimeClass, f, rcName)
 | |
| 		pod := e2epod.NewPodClient(f).Create(ctx, e2enode.NewRuntimeClassPod(rcName))
 | |
| 		// there is only one pod in the namespace
 | |
| 		label := labels.SelectorFromSet(labels.Set(map[string]string{}))
 | |
| 		pods, err := e2epod.WaitForPodsWithLabelScheduled(ctx, f.ClientSet, f.Namespace.Name, label)
 | |
| 		framework.ExpectNoError(err, "Failed to schedule Pod with the RuntimeClass")
 | |
| 
 | |
| 		framework.ExpectEqual(len(pods.Items), 1)
 | |
| 		scheduledPod := &pods.Items[0]
 | |
| 		framework.ExpectEqual(scheduledPod.Name, pod.Name)
 | |
| 
 | |
| 		framework.ExpectEqual(scheduledPod.Spec.Overhead[v1.ResourceCPU], resource.MustParse("10m"))
 | |
| 		framework.ExpectEqual(scheduledPod.Spec.Overhead[v1.ResourceMemory], resource.MustParse("1Mi"))
 | |
| 	})
 | |
| 
 | |
| 	/*
 | |
| 		Release: v1.20
 | |
| 		Testname: Pod with the deleted RuntimeClass is rejected.
 | |
| 		Description: Pod requesting the deleted RuntimeClass must be rejected.
 | |
| 	*/
 | |
| 	framework.ConformanceIt("should reject a Pod requesting a deleted RuntimeClass [NodeConformance]", func(ctx context.Context) {
 | |
| 		rcName := createRuntimeClass(ctx, f, "delete-me", "runc", nil)
 | |
| 		rcClient := f.ClientSet.NodeV1().RuntimeClasses()
 | |
| 
 | |
| 		ginkgo.By("Deleting RuntimeClass "+rcName, func() {
 | |
| 			err := rcClient.Delete(ctx, rcName, metav1.DeleteOptions{})
 | |
| 			framework.ExpectNoError(err, "failed to delete RuntimeClass %s", rcName)
 | |
| 
 | |
| 			ginkgo.By("Waiting for the RuntimeClass to disappear")
 | |
| 			framework.ExpectNoError(wait.PollImmediate(framework.Poll, time.Minute, func() (bool, error) {
 | |
| 				_, err := rcClient.Get(ctx, rcName, metav1.GetOptions{})
 | |
| 				if apierrors.IsNotFound(err) {
 | |
| 					return true, nil // done
 | |
| 				}
 | |
| 				if err != nil {
 | |
| 					return true, err // stop wait with error
 | |
| 				}
 | |
| 				return false, nil
 | |
| 			}))
 | |
| 		})
 | |
| 
 | |
| 		expectPodRejection(ctx, f, e2enode.NewRuntimeClassPod(rcName))
 | |
| 	})
 | |
| 
 | |
| 	/*
 | |
| 		Release: v1.20
 | |
| 		Testname: RuntimeClass API
 | |
| 		Description:
 | |
| 		The node.k8s.io API group MUST exist in the /apis discovery document.
 | |
| 		The node.k8s.io/v1 API group/version MUST exist in the /apis/mode.k8s.io discovery document.
 | |
| 		The runtimeclasses resource MUST exist in the /apis/node.k8s.io/v1 discovery document.
 | |
| 		The runtimeclasses resource must support create, get, list, watch, update, patch, delete, and deletecollection.
 | |
| 	*/
 | |
| 	framework.ConformanceIt(" should support RuntimeClasses API operations", func(ctx context.Context) {
 | |
| 		// Setup
 | |
| 		rcVersion := "v1"
 | |
| 		rcClient := f.ClientSet.NodeV1().RuntimeClasses()
 | |
| 
 | |
| 		// This is a conformance test that must configure opaque handlers to validate CRUD operations.
 | |
| 		// Test should not use any existing handler like gVisor or runc
 | |
| 		//
 | |
| 		// All CRUD operations in this test are limited to the objects with the label test=f.UniqueName
 | |
| 		rc := runtimeclasstest.NewRuntimeClass(f.UniqueName+"-handler", f.UniqueName+"-conformance-runtime-class")
 | |
| 		rc.SetLabels(map[string]string{"test": f.UniqueName})
 | |
| 		rc2 := runtimeclasstest.NewRuntimeClass(f.UniqueName+"-handler2", f.UniqueName+"-conformance-runtime-class2")
 | |
| 		rc2.SetLabels(map[string]string{"test": f.UniqueName})
 | |
| 		rc3 := runtimeclasstest.NewRuntimeClass(f.UniqueName+"-handler3", f.UniqueName+"-conformance-runtime-class3")
 | |
| 		rc3.SetLabels(map[string]string{"test": f.UniqueName})
 | |
| 
 | |
| 		// Discovery
 | |
| 
 | |
| 		ginkgo.By("getting /apis")
 | |
| 		{
 | |
| 			discoveryGroups, err := f.ClientSet.Discovery().ServerGroups()
 | |
| 			framework.ExpectNoError(err)
 | |
| 			found := false
 | |
| 			for _, group := range discoveryGroups.Groups {
 | |
| 				if group.Name == nodev1.GroupName {
 | |
| 					for _, version := range group.Versions {
 | |
| 						if version.Version == rcVersion {
 | |
| 							found = true
 | |
| 							break
 | |
| 						}
 | |
| 					}
 | |
| 				}
 | |
| 			}
 | |
| 			if !found {
 | |
| 				framework.Failf("expected RuntimeClass API group/version, got %#v", discoveryGroups.Groups)
 | |
| 			}
 | |
| 		}
 | |
| 
 | |
| 		ginkgo.By("getting /apis/node.k8s.io")
 | |
| 		{
 | |
| 			group := &metav1.APIGroup{}
 | |
| 			err := f.ClientSet.Discovery().RESTClient().Get().AbsPath("/apis/node.k8s.io").Do(ctx).Into(group)
 | |
| 			framework.ExpectNoError(err)
 | |
| 			found := false
 | |
| 			for _, version := range group.Versions {
 | |
| 				if version.Version == rcVersion {
 | |
| 					found = true
 | |
| 					break
 | |
| 				}
 | |
| 			}
 | |
| 			if !found {
 | |
| 				framework.Failf("expected RuntimeClass API version, got %#v", group.Versions)
 | |
| 			}
 | |
| 		}
 | |
| 
 | |
| 		ginkgo.By("getting /apis/node.k8s.io/" + rcVersion)
 | |
| 		{
 | |
| 			resources, err := f.ClientSet.Discovery().ServerResourcesForGroupVersion(nodev1.SchemeGroupVersion.String())
 | |
| 			framework.ExpectNoError(err)
 | |
| 			found := false
 | |
| 			for _, resource := range resources.APIResources {
 | |
| 				switch resource.Name {
 | |
| 				case "runtimeclasses":
 | |
| 					found = true
 | |
| 				}
 | |
| 			}
 | |
| 			if !found {
 | |
| 				framework.Failf("expected runtimeclasses, got %#v", resources.APIResources)
 | |
| 			}
 | |
| 		}
 | |
| 
 | |
| 		// Main resource create/read/update/watch operations
 | |
| 
 | |
| 		ginkgo.By("creating")
 | |
| 		createdRC, err := rcClient.Create(ctx, rc, metav1.CreateOptions{})
 | |
| 		framework.ExpectNoError(err)
 | |
| 		_, err = rcClient.Create(ctx, rc, metav1.CreateOptions{})
 | |
| 		if !apierrors.IsAlreadyExists(err) {
 | |
| 			framework.Failf("expected 409, got %#v", err)
 | |
| 		}
 | |
| 		_, err = rcClient.Create(ctx, rc2, metav1.CreateOptions{})
 | |
| 		framework.ExpectNoError(err)
 | |
| 
 | |
| 		ginkgo.By("watching")
 | |
| 		framework.Logf("starting watch")
 | |
| 		rcWatch, err := rcClient.Watch(ctx, metav1.ListOptions{LabelSelector: "test=" + f.UniqueName})
 | |
| 		framework.ExpectNoError(err)
 | |
| 
 | |
| 		// added for a watch
 | |
| 		_, err = rcClient.Create(ctx, rc3, metav1.CreateOptions{})
 | |
| 		framework.ExpectNoError(err)
 | |
| 
 | |
| 		ginkgo.By("getting")
 | |
| 		gottenRC, err := rcClient.Get(ctx, rc.Name, metav1.GetOptions{})
 | |
| 		framework.ExpectNoError(err)
 | |
| 		framework.ExpectEqual(gottenRC.UID, createdRC.UID)
 | |
| 
 | |
| 		ginkgo.By("listing")
 | |
| 		rcs, err := rcClient.List(ctx, metav1.ListOptions{LabelSelector: "test=" + f.UniqueName})
 | |
| 		framework.ExpectNoError(err)
 | |
| 		framework.ExpectEqual(len(rcs.Items), 3, "filtered list should have 3 items")
 | |
| 
 | |
| 		ginkgo.By("patching")
 | |
| 		patchedRC, err := rcClient.Patch(ctx, createdRC.Name, types.MergePatchType, []byte(`{"metadata":{"annotations":{"patched":"true"}}}`), metav1.PatchOptions{})
 | |
| 		framework.ExpectNoError(err)
 | |
| 		framework.ExpectEqual(patchedRC.Annotations["patched"], "true", "patched object should have the applied annotation")
 | |
| 
 | |
| 		ginkgo.By("updating")
 | |
| 		csrToUpdate := patchedRC.DeepCopy()
 | |
| 		csrToUpdate.Annotations["updated"] = "true"
 | |
| 		updatedRC, err := rcClient.Update(ctx, csrToUpdate, metav1.UpdateOptions{})
 | |
| 		framework.ExpectNoError(err)
 | |
| 		framework.ExpectEqual(updatedRC.Annotations["updated"], "true", "updated object should have the applied annotation")
 | |
| 
 | |
| 		framework.Logf("waiting for watch events with expected annotations")
 | |
| 		for sawAdded, sawPatched, sawUpdated := false, false, false; !sawAdded && !sawPatched && !sawUpdated; {
 | |
| 			select {
 | |
| 			case evt, ok := <-rcWatch.ResultChan():
 | |
| 				if !ok {
 | |
| 					framework.Fail("watch channel should not close")
 | |
| 				}
 | |
| 				if evt.Type == watch.Modified {
 | |
| 					watchedRC, isRC := evt.Object.(*nodev1.RuntimeClass)
 | |
| 					if !isRC {
 | |
| 						framework.Failf("expected RC, got %T", evt.Object)
 | |
| 					}
 | |
| 					if watchedRC.Annotations["patched"] == "true" {
 | |
| 						framework.Logf("saw patched annotations")
 | |
| 						sawPatched = true
 | |
| 					} else if watchedRC.Annotations["updated"] == "true" {
 | |
| 						framework.Logf("saw updated annotations")
 | |
| 						sawUpdated = true
 | |
| 					} else {
 | |
| 						framework.Logf("missing expected annotations, waiting: %#v", watchedRC.Annotations)
 | |
| 					}
 | |
| 				} else if evt.Type == watch.Added {
 | |
| 					_, isRC := evt.Object.(*nodev1.RuntimeClass)
 | |
| 					if !isRC {
 | |
| 						framework.Failf("expected RC, got %T", evt.Object)
 | |
| 					}
 | |
| 					sawAdded = true
 | |
| 				}
 | |
| 
 | |
| 			case <-time.After(wait.ForeverTestTimeout):
 | |
| 				framework.Fail("timed out waiting for watch event")
 | |
| 			}
 | |
| 		}
 | |
| 		rcWatch.Stop()
 | |
| 
 | |
| 		// main resource delete operations
 | |
| 
 | |
| 		ginkgo.By("deleting")
 | |
| 		err = rcClient.Delete(ctx, createdRC.Name, metav1.DeleteOptions{})
 | |
| 		framework.ExpectNoError(err)
 | |
| 		_, err = rcClient.Get(ctx, createdRC.Name, metav1.GetOptions{})
 | |
| 		if !apierrors.IsNotFound(err) {
 | |
| 			framework.Failf("expected 404, got %#v", err)
 | |
| 		}
 | |
| 		rcs, err = rcClient.List(ctx, metav1.ListOptions{LabelSelector: "test=" + f.UniqueName})
 | |
| 		framework.ExpectNoError(err)
 | |
| 		framework.ExpectEqual(len(rcs.Items), 2, "filtered list should have 2 items")
 | |
| 
 | |
| 		ginkgo.By("deleting a collection")
 | |
| 		err = rcClient.DeleteCollection(ctx, metav1.DeleteOptions{}, metav1.ListOptions{LabelSelector: "test=" + f.UniqueName})
 | |
| 		framework.ExpectNoError(err)
 | |
| 		rcs, err = rcClient.List(ctx, metav1.ListOptions{LabelSelector: "test=" + f.UniqueName})
 | |
| 		framework.ExpectNoError(err)
 | |
| 		framework.ExpectEqual(len(rcs.Items), 0, "filtered list should have 0 items")
 | |
| 	})
 | |
| })
 | |
| 
 | |
| func deleteRuntimeClass(ctx context.Context, f *framework.Framework, name string) {
 | |
| 	err := f.ClientSet.NodeV1().RuntimeClasses().Delete(ctx, name, metav1.DeleteOptions{})
 | |
| 	framework.ExpectNoError(err, "failed to delete RuntimeClass resource")
 | |
| }
 | |
| 
 | |
| // createRuntimeClass generates a RuntimeClass with the desired handler and a "namespaced" name,
 | |
| // synchronously creates it, and returns the generated name.
 | |
| func createRuntimeClass(ctx context.Context, f *framework.Framework, name, handler string, overhead *nodev1.Overhead) string {
 | |
| 	uniqueName := fmt.Sprintf("%s-%s", f.Namespace.Name, name)
 | |
| 	rc := runtimeclasstest.NewRuntimeClass(uniqueName, handler)
 | |
| 	rc.Overhead = overhead
 | |
| 	rc, err := f.ClientSet.NodeV1().RuntimeClasses().Create(ctx, rc, metav1.CreateOptions{})
 | |
| 	framework.ExpectNoError(err, "failed to create RuntimeClass resource")
 | |
| 	return rc.GetName()
 | |
| }
 | |
| 
 | |
| func expectPodRejection(ctx context.Context, f *framework.Framework, pod *v1.Pod) {
 | |
| 	_, err := f.ClientSet.CoreV1().Pods(f.Namespace.Name).Create(ctx, pod, metav1.CreateOptions{})
 | |
| 	framework.ExpectError(err, "should be forbidden")
 | |
| 	if !apierrors.IsForbidden(err) {
 | |
| 		framework.Failf("expected forbidden error, got %#v", err)
 | |
| 	}
 | |
| }
 | |
| 
 | |
| // expectPodSuccess waits for the given pod to terminate successfully.
 | |
| func expectPodSuccess(ctx context.Context, f *framework.Framework, pod *v1.Pod) {
 | |
| 	framework.ExpectNoError(e2epod.WaitForPodSuccessInNamespace(
 | |
| 		ctx, f.ClientSet, pod.Name, f.Namespace.Name))
 | |
| }
 |