From 910bd000ff77c7af3232b7647bbb5c87853c019b Mon Sep 17 00:00:00 2001 From: carlory Date: Tue, 27 Aug 2019 11:15:16 +0800 Subject: [PATCH] use log functions of core framework in the 'test/e2e/storage/*' --- test/e2e/storage/BUILD | 1 - test/e2e/storage/csi_mock_volume.go | 19 +++--- test/e2e/storage/drivers/BUILD | 1 - test/e2e/storage/drivers/csi.go | 7 +- test/e2e/storage/drivers/csi_objects.go | 13 ++-- test/e2e/storage/drivers/in_tree.go | 31 +++++---- test/e2e/storage/empty_dir_wrapper.go | 17 +++-- test/e2e/storage/ephemeral_volume.go | 3 +- test/e2e/storage/flexvolume.go | 5 +- .../flexvolume_mounted_volume_resize.go | 9 ++- test/e2e/storage/flexvolume_online_resize.go | 9 ++- .../generic_persistent_volume-disruptive.go | 5 +- test/e2e/storage/mounted_volume_resize.go | 9 ++- .../nfs_persistent_volume-disruptive.go | 17 +++-- test/e2e/storage/pd.go | 65 +++++++++---------- test/e2e/storage/persistent_volumes-gce.go | 5 +- test/e2e/storage/persistent_volumes-local.go | 17 +++-- test/e2e/storage/persistent_volumes.go | 16 ++--- test/e2e/storage/pv_protection.go | 5 +- test/e2e/storage/regional_pd.go | 23 ++++--- test/e2e/storage/testsuites/BUILD | 1 - test/e2e/storage/testsuites/base.go | 31 +++++---- .../storage/testsuites/driveroperations.go | 4 +- test/e2e/storage/testsuites/multivolume.go | 3 +- test/e2e/storage/testsuites/provisioning.go | 31 +++++---- test/e2e/storage/testsuites/snapshottable.go | 27 ++++---- test/e2e/storage/testsuites/subpath.go | 15 ++--- test/e2e/storage/testsuites/volume_expand.go | 13 ++-- test/e2e/storage/testsuites/volume_io.go | 7 +- test/e2e/storage/testsuites/volumemode.go | 11 ++-- test/e2e/storage/utils/BUILD | 1 - test/e2e/storage/utils/local.go | 7 +- test/e2e/storage/utils/utils.go | 25 ++++--- test/e2e/storage/volume_limits.go | 5 +- test/e2e/storage/volume_metrics.go | 49 +++++++------- test/e2e/storage/volume_provisioning.go | 63 +++++++++--------- test/e2e/storage/volumes.go | 5 +- test/e2e/storage/vsphere/BUILD | 1 - test/e2e/storage/vsphere/bootstrap.go | 9 ++- test/e2e/storage/vsphere/config.go | 14 ++-- test/e2e/storage/vsphere/nodemapper.go | 20 +++--- .../vsphere/persistent_volumes-vsphere.go | 3 +- test/e2e/storage/vsphere/pv_reclaimpolicy.go | 5 +- test/e2e/storage/vsphere/vsphere.go | 20 +++--- .../storage/vsphere/vsphere_statefulsets.go | 7 +- test/e2e/storage/vsphere/vsphere_utils.go | 45 +++++++------ .../vsphere/vsphere_volume_datastore.go | 5 +- .../vsphere/vsphere_volume_diskformat.go | 3 +- .../storage/vsphere/vsphere_volume_fstype.go | 3 +- .../vsphere/vsphere_volume_node_poweroff.go | 9 ++- .../storage/vsphere/vsphere_volume_perf.go | 11 ++-- .../vsphere/vsphere_volume_vpxd_restart.go | 3 +- .../vsphere/vsphere_volume_vsan_policy.go | 33 +++++----- .../storage/vsphere/vsphere_zone_support.go | 5 +- 54 files changed, 361 insertions(+), 410 deletions(-) diff --git a/test/e2e/storage/BUILD b/test/e2e/storage/BUILD index 488094990d3..427418ec820 100644 --- a/test/e2e/storage/BUILD +++ b/test/e2e/storage/BUILD @@ -65,7 +65,6 @@ go_library( "//test/e2e/framework:go_default_library", "//test/e2e/framework/auth:go_default_library", "//test/e2e/framework/deployment:go_default_library", - "//test/e2e/framework/log:go_default_library", "//test/e2e/framework/metrics:go_default_library", "//test/e2e/framework/node:go_default_library", "//test/e2e/framework/pod:go_default_library", diff --git a/test/e2e/storage/csi_mock_volume.go b/test/e2e/storage/csi_mock_volume.go index 4583eb1bd19..38e7da2f22f 100644 --- a/test/e2e/storage/csi_mock_volume.go +++ b/test/e2e/storage/csi_mock_volume.go @@ -36,7 +36,6 @@ import ( "k8s.io/apimachinery/pkg/util/wait" clientset "k8s.io/client-go/kubernetes" "k8s.io/kubernetes/test/e2e/framework" - e2elog "k8s.io/kubernetes/test/e2e/framework/log" e2epod "k8s.io/kubernetes/test/e2e/framework/pod" "k8s.io/kubernetes/test/e2e/storage/drivers" "k8s.io/kubernetes/test/e2e/storage/testsuites" @@ -480,7 +479,7 @@ var _ = utils.SIGDescribe("CSI mock volume", func() { pvcSize := pvc.Spec.Resources.Requests[v1.ResourceStorage] if pvcSize.Cmp(newSize) != 0 { - e2elog.Failf("error updating pvc size %q", pvc.Name) + framework.Failf("error updating pvc size %q", pvc.Name) } if test.expectFailure { err = testsuites.WaitForResizingCondition(pvc, m.cs, csiResizingConditionWait) @@ -572,7 +571,7 @@ var _ = utils.SIGDescribe("CSI mock volume", func() { pvcSize := pvc.Spec.Resources.Requests[v1.ResourceStorage] if pvcSize.Cmp(newSize) != 0 { - e2elog.Failf("error updating pvc size %q", pvc.Name) + framework.Failf("error updating pvc size %q", pvc.Name) } ginkgo.By("Waiting for persistent volume resize to finish") @@ -741,7 +740,7 @@ func checkPodLogs(cs clientset.Interface, namespace, driverPodName, driverContai if err != nil { return fmt.Errorf("could not load CSI driver logs: %s", err) } - e2elog.Logf("CSI driver logs:\n%s", log) + framework.Logf("CSI driver logs:\n%s", log) // Find NodePublish in the logs foundAttributes := sets.NewString() logLines := strings.Split(log, "\n") @@ -762,7 +761,7 @@ func checkPodLogs(cs clientset.Interface, namespace, driverPodName, driverContai var call MockCSICall err := json.Unmarshal([]byte(line), &call) if err != nil { - e2elog.Logf("Could not parse CSI driver log line %q: %s", line, err) + framework.Logf("Could not parse CSI driver log line %q: %s", line, err) continue } switch call.Method { @@ -774,12 +773,12 @@ func checkPodLogs(cs clientset.Interface, namespace, driverPodName, driverContai vv, found := call.Request.VolumeContext[k] if found && v == vv { foundAttributes.Insert(k) - e2elog.Logf("Found volume attribute %s: %s", k, v) + framework.Logf("Found volume attribute %s: %s", k, v) } } } case "/csi.v1.Node/NodeUnpublishVolume": - e2elog.Logf("Found NodeUnpublishVolume: %+v", call) + framework.Logf("Found NodeUnpublishVolume: %+v", call) numNodeUnpublishVolume++ } } @@ -801,7 +800,7 @@ func checkPodLogs(cs clientset.Interface, namespace, driverPodName, driverContai func waitForCSIDriver(cs clientset.Interface, driverName string) error { timeout := 4 * time.Minute - e2elog.Logf("waiting up to %v for CSIDriver %q", timeout, driverName) + framework.Logf("waiting up to %v for CSIDriver %q", timeout, driverName) for start := time.Now(); time.Since(start) < timeout; time.Sleep(framework.Poll) { _, err := cs.StorageV1beta1().CSIDrivers().Get(driverName, metav1.GetOptions{}) if !errors.IsNotFound(err) { @@ -814,9 +813,9 @@ func waitForCSIDriver(cs clientset.Interface, driverName string) error { func destroyCSIDriver(cs clientset.Interface, driverName string) { driverGet, err := cs.StorageV1beta1().CSIDrivers().Get(driverName, metav1.GetOptions{}) if err == nil { - e2elog.Logf("deleting %s.%s: %s", driverGet.TypeMeta.APIVersion, driverGet.TypeMeta.Kind, driverGet.ObjectMeta.Name) + framework.Logf("deleting %s.%s: %s", driverGet.TypeMeta.APIVersion, driverGet.TypeMeta.Kind, driverGet.ObjectMeta.Name) // Uncomment the following line to get full dump of CSIDriver object - // e2elog.Logf("%s", framework.PrettyPrint(driverGet)) + // framework.Logf("%s", framework.PrettyPrint(driverGet)) cs.StorageV1beta1().CSIDrivers().Delete(driverName, nil) } } diff --git a/test/e2e/storage/drivers/BUILD b/test/e2e/storage/drivers/BUILD index 5dbd9054686..0769b1772b7 100644 --- a/test/e2e/storage/drivers/BUILD +++ b/test/e2e/storage/drivers/BUILD @@ -24,7 +24,6 @@ go_library( "//staging/src/k8s.io/client-go/kubernetes:go_default_library", "//test/e2e/framework:go_default_library", "//test/e2e/framework/auth:go_default_library", - "//test/e2e/framework/log:go_default_library", "//test/e2e/framework/pod:go_default_library", "//test/e2e/framework/volume:go_default_library", "//test/e2e/storage/testpatterns:go_default_library", diff --git a/test/e2e/storage/drivers/csi.go b/test/e2e/storage/drivers/csi.go index ea6413a88b3..d2773ca84b8 100644 --- a/test/e2e/storage/drivers/csi.go +++ b/test/e2e/storage/drivers/csi.go @@ -47,7 +47,6 @@ import ( "k8s.io/apimachinery/pkg/apis/meta/v1/unstructured" "k8s.io/apimachinery/pkg/util/sets" "k8s.io/kubernetes/test/e2e/framework" - e2elog "k8s.io/kubernetes/test/e2e/framework/log" "k8s.io/kubernetes/test/e2e/storage/testpatterns" "k8s.io/kubernetes/test/e2e/storage/testsuites" "k8s.io/kubernetes/test/e2e/storage/utils" @@ -187,7 +186,7 @@ func (h *hostpathCSIDriver) PrepareTest(f *framework.Framework) (*testsuites.Per }, h.manifests...) if err != nil { - e2elog.Failf("deploying %s driver: %v", h.driverInfo.Name, err) + framework.Failf("deploying %s driver: %v", h.driverInfo.Name, err) } return config, func() { @@ -333,7 +332,7 @@ func (m *mockCSIDriver) PrepareTest(f *framework.Framework) (*testsuites.PerTest }, m.manifests...) if err != nil { - e2elog.Failf("deploying csi mock driver: %v", err) + framework.Failf("deploying csi mock driver: %v", err) } return config, func() { @@ -456,7 +455,7 @@ func (g *gcePDCSIDriver) PrepareTest(f *framework.Framework) (*testsuites.PerTes cleanup, err := f.CreateFromManifests(nil, manifests...) if err != nil { - e2elog.Failf("deploying csi gce-pd driver: %v", err) + framework.Failf("deploying csi gce-pd driver: %v", err) } return &testsuites.PerTestConfig{ diff --git a/test/e2e/storage/drivers/csi_objects.go b/test/e2e/storage/drivers/csi_objects.go index 8d2e2a389b8..5935a71a791 100644 --- a/test/e2e/storage/drivers/csi_objects.go +++ b/test/e2e/storage/drivers/csi_objects.go @@ -33,21 +33,20 @@ import ( clientset "k8s.io/client-go/kubernetes" "k8s.io/kubernetes/test/e2e/framework" - e2elog "k8s.io/kubernetes/test/e2e/framework/log" ) func shredFile(filePath string) { if _, err := os.Stat(filePath); os.IsNotExist(err) { - e2elog.Logf("File %v was not found, skipping shredding", filePath) + framework.Logf("File %v was not found, skipping shredding", filePath) return } - e2elog.Logf("Shredding file %v", filePath) + framework.Logf("Shredding file %v", filePath) _, _, err := framework.RunCmd("shred", "--remove", filePath) if err != nil { - e2elog.Logf("Failed to shred file %v: %v", filePath, err) + framework.Logf("Failed to shred file %v: %v", filePath, err) } if _, err := os.Stat(filePath); os.IsNotExist(err) { - e2elog.Logf("File %v successfully shredded", filePath) + framework.Logf("File %v successfully shredded", filePath) return } // Shred failed Try to remove the file for good meausure @@ -67,13 +66,13 @@ func createGCESecrets(client clientset.Interface, ns string) { premadeSAFile, ok := os.LookupEnv(saEnv) if !ok { - e2elog.Logf("Could not find env var %v, please either create cloud-sa"+ + framework.Logf("Could not find env var %v, please either create cloud-sa"+ " secret manually or rerun test after setting %v to the filepath of"+ " the GCP Service Account to give to the GCE Persistent Disk CSI Driver", saEnv, saEnv) return } - e2elog.Logf("Found CI service account key at %v", premadeSAFile) + framework.Logf("Found CI service account key at %v", premadeSAFile) // Need to copy it saFile stdout, stderr, err := framework.RunCmd("cp", premadeSAFile, saFile) framework.ExpectNoError(err, "error copying service account key: %s\nstdout: %s\nstderr: %s", err, stdout, stderr) diff --git a/test/e2e/storage/drivers/in_tree.go b/test/e2e/storage/drivers/in_tree.go index cf66453539b..d4602ace63d 100644 --- a/test/e2e/storage/drivers/in_tree.go +++ b/test/e2e/storage/drivers/in_tree.go @@ -55,7 +55,6 @@ import ( "k8s.io/apiserver/pkg/authentication/serviceaccount" "k8s.io/kubernetes/test/e2e/framework" "k8s.io/kubernetes/test/e2e/framework/auth" - e2elog "k8s.io/kubernetes/test/e2e/framework/log" e2epod "k8s.io/kubernetes/test/e2e/framework/pod" "k8s.io/kubernetes/test/e2e/framework/volume" "k8s.io/kubernetes/test/e2e/storage/testpatterns" @@ -203,7 +202,7 @@ func (n *nfsDriver) CreateVolume(config *testsuites.PerTestConfig, volType testp case testpatterns.DynamicPV: // Do nothing default: - e2elog.Failf("Unsupported volType:%v is specified", volType) + framework.Failf("Unsupported volType:%v is specified", volType) } return nil } @@ -315,18 +314,18 @@ func (v *glusterVolume) DeleteVolume() { name := v.prefix + "-server" - e2elog.Logf("Deleting Gluster endpoints %q...", name) + framework.Logf("Deleting Gluster endpoints %q...", name) err := cs.CoreV1().Endpoints(ns.Name).Delete(name, nil) if err != nil { if !errors.IsNotFound(err) { - e2elog.Failf("Gluster delete endpoints failed: %v", err) + framework.Failf("Gluster delete endpoints failed: %v", err) } - e2elog.Logf("Gluster endpoints %q not found, assuming deleted", name) + framework.Logf("Gluster endpoints %q not found, assuming deleted", name) } - e2elog.Logf("Deleting Gluster server pod %q...", v.serverPod.Name) + framework.Logf("Deleting Gluster server pod %q...", v.serverPod.Name) err = e2epod.DeletePodWithWait(cs, v.serverPod) if err != nil { - e2elog.Failf("Gluster server pod delete failed: %v", err) + framework.Failf("Gluster server pod delete failed: %v", err) } } @@ -1069,7 +1068,7 @@ func (c *cinderDriver) CreateVolume(config *testsuites.PerTestConfig, volType te ginkgo.By("creating a test Cinder volume") output, err := exec.Command("cinder", "create", "--display-name="+volumeName, "1").CombinedOutput() outputString := string(output[:]) - e2elog.Logf("cinder output:\n%s", outputString) + framework.Logf("cinder output:\n%s", outputString) framework.ExpectNoError(err) // Parse 'id'' from stdout. Expected format: @@ -1089,7 +1088,7 @@ func (c *cinderDriver) CreateVolume(config *testsuites.PerTestConfig, volType te volumeID = fields[3] break } - e2elog.Logf("Volume ID: %s", volumeID) + framework.Logf("Volume ID: %s", volumeID) framework.ExpectNotEqual(volumeID, "") return &cinderVolume{ volumeName: volumeName, @@ -1106,16 +1105,16 @@ func (v *cinderVolume) DeleteVolume() { var err error timeout := time.Second * 120 - e2elog.Logf("Waiting up to %v for removal of cinder volume %s", timeout, name) + framework.Logf("Waiting up to %v for removal of cinder volume %s", timeout, name) for start := time.Now(); time.Since(start) < timeout; time.Sleep(5 * time.Second) { output, err = exec.Command("cinder", "delete", name).CombinedOutput() if err == nil { - e2elog.Logf("Cinder volume %s deleted", name) + framework.Logf("Cinder volume %s deleted", name) return } - e2elog.Logf("Failed to delete volume %s: %v", name, err) + framework.Logf("Failed to delete volume %s: %v", name, err) } - e2elog.Logf("Giving up deleting volume %s: %v\n%s", name, err, string(output[:])) + framework.Logf("Giving up deleting volume %s: %v\n%s", name, err, string(output[:])) } // GCE @@ -1786,7 +1785,7 @@ func (l *localDriver) CreateVolume(config *testsuites.PerTestConfig, volType tes ltr: l.ltrMgr.Create(node, l.volumeType, nil), } default: - e2elog.Failf("Unsupported volType: %v is specified", volType) + framework.Failf("Unsupported volType: %v is specified", volType) } return nil } @@ -1798,11 +1797,11 @@ func (v *localVolume) DeleteVolume() { func (l *localDriver) nodeAffinityForNode(node *v1.Node) *v1.VolumeNodeAffinity { nodeKey := "kubernetes.io/hostname" if node.Labels == nil { - e2elog.Failf("Node does not have labels") + framework.Failf("Node does not have labels") } nodeValue, found := node.Labels[nodeKey] if !found { - e2elog.Failf("Node does not have required label %q", nodeKey) + framework.Failf("Node does not have required label %q", nodeKey) } return &v1.VolumeNodeAffinity{ Required: &v1.NodeSelector{ diff --git a/test/e2e/storage/empty_dir_wrapper.go b/test/e2e/storage/empty_dir_wrapper.go index c1f557eb070..a02116503f6 100644 --- a/test/e2e/storage/empty_dir_wrapper.go +++ b/test/e2e/storage/empty_dir_wrapper.go @@ -26,7 +26,6 @@ import ( "k8s.io/apimachinery/pkg/util/intstr" "k8s.io/apimachinery/pkg/util/uuid" "k8s.io/kubernetes/test/e2e/framework" - e2elog "k8s.io/kubernetes/test/e2e/framework/log" e2epod "k8s.io/kubernetes/test/e2e/framework/pod" "k8s.io/kubernetes/test/e2e/storage/utils" imageutils "k8s.io/kubernetes/test/utils/image" @@ -79,7 +78,7 @@ var _ = utils.SIGDescribe("EmptyDir wrapper volumes", func() { var err error if secret, err = f.ClientSet.CoreV1().Secrets(f.Namespace.Name).Create(secret); err != nil { - e2elog.Failf("unable to create test secret %s: %v", secret.Name, err) + framework.Failf("unable to create test secret %s: %v", secret.Name, err) } configMapVolumeName := "configmap-volume" @@ -96,7 +95,7 @@ var _ = utils.SIGDescribe("EmptyDir wrapper volumes", func() { } if configMap, err = f.ClientSet.CoreV1().ConfigMaps(f.Namespace.Name).Create(configMap); err != nil { - e2elog.Failf("unable to create test configMap %s: %v", configMap.Name, err) + framework.Failf("unable to create test configMap %s: %v", configMap.Name, err) } pod := &v1.Pod{ @@ -148,15 +147,15 @@ var _ = utils.SIGDescribe("EmptyDir wrapper volumes", func() { defer func() { ginkgo.By("Cleaning up the secret") if err := f.ClientSet.CoreV1().Secrets(f.Namespace.Name).Delete(secret.Name, nil); err != nil { - e2elog.Failf("unable to delete secret %v: %v", secret.Name, err) + framework.Failf("unable to delete secret %v: %v", secret.Name, err) } ginkgo.By("Cleaning up the configmap") if err := f.ClientSet.CoreV1().ConfigMaps(f.Namespace.Name).Delete(configMap.Name, nil); err != nil { - e2elog.Failf("unable to delete configmap %v: %v", configMap.Name, err) + framework.Failf("unable to delete configmap %v: %v", configMap.Name, err) } ginkgo.By("Cleaning up the pod") if err = f.ClientSet.CoreV1().Pods(f.Namespace.Name).Delete(pod.Name, metav1.NewDeleteOptions(0)); err != nil { - e2elog.Failf("unable to delete pod %v: %v", pod.Name, err) + framework.Failf("unable to delete pod %v: %v", pod.Name, err) } }() }) @@ -254,17 +253,17 @@ func createGitServer(f *framework.Framework) (gitURL string, gitRepo string, cle } if gitServerSvc, err = f.ClientSet.CoreV1().Services(f.Namespace.Name).Create(gitServerSvc); err != nil { - e2elog.Failf("unable to create test git server service %s: %v", gitServerSvc.Name, err) + framework.Failf("unable to create test git server service %s: %v", gitServerSvc.Name, err) } return "http://" + gitServerSvc.Spec.ClusterIP + ":" + strconv.Itoa(httpPort), "test", func() { ginkgo.By("Cleaning up the git server pod") if err := f.ClientSet.CoreV1().Pods(f.Namespace.Name).Delete(gitServerPod.Name, metav1.NewDeleteOptions(0)); err != nil { - e2elog.Failf("unable to delete git server pod %v: %v", gitServerPod.Name, err) + framework.Failf("unable to delete git server pod %v: %v", gitServerPod.Name, err) } ginkgo.By("Cleaning up the git server svc") if err := f.ClientSet.CoreV1().Services(f.Namespace.Name).Delete(gitServerSvc.Name, nil); err != nil { - e2elog.Failf("unable to delete git server svc %v: %v", gitServerSvc.Name, err) + framework.Failf("unable to delete git server svc %v: %v", gitServerSvc.Name, err) } } } diff --git a/test/e2e/storage/ephemeral_volume.go b/test/e2e/storage/ephemeral_volume.go index 45c98ad9e40..057d21e9c78 100644 --- a/test/e2e/storage/ephemeral_volume.go +++ b/test/e2e/storage/ephemeral_volume.go @@ -26,7 +26,6 @@ import ( "k8s.io/apimachinery/pkg/util/rand" clientset "k8s.io/client-go/kubernetes" "k8s.io/kubernetes/test/e2e/framework" - e2elog "k8s.io/kubernetes/test/e2e/framework/log" e2epod "k8s.io/kubernetes/test/e2e/framework/pod" "k8s.io/kubernetes/test/e2e/storage/utils" imageutils "k8s.io/kubernetes/test/utils/image" @@ -60,7 +59,7 @@ var _ = utils.SIGDescribe("Ephemeralstorage", func() { // Allow it to sleep for 30 seconds time.Sleep(30 * time.Second) - e2elog.Logf("Deleting pod %q/%q", pod.Namespace, pod.Name) + framework.Logf("Deleting pod %q/%q", pod.Namespace, pod.Name) framework.ExpectNoError(e2epod.DeletePodWithWait(c, pod)) }) } diff --git a/test/e2e/storage/flexvolume.go b/test/e2e/storage/flexvolume.go index 8ae2ad99048..736c245d99d 100644 --- a/test/e2e/storage/flexvolume.go +++ b/test/e2e/storage/flexvolume.go @@ -29,7 +29,6 @@ import ( apierrs "k8s.io/apimachinery/pkg/api/errors" clientset "k8s.io/client-go/kubernetes" "k8s.io/kubernetes/test/e2e/framework" - e2elog "k8s.io/kubernetes/test/e2e/framework/log" e2enode "k8s.io/kubernetes/test/e2e/framework/node" e2essh "k8s.io/kubernetes/test/e2e/framework/ssh" "k8s.io/kubernetes/test/e2e/framework/testfiles" @@ -119,7 +118,7 @@ func uninstallFlex(c clientset.Interface, node *v1.Node, vendor, driver string) } if host == "" { - e2elog.Failf("Error getting node ip : %v", err) + framework.Failf("Error getting node ip : %v", err) } cmd := fmt.Sprintf("sudo rm -r %s", flexDir) @@ -140,7 +139,7 @@ func sshAndLog(cmd, host string, failOnError bool) { e2essh.LogResult(result) framework.ExpectNoError(err) if result.Code != 0 && failOnError { - e2elog.Failf("%s returned non-zero, stderr: %s", cmd, result.Stderr) + framework.Failf("%s returned non-zero, stderr: %s", cmd, result.Stderr) } } diff --git a/test/e2e/storage/flexvolume_mounted_volume_resize.go b/test/e2e/storage/flexvolume_mounted_volume_resize.go index b05c4be426d..718c41e922e 100644 --- a/test/e2e/storage/flexvolume_mounted_volume_resize.go +++ b/test/e2e/storage/flexvolume_mounted_volume_resize.go @@ -31,7 +31,6 @@ import ( clientset "k8s.io/client-go/kubernetes" "k8s.io/kubernetes/test/e2e/framework" e2edeploy "k8s.io/kubernetes/test/e2e/framework/deployment" - e2elog "k8s.io/kubernetes/test/e2e/framework/log" e2epod "k8s.io/kubernetes/test/e2e/framework/pod" "k8s.io/kubernetes/test/e2e/storage/testsuites" "k8s.io/kubernetes/test/e2e/storage/utils" @@ -70,7 +69,7 @@ var _ = utils.SIGDescribe("Mounted flexvolume expand[Slow]", func() { if len(nodeList.Items) != 0 { nodeName = nodeList.Items[0].Name } else { - e2elog.Failf("Unable to find ready and schedulable Node") + framework.Failf("Unable to find ready and schedulable Node") } nodeKey = "mounted_flexvolume_expand" @@ -112,11 +111,11 @@ var _ = utils.SIGDescribe("Mounted flexvolume expand[Slow]", func() { }) ginkgo.AfterEach(func() { - e2elog.Logf("AfterEach: Cleaning up resources for mounted volume resize") + framework.Logf("AfterEach: Cleaning up resources for mounted volume resize") if c != nil { if errs := framework.PVPVCCleanup(c, ns, nil, pvc); len(errs) > 0 { - e2elog.Failf("AfterEach: Failed to delete PVC and/or PV. Errors: %v", utilerrors.NewAggregate(errs)) + framework.Failf("AfterEach: Failed to delete PVC and/or PV. Errors: %v", utilerrors.NewAggregate(errs)) } pvc, nodeName, isNodeLabeled, nodeLabelValue = nil, "", false, "" nodeKeyValueLabel = make(map[string]string) @@ -167,7 +166,7 @@ var _ = utils.SIGDescribe("Mounted flexvolume expand[Slow]", func() { pvcSize := pvc.Spec.Resources.Requests[v1.ResourceStorage] if pvcSize.Cmp(newSize) != 0 { - e2elog.Failf("error updating pvc size %q", pvc.Name) + framework.Failf("error updating pvc size %q", pvc.Name) } ginkgo.By("Waiting for cloudprovider resize to finish") diff --git a/test/e2e/storage/flexvolume_online_resize.go b/test/e2e/storage/flexvolume_online_resize.go index ad1c92235bb..da8be137e55 100644 --- a/test/e2e/storage/flexvolume_online_resize.go +++ b/test/e2e/storage/flexvolume_online_resize.go @@ -28,7 +28,6 @@ import ( utilerrors "k8s.io/apimachinery/pkg/util/errors" clientset "k8s.io/client-go/kubernetes" "k8s.io/kubernetes/test/e2e/framework" - e2elog "k8s.io/kubernetes/test/e2e/framework/log" e2epod "k8s.io/kubernetes/test/e2e/framework/pod" "k8s.io/kubernetes/test/e2e/storage/testsuites" "k8s.io/kubernetes/test/e2e/storage/utils" @@ -61,7 +60,7 @@ var _ = utils.SIGDescribe("Mounted flexvolume volume expand [Slow] [Feature:Expa nodeList = framework.GetReadySchedulableNodesOrDie(f.ClientSet) if len(nodeList.Items) == 0 { - e2elog.Failf("unable to find ready and schedulable Node") + framework.Failf("unable to find ready and schedulable Node") } nodeName = nodeList.Items[0].Name @@ -105,11 +104,11 @@ var _ = utils.SIGDescribe("Mounted flexvolume volume expand [Slow] [Feature:Expa }) ginkgo.AfterEach(func() { - e2elog.Logf("AfterEach: Cleaning up resources for mounted volume resize") + framework.Logf("AfterEach: Cleaning up resources for mounted volume resize") if c != nil { if errs := framework.PVPVCCleanup(c, ns, nil, pvc); len(errs) > 0 { - e2elog.Failf("AfterEach: Failed to delete PVC and/or PV. Errors: %v", utilerrors.NewAggregate(errs)) + framework.Failf("AfterEach: Failed to delete PVC and/or PV. Errors: %v", utilerrors.NewAggregate(errs)) } pvc, nodeName, isNodeLabeled, nodeLabelValue = nil, "", false, "" nodeKeyValueLabel = make(map[string]string) @@ -167,7 +166,7 @@ var _ = utils.SIGDescribe("Mounted flexvolume volume expand [Slow] [Feature:Expa pvcSize := pvc.Spec.Resources.Requests[v1.ResourceStorage] if pvcSize.Cmp(newSize) != 0 { - e2elog.Failf("error updating pvc size %q", pvc.Name) + framework.Failf("error updating pvc size %q", pvc.Name) } ginkgo.By("Waiting for cloudprovider resize to finish") diff --git a/test/e2e/storage/generic_persistent_volume-disruptive.go b/test/e2e/storage/generic_persistent_volume-disruptive.go index 2a6a3b4c9dc..66210917bc9 100644 --- a/test/e2e/storage/generic_persistent_volume-disruptive.go +++ b/test/e2e/storage/generic_persistent_volume-disruptive.go @@ -22,7 +22,6 @@ import ( v1 "k8s.io/api/core/v1" clientset "k8s.io/client-go/kubernetes" "k8s.io/kubernetes/test/e2e/framework" - e2elog "k8s.io/kubernetes/test/e2e/framework/log" e2epod "k8s.io/kubernetes/test/e2e/framework/pod" "k8s.io/kubernetes/test/e2e/storage/testsuites" "k8s.io/kubernetes/test/e2e/storage/utils" @@ -66,7 +65,7 @@ var _ = utils.SIGDescribe("GenericPersistentVolume[Disruptive]", func() { pv *v1.PersistentVolume ) ginkgo.BeforeEach(func() { - e2elog.Logf("Initializing pod and pvcs for test") + framework.Logf("Initializing pod and pvcs for test") clientPod, pvc, pv = createPodPVCFromSC(f, c, ns) }) for _, test := range disruptiveTestTable { @@ -78,7 +77,7 @@ var _ = utils.SIGDescribe("GenericPersistentVolume[Disruptive]", func() { }(test) } ginkgo.AfterEach(func() { - e2elog.Logf("Tearing down test spec") + framework.Logf("Tearing down test spec") tearDownTestCase(c, f, ns, clientPod, pvc, pv, false) pvc, clientPod = nil, nil }) diff --git a/test/e2e/storage/mounted_volume_resize.go b/test/e2e/storage/mounted_volume_resize.go index 4f6911a33e2..a29d9296571 100644 --- a/test/e2e/storage/mounted_volume_resize.go +++ b/test/e2e/storage/mounted_volume_resize.go @@ -32,7 +32,6 @@ import ( "k8s.io/kubernetes/pkg/client/conditions" "k8s.io/kubernetes/test/e2e/framework" e2edeploy "k8s.io/kubernetes/test/e2e/framework/deployment" - e2elog "k8s.io/kubernetes/test/e2e/framework/log" e2epod "k8s.io/kubernetes/test/e2e/framework/pod" "k8s.io/kubernetes/test/e2e/storage/testsuites" "k8s.io/kubernetes/test/e2e/storage/utils" @@ -63,7 +62,7 @@ var _ = utils.SIGDescribe("Mounted volume expand", func() { if len(nodeList.Items) != 0 { nodeName = nodeList.Items[0].Name } else { - e2elog.Failf("Unable to find ready and schedulable Node") + framework.Failf("Unable to find ready and schedulable Node") } nodeKey = "mounted_volume_expand" @@ -102,11 +101,11 @@ var _ = utils.SIGDescribe("Mounted volume expand", func() { }) ginkgo.AfterEach(func() { - e2elog.Logf("AfterEach: Cleaning up resources for mounted volume resize") + framework.Logf("AfterEach: Cleaning up resources for mounted volume resize") if c != nil { if errs := framework.PVPVCCleanup(c, ns, nil, pvc); len(errs) > 0 { - e2elog.Failf("AfterEach: Failed to delete PVC and/or PV. Errors: %v", utilerrors.NewAggregate(errs)) + framework.Failf("AfterEach: Failed to delete PVC and/or PV. Errors: %v", utilerrors.NewAggregate(errs)) } pvc, nodeName, isNodeLabeled, nodeLabelValue = nil, "", false, "" nodeKeyValueLabel = make(map[string]string) @@ -139,7 +138,7 @@ var _ = utils.SIGDescribe("Mounted volume expand", func() { pvcSize := pvc.Spec.Resources.Requests[v1.ResourceStorage] if pvcSize.Cmp(newSize) != 0 { - e2elog.Failf("error updating pvc size %q", pvc.Name) + framework.Failf("error updating pvc size %q", pvc.Name) } ginkgo.By("Waiting for cloudprovider resize to finish") diff --git a/test/e2e/storage/nfs_persistent_volume-disruptive.go b/test/e2e/storage/nfs_persistent_volume-disruptive.go index 3894a5c430f..988b9fbfd5d 100644 --- a/test/e2e/storage/nfs_persistent_volume-disruptive.go +++ b/test/e2e/storage/nfs_persistent_volume-disruptive.go @@ -29,7 +29,6 @@ import ( utilerrors "k8s.io/apimachinery/pkg/util/errors" clientset "k8s.io/client-go/kubernetes" "k8s.io/kubernetes/test/e2e/framework" - e2elog "k8s.io/kubernetes/test/e2e/framework/log" e2enode "k8s.io/kubernetes/test/e2e/framework/node" e2epod "k8s.io/kubernetes/test/e2e/framework/pod" "k8s.io/kubernetes/test/e2e/framework/volume" @@ -87,7 +86,7 @@ var _ = utils.SIGDescribe("NFSPersistentVolumes[Disruptive][Flaky]", func() { // Get the first ready node IP that is not hosting the NFS pod. var err error if clientNodeIP == "" { - e2elog.Logf("Designating test node") + framework.Logf("Designating test node") nodes := framework.GetReadySchedulableNodesOrDie(c) for _, node := range nodes.Items { if node.Name != nfsServerPod.Spec.NodeName { @@ -157,11 +156,11 @@ var _ = utils.SIGDescribe("NFSPersistentVolumes[Disruptive][Flaky]", func() { // Delete PV and PVCs if errs := framework.PVPVCCleanup(c, ns, pv1, pvc1); len(errs) > 0 { - e2elog.Failf("AfterEach: Failed to delete PVC and/or PV. Errors: %v", utilerrors.NewAggregate(errs)) + framework.Failf("AfterEach: Failed to delete PVC and/or PV. Errors: %v", utilerrors.NewAggregate(errs)) } pv1, pvc1 = nil, nil if errs := framework.PVPVCCleanup(c, ns, pv2, pvc2); len(errs) > 0 { - e2elog.Failf("AfterEach: Failed to delete PVC and/or PV. Errors: %v", utilerrors.NewAggregate(errs)) + framework.Failf("AfterEach: Failed to delete PVC and/or PV. Errors: %v", utilerrors.NewAggregate(errs)) } pv2, pvc2 = nil, nil @@ -187,7 +186,7 @@ var _ = utils.SIGDescribe("NFSPersistentVolumes[Disruptive][Flaky]", func() { framework.ExpectNoError(err) err = framework.WaitForControllerManagerUp() framework.ExpectNoError(err) - e2elog.Logf("kube-controller-manager restarted") + framework.Logf("kube-controller-manager restarted") ginkgo.By("Observing the kube-controller-manager healthy for at least 2 minutes") // Continue checking for 2 minutes to make sure kube-controller-manager is healthy @@ -205,12 +204,12 @@ var _ = utils.SIGDescribe("NFSPersistentVolumes[Disruptive][Flaky]", func() { ) ginkgo.BeforeEach(func() { - e2elog.Logf("Initializing test spec") + framework.Logf("Initializing test spec") clientPod, pv, pvc = initTestCase(f, c, nfsPVconfig, pvcConfig, ns, clientNode.Name) }) ginkgo.AfterEach(func() { - e2elog.Logf("Tearing down test spec") + framework.Logf("Tearing down test spec") tearDownTestCase(c, f, ns, clientPod, pvc, pv, true /* force PV delete */) pv, pvc, clientPod = nil, nil, nil }) @@ -258,9 +257,9 @@ func initTestCase(f *framework.Framework, c clientset.Interface, pvConfig framew framework.ExpectNoError(err) pod := e2epod.MakePod(ns, nil, []*v1.PersistentVolumeClaim{pvc}, true, "") pod.Spec.NodeName = nodeName - e2elog.Logf("Creating NFS client pod.") + framework.Logf("Creating NFS client pod.") pod, err = c.CoreV1().Pods(ns).Create(pod) - e2elog.Logf("NFS client Pod %q created on Node %q", pod.Name, nodeName) + framework.Logf("NFS client Pod %q created on Node %q", pod.Name, nodeName) framework.ExpectNoError(err) defer func() { if err != nil { diff --git a/test/e2e/storage/pd.go b/test/e2e/storage/pd.go index e89978477fa..ab88786f97e 100644 --- a/test/e2e/storage/pd.go +++ b/test/e2e/storage/pd.go @@ -39,7 +39,6 @@ import ( clientset "k8s.io/client-go/kubernetes" v1core "k8s.io/client-go/kubernetes/typed/core/v1" "k8s.io/kubernetes/test/e2e/framework" - e2elog "k8s.io/kubernetes/test/e2e/framework/log" e2enode "k8s.io/kubernetes/test/e2e/framework/node" "k8s.io/kubernetes/test/e2e/framework/providers/gce" "k8s.io/kubernetes/test/e2e/storage/utils" @@ -146,7 +145,7 @@ var _ = utils.SIGDescribe("Pod Disks", func() { ginkgo.By("deleting the fmtPod") framework.ExpectNoError(podClient.Delete(fmtPod.Name, metav1.NewDeleteOptions(0)), "Failed to delete fmtPod") - e2elog.Logf("deleted fmtPod %q", fmtPod.Name) + framework.Logf("deleted fmtPod %q", fmtPod.Name) ginkgo.By("waiting for PD to detach") framework.ExpectNoError(waitForPDDetach(diskName, host0Name)) } @@ -158,7 +157,7 @@ var _ = utils.SIGDescribe("Pod Disks", func() { defer func() { // Teardown should do nothing unless test failed ginkgo.By("defer: cleaning up PD-RW test environment") - e2elog.Logf("defer cleanup errors can usually be ignored") + framework.Logf("defer cleanup errors can usually be ignored") if fmtPod != nil { podClient.Delete(fmtPod.Name, podDelOpt) } @@ -171,7 +170,7 @@ var _ = utils.SIGDescribe("Pod Disks", func() { _, err = podClient.Create(host0Pod) framework.ExpectNoError(err, fmt.Sprintf("Failed to create host0Pod: %v", err)) framework.ExpectNoError(f.WaitForPodRunningSlow(host0Pod.Name)) - e2elog.Logf("host0Pod: %q, node0: %q", host0Pod.Name, host0Name) + framework.Logf("host0Pod: %q, node0: %q", host0Pod.Name, host0Name) var containerName, testFile, testFileContents string if !readOnly { @@ -180,36 +179,36 @@ var _ = utils.SIGDescribe("Pod Disks", func() { testFile = "/testpd1/tracker" testFileContents = fmt.Sprintf("%v", rand.Int()) framework.ExpectNoError(f.WriteFileViaContainer(host0Pod.Name, containerName, testFile, testFileContents)) - e2elog.Logf("wrote %q to file %q in pod %q on node %q", testFileContents, testFile, host0Pod.Name, host0Name) + framework.Logf("wrote %q to file %q in pod %q on node %q", testFileContents, testFile, host0Pod.Name, host0Name) ginkgo.By("verifying PD is present in node0's VolumeInUse list") framework.ExpectNoError(waitForPDInVolumesInUse(nodeClient, diskName, host0Name, nodeStatusTimeout, true /* shouldExist */)) ginkgo.By("deleting host0Pod") // delete this pod before creating next pod framework.ExpectNoError(podClient.Delete(host0Pod.Name, podDelOpt), "Failed to delete host0Pod") - e2elog.Logf("deleted host0Pod %q", host0Pod.Name) + framework.Logf("deleted host0Pod %q", host0Pod.Name) } ginkgo.By("creating host1Pod on node1") _, err = podClient.Create(host1Pod) framework.ExpectNoError(err, "Failed to create host1Pod") framework.ExpectNoError(f.WaitForPodRunningSlow(host1Pod.Name)) - e2elog.Logf("host1Pod: %q, node1: %q", host1Pod.Name, host1Name) + framework.Logf("host1Pod: %q, node1: %q", host1Pod.Name, host1Name) if readOnly { ginkgo.By("deleting host0Pod") framework.ExpectNoError(podClient.Delete(host0Pod.Name, podDelOpt), "Failed to delete host0Pod") - e2elog.Logf("deleted host0Pod %q", host0Pod.Name) + framework.Logf("deleted host0Pod %q", host0Pod.Name) } else { ginkgo.By("verifying PD contents in host1Pod") verifyPDContentsViaContainer(f, host1Pod.Name, containerName, map[string]string{testFile: testFileContents}) - e2elog.Logf("verified PD contents in pod %q", host1Pod.Name) + framework.Logf("verified PD contents in pod %q", host1Pod.Name) ginkgo.By("verifying PD is removed from node0") framework.ExpectNoError(waitForPDInVolumesInUse(nodeClient, diskName, host0Name, nodeStatusTimeout, false /* shouldExist */)) - e2elog.Logf("PD %q removed from node %q's VolumeInUse list", diskName, host1Pod.Name) + framework.Logf("PD %q removed from node %q's VolumeInUse list", diskName, host1Pod.Name) } ginkgo.By("deleting host1Pod") framework.ExpectNoError(podClient.Delete(host1Pod.Name, podDelOpt), "Failed to delete host1Pod") - e2elog.Logf("deleted host1Pod %q", host1Pod.Name) + framework.Logf("deleted host1Pod %q", host1Pod.Name) ginkgo.By("Test completed successfully, waiting for PD to detach from both nodes") waitForPDDetach(diskName, host0Name) @@ -258,7 +257,7 @@ var _ = utils.SIGDescribe("Pod Disks", func() { defer func() { // Teardown should do nothing unless test failed. ginkgo.By("defer: cleaning up PD-RW test environment") - e2elog.Logf("defer cleanup errors can usually be ignored") + framework.Logf("defer cleanup errors can usually be ignored") if host0Pod != nil { podClient.Delete(host0Pod.Name, metav1.NewDeleteOptions(0)) } @@ -268,7 +267,7 @@ var _ = utils.SIGDescribe("Pod Disks", func() { }() for i := 0; i < t.repeatCnt; i++ { // "rapid" repeat loop - e2elog.Logf("PD Read/Writer Iteration #%v", i) + framework.Logf("PD Read/Writer Iteration #%v", i) ginkgo.By(fmt.Sprintf("creating host0Pod with %d containers on node0", numContainers)) host0Pod = testPDPod(diskNames, host0Name, false /* readOnly */, numContainers) _, err = podClient.Create(host0Pod) @@ -285,7 +284,7 @@ var _ = utils.SIGDescribe("Pod Disks", func() { testFileContents := fmt.Sprintf("%v", rand.Int()) fileAndContentToVerify[testFile] = testFileContents framework.ExpectNoError(f.WriteFileViaContainer(host0Pod.Name, containerName, testFile, testFileContents)) - e2elog.Logf("wrote %q to file %q in pod %q (container %q) on node %q", testFileContents, testFile, host0Pod.Name, containerName, host0Name) + framework.Logf("wrote %q to file %q in pod %q (container %q) on node %q", testFileContents, testFile, host0Pod.Name, containerName, host0Name) } ginkgo.By("verifying PD contents via a container") @@ -346,7 +345,7 @@ var _ = utils.SIGDescribe("Pod Disks", func() { defer func() { ginkgo.By("defer: cleaning up PD-RW test env") - e2elog.Logf("defer cleanup errors can usually be ignored") + framework.Logf("defer cleanup errors can usually be ignored") ginkgo.By("defer: delete host0Pod") podClient.Delete(host0Pod.Name, metav1.NewDeleteOptions(0)) ginkgo.By("defer: detach and delete PDs") @@ -364,7 +363,7 @@ var _ = utils.SIGDescribe("Pod Disks", func() { // if this defer is reached due to an Expect then nested // Expects are lost, so use Failf here if numNodes != origNodeCnt { - e2elog.Failf("defer: Requires current node count (%d) to return to original node count (%d)", numNodes, origNodeCnt) + framework.Failf("defer: Requires current node count (%d) to return to original node count (%d)", numNodes, origNodeCnt) } } }() @@ -379,7 +378,7 @@ var _ = utils.SIGDescribe("Pod Disks", func() { testFile := "/testpd1/tracker" testFileContents := fmt.Sprintf("%v", rand.Int()) framework.ExpectNoError(f.WriteFileViaContainer(host0Pod.Name, containerName, testFile, testFileContents)) - e2elog.Logf("wrote %q to file %q in pod %q on node %q", testFileContents, testFile, host0Pod.Name, host0Name) + framework.Logf("wrote %q to file %q in pod %q on node %q", testFileContents, testFile, host0Pod.Name, host0Name) ginkgo.By("verifying PD is present in node0's VolumeInUse list") framework.ExpectNoError(waitForPDInVolumesInUse(nodeClient, diskName, host0Name, nodeStatusTimeout, true /* should exist*/)) @@ -455,17 +454,17 @@ func verifyPDContentsViaContainer(f *framework.Framework, podName, containerName v, err := f.ReadFileViaContainer(podName, containerName, filePath) value = v if err != nil { - e2elog.Logf("Error reading file: %v", err) + framework.Logf("Error reading file: %v", err) } framework.ExpectNoError(err) - e2elog.Logf("Read file %q with content: %v (iteration %d)", filePath, v, i) + framework.Logf("Read file %q with content: %v (iteration %d)", filePath, v, i) if strings.TrimSpace(v) != strings.TrimSpace(expectedContents) { - e2elog.Logf("Warning: read content <%q> does not match execpted content <%q>.", v, expectedContents) + framework.Logf("Warning: read content <%q> does not match execpted content <%q>.", v, expectedContents) size, err := f.CheckFileSizeViaContainer(podName, containerName, filePath) if err != nil { - e2elog.Logf("Error checking file size: %v", err) + framework.Logf("Error checking file size: %v", err) } - e2elog.Logf("Check file %q size: %q", filePath, size) + framework.Logf("Check file %q size: %q", filePath, size) } else { break } @@ -486,7 +485,7 @@ func detachPD(nodeName types.NodeName, pdName string) error { // PD already detached, ignore error. return nil } - e2elog.Logf("Error detaching PD %q: %v", pdName, err) + framework.Logf("Error detaching PD %q: %v", pdName, err) } return err @@ -518,7 +517,7 @@ func testPDPod(diskNames []string, targetNode types.NodeName, readOnly bool, num // escape if not a supported provider if !(framework.TestContext.Provider == "gce" || framework.TestContext.Provider == "gke" || framework.TestContext.Provider == "aws") { - e2elog.Failf(fmt.Sprintf("func `testPDPod` only supports gce, gke, and aws providers, not %v", framework.TestContext.Provider)) + framework.Failf(fmt.Sprintf("func `testPDPod` only supports gce, gke, and aws providers, not %v", framework.TestContext.Provider)) } containers := make([]v1.Container, numContainers) @@ -579,7 +578,7 @@ func testPDPod(diskNames []string, targetNode types.NodeName, readOnly bool, num // Waits for specified PD to detach from specified hostName func waitForPDDetach(diskName string, nodeName types.NodeName) error { if framework.TestContext.Provider == "gce" || framework.TestContext.Provider == "gke" { - e2elog.Logf("Waiting for GCE PD %q to detach from node %q.", diskName, nodeName) + framework.Logf("Waiting for GCE PD %q to detach from node %q.", diskName, nodeName) gceCloud, err := gce.GetGCECloud() if err != nil { return err @@ -587,15 +586,15 @@ func waitForPDDetach(diskName string, nodeName types.NodeName) error { for start := time.Now(); time.Since(start) < gcePDDetachTimeout; time.Sleep(gcePDDetachPollTime) { diskAttached, err := gceCloud.DiskIsAttached(diskName, nodeName) if err != nil { - e2elog.Logf("Error waiting for PD %q to detach from node %q. 'DiskIsAttached(...)' failed with %v", diskName, nodeName, err) + framework.Logf("Error waiting for PD %q to detach from node %q. 'DiskIsAttached(...)' failed with %v", diskName, nodeName, err) return err } if !diskAttached { // Specified disk does not appear to be attached to specified node - e2elog.Logf("GCE PD %q appears to have successfully detached from %q.", diskName, nodeName) + framework.Logf("GCE PD %q appears to have successfully detached from %q.", diskName, nodeName) return nil } - e2elog.Logf("Waiting for GCE PD %q to detach from %q.", diskName, nodeName) + framework.Logf("Waiting for GCE PD %q to detach from %q.", diskName, nodeName) } return fmt.Errorf("Gave up waiting for GCE PD %q to detach from %q after %v", diskName, nodeName, gcePDDetachTimeout) } @@ -604,7 +603,7 @@ func waitForPDDetach(diskName string, nodeName types.NodeName) error { func detachAndDeletePDs(diskName string, hosts []types.NodeName) { for _, host := range hosts { - e2elog.Logf("Detaching GCE PD %q from node %q.", diskName, host) + framework.Logf("Detaching GCE PD %q from node %q.", diskName, host) detachPD(host, diskName) ginkgo.By(fmt.Sprintf("Waiting for PD %q to detach from %q", diskName, host)) waitForPDDetach(diskName, host) @@ -623,11 +622,11 @@ func waitForPDInVolumesInUse( if !shouldExist { logStr = "to NOT contain" } - e2elog.Logf("Waiting for node %s's VolumesInUse Status %s PD %q", nodeName, logStr, diskName) + framework.Logf("Waiting for node %s's VolumesInUse Status %s PD %q", nodeName, logStr, diskName) for start := time.Now(); time.Since(start) < timeout; time.Sleep(nodeStatusPollTime) { nodeObj, err := nodeClient.Get(string(nodeName), metav1.GetOptions{}) if err != nil || nodeObj == nil { - e2elog.Logf("Failed to fetch node object %q from API server. err=%v", nodeName, err) + framework.Logf("Failed to fetch node object %q from API server. err=%v", nodeName, err) continue } exists := false @@ -635,14 +634,14 @@ func waitForPDInVolumesInUse( volumeInUseStr := string(volumeInUse) if strings.Contains(volumeInUseStr, diskName) { if shouldExist { - e2elog.Logf("Found PD %q in node %q's VolumesInUse Status: %q", diskName, nodeName, volumeInUseStr) + framework.Logf("Found PD %q in node %q's VolumesInUse Status: %q", diskName, nodeName, volumeInUseStr) return nil } exists = true } } if !shouldExist && !exists { - e2elog.Logf("Verified PD %q does not exist in node %q's VolumesInUse Status.", diskName, nodeName) + framework.Logf("Verified PD %q does not exist in node %q's VolumesInUse Status.", diskName, nodeName) return nil } } diff --git a/test/e2e/storage/persistent_volumes-gce.go b/test/e2e/storage/persistent_volumes-gce.go index 36454da9222..1612f9a97cd 100644 --- a/test/e2e/storage/persistent_volumes-gce.go +++ b/test/e2e/storage/persistent_volumes-gce.go @@ -26,7 +26,6 @@ import ( utilerrors "k8s.io/apimachinery/pkg/util/errors" clientset "k8s.io/client-go/kubernetes" "k8s.io/kubernetes/test/e2e/framework" - e2elog "k8s.io/kubernetes/test/e2e/framework/log" e2epod "k8s.io/kubernetes/test/e2e/framework/pod" "k8s.io/kubernetes/test/e2e/framework/providers/gce" "k8s.io/kubernetes/test/e2e/storage/utils" @@ -106,11 +105,11 @@ var _ = utils.SIGDescribe("PersistentVolumes GCEPD", func() { }) ginkgo.AfterEach(func() { - e2elog.Logf("AfterEach: Cleaning up test resources") + framework.Logf("AfterEach: Cleaning up test resources") if c != nil { framework.ExpectNoError(e2epod.DeletePodWithWait(c, clientPod)) if errs := framework.PVPVCCleanup(c, ns, pv, pvc); len(errs) > 0 { - e2elog.Failf("AfterEach: Failed to delete PVC and/or PV. Errors: %v", utilerrors.NewAggregate(errs)) + framework.Failf("AfterEach: Failed to delete PVC and/or PV. Errors: %v", utilerrors.NewAggregate(errs)) } clientPod, pv, pvc, node = nil, nil, nil, "" if diskName != "" { diff --git a/test/e2e/storage/persistent_volumes-local.go b/test/e2e/storage/persistent_volumes-local.go index ab9df34473a..892b5239a4f 100644 --- a/test/e2e/storage/persistent_volumes-local.go +++ b/test/e2e/storage/persistent_volumes-local.go @@ -39,7 +39,6 @@ import ( "k8s.io/apimachinery/pkg/watch" clientset "k8s.io/client-go/kubernetes" "k8s.io/kubernetes/test/e2e/framework" - e2elog "k8s.io/kubernetes/test/e2e/framework/log" e2epod "k8s.io/kubernetes/test/e2e/framework/pod" e2esset "k8s.io/kubernetes/test/e2e/framework/statefulset" "k8s.io/kubernetes/test/e2e/storage/utils" @@ -570,7 +569,7 @@ var _ = utils.SIGDescribe("PersistentVolumes-local ", func() { for _, pod := range pods { if err := deletePodAndPVCs(config, pod); err != nil { - e2elog.Logf("Deleting pod %v failed: %v", pod.Name, err) + framework.Logf("Deleting pod %v failed: %v", pod.Name, err) } } }() @@ -594,7 +593,7 @@ var _ = utils.SIGDescribe("PersistentVolumes-local ", func() { } delete(pods, pod.Name) numFinished++ - e2elog.Logf("%v/%v pods finished", numFinished, totalPods) + framework.Logf("%v/%v pods finished", numFinished, totalPods) case v1.PodFailed: case v1.PodUnknown: return false, fmt.Errorf("pod %v is in %v phase", pod.Name, pod.Status.Phase) @@ -674,7 +673,7 @@ var _ = utils.SIGDescribe("PersistentVolumes-local ", func() { }) func deletePodAndPVCs(config *localTestConfig, pod *v1.Pod) error { - e2elog.Logf("Deleting pod %v", pod.Name) + framework.Logf("Deleting pod %v", pod.Name) if err := config.client.CoreV1().Pods(config.ns).Delete(pod.Name, nil); err != nil { return err } @@ -827,7 +826,7 @@ func cleanupLocalPVCsPVs(config *localTestConfig, volumes []*localTestVolume) { ginkgo.By("Cleaning up PVC and PV") errs := framework.PVPVCCleanup(config.client, config.ns, volume.pv, volume.pvc) if len(errs) > 0 { - e2elog.Failf("Failed to delete PV and/or PVC: %v", utilerrors.NewAggregate(errs)) + framework.Failf("Failed to delete PV and/or PVC: %v", utilerrors.NewAggregate(errs)) } } } @@ -848,7 +847,7 @@ func verifyLocalVolume(config *localTestConfig, volume *localTestVolume) { func verifyLocalPod(config *localTestConfig, volume *localTestVolume, pod *v1.Pod, expectedNodeName string) { podNodeName, err := podNodeName(config, pod) framework.ExpectNoError(err) - e2elog.Logf("pod %q created on Node %q", pod.Name, podNodeName) + framework.Logf("pod %q created on Node %q", pod.Name, podNodeName) framework.ExpectEqual(podNodeName, expectedNodeName) } @@ -868,11 +867,11 @@ func makeLocalPVConfig(config *localTestConfig, volume *localTestVolume) framewo // TODO: hostname may not be the best option nodeKey := "kubernetes.io/hostname" if volume.ltr.Node.Labels == nil { - e2elog.Failf("Node does not have labels") + framework.Failf("Node does not have labels") } nodeValue, found := volume.ltr.Node.Labels[nodeKey] if !found { - e2elog.Failf("Node does not have required label %q", nodeKey) + framework.Failf("Node does not have required label %q", nodeKey) } pvConfig := framework.PersistentVolumeConfig{ @@ -1031,7 +1030,7 @@ func testReadFileContent(testFileDir string, testFile string, testFileContent st // Fail on error func podRWCmdExec(pod *v1.Pod, cmd string) string { out, err := utils.PodExec(pod, cmd) - e2elog.Logf("podRWCmdExec out: %q err: %v", out, err) + framework.Logf("podRWCmdExec out: %q err: %v", out, err) framework.ExpectNoError(err) return out } diff --git a/test/e2e/storage/persistent_volumes.go b/test/e2e/storage/persistent_volumes.go index 27564b98ac8..614c40cf9d4 100644 --- a/test/e2e/storage/persistent_volumes.go +++ b/test/e2e/storage/persistent_volumes.go @@ -29,7 +29,6 @@ import ( utilerrors "k8s.io/apimachinery/pkg/util/errors" clientset "k8s.io/client-go/kubernetes" "k8s.io/kubernetes/test/e2e/framework" - e2elog "k8s.io/kubernetes/test/e2e/framework/log" e2epod "k8s.io/kubernetes/test/e2e/framework/pod" e2esset "k8s.io/kubernetes/test/e2e/framework/statefulset" "k8s.io/kubernetes/test/e2e/framework/volume" @@ -153,9 +152,9 @@ var _ = utils.SIGDescribe("PersistentVolumes", func() { ginkgo.Context("with Single PV - PVC pairs", func() { // Note: this is the only code where the pv is deleted. ginkgo.AfterEach(func() { - e2elog.Logf("AfterEach: Cleaning up test resources.") + framework.Logf("AfterEach: Cleaning up test resources.") if errs := framework.PVPVCCleanup(c, ns, pv, pvc); len(errs) > 0 { - e2elog.Failf("AfterEach: Failed to delete PVC and/or PV. Errors: %v", utilerrors.NewAggregate(errs)) + framework.Failf("AfterEach: Failed to delete PVC and/or PV. Errors: %v", utilerrors.NewAggregate(errs)) } }) @@ -215,14 +214,14 @@ var _ = utils.SIGDescribe("PersistentVolumes", func() { var claims framework.PVCMap ginkgo.AfterEach(func() { - e2elog.Logf("AfterEach: deleting %v PVCs and %v PVs...", len(claims), len(pvols)) + framework.Logf("AfterEach: deleting %v PVCs and %v PVs...", len(claims), len(pvols)) errs := framework.PVPVCMapCleanup(c, ns, pvols, claims) if len(errs) > 0 { errmsg := []string{} for _, e := range errs { errmsg = append(errmsg, e.Error()) } - e2elog.Failf("AfterEach: Failed to delete 1 or more PVs/PVCs. Errors: %v", strings.Join(errmsg, "; ")) + framework.Failf("AfterEach: Failed to delete 1 or more PVs/PVCs. Errors: %v", strings.Join(errmsg, "; ")) } }) @@ -269,9 +268,9 @@ var _ = utils.SIGDescribe("PersistentVolumes", func() { }) ginkgo.AfterEach(func() { - e2elog.Logf("AfterEach: Cleaning up test resources.") + framework.Logf("AfterEach: Cleaning up test resources.") if errs := framework.PVPVCCleanup(c, ns, pv, pvc); len(errs) > 0 { - e2elog.Failf("AfterEach: Failed to delete PVC and/or PV. Errors: %v", utilerrors.NewAggregate(errs)) + framework.Failf("AfterEach: Failed to delete PVC and/or PV. Errors: %v", utilerrors.NewAggregate(errs)) } }) @@ -302,8 +301,9 @@ var _ = utils.SIGDescribe("PersistentVolumes", func() { pod, err = c.CoreV1().Pods(ns).Create(pod) framework.ExpectNoError(err) framework.ExpectNoError(e2epod.WaitForPodSuccessInNamespace(c, pod.Name, ns)) + framework.ExpectNoError(e2epod.DeletePodWithWait(c, pod)) - e2elog.Logf("Pod exited without failure; the volume has been recycled.") + framework.Logf("Pod exited without failure; the volume has been recycled.") }) }) }) diff --git a/test/e2e/storage/pv_protection.go b/test/e2e/storage/pv_protection.go index dc708d9b8e1..4050e30852e 100644 --- a/test/e2e/storage/pv_protection.go +++ b/test/e2e/storage/pv_protection.go @@ -30,7 +30,6 @@ import ( "k8s.io/kubernetes/pkg/util/slice" volumeutil "k8s.io/kubernetes/pkg/volume/util" "k8s.io/kubernetes/test/e2e/framework" - e2elog "k8s.io/kubernetes/test/e2e/framework/log" "k8s.io/kubernetes/test/e2e/storage/utils" ) @@ -90,9 +89,9 @@ var _ = utils.SIGDescribe("PV Protection", func() { }) ginkgo.AfterEach(func() { - e2elog.Logf("AfterEach: Cleaning up test resources.") + framework.Logf("AfterEach: Cleaning up test resources.") if errs := framework.PVPVCCleanup(client, nameSpace, pv, pvc); len(errs) > 0 { - e2elog.Failf("AfterEach: Failed to delete PVC and/or PV. Errors: %v", utilerrors.NewAggregate(errs)) + framework.Failf("AfterEach: Failed to delete PVC and/or PV. Errors: %v", utilerrors.NewAggregate(errs)) } }) diff --git a/test/e2e/storage/regional_pd.go b/test/e2e/storage/regional_pd.go index 10e592d629a..28f16376193 100644 --- a/test/e2e/storage/regional_pd.go +++ b/test/e2e/storage/regional_pd.go @@ -39,7 +39,6 @@ import ( volumehelpers "k8s.io/cloud-provider/volume/helpers" podutil "k8s.io/kubernetes/pkg/api/v1/pod" "k8s.io/kubernetes/test/e2e/framework" - e2elog "k8s.io/kubernetes/test/e2e/framework/log" e2epod "k8s.io/kubernetes/test/e2e/framework/pod" "k8s.io/kubernetes/test/e2e/storage/testsuites" "k8s.io/kubernetes/test/e2e/storage/utils" @@ -185,7 +184,7 @@ func testZonalFailover(c clientset.Interface, ns string) { _, err := c.StorageV1().StorageClasses().Create(class) framework.ExpectNoError(err) defer func() { - e2elog.Logf("deleting storage class %s", class.Name) + framework.Logf("deleting storage class %s", class.Name) framework.ExpectNoError(c.StorageV1().StorageClasses().Delete(class.Name, nil), "Error deleting StorageClass %s", class.Name) }() @@ -197,19 +196,19 @@ func testZonalFailover(c clientset.Interface, ns string) { framework.ExpectNoError(err) defer func() { - e2elog.Logf("deleting statefulset%q/%q", statefulSet.Namespace, statefulSet.Name) + framework.Logf("deleting statefulset%q/%q", statefulSet.Namespace, statefulSet.Name) // typically this claim has already been deleted framework.ExpectNoError(c.AppsV1().StatefulSets(ns).Delete(statefulSet.Name, nil /* options */), "Error deleting StatefulSet %s", statefulSet.Name) - e2elog.Logf("deleting claims in namespace %s", ns) + framework.Logf("deleting claims in namespace %s", ns) pvc := getPVC(c, ns, regionalPDLabels) framework.ExpectNoError(c.CoreV1().PersistentVolumeClaims(pvc.Namespace).Delete(pvc.Name, nil), "Error deleting claim %s.", pvc.Name) if pvc.Spec.VolumeName != "" { err = framework.WaitForPersistentVolumeDeleted(c, pvc.Spec.VolumeName, framework.Poll, pvDeletionTimeout) if err != nil { - e2elog.Logf("WARNING: PV %s is not yet deleted, and subsequent tests may be affected.", pvc.Spec.VolumeName) + framework.Logf("WARNING: PV %s is not yet deleted, and subsequent tests may be affected.", pvc.Spec.VolumeName) } } }() @@ -238,7 +237,7 @@ func testZonalFailover(c clientset.Interface, ns string) { removeTaintFunc := addTaint(c, ns, nodesInZone.Items, podZone) defer func() { - e2elog.Logf("removing previously added node taints") + framework.Logf("removing previously added node taints") removeTaintFunc() }() @@ -254,7 +253,7 @@ func testZonalFailover(c clientset.Interface, ns string) { otherZone = cloudZones[0] } err = wait.PollImmediate(framework.Poll, statefulSetReadyTimeout, func() (bool, error) { - e2elog.Logf("checking whether new pod is scheduled in zone %q", otherZone) + framework.Logf("checking whether new pod is scheduled in zone %q", otherZone) pod = getPod(c, ns, regionalPDLabels) nodeName = pod.Spec.NodeName node, err = c.CoreV1().Nodes().Get(nodeName, metav1.GetOptions{}) @@ -347,11 +346,11 @@ func testRegionalDelayedBinding(c clientset.Interface, ns string, pvcCount int) } pvs, node := test.TestBindingWaitForFirstConsumerMultiPVC(claims, nil /* node selector */, false /* expect unschedulable */) if node == nil { - e2elog.Failf("unexpected nil node found") + framework.Failf("unexpected nil node found") } zone, ok := node.Labels[v1.LabelZoneFailureDomain] if !ok { - e2elog.Failf("label %s not found on Node", v1.LabelZoneFailureDomain) + framework.Failf("label %s not found on Node", v1.LabelZoneFailureDomain) } for _, pv := range pvs { checkZoneFromLabelAndAffinity(pv, zone, false) @@ -414,11 +413,11 @@ func testRegionalAllowedTopologiesWithDelayedBinding(c clientset.Interface, ns s } pvs, node := test.TestBindingWaitForFirstConsumerMultiPVC(claims, nil /* node selector */, false /* expect unschedulable */) if node == nil { - e2elog.Failf("unexpected nil node found") + framework.Failf("unexpected nil node found") } nodeZone, ok := node.Labels[v1.LabelZoneFailureDomain] if !ok { - e2elog.Failf("label %s not found on Node", v1.LabelZoneFailureDomain) + framework.Failf("label %s not found on Node", v1.LabelZoneFailureDomain) } zoneFound := false for _, zone := range topoZones { @@ -428,7 +427,7 @@ func testRegionalAllowedTopologiesWithDelayedBinding(c clientset.Interface, ns s } } if !zoneFound { - e2elog.Failf("zones specified in AllowedTopologies: %v does not contain zone of node where PV got provisioned: %s", topoZones, nodeZone) + framework.Failf("zones specified in AllowedTopologies: %v does not contain zone of node where PV got provisioned: %s", topoZones, nodeZone) } for _, pv := range pvs { checkZonesFromLabelAndAffinity(pv, sets.NewString(topoZones...), true) diff --git a/test/e2e/storage/testsuites/BUILD b/test/e2e/storage/testsuites/BUILD index b2f713a1689..66389694a61 100644 --- a/test/e2e/storage/testsuites/BUILD +++ b/test/e2e/storage/testsuites/BUILD @@ -45,7 +45,6 @@ go_library( "//staging/src/k8s.io/csi-translation-lib/plugins:go_default_library", "//test/e2e/common:go_default_library", "//test/e2e/framework:go_default_library", - "//test/e2e/framework/log:go_default_library", "//test/e2e/framework/metrics:go_default_library", "//test/e2e/framework/pod:go_default_library", "//test/e2e/framework/podlogs:go_default_library", diff --git a/test/e2e/storage/testsuites/base.go b/test/e2e/storage/testsuites/base.go index 9177483fe74..dcb6cf06cf7 100644 --- a/test/e2e/storage/testsuites/base.go +++ b/test/e2e/storage/testsuites/base.go @@ -36,7 +36,6 @@ import ( clientset "k8s.io/client-go/kubernetes" csilib "k8s.io/csi-translation-lib" "k8s.io/kubernetes/test/e2e/framework" - e2elog "k8s.io/kubernetes/test/e2e/framework/log" "k8s.io/kubernetes/test/e2e/framework/metrics" "k8s.io/kubernetes/test/e2e/framework/podlogs" "k8s.io/kubernetes/test/e2e/framework/volume" @@ -196,13 +195,13 @@ func createGenericVolumeTestResource(driver TestDriver, config *PerTestConfig, p switch pattern.VolType { case testpatterns.InlineVolume: - e2elog.Logf("Creating resource for inline volume") + framework.Logf("Creating resource for inline volume") if iDriver, ok := driver.(InlineVolumeTestDriver); ok { r.volSource = iDriver.GetVolumeSource(false, pattern.FsType, r.volume) r.volType = dInfo.Name } case testpatterns.PreprovisionedPV: - e2elog.Logf("Creating resource for pre-provisioned PV") + framework.Logf("Creating resource for pre-provisioned PV") if pDriver, ok := driver.(PreprovisionedPVTestDriver); ok { pvSource, volumeNodeAffinity := pDriver.GetPersistentVolumeSource(false, pattern.FsType, r.volume) if pvSource != nil { @@ -212,7 +211,7 @@ func createGenericVolumeTestResource(driver TestDriver, config *PerTestConfig, p r.volType = fmt.Sprintf("%s-preprovisionedPV", dInfo.Name) } case testpatterns.DynamicPV: - e2elog.Logf("Creating resource for dynamic PV") + framework.Logf("Creating resource for dynamic PV") if dDriver, ok := driver.(DynamicPVTestDriver); ok { claimSize := dDriver.GetClaimSize() r.sc = dDriver.GetDynamicProvisionStorageClass(r.config, pattern.FsType) @@ -237,7 +236,7 @@ func createGenericVolumeTestResource(driver TestDriver, config *PerTestConfig, p r.volType = fmt.Sprintf("%s-dynamicPV", dInfo.Name) } default: - e2elog.Failf("genericVolumeTestResource doesn't support: %s", pattern.VolType) + framework.Failf("genericVolumeTestResource doesn't support: %s", pattern.VolType) } if r.volSource == nil { @@ -266,13 +265,13 @@ func (r *genericVolumeTestResource) cleanupResource() { case testpatterns.PreprovisionedPV: ginkgo.By("Deleting pv and pvc") if errs := framework.PVPVCCleanup(f.ClientSet, f.Namespace.Name, r.pv, r.pvc); len(errs) != 0 { - e2elog.Failf("Failed to delete PVC or PV: %v", utilerrors.NewAggregate(errs)) + framework.Failf("Failed to delete PVC or PV: %v", utilerrors.NewAggregate(errs)) } case testpatterns.DynamicPV: ginkgo.By("Deleting pvc") // We only delete the PVC so that PV (and disk) can be cleaned up by dynamic provisioner if r.pv != nil && r.pv.Spec.PersistentVolumeReclaimPolicy != v1.PersistentVolumeReclaimDelete { - e2elog.Failf("Test framework does not currently support Dynamically Provisioned Persistent Volume %v specified with reclaim policy that isnt %v", + framework.Failf("Test framework does not currently support Dynamically Provisioned Persistent Volume %v specified with reclaim policy that isnt %v", r.pv.Name, v1.PersistentVolumeReclaimDelete) } if r.pvc != nil { @@ -284,7 +283,7 @@ func (r *genericVolumeTestResource) cleanupResource() { } } default: - e2elog.Failf("Found PVC (%v) or PV (%v) but not running Preprovisioned or Dynamic test pattern", r.pvc, r.pv) + framework.Failf("Found PVC (%v) or PV (%v) but not running Preprovisioned or Dynamic test pattern", r.pvc, r.pv) } } @@ -325,7 +324,7 @@ func createPVCPV( pvcConfig.VolumeMode = &volMode } - e2elog.Logf("Creating PVC and PV") + framework.Logf("Creating PVC and PV") pv, pvc, err := framework.CreatePVCPV(f.ClientSet, pvConfig, pvcConfig, f.Namespace.Name, false) framework.ExpectNoError(err, "PVC, PV creation failed") @@ -513,7 +512,7 @@ func getVolumeOpCounts(c clientset.Interface, pluginName string) opCounts { } if !metricsGrabber.HasRegisteredMaster() { - e2elog.Logf("Warning: Environment does not support getting controller-manager metrics") + framework.Logf("Warning: Environment does not support getting controller-manager metrics") return opCounts{} } @@ -521,7 +520,7 @@ func getVolumeOpCounts(c clientset.Interface, pluginName string) opCounts { framework.ExpectNoError(err, "Error getting c-m metrics : %v", err) totOps := getVolumeOpsFromMetricsForPlugin(metrics.Metrics(controllerMetrics), pluginName) - e2elog.Logf("Node name not specified for getVolumeOpCounts, falling back to listing nodes from API Server") + framework.Logf("Node name not specified for getVolumeOpCounts, falling back to listing nodes from API Server") nodes, err := c.CoreV1().Nodes().List(metav1.ListOptions{}) framework.ExpectNoError(err, "Error listing nodes: %v", err) if len(nodes.Items) <= nodeLimit { @@ -534,7 +533,7 @@ func getVolumeOpCounts(c clientset.Interface, pluginName string) opCounts { totOps = addOpCounts(totOps, getVolumeOpsFromMetricsForPlugin(metrics.Metrics(nodeMetrics), pluginName)) } } else { - e2elog.Logf("Skipping operation metrics gathering from nodes in getVolumeOpCounts, greater than %v nodes", nodeLimit) + framework.Logf("Skipping operation metrics gathering from nodes in getVolumeOpCounts, greater than %v nodes", nodeLimit) } return totOps @@ -560,7 +559,7 @@ func getMigrationVolumeOpCounts(cs clientset.Interface, pluginName string) (opCo var migratedOps opCounts csiName, err := csilib.GetCSINameFromInTreeName(pluginName) if err != nil { - e2elog.Logf("Could not find CSI Name for in-tree plugin %v", pluginName) + framework.Logf("Could not find CSI Name for in-tree plugin %v", pluginName) migratedOps = opCounts{} } else { csiName = "kubernetes.io/csi:" + csiName @@ -569,7 +568,7 @@ func getMigrationVolumeOpCounts(cs clientset.Interface, pluginName string) (opCo return getVolumeOpCounts(cs, pluginName), migratedOps } // Not an in-tree driver - e2elog.Logf("Test running for native CSI Driver, not checking metrics") + framework.Logf("Test running for native CSI Driver, not checking metrics") return opCounts{}, opCounts{} } @@ -593,14 +592,14 @@ func validateMigrationVolumeOpCounts(cs clientset.Interface, pluginName string, for op, count := range newInTreeOps { if count != oldInTreeOps[op] { - e2elog.Failf("In-tree plugin %v migrated to CSI Driver, however found %v %v metrics for in-tree plugin", pluginName, count-oldInTreeOps[op], op) + framework.Failf("In-tree plugin %v migrated to CSI Driver, however found %v %v metrics for in-tree plugin", pluginName, count-oldInTreeOps[op], op) } } // We don't check for migrated metrics because some negative test cases // may not do any volume operations and therefore not emit any metrics } else { // In-tree plugin is not migrated - e2elog.Logf("In-tree plugin %v is not migrated, not validating any metrics", pluginName) + framework.Logf("In-tree plugin %v is not migrated, not validating any metrics", pluginName) // We don't check in-tree plugin metrics because some negative test // cases may not do any volume operations and therefore not emit any diff --git a/test/e2e/storage/testsuites/driveroperations.go b/test/e2e/storage/testsuites/driveroperations.go index 54be67b64c6..4d4298e61a7 100644 --- a/test/e2e/storage/testsuites/driveroperations.go +++ b/test/e2e/storage/testsuites/driveroperations.go @@ -23,7 +23,7 @@ import ( metav1 "k8s.io/apimachinery/pkg/apis/meta/v1" "k8s.io/apimachinery/pkg/apis/meta/v1/unstructured" "k8s.io/apiserver/pkg/storage/names" - e2elog "k8s.io/kubernetes/test/e2e/framework/log" + "k8s.io/kubernetes/test/e2e/framework" "k8s.io/kubernetes/test/e2e/storage/testpatterns" ) @@ -49,7 +49,7 @@ func CreateVolume(driver TestDriver, config *PerTestConfig, volType testpatterns case testpatterns.DynamicPV: // No need to create volume default: - e2elog.Failf("Invalid volType specified: %v", volType) + framework.Failf("Invalid volType specified: %v", volType) } return nil } diff --git a/test/e2e/storage/testsuites/multivolume.go b/test/e2e/storage/testsuites/multivolume.go index 875f59fbe11..06178bd42c0 100644 --- a/test/e2e/storage/testsuites/multivolume.go +++ b/test/e2e/storage/testsuites/multivolume.go @@ -26,7 +26,6 @@ import ( metav1 "k8s.io/apimachinery/pkg/apis/meta/v1" clientset "k8s.io/client-go/kubernetes" "k8s.io/kubernetes/test/e2e/framework" - e2elog "k8s.io/kubernetes/test/e2e/framework/log" e2epod "k8s.io/kubernetes/test/e2e/framework/pod" "k8s.io/kubernetes/test/e2e/storage/testpatterns" "k8s.io/kubernetes/test/e2e/storage/utils" @@ -452,7 +451,7 @@ func TestConcurrentAccessToSingleVolume(f *framework.Framework, cs clientset.Int // Delete the last pod and remove from slice of pods if len(pods) < 2 { - e2elog.Failf("Number of pods shouldn't be less than 2, but got %d", len(pods)) + framework.Failf("Number of pods shouldn't be less than 2, but got %d", len(pods)) } lastPod := pods[len(pods)-1] framework.ExpectNoError(e2epod.DeletePodWithWait(cs, lastPod)) diff --git a/test/e2e/storage/testsuites/provisioning.go b/test/e2e/storage/testsuites/provisioning.go index 58d5da2d810..e2b9f563641 100644 --- a/test/e2e/storage/testsuites/provisioning.go +++ b/test/e2e/storage/testsuites/provisioning.go @@ -33,7 +33,6 @@ import ( "k8s.io/client-go/dynamic" clientset "k8s.io/client-go/kubernetes" "k8s.io/kubernetes/test/e2e/framework" - e2elog "k8s.io/kubernetes/test/e2e/framework/log" e2epod "k8s.io/kubernetes/test/e2e/framework/pod" "k8s.io/kubernetes/test/e2e/framework/volume" "k8s.io/kubernetes/test/e2e/storage/testpatterns" @@ -143,7 +142,7 @@ func (p *provisioningTestSuite) defineTests(driver TestDriver, pattern testpatte ClaimSize: claimSize, StorageClassName: &(l.sc.Name), }, l.config.Framework.Namespace.Name) - e2elog.Logf("In creating storage class object and pvc objects for driver - sc: %v, pvc: %v, src-pvc: %v", l.sc, l.pvc, l.sourcePVC) + framework.Logf("In creating storage class object and pvc objects for driver - sc: %v, pvc: %v, src-pvc: %v", l.sc, l.pvc, l.sourcePVC) l.testCase = &StorageClassTest{ Client: l.config.Framework.ClientSet, Claim: l.pvc, @@ -185,7 +184,7 @@ func (p *provisioningTestSuite) defineTests(driver TestDriver, pattern testpatte sDriver, ok := driver.(SnapshottableTestDriver) if !ok { - e2elog.Failf("Driver %q has CapSnapshotDataSource but does not implement SnapshottableTestDriver", dInfo.Name) + framework.Failf("Driver %q has CapSnapshotDataSource but does not implement SnapshottableTestDriver", dInfo.Name) } init() @@ -245,7 +244,7 @@ func (t StorageClassTest) TestDynamicProvisioning() *v1.PersistentVolume { class, err = client.StorageV1().StorageClasses().Get(class.Name, metav1.GetOptions{}) framework.ExpectNoError(err) defer func() { - e2elog.Logf("deleting storage class %s", class.Name) + framework.Logf("deleting storage class %s", class.Name) framework.ExpectNoError(client.StorageV1().StorageClasses().Delete(class.Name, nil)) }() } @@ -254,11 +253,11 @@ func (t StorageClassTest) TestDynamicProvisioning() *v1.PersistentVolume { claim, err = client.CoreV1().PersistentVolumeClaims(claim.Namespace).Create(claim) framework.ExpectNoError(err) defer func() { - e2elog.Logf("deleting claim %q/%q", claim.Namespace, claim.Name) + framework.Logf("deleting claim %q/%q", claim.Namespace, claim.Name) // typically this claim has already been deleted err = client.CoreV1().PersistentVolumeClaims(claim.Namespace).Delete(claim.Name, nil) if err != nil && !apierrs.IsNotFound(err) { - e2elog.Failf("Error deleting claim %q. Error: %v", claim.Name, err) + framework.Failf("Error deleting claim %q. Error: %v", claim.Name, err) } }() @@ -477,7 +476,7 @@ func (t StorageClassTest) TestBindingWaitForFirstConsumerMultiPVC(claims []*v1.P } if len(errors) > 0 { for claimName, err := range errors { - e2elog.Logf("Failed to delete PVC: %s due to error: %v", claimName, err) + framework.Logf("Failed to delete PVC: %s due to error: %v", claimName, err) } } }() @@ -596,9 +595,9 @@ func StopPod(c clientset.Interface, pod *v1.Pod) { } body, err := c.CoreV1().Pods(pod.Namespace).GetLogs(pod.Name, &v1.PodLogOptions{}).Do().Raw() if err != nil { - e2elog.Logf("Error getting logs for pod %s: %v", pod.Name, err) + framework.Logf("Error getting logs for pod %s: %v", pod.Name, err) } else { - e2elog.Logf("Pod %s has the following logs: %s", pod.Name, body) + framework.Logf("Pod %s has the following logs: %s", pod.Name, body) } e2epod.DeletePodOrFail(c, pod.Namespace, pod.Name) e2epod.WaitForPodNoLongerRunningInNamespace(c, pod.Name, pod.Namespace) @@ -667,19 +666,19 @@ func prepareSnapshotDataSourceForProvisioning( } cleanupFunc := func() { - e2elog.Logf("deleting snapshot %q/%q", snapshot.GetNamespace(), snapshot.GetName()) + framework.Logf("deleting snapshot %q/%q", snapshot.GetNamespace(), snapshot.GetName()) err = dynamicClient.Resource(snapshotGVR).Namespace(updatedClaim.Namespace).Delete(snapshot.GetName(), nil) if err != nil && !apierrs.IsNotFound(err) { - e2elog.Failf("Error deleting snapshot %q. Error: %v", snapshot.GetName(), err) + framework.Failf("Error deleting snapshot %q. Error: %v", snapshot.GetName(), err) } - e2elog.Logf("deleting initClaim %q/%q", updatedClaim.Namespace, updatedClaim.Name) + framework.Logf("deleting initClaim %q/%q", updatedClaim.Namespace, updatedClaim.Name) err = client.CoreV1().PersistentVolumeClaims(updatedClaim.Namespace).Delete(updatedClaim.Name, nil) if err != nil && !apierrs.IsNotFound(err) { - e2elog.Failf("Error deleting initClaim %q. Error: %v", updatedClaim.Name, err) + framework.Failf("Error deleting initClaim %q. Error: %v", updatedClaim.Name, err) } - e2elog.Logf("deleting SnapshotClass %s", snapshotClass.GetName()) + framework.Logf("deleting SnapshotClass %s", snapshotClass.GetName()) framework.ExpectNoError(dynamicClient.Resource(snapshotClassGVR).Delete(snapshotClass.GetName(), nil)) } @@ -715,10 +714,10 @@ func preparePVCDataSourceForProvisioning( } cleanupFunc := func() { - e2elog.Logf("deleting source PVC %q/%q", sourcePVC.Namespace, sourcePVC.Name) + framework.Logf("deleting source PVC %q/%q", sourcePVC.Namespace, sourcePVC.Name) err = client.CoreV1().PersistentVolumeClaims(sourcePVC.Namespace).Delete(sourcePVC.Name, nil) if err != nil && !apierrs.IsNotFound(err) { - e2elog.Failf("Error deleting source PVC %q. Error: %v", sourcePVC.Name, err) + framework.Failf("Error deleting source PVC %q. Error: %v", sourcePVC.Name, err) } } diff --git a/test/e2e/storage/testsuites/snapshottable.go b/test/e2e/storage/testsuites/snapshottable.go index 5dae3e14a63..3c650f76ad2 100644 --- a/test/e2e/storage/testsuites/snapshottable.go +++ b/test/e2e/storage/testsuites/snapshottable.go @@ -28,7 +28,6 @@ import ( "k8s.io/apimachinery/pkg/runtime/schema" "k8s.io/client-go/dynamic" "k8s.io/kubernetes/test/e2e/framework" - e2elog "k8s.io/kubernetes/test/e2e/framework/log" "k8s.io/kubernetes/test/e2e/storage/testpatterns" ) @@ -115,13 +114,13 @@ func (s *snapshottableTestSuite) defineTests(driver TestDriver, pattern testpatt StorageClassName: &(class.Name), }, config.Framework.Namespace.Name) - e2elog.Logf("In creating storage class object and pvc object for driver - sc: %v, pvc: %v", class, pvc) + framework.Logf("In creating storage class object and pvc object for driver - sc: %v, pvc: %v", class, pvc) ginkgo.By("creating a StorageClass " + class.Name) class, err := cs.StorageV1().StorageClasses().Create(class) framework.ExpectNoError(err) defer func() { - e2elog.Logf("deleting storage class %s", class.Name) + framework.Logf("deleting storage class %s", class.Name) framework.ExpectNoError(cs.StorageV1().StorageClasses().Delete(class.Name, nil)) }() @@ -129,11 +128,11 @@ func (s *snapshottableTestSuite) defineTests(driver TestDriver, pattern testpatt pvc, err = cs.CoreV1().PersistentVolumeClaims(pvc.Namespace).Create(pvc) framework.ExpectNoError(err) defer func() { - e2elog.Logf("deleting claim %q/%q", pvc.Namespace, pvc.Name) + framework.Logf("deleting claim %q/%q", pvc.Namespace, pvc.Name) // typically this claim has already been deleted err = cs.CoreV1().PersistentVolumeClaims(pvc.Namespace).Delete(pvc.Name, nil) if err != nil && !apierrs.IsNotFound(err) { - e2elog.Failf("Error deleting claim %q. Error: %v", pvc.Name, err) + framework.Failf("Error deleting claim %q. Error: %v", pvc.Name, err) } }() err = framework.WaitForPersistentVolumeClaimPhase(v1.ClaimBound, cs, pvc.Namespace, pvc.Name, framework.Poll, framework.ClaimProvisionTimeout) @@ -152,7 +151,7 @@ func (s *snapshottableTestSuite) defineTests(driver TestDriver, pattern testpatt vsc, err = dc.Resource(snapshotClassGVR).Create(vsc, metav1.CreateOptions{}) framework.ExpectNoError(err) defer func() { - e2elog.Logf("deleting SnapshotClass %s", vsc.GetName()) + framework.Logf("deleting SnapshotClass %s", vsc.GetName()) framework.ExpectNoError(dc.Resource(snapshotClassGVR).Delete(vsc.GetName(), nil)) }() @@ -162,11 +161,11 @@ func (s *snapshottableTestSuite) defineTests(driver TestDriver, pattern testpatt snapshot, err = dc.Resource(snapshotGVR).Namespace(snapshot.GetNamespace()).Create(snapshot, metav1.CreateOptions{}) framework.ExpectNoError(err) defer func() { - e2elog.Logf("deleting snapshot %q/%q", snapshot.GetNamespace(), snapshot.GetName()) + framework.Logf("deleting snapshot %q/%q", snapshot.GetNamespace(), snapshot.GetName()) // typically this snapshot has already been deleted err = dc.Resource(snapshotGVR).Namespace(snapshot.GetNamespace()).Delete(snapshot.GetName(), nil) if err != nil && !apierrs.IsNotFound(err) { - e2elog.Failf("Error deleting snapshot %q. Error: %v", pvc.Name, err) + framework.Failf("Error deleting snapshot %q. Error: %v", pvc.Name, err) } }() err = WaitForSnapshotReady(dc, snapshot.GetNamespace(), snapshot.GetName(), framework.Poll, framework.SnapshotCreateTimeout) @@ -198,27 +197,27 @@ func (s *snapshottableTestSuite) defineTests(driver TestDriver, pattern testpatt // WaitForSnapshotReady waits for a VolumeSnapshot to be ready to use or until timeout occurs, whichever comes first. func WaitForSnapshotReady(c dynamic.Interface, ns string, snapshotName string, Poll, timeout time.Duration) error { - e2elog.Logf("Waiting up to %v for VolumeSnapshot %s to become ready", timeout, snapshotName) + framework.Logf("Waiting up to %v for VolumeSnapshot %s to become ready", timeout, snapshotName) for start := time.Now(); time.Since(start) < timeout; time.Sleep(Poll) { snapshot, err := c.Resource(snapshotGVR).Namespace(ns).Get(snapshotName, metav1.GetOptions{}) if err != nil { - e2elog.Logf("Failed to get claim %q, retrying in %v. Error: %v", snapshotName, Poll, err) + framework.Logf("Failed to get claim %q, retrying in %v. Error: %v", snapshotName, Poll, err) continue } else { status := snapshot.Object["status"] if status == nil { - e2elog.Logf("VolumeSnapshot %s found but is not ready.", snapshotName) + framework.Logf("VolumeSnapshot %s found but is not ready.", snapshotName) continue } value := status.(map[string]interface{}) if value["readyToUse"] == true { - e2elog.Logf("VolumeSnapshot %s found and is ready", snapshotName, time.Since(start)) + framework.Logf("VolumeSnapshot %s found and is ready", snapshotName, time.Since(start)) return nil } else if value["ready"] == true { - e2elog.Logf("VolumeSnapshot %s found and is ready", snapshotName, time.Since(start)) + framework.Logf("VolumeSnapshot %s found and is ready", snapshotName, time.Since(start)) return nil } else { - e2elog.Logf("VolumeSnapshot %s found but is not ready.", snapshotName) + framework.Logf("VolumeSnapshot %s found but is not ready.", snapshotName) } } } diff --git a/test/e2e/storage/testsuites/subpath.go b/test/e2e/storage/testsuites/subpath.go index 8f442e8d9fc..b2b5186fb92 100644 --- a/test/e2e/storage/testsuites/subpath.go +++ b/test/e2e/storage/testsuites/subpath.go @@ -27,7 +27,6 @@ import ( "k8s.io/apimachinery/pkg/util/rand" "k8s.io/apimachinery/pkg/util/wait" "k8s.io/kubernetes/test/e2e/framework" - e2elog "k8s.io/kubernetes/test/e2e/framework/log" e2epod "k8s.io/kubernetes/test/e2e/framework/pod" "k8s.io/kubernetes/test/e2e/framework/volume" "k8s.io/kubernetes/test/e2e/storage/testpatterns" @@ -136,7 +135,7 @@ func (s *subPathTestSuite) defineTests(driver TestDriver, pattern testpatterns.T }, } default: - e2elog.Failf("SubPath test doesn't support: %s", volType) + framework.Failf("SubPath test doesn't support: %s", volType) } subPath := f.Namespace.Name @@ -807,7 +806,7 @@ func testPodContainerRestart(f *framework.Framework, pod *v1.Pod) { ginkgo.By("Failing liveness probe") out, err := podContainerExec(pod, 1, fmt.Sprintf("rm %v", probeFilePath)) - e2elog.Logf("Pod exec output: %v", out) + framework.Logf("Pod exec output: %v", out) framework.ExpectNoError(err, "while failing liveness probe") // Check that container has restarted @@ -820,10 +819,10 @@ func testPodContainerRestart(f *framework.Framework, pod *v1.Pod) { } for _, status := range pod.Status.ContainerStatuses { if status.Name == pod.Spec.Containers[0].Name { - e2elog.Logf("Container %v, restarts: %v", status.Name, status.RestartCount) + framework.Logf("Container %v, restarts: %v", status.Name, status.RestartCount) restarts = status.RestartCount if restarts > 0 { - e2elog.Logf("Container has restart count: %v", restarts) + framework.Logf("Container has restart count: %v", restarts) return true, nil } } @@ -841,7 +840,7 @@ func testPodContainerRestart(f *framework.Framework, pod *v1.Pod) { writeCmd = fmt.Sprintf("echo test-after > %v", probeFilePath) } out, err = podContainerExec(pod, 1, writeCmd) - e2elog.Logf("Pod exec output: %v", out) + framework.Logf("Pod exec output: %v", out) framework.ExpectNoError(err, "while rewriting the probe file") // Wait for container restarts to stabilize @@ -858,13 +857,13 @@ func testPodContainerRestart(f *framework.Framework, pod *v1.Pod) { if status.RestartCount == restarts { stableCount++ if stableCount > stableThreshold { - e2elog.Logf("Container restart has stabilized") + framework.Logf("Container restart has stabilized") return true, nil } } else { restarts = status.RestartCount stableCount = 0 - e2elog.Logf("Container has restart count: %v", restarts) + framework.Logf("Container has restart count: %v", restarts) } break } diff --git a/test/e2e/storage/testsuites/volume_expand.go b/test/e2e/storage/testsuites/volume_expand.go index ba95fa6294d..fbee4871078 100644 --- a/test/e2e/storage/testsuites/volume_expand.go +++ b/test/e2e/storage/testsuites/volume_expand.go @@ -29,7 +29,6 @@ import ( "k8s.io/apimachinery/pkg/util/wait" clientset "k8s.io/client-go/kubernetes" "k8s.io/kubernetes/test/e2e/framework" - e2elog "k8s.io/kubernetes/test/e2e/framework/log" e2epod "k8s.io/kubernetes/test/e2e/framework/pod" "k8s.io/kubernetes/test/e2e/storage/testpatterns" ) @@ -146,7 +145,7 @@ func (v *volumeExpandTestSuite) defineTests(driver TestDriver, pattern testpatte currentPvcSize := l.resource.pvc.Spec.Resources.Requests[v1.ResourceStorage] newSize := currentPvcSize.DeepCopy() newSize.Add(resource.MustParse("1Gi")) - e2elog.Logf("currentPvcSize %v, newSize %v", currentPvcSize, newSize) + framework.Logf("currentPvcSize %v, newSize %v", currentPvcSize, newSize) _, err = ExpandPVCSize(l.resource.pvc, newSize, f.ClientSet) framework.ExpectError(err, "While updating non-expandable PVC") }) @@ -173,7 +172,7 @@ func (v *volumeExpandTestSuite) defineTests(driver TestDriver, pattern testpatte currentPvcSize := l.resource.pvc.Spec.Resources.Requests[v1.ResourceStorage] newSize := currentPvcSize.DeepCopy() newSize.Add(resource.MustParse("1Gi")) - e2elog.Logf("currentPvcSize %v, newSize %v", currentPvcSize, newSize) + framework.Logf("currentPvcSize %v, newSize %v", currentPvcSize, newSize) newPVC, err := ExpandPVCSize(l.resource.pvc, newSize, f.ClientSet) framework.ExpectNoError(err, "While updating pvc for more size") l.resource.pvc = newPVC @@ -181,7 +180,7 @@ func (v *volumeExpandTestSuite) defineTests(driver TestDriver, pattern testpatte pvcSize := l.resource.pvc.Spec.Resources.Requests[v1.ResourceStorage] if pvcSize.Cmp(newSize) != 0 { - e2elog.Failf("error updating pvc size %q", l.resource.pvc.Name) + framework.Failf("error updating pvc size %q", l.resource.pvc.Name) } ginkgo.By("Waiting for cloudprovider resize to finish") @@ -233,7 +232,7 @@ func (v *volumeExpandTestSuite) defineTests(driver TestDriver, pattern testpatte currentPvcSize := l.resource.pvc.Spec.Resources.Requests[v1.ResourceStorage] newSize := currentPvcSize.DeepCopy() newSize.Add(resource.MustParse("1Gi")) - e2elog.Logf("currentPvcSize %v, newSize %v", currentPvcSize, newSize) + framework.Logf("currentPvcSize %v, newSize %v", currentPvcSize, newSize) newPVC, err := ExpandPVCSize(l.resource.pvc, newSize, f.ClientSet) framework.ExpectNoError(err, "While updating pvc for more size") l.resource.pvc = newPVC @@ -241,7 +240,7 @@ func (v *volumeExpandTestSuite) defineTests(driver TestDriver, pattern testpatte pvcSize := l.resource.pvc.Spec.Resources.Requests[v1.ResourceStorage] if pvcSize.Cmp(newSize) != 0 { - e2elog.Failf("error updating pvc size %q", l.resource.pvc.Name) + framework.Failf("error updating pvc size %q", l.resource.pvc.Name) } ginkgo.By("Waiting for cloudprovider resize to finish") @@ -276,7 +275,7 @@ func ExpandPVCSize(origPVC *v1.PersistentVolumeClaim, size resource.Quantity, c if err == nil { return true, nil } - e2elog.Logf("Error updating pvc %s with %v", pvcName, err) + framework.Logf("Error updating pvc %s with %v", pvcName, err) return false, nil }) return updatedPVC, waitErr diff --git a/test/e2e/storage/testsuites/volume_io.go b/test/e2e/storage/testsuites/volume_io.go index 9def24dbf0f..71d29acfd5b 100644 --- a/test/e2e/storage/testsuites/volume_io.go +++ b/test/e2e/storage/testsuites/volume_io.go @@ -35,7 +35,6 @@ import ( metav1 "k8s.io/apimachinery/pkg/apis/meta/v1" clientset "k8s.io/client-go/kubernetes" "k8s.io/kubernetes/test/e2e/framework" - e2elog "k8s.io/kubernetes/test/e2e/framework/log" e2epod "k8s.io/kubernetes/test/e2e/framework/pod" "k8s.io/kubernetes/test/e2e/framework/volume" "k8s.io/kubernetes/test/e2e/storage/testpatterns" @@ -286,7 +285,7 @@ func deleteFile(pod *v1.Pod, fpath string) { _, err := utils.PodExec(pod, fmt.Sprintf("rm -f %s", fpath)) if err != nil { // keep going, the test dir will be deleted when the volume is unmounted - e2elog.Logf("unable to delete test file %s: %v\nerror ignored, continuing test", fpath, err) + framework.Logf("unable to delete test file %s: %v\nerror ignored, continuing test", fpath, err) } } @@ -318,12 +317,12 @@ func testVolumeIO(f *framework.Framework, cs clientset.Interface, config volume. ginkgo.By(fmt.Sprintf("deleting client pod %q...", clientPod.Name)) e := e2epod.DeletePodWithWait(cs, clientPod) if e != nil { - e2elog.Logf("client pod failed to delete: %v", e) + framework.Logf("client pod failed to delete: %v", e) if err == nil { // delete err is returned if err is not set err = e } } else { - e2elog.Logf("sleeping a bit so kubelet can unmount and detach the volume") + framework.Logf("sleeping a bit so kubelet can unmount and detach the volume") time.Sleep(volume.PodCleanupTimeout) } }() diff --git a/test/e2e/storage/testsuites/volumemode.go b/test/e2e/storage/testsuites/volumemode.go index 12735181da1..da4b5846ca9 100644 --- a/test/e2e/storage/testsuites/volumemode.go +++ b/test/e2e/storage/testsuites/volumemode.go @@ -32,7 +32,6 @@ import ( "k8s.io/kubernetes/pkg/kubelet/events" "k8s.io/kubernetes/test/e2e/common" "k8s.io/kubernetes/test/e2e/framework" - e2elog "k8s.io/kubernetes/test/e2e/framework/log" e2epod "k8s.io/kubernetes/test/e2e/framework/pod" "k8s.io/kubernetes/test/e2e/storage/testpatterns" "k8s.io/kubernetes/test/e2e/storage/utils" @@ -162,7 +161,7 @@ func (t *volumeModeTestSuite) defineTests(driver TestDriver, pattern testpattern }, l.ns.Name) } default: - e2elog.Failf("Volume mode test doesn't support: %s", pattern.VolType) + framework.Failf("Volume mode test doesn't support: %s", pattern.VolType) } } @@ -224,7 +223,7 @@ func (t *volumeModeTestSuite) defineTests(driver TestDriver, pattern testpattern err = common.WaitTimeoutForEvent(l.cs, l.ns.Name, eventSelector, msg, framework.PodStartTimeout) // Events are unreliable, don't depend on the event. It's used only to speed up the test. if err != nil { - e2elog.Logf("Warning: did not get event about FailedMountVolume") + framework.Logf("Warning: did not get event about FailedMountVolume") } // Check the pod is still not running @@ -261,7 +260,7 @@ func (t *volumeModeTestSuite) defineTests(driver TestDriver, pattern testpattern err = common.WaitTimeoutForEvent(l.cs, l.ns.Name, eventSelector, msg, framework.ClaimProvisionTimeout) // Events are unreliable, don't depend on the event. It's used only to speed up the test. if err != nil { - e2elog.Logf("Warning: did not get event about provisioing failed") + framework.Logf("Warning: did not get event about provisioing failed") } // Check the pvc is still pending @@ -271,7 +270,7 @@ func (t *volumeModeTestSuite) defineTests(driver TestDriver, pattern testpattern }) } default: - e2elog.Failf("Volume mode test doesn't support volType: %v", pattern.VolType) + framework.Failf("Volume mode test doesn't support volType: %v", pattern.VolType) } ginkgo.It("should fail to use a volume in a pod with mismatched mode [Slow]", func() { @@ -311,7 +310,7 @@ func (t *volumeModeTestSuite) defineTests(driver TestDriver, pattern testpattern err = common.WaitTimeoutForEvent(l.cs, l.ns.Name, eventSelector, msg, framework.PodStartTimeout) // Events are unreliable, don't depend on them. They're used only to speed up the test. if err != nil { - e2elog.Logf("Warning: did not get event about mismatched volume use") + framework.Logf("Warning: did not get event about mismatched volume use") } // Check the pod is still not running diff --git a/test/e2e/storage/utils/BUILD b/test/e2e/storage/utils/BUILD index 0e63b5d0c60..f06c92ddc41 100644 --- a/test/e2e/storage/utils/BUILD +++ b/test/e2e/storage/utils/BUILD @@ -27,7 +27,6 @@ go_library( "//staging/src/k8s.io/apimachinery/pkg/util/wait:go_default_library", "//staging/src/k8s.io/client-go/kubernetes:go_default_library", "//test/e2e/framework:go_default_library", - "//test/e2e/framework/log:go_default_library", "//test/e2e/framework/node:go_default_library", "//test/e2e/framework/pod:go_default_library", "//test/e2e/framework/ssh:go_default_library", diff --git a/test/e2e/storage/utils/local.go b/test/e2e/storage/utils/local.go index ff05e98af78..9050d25ffef 100644 --- a/test/e2e/storage/utils/local.go +++ b/test/e2e/storage/utils/local.go @@ -29,7 +29,6 @@ import ( "k8s.io/api/core/v1" "k8s.io/apimachinery/pkg/util/uuid" "k8s.io/kubernetes/test/e2e/framework" - e2elog "k8s.io/kubernetes/test/e2e/framework/log" ) // LocalVolumeType represents type of local volume, e.g. tmpfs, directory, @@ -310,11 +309,11 @@ func (l *ltrMgr) Create(node *v1.Node, volumeType LocalVolumeType, parameters ma case LocalVolumeGCELocalSSD: ltr = l.setupLocalVolumeGCELocalSSD(node, parameters) default: - e2elog.Failf("Failed to create local test resource on node %q, unsupported volume type: %v is specified", node.Name, volumeType) + framework.Failf("Failed to create local test resource on node %q, unsupported volume type: %v is specified", node.Name, volumeType) return nil } if ltr == nil { - e2elog.Failf("Failed to create local test resource on node %q, volume type: %v, parameters: %v", node.Name, volumeType, parameters) + framework.Failf("Failed to create local test resource on node %q, volume type: %v, parameters: %v", node.Name, volumeType, parameters) } ltr.VolumeType = volumeType return ltr @@ -339,7 +338,7 @@ func (l *ltrMgr) Remove(ltr *LocalTestResource) { case LocalVolumeGCELocalSSD: l.cleanupLocalVolumeGCELocalSSD(ltr) default: - e2elog.Failf("Failed to remove local test resource, unsupported volume type: %v is specified", ltr.VolumeType) + framework.Failf("Failed to remove local test resource, unsupported volume type: %v is specified", ltr.VolumeType) } return } diff --git a/test/e2e/storage/utils/utils.go b/test/e2e/storage/utils/utils.go index 217489bc5c4..4fe31a2414c 100644 --- a/test/e2e/storage/utils/utils.go +++ b/test/e2e/storage/utils/utils.go @@ -35,7 +35,6 @@ import ( "k8s.io/apimachinery/pkg/util/wait" clientset "k8s.io/client-go/kubernetes" "k8s.io/kubernetes/test/e2e/framework" - e2elog "k8s.io/kubernetes/test/e2e/framework/log" e2enode "k8s.io/kubernetes/test/e2e/framework/node" e2epod "k8s.io/kubernetes/test/e2e/framework/pod" e2essh "k8s.io/kubernetes/test/e2e/framework/ssh" @@ -103,7 +102,7 @@ func VerifyExecInPodFail(pod *v1.Pod, bashExec string, exitCode int) { } func isSudoPresent(nodeIP string, provider string) bool { - e2elog.Logf("Checking if sudo command is present") + framework.Logf("Checking if sudo command is present") sshResult, err := e2essh.SSH("sudo --version", nodeIP, provider) framework.ExpectNoError(err, "SSH to %q errored.", nodeIP) if !strings.Contains(sshResult.Stderr, "command not found") { @@ -127,7 +126,7 @@ func KubeletCommand(kOp KubeletOpt, c clientset.Interface, pod *v1.Pod) { framework.ExpectNoError(err) nodeIP = nodeIP + ":22" - e2elog.Logf("Checking if systemctl command is present") + framework.Logf("Checking if systemctl command is present") sshResult, err := e2essh.SSH("systemctl --version", nodeIP, framework.TestContext.Provider) framework.ExpectNoError(err, fmt.Sprintf("SSH to Node %q errored.", pod.Spec.NodeName)) if !strings.Contains(sshResult.Stderr, "command not found") { @@ -146,7 +145,7 @@ func KubeletCommand(kOp KubeletOpt, c clientset.Interface, pod *v1.Pod) { kubeletPid = getKubeletMainPid(nodeIP, sudoPresent, systemctlPresent) } - e2elog.Logf("Attempting `%s`", command) + framework.Logf("Attempting `%s`", command) sshResult, err = e2essh.SSH(command, nodeIP, framework.TestContext.Provider) framework.ExpectNoError(err, fmt.Sprintf("SSH to Node %q errored.", pod.Spec.NodeName)) e2essh.LogResult(sshResult) @@ -154,7 +153,7 @@ func KubeletCommand(kOp KubeletOpt, c clientset.Interface, pod *v1.Pod) { if kOp == KStop { if ok := e2enode.WaitForNodeToBeNotReady(c, pod.Spec.NodeName, NodeStateTimeout); !ok { - e2elog.Failf("Node %s failed to enter NotReady state", pod.Spec.NodeName) + framework.Failf("Node %s failed to enter NotReady state", pod.Spec.NodeName) } } if kOp == KRestart { @@ -168,13 +167,13 @@ func KubeletCommand(kOp KubeletOpt, c clientset.Interface, pod *v1.Pod) { } } gomega.Expect(isPidChanged).To(gomega.BeTrue(), "Kubelet PID remained unchanged after restarting Kubelet") - e2elog.Logf("Noticed that kubelet PID is changed. Waiting for 30 Seconds for Kubelet to come back") + framework.Logf("Noticed that kubelet PID is changed. Waiting for 30 Seconds for Kubelet to come back") time.Sleep(30 * time.Second) } if kOp == KStart || kOp == KRestart { // For kubelet start and restart operations, Wait until Node becomes Ready if ok := e2enode.WaitForNodeToBeReady(c, pod.Spec.NodeName, NodeStateTimeout); !ok { - e2elog.Failf("Node %s failed to enter Ready state", pod.Spec.NodeName) + framework.Failf("Node %s failed to enter Ready state", pod.Spec.NodeName) } } } @@ -190,7 +189,7 @@ func getKubeletMainPid(nodeIP string, sudoPresent bool, systemctlPresent bool) s if sudoPresent { command = fmt.Sprintf("sudo %s", command) } - e2elog.Logf("Attempting `%s`", command) + framework.Logf("Attempting `%s`", command) sshResult, err := e2essh.SSH(command, nodeIP, framework.TestContext.Provider) framework.ExpectNoError(err, fmt.Sprintf("SSH to Node %q errored.", nodeIP)) e2essh.LogResult(sshResult) @@ -214,7 +213,7 @@ func TestKubeletRestartsAndRestoresMount(c clientset.Interface, f *framework.Fra ginkgo.By("Testing that written file is accessible.") CheckReadFromPath(clientPod, v1.PersistentVolumeFilesystem, path, byteLen, seed) - e2elog.Logf("Volume mount detected on pod %s and written file %s is readable post-restart.", clientPod.Name, path) + framework.Logf("Volume mount detected on pod %s and written file %s is readable post-restart.", clientPod.Name, path) } // TestKubeletRestartsAndRestoresMap tests that a volume mapped to a pod remains mapped after a kubelet restarts @@ -232,7 +231,7 @@ func TestKubeletRestartsAndRestoresMap(c clientset.Interface, f *framework.Frame ginkgo.By("Testing that written pv is accessible.") CheckReadFromPath(clientPod, v1.PersistentVolumeBlock, path, byteLen, seed) - e2elog.Logf("Volume map detected on pod %s and written data %s is readable post-restart.", clientPod.Name, path) + framework.Logf("Volume map detected on pod %s and written data %s is readable post-restart.", clientPod.Name, path) } // TestVolumeUnmountsFromDeletedPodWithForceOption tests that a volume unmounts if the client pod was deleted while the kubelet was down. @@ -289,7 +288,7 @@ func TestVolumeUnmountsFromDeletedPodWithForceOption(c clientset.Interface, f *f e2essh.LogResult(result) framework.ExpectNoError(err, "Encountered SSH error.") gomega.Expect(result.Stdout).To(gomega.BeEmpty(), "Expected grep stdout to be empty (i.e. no mount found).") - e2elog.Logf("Volume unmounted on node %s", clientPod.Spec.NodeName) + framework.Logf("Volume unmounted on node %s", clientPod.Spec.NodeName) if checkSubpath { ginkgo.By("Expecting the volume subpath mount not to be found.") @@ -297,7 +296,7 @@ func TestVolumeUnmountsFromDeletedPodWithForceOption(c clientset.Interface, f *f e2essh.LogResult(result) framework.ExpectNoError(err, "Encountered SSH error.") gomega.Expect(result.Stdout).To(gomega.BeEmpty(), "Expected grep stdout to be empty (i.e. no subpath mount found).") - e2elog.Logf("Subpath volume unmounted on node %s", clientPod.Spec.NodeName) + framework.Logf("Subpath volume unmounted on node %s", clientPod.Spec.NodeName) } } @@ -366,7 +365,7 @@ func TestVolumeUnmapsFromDeletedPodWithForceOption(c clientset.Interface, f *fra // TODO: Needs to check GetGlobalMapPath and descriptor lock, as well. - e2elog.Logf("Volume unmaped on node %s", clientPod.Spec.NodeName) + framework.Logf("Volume unmaped on node %s", clientPod.Spec.NodeName) } // TestVolumeUnmapsFromDeletedPod tests that a volume unmaps if the client pod was deleted while the kubelet was down. diff --git a/test/e2e/storage/volume_limits.go b/test/e2e/storage/volume_limits.go index 97cc7e06c64..97c82b25907 100644 --- a/test/e2e/storage/volume_limits.go +++ b/test/e2e/storage/volume_limits.go @@ -22,7 +22,6 @@ import ( clientset "k8s.io/client-go/kubernetes" v1helper "k8s.io/kubernetes/pkg/apis/core/v1/helper" "k8s.io/kubernetes/test/e2e/framework" - e2elog "k8s.io/kubernetes/test/e2e/framework/log" "k8s.io/kubernetes/test/e2e/storage/utils" ) @@ -40,12 +39,12 @@ var _ = utils.SIGDescribe("Volume limits", func() { ginkgo.It("should verify that all nodes have volume limits", func() { nodeList := framework.GetReadySchedulableNodesOrDie(f.ClientSet) if len(nodeList.Items) == 0 { - e2elog.Failf("Unable to find ready and schedulable Node") + framework.Failf("Unable to find ready and schedulable Node") } for _, node := range nodeList.Items { volumeLimits := getVolumeLimit(&node) if len(volumeLimits) == 0 { - e2elog.Failf("Expected volume limits to be set") + framework.Failf("Expected volume limits to be set") } } }) diff --git a/test/e2e/storage/volume_metrics.go b/test/e2e/storage/volume_metrics.go index 71bae395773..b7d59918c75 100644 --- a/test/e2e/storage/volume_metrics.go +++ b/test/e2e/storage/volume_metrics.go @@ -31,7 +31,6 @@ import ( clientset "k8s.io/client-go/kubernetes" kubeletmetrics "k8s.io/kubernetes/pkg/kubelet/metrics" "k8s.io/kubernetes/test/e2e/framework" - e2elog "k8s.io/kubernetes/test/e2e/framework/log" "k8s.io/kubernetes/test/e2e/framework/metrics" e2epod "k8s.io/kubernetes/test/e2e/framework/pod" "k8s.io/kubernetes/test/e2e/storage/testsuites" @@ -58,7 +57,7 @@ var _ = utils.SIGDescribe("[Serial] Volume metrics", func() { framework.SkipUnlessProviderIs("gce", "gke", "aws") defaultScName, err = framework.GetDefaultStorageClassName(c) if err != nil { - e2elog.Failf(err.Error()) + framework.Failf(err.Error()) } test := testsuites.StorageClassTest{ Name: "default", @@ -73,14 +72,14 @@ var _ = utils.SIGDescribe("[Serial] Volume metrics", func() { metricsGrabber, err = metrics.NewMetricsGrabber(c, nil, true, false, true, false, false) if err != nil { - e2elog.Failf("Error creating metrics grabber : %v", err) + framework.Failf("Error creating metrics grabber : %v", err) } }) ginkgo.AfterEach(func() { newPvc, err := c.CoreV1().PersistentVolumeClaims(pvc.Namespace).Get(pvc.Name, metav1.GetOptions{}) if err != nil { - e2elog.Logf("Failed to get pvc %s/%s: %v", pvc.Namespace, pvc.Name, err) + framework.Logf("Failed to get pvc %s/%s: %v", pvc.Namespace, pvc.Name, err) } else { framework.DeletePersistentVolumeClaim(c, newPvc.Name, newPvc.Namespace) if newPvc.Spec.VolumeName != "" { @@ -122,7 +121,7 @@ var _ = utils.SIGDescribe("[Serial] Volume metrics", func() { err = e2epod.WaitForPodRunningInNamespace(c, pod) framework.ExpectNoError(e2epod.WaitForPodRunningInNamespace(c, pod), "Error starting pod %s", pod.Name) - e2elog.Logf("Deleting pod %q/%q", pod.Namespace, pod.Name) + framework.Logf("Deleting pod %q/%q", pod.Namespace, pod.Name) framework.ExpectNoError(e2epod.DeletePodWithWait(c, pod)) updatedStorageMetrics := waitForDetachAndGrabMetrics(storageOpMetrics, metricsGrabber) @@ -181,7 +180,7 @@ var _ = utils.SIGDescribe("[Serial] Volume metrics", func() { err = e2epod.WaitTimeoutForPodRunningInNamespace(c, pod.Name, pod.Namespace, framework.PodStartShortTimeout) framework.ExpectError(err) - e2elog.Logf("Deleting pod %q/%q", pod.Namespace, pod.Name) + framework.Logf("Deleting pod %q/%q", pod.Namespace, pod.Name) framework.ExpectNoError(e2epod.DeletePodWithWait(c, pod)) ginkgo.By("Checking failure metrics") @@ -223,12 +222,12 @@ var _ = utils.SIGDescribe("[Serial] Volume metrics", func() { // by the volume stats collector var kubeMetrics metrics.KubeletMetrics waitErr := wait.Poll(30*time.Second, 5*time.Minute, func() (bool, error) { - e2elog.Logf("Grabbing Kubelet metrics") + framework.Logf("Grabbing Kubelet metrics") // Grab kubelet metrics from the node the pod was scheduled on var err error kubeMetrics, err = metricsGrabber.GrabFromKubelet(pod.Spec.NodeName) if err != nil { - e2elog.Logf("Error fetching kubelet metrics") + framework.Logf("Error fetching kubelet metrics") return false, err } key := volumeStatKeys[0] @@ -246,7 +245,7 @@ var _ = utils.SIGDescribe("[Serial] Volume metrics", func() { gomega.Expect(found).To(gomega.BeTrue(), "PVC %s, Namespace %s not found for %s", pvc.Name, pvc.Namespace, kubeletKeyName) } - e2elog.Logf("Deleting pod %q/%q", pod.Namespace, pod.Name) + framework.Logf("Deleting pod %q/%q", pod.Namespace, pod.Name) framework.ExpectNoError(e2epod.DeletePodWithWait(c, pod)) }) @@ -277,7 +276,7 @@ var _ = utils.SIGDescribe("[Serial] Volume metrics", func() { valid := hasValidMetrics(metrics.Metrics(controllerMetrics), metricKey, dimensions...) gomega.Expect(valid).To(gomega.BeTrue(), "Invalid metric in P/V Controller metrics: %q", metricKey) - e2elog.Logf("Deleting pod %q/%q", pod.Namespace, pod.Name) + framework.Logf("Deleting pod %q/%q", pod.Namespace, pod.Name) framework.ExpectNoError(e2epod.DeletePodWithWait(c, pod)) }) @@ -307,7 +306,7 @@ var _ = utils.SIGDescribe("[Serial] Volume metrics", func() { valid := hasValidMetrics(metrics.Metrics(kubeMetrics), totalVolumesKey, dimensions...) gomega.Expect(valid).To(gomega.BeTrue(), "Invalid metric in Volume Manager metrics: %q", totalVolumesKey) - e2elog.Logf("Deleting pod %q/%q", pod.Namespace, pod.Name) + framework.Logf("Deleting pod %q/%q", pod.Namespace, pod.Name) framework.ExpectNoError(e2epod.DeletePodWithWait(c, pod)) }) @@ -367,7 +366,7 @@ var _ = utils.SIGDescribe("[Serial] Volume metrics", func() { } } - e2elog.Logf("Deleting pod %q/%q", pod.Namespace, pod.Name) + framework.Logf("Deleting pod %q/%q", pod.Namespace, pod.Name) framework.ExpectNoError(e2epod.DeletePodWithWait(c, pod)) }) @@ -457,10 +456,10 @@ var _ = utils.SIGDescribe("[Serial] Volume metrics", func() { ginkgo.AfterEach(func() { if err := framework.DeletePersistentVolume(c, pv.Name); err != nil { - e2elog.Failf("Error deleting pv: %v", err) + framework.Failf("Error deleting pv: %v", err) } if err := framework.DeletePersistentVolumeClaim(c, pvc.Name, pvc.Namespace); err != nil { - e2elog.Failf("Error deleting pvc: %v", err) + framework.Failf("Error deleting pvc: %v", err) } // Clear original metric values. @@ -537,7 +536,7 @@ func waitForDetachAndGrabMetrics(oldMetrics *storageControllerMetrics, metricsGr updatedMetrics, err := metricsGrabber.GrabFromControllerManager() if err != nil { - e2elog.Logf("Error fetching controller-manager metrics") + framework.Logf("Error fetching controller-manager metrics") return false, err } @@ -640,18 +639,18 @@ func getControllerStorageMetrics(ms metrics.ControllerManagerMetrics) *storageCo func findVolumeStatMetric(metricKeyName string, namespace string, pvcName string, kubeletMetrics metrics.KubeletMetrics) bool { found := false errCount := 0 - e2elog.Logf("Looking for sample in metric `%s` tagged with namespace `%s`, PVC `%s`", metricKeyName, namespace, pvcName) + framework.Logf("Looking for sample in metric `%s` tagged with namespace `%s`, PVC `%s`", metricKeyName, namespace, pvcName) if samples, ok := kubeletMetrics[metricKeyName]; ok { for _, sample := range samples { - e2elog.Logf("Found sample %s", sample.String()) + framework.Logf("Found sample %s", sample.String()) samplePVC, ok := sample.Metric["persistentvolumeclaim"] if !ok { - e2elog.Logf("Error getting pvc for metric %s, sample %s", metricKeyName, sample.String()) + framework.Logf("Error getting pvc for metric %s, sample %s", metricKeyName, sample.String()) errCount++ } sampleNS, ok := sample.Metric["namespace"] if !ok { - e2elog.Logf("Error getting namespace for metric %s, sample %s", metricKeyName, sample.String()) + framework.Logf("Error getting namespace for metric %s, sample %s", metricKeyName, sample.String()) errCount++ } @@ -675,7 +674,7 @@ func waitForPVControllerSync(metricsGrabber *metrics.Grabber, metricName, dimens verifyMetricFunc := func() (bool, error) { updatedMetrics, err := metricsGrabber.GrabFromControllerManager() if err != nil { - e2elog.Logf("Error fetching controller-manager metrics") + framework.Logf("Error fetching controller-manager metrics") return false, err } return len(getPVControllerMetrics(updatedMetrics, metricName, dimension)) > 0, nil @@ -718,17 +717,17 @@ func calculateRelativeValues(originValues, updatedValues map[string]int64) map[s func hasValidMetrics(metrics metrics.Metrics, metricKey string, dimensions ...string) bool { var errCount int - e2elog.Logf("Looking for sample in metric %q", metricKey) + framework.Logf("Looking for sample in metric %q", metricKey) samples, ok := metrics[metricKey] if !ok { - e2elog.Logf("Key %q was not found in metrics", metricKey) + framework.Logf("Key %q was not found in metrics", metricKey) return false } for _, sample := range samples { - e2elog.Logf("Found sample %q", sample.String()) + framework.Logf("Found sample %q", sample.String()) for _, d := range dimensions { if _, ok := sample.Metric[model.LabelName(d)]; !ok { - e2elog.Logf("Error getting dimension %q for metric %q, sample %q", d, metricKey, sample.String()) + framework.Logf("Error getting dimension %q for metric %q, sample %q", d, metricKey, sample.String()) errCount++ } } @@ -739,7 +738,7 @@ func hasValidMetrics(metrics metrics.Metrics, metricKey string, dimensions ...st func getStatesMetrics(metricKey string, givenMetrics metrics.Metrics) map[string]map[string]int64 { states := make(map[string]map[string]int64) for _, sample := range givenMetrics[metricKey] { - e2elog.Logf("Found sample %q", sample.String()) + framework.Logf("Found sample %q", sample.String()) state := string(sample.Metric["state"]) pluginName := string(sample.Metric["plugin_name"]) states[state] = map[string]int64{pluginName: int64(sample.Value)} diff --git a/test/e2e/storage/volume_provisioning.go b/test/e2e/storage/volume_provisioning.go index cf9034673c6..76377c8dac5 100644 --- a/test/e2e/storage/volume_provisioning.go +++ b/test/e2e/storage/volume_provisioning.go @@ -46,7 +46,6 @@ import ( storageutil "k8s.io/kubernetes/pkg/apis/storage/v1/util" "k8s.io/kubernetes/test/e2e/framework" "k8s.io/kubernetes/test/e2e/framework/auth" - e2elog "k8s.io/kubernetes/test/e2e/framework/log" e2epod "k8s.io/kubernetes/test/e2e/framework/pod" "k8s.io/kubernetes/test/e2e/framework/providers/gce" "k8s.io/kubernetes/test/e2e/storage/testsuites" @@ -70,28 +69,28 @@ func checkZoneFromLabelAndAffinity(pv *v1.PersistentVolume, zone string, matchZo func checkZonesFromLabelAndAffinity(pv *v1.PersistentVolume, zones sets.String, matchZones bool) { ginkgo.By("checking PV's zone label and node affinity terms match expected zone") if pv == nil { - e2elog.Failf("nil pv passed") + framework.Failf("nil pv passed") } pvLabel, ok := pv.Labels[v1.LabelZoneFailureDomain] if !ok { - e2elog.Failf("label %s not found on PV", v1.LabelZoneFailureDomain) + framework.Failf("label %s not found on PV", v1.LabelZoneFailureDomain) } zonesFromLabel, err := volumehelpers.LabelZonesToSet(pvLabel) if err != nil { - e2elog.Failf("unable to parse zone labels %s: %v", pvLabel, err) + framework.Failf("unable to parse zone labels %s: %v", pvLabel, err) } if matchZones && !zonesFromLabel.Equal(zones) { - e2elog.Failf("value[s] of %s label for PV: %v does not match expected zone[s]: %v", v1.LabelZoneFailureDomain, zonesFromLabel, zones) + framework.Failf("value[s] of %s label for PV: %v does not match expected zone[s]: %v", v1.LabelZoneFailureDomain, zonesFromLabel, zones) } if !matchZones && !zonesFromLabel.IsSuperset(zones) { - e2elog.Failf("value[s] of %s label for PV: %v does not contain expected zone[s]: %v", v1.LabelZoneFailureDomain, zonesFromLabel, zones) + framework.Failf("value[s] of %s label for PV: %v does not contain expected zone[s]: %v", v1.LabelZoneFailureDomain, zonesFromLabel, zones) } if pv.Spec.NodeAffinity == nil { - e2elog.Failf("node affinity not found in PV spec %v", pv.Spec) + framework.Failf("node affinity not found in PV spec %v", pv.Spec) } if len(pv.Spec.NodeAffinity.Required.NodeSelectorTerms) == 0 { - e2elog.Failf("node selector terms not found in PV spec %v", pv.Spec) + framework.Failf("node selector terms not found in PV spec %v", pv.Spec) } for _, term := range pv.Spec.NodeAffinity.Required.NodeSelectorTerms { @@ -103,15 +102,15 @@ func checkZonesFromLabelAndAffinity(pv *v1.PersistentVolume, zones sets.String, keyFound = true zonesFromNodeAffinity := sets.NewString(r.Values...) if matchZones && !zonesFromNodeAffinity.Equal(zones) { - e2elog.Failf("zones from NodeAffinity of PV: %v does not equal expected zone[s]: %v", zonesFromNodeAffinity, zones) + framework.Failf("zones from NodeAffinity of PV: %v does not equal expected zone[s]: %v", zonesFromNodeAffinity, zones) } if !matchZones && !zonesFromNodeAffinity.IsSuperset(zones) { - e2elog.Failf("zones from NodeAffinity of PV: %v does not contain expected zone[s]: %v", zonesFromNodeAffinity, zones) + framework.Failf("zones from NodeAffinity of PV: %v does not contain expected zone[s]: %v", zonesFromNodeAffinity, zones) } break } if !keyFound { - e2elog.Failf("label %s not found in term %v", v1.LabelZoneFailureDomain, term) + framework.Failf("label %s not found in term %v", v1.LabelZoneFailureDomain, term) } } } @@ -130,10 +129,10 @@ func checkAWSEBS(volume *v1.PersistentVolume, volumeType string, encrypted bool) if len(zone) > 0 { region := zone[:len(zone)-1] cfg := aws.Config{Region: ®ion} - e2elog.Logf("using region %s", region) + framework.Logf("using region %s", region) client = ec2.New(session.New(), &cfg) } else { - e2elog.Logf("no region configured") + framework.Logf("no region configured") client = ec2.New(session.New()) } @@ -208,7 +207,7 @@ func testZonalDelayedBinding(c clientset.Interface, ns string, specifyAllowedTop } for _, test := range tests { if !framework.ProviderIs(test.CloudProviders...) { - e2elog.Logf("Skipping %q: cloud providers is not %v", test.Name, test.CloudProviders) + framework.Logf("Skipping %q: cloud providers is not %v", test.Name, test.CloudProviders) continue } action := "creating claims with class with waitForFirstConsumer" @@ -234,14 +233,14 @@ func testZonalDelayedBinding(c clientset.Interface, ns string, specifyAllowedTop } pvs, node := test.TestBindingWaitForFirstConsumerMultiPVC(claims, nil /* node selector */, false /* expect unschedulable */) if node == nil { - e2elog.Failf("unexpected nil node found") + framework.Failf("unexpected nil node found") } zone, ok := node.Labels[v1.LabelZoneFailureDomain] if !ok { - e2elog.Failf("label %s not found on Node", v1.LabelZoneFailureDomain) + framework.Failf("label %s not found on Node", v1.LabelZoneFailureDomain) } if specifyAllowedTopology && topoZone != zone { - e2elog.Failf("zone specified in allowedTopologies: %s does not match zone of node where PV got provisioned: %s", topoZone, zone) + framework.Failf("zone specified in allowedTopologies: %s does not match zone of node where PV got provisioned: %s", topoZone, zone) } for _, pv := range pvs { checkZoneFromLabelAndAffinity(pv, zone, true) @@ -450,7 +449,7 @@ var _ = utils.SIGDescribe("Dynamic Provisioning", func() { test := t if !framework.ProviderIs(test.CloudProviders...) { - e2elog.Logf("Skipping %q: cloud providers is not %v", test.Name, test.CloudProviders) + framework.Logf("Skipping %q: cloud providers is not %v", test.Name, test.CloudProviders) continue } @@ -590,7 +589,7 @@ var _ = utils.SIGDescribe("Dynamic Provisioning", func() { // The claim should timeout phase:Pending err = framework.WaitForPersistentVolumeClaimPhase(v1.ClaimBound, c, ns, pvc.Name, 2*time.Second, framework.ClaimProvisionShortTimeout) framework.ExpectError(err) - e2elog.Logf(err.Error()) + framework.Logf(err.Error()) }) ginkgo.It("should test that deleting a claim before the volume is provisioned deletes the volume.", func() { @@ -637,13 +636,13 @@ var _ = utils.SIGDescribe("Dynamic Provisioning", func() { // Report indicators of regression if len(residualPVs) > 0 { - e2elog.Logf("Remaining PersistentVolumes:") + framework.Logf("Remaining PersistentVolumes:") for i, pv := range residualPVs { - e2elog.Logf("\t%d) %s", i+1, pv.Name) + framework.Logf("\t%d) %s", i+1, pv.Name) } - e2elog.Failf("Expected 0 PersistentVolumes remaining. Found %d", len(residualPVs)) + framework.Failf("Expected 0 PersistentVolumes remaining. Found %d", len(residualPVs)) } - e2elog.Logf("0 PersistentVolumes remain.") + framework.Logf("0 PersistentVolumes remain.") }) ginkgo.It("deletion should be idempotent", func() { @@ -798,7 +797,7 @@ var _ = utils.SIGDescribe("Dynamic Provisioning", func() { framework.SkipUnlessProviderIs("openstack", "gce", "aws", "gke", "vsphere", "azure") scName, scErr := framework.GetDefaultStorageClassName(c) if scErr != nil { - e2elog.Failf(scErr.Error()) + framework.Failf(scErr.Error()) } test := testsuites.StorageClassTest{ Name: "default", @@ -824,7 +823,7 @@ var _ = utils.SIGDescribe("Dynamic Provisioning", func() { // The claim should timeout phase:Pending err = framework.WaitForPersistentVolumeClaimPhase(v1.ClaimBound, c, ns, claim.Name, 2*time.Second, framework.ClaimProvisionShortTimeout) framework.ExpectError(err) - e2elog.Logf(err.Error()) + framework.Logf(err.Error()) claim, err = c.CoreV1().PersistentVolumeClaims(ns).Get(claim.Name, metav1.GetOptions{}) framework.ExpectNoError(err) framework.ExpectEqual(claim.Status.Phase, v1.ClaimPending) @@ -835,7 +834,7 @@ var _ = utils.SIGDescribe("Dynamic Provisioning", func() { framework.SkipUnlessProviderIs("openstack", "gce", "aws", "gke", "vsphere", "azure") scName, scErr := framework.GetDefaultStorageClassName(c) if scErr != nil { - e2elog.Failf(scErr.Error()) + framework.Failf(scErr.Error()) } test := testsuites.StorageClassTest{ Name: "default", @@ -861,7 +860,7 @@ var _ = utils.SIGDescribe("Dynamic Provisioning", func() { // The claim should timeout phase:Pending err = framework.WaitForPersistentVolumeClaimPhase(v1.ClaimBound, c, ns, claim.Name, 2*time.Second, framework.ClaimProvisionShortTimeout) framework.ExpectError(err) - e2elog.Logf(err.Error()) + framework.Logf(err.Error()) claim, err = c.CoreV1().PersistentVolumeClaims(ns).Get(claim.Name, metav1.GetOptions{}) framework.ExpectNoError(err) framework.ExpectEqual(claim.Status.Phase, v1.ClaimPending) @@ -913,7 +912,7 @@ var _ = utils.SIGDescribe("Dynamic Provisioning", func() { class, err := c.StorageV1().StorageClasses().Create(class) framework.ExpectNoError(err) defer func() { - e2elog.Logf("deleting storage class %s", class.Name) + framework.Logf("deleting storage class %s", class.Name) framework.ExpectNoError(c.StorageV1().StorageClasses().Delete(class.Name, nil)) }() @@ -926,10 +925,10 @@ var _ = utils.SIGDescribe("Dynamic Provisioning", func() { claim, err = c.CoreV1().PersistentVolumeClaims(claim.Namespace).Create(claim) framework.ExpectNoError(err) defer func() { - e2elog.Logf("deleting claim %q/%q", claim.Namespace, claim.Name) + framework.Logf("deleting claim %q/%q", claim.Namespace, claim.Name) err = c.CoreV1().PersistentVolumeClaims(claim.Namespace).Delete(claim.Name, nil) if err != nil && !apierrs.IsNotFound(err) { - e2elog.Failf("Error deleting claim %q. Error: %v", claim.Name, err) + framework.Failf("Error deleting claim %q. Error: %v", claim.Name, err) } }() @@ -958,7 +957,7 @@ var _ = utils.SIGDescribe("Dynamic Provisioning", func() { return false, nil }) if err == wait.ErrWaitTimeout { - e2elog.Logf("The test missed event about failed provisioning, but checked that no volume was provisioned for %v", framework.ClaimProvisionTimeout) + framework.Logf("The test missed event about failed provisioning, but checked that no volume was provisioned for %v", framework.ClaimProvisionTimeout) err = nil } framework.ExpectNoError(err) @@ -990,7 +989,7 @@ var _ = utils.SIGDescribe("Dynamic Provisioning", func() { } for _, test := range tests { if !framework.ProviderIs(test.CloudProviders...) { - e2elog.Logf("Skipping %q: cloud providers is not %v", test.Name, test.CloudProviders) + framework.Logf("Skipping %q: cloud providers is not %v", test.Name, test.CloudProviders) continue } ginkgo.By("creating a claim with class with allowedTopologies set") diff --git a/test/e2e/storage/volumes.go b/test/e2e/storage/volumes.go index 09485f7692e..aa1eb44a1d6 100644 --- a/test/e2e/storage/volumes.go +++ b/test/e2e/storage/volumes.go @@ -20,11 +20,10 @@ package storage import ( "github.com/onsi/ginkgo" - "k8s.io/api/core/v1" + v1 "k8s.io/api/core/v1" metav1 "k8s.io/apimachinery/pkg/apis/meta/v1" clientset "k8s.io/client-go/kubernetes" "k8s.io/kubernetes/test/e2e/framework" - e2elog "k8s.io/kubernetes/test/e2e/framework/log" "k8s.io/kubernetes/test/e2e/framework/volume" "k8s.io/kubernetes/test/e2e/storage/utils" ) @@ -66,7 +65,7 @@ var _ = utils.SIGDescribe("Volumes", func() { }, } if _, err := cs.CoreV1().ConfigMaps(namespace.Name).Create(configMap); err != nil { - e2elog.Failf("unable to create test configmap: %v", err) + framework.Failf("unable to create test configmap: %v", err) } defer func() { _ = cs.CoreV1().ConfigMaps(namespace.Name).Delete(configMap.Name, nil) diff --git a/test/e2e/storage/vsphere/BUILD b/test/e2e/storage/vsphere/BUILD index 88c67b22b0c..f4837636069 100644 --- a/test/e2e/storage/vsphere/BUILD +++ b/test/e2e/storage/vsphere/BUILD @@ -53,7 +53,6 @@ go_library( "//staging/src/k8s.io/client-go/kubernetes:go_default_library", "//test/e2e/framework:go_default_library", "//test/e2e/framework/deployment:go_default_library", - "//test/e2e/framework/log:go_default_library", "//test/e2e/framework/node:go_default_library", "//test/e2e/framework/pod:go_default_library", "//test/e2e/framework/ssh:go_default_library", diff --git a/test/e2e/storage/vsphere/bootstrap.go b/test/e2e/storage/vsphere/bootstrap.go index 39eb39f36f8..1775ec40c47 100644 --- a/test/e2e/storage/vsphere/bootstrap.go +++ b/test/e2e/storage/vsphere/bootstrap.go @@ -19,7 +19,6 @@ package vsphere import ( metav1 "k8s.io/apimachinery/pkg/apis/meta/v1" "k8s.io/kubernetes/test/e2e/framework" - e2elog "k8s.io/kubernetes/test/e2e/framework/log" "sync" ) @@ -43,23 +42,23 @@ func bootstrapOnce() { // 1. Read vSphere conf and get VSphere instances vsphereInstances, err := GetVSphereInstances() if err != nil { - e2elog.Failf("Failed to bootstrap vSphere with error: %v", err) + framework.Failf("Failed to bootstrap vSphere with error: %v", err) } // 2. Get all nodes nodeList, err := f.ClientSet.CoreV1().Nodes().List(metav1.ListOptions{}) if err != nil { - e2elog.Failf("Failed to get nodes: %v", err) + framework.Failf("Failed to get nodes: %v", err) } TestContext = VSphereContext{NodeMapper: &NodeMapper{}, VSphereInstances: vsphereInstances} // 3. Get Node to VSphere mapping err = TestContext.NodeMapper.GenerateNodeMap(vsphereInstances, *nodeList) if err != nil { - e2elog.Failf("Failed to bootstrap vSphere with error: %v", err) + framework.Failf("Failed to bootstrap vSphere with error: %v", err) } // 4. Generate Zone to Datastore mapping err = TestContext.NodeMapper.GenerateZoneToDatastoreMap() if err != nil { - e2elog.Failf("Failed to generate zone to datastore mapping with error: %v", err) + framework.Failf("Failed to generate zone to datastore mapping with error: %v", err) } close(waiting) } diff --git a/test/e2e/storage/vsphere/config.go b/test/e2e/storage/vsphere/config.go index e8abbb0ce12..98f3feb0c35 100644 --- a/test/e2e/storage/vsphere/config.go +++ b/test/e2e/storage/vsphere/config.go @@ -23,7 +23,7 @@ import ( "os" "gopkg.in/gcfg.v1" - e2elog "k8s.io/kubernetes/test/e2e/framework/log" + "k8s.io/kubernetes/test/e2e/framework" ) const ( @@ -130,13 +130,13 @@ func populateInstanceMap(cfg *ConfigFile) (map[string]*VSphere, error) { if cfg.Workspace.VCenterIP == "" || cfg.Workspace.DefaultDatastore == "" || cfg.Workspace.Folder == "" || cfg.Workspace.Datacenter == "" { msg := fmt.Sprintf("All fields in workspace are mandatory."+ " vsphere.conf does not have the workspace specified correctly. cfg.Workspace: %+v", cfg.Workspace) - e2elog.Logf(msg) + framework.Logf(msg) return nil, errors.New(msg) } for vcServer, vcConfig := range cfg.VirtualCenter { - e2elog.Logf("Initializing vc server %s", vcServer) + framework.Logf("Initializing vc server %s", vcServer) if vcServer == "" { - e2elog.Logf("vsphere.conf does not have the VirtualCenter IP address specified") + framework.Logf("vsphere.conf does not have the VirtualCenter IP address specified") return nil, errors.New("vsphere.conf does not have the VirtualCenter IP address specified") } vcConfig.Hostname = vcServer @@ -149,12 +149,12 @@ func populateInstanceMap(cfg *ConfigFile) (map[string]*VSphere, error) { } if vcConfig.Username == "" { msg := fmt.Sprintf("vcConfig.Username is empty for vc %s!", vcServer) - e2elog.Logf(msg) + framework.Logf(msg) return nil, errors.New(msg) } if vcConfig.Password == "" { msg := fmt.Sprintf("vcConfig.Password is empty for vc %s!", vcServer) - e2elog.Logf(msg) + framework.Logf(msg) return nil, errors.New(msg) } if vcConfig.Port == "" { @@ -176,6 +176,6 @@ func populateInstanceMap(cfg *ConfigFile) (map[string]*VSphere, error) { vsphereInstances[vcServer] = &vsphereIns } - e2elog.Logf("ConfigFile %v \n vSphere instances %v", cfg, vsphereInstances) + framework.Logf("ConfigFile %v \n vSphere instances %v", cfg, vsphereInstances) return vsphereInstances, nil } diff --git a/test/e2e/storage/vsphere/nodemapper.go b/test/e2e/storage/vsphere/nodemapper.go index 19854308d74..7694a586450 100644 --- a/test/e2e/storage/vsphere/nodemapper.go +++ b/test/e2e/storage/vsphere/nodemapper.go @@ -28,7 +28,7 @@ import ( "github.com/vmware/govmomi/vim25/mo" "github.com/vmware/govmomi/vim25/types" v1 "k8s.io/api/core/v1" - e2elog "k8s.io/kubernetes/test/e2e/framework/log" + "k8s.io/kubernetes/test/e2e/framework" neturl "net/url" ) @@ -77,7 +77,7 @@ func (nm *NodeMapper) GenerateNodeMap(vSphereInstances map[string]*VSphere, node if vs.Config.Datacenters == "" { datacenters, err = vs.GetAllDatacenter(ctx) if err != nil { - e2elog.Logf("NodeMapper error: %v", err) + framework.Logf("NodeMapper error: %v", err) continue } } else { @@ -89,7 +89,7 @@ func (nm *NodeMapper) GenerateNodeMap(vSphereInstances map[string]*VSphere, node } datacenter, err := vs.GetDatacenter(ctx, dc) if err != nil { - e2elog.Logf("NodeMapper error dc: %s \n err: %v", dc, err) + framework.Logf("NodeMapper error dc: %s \n err: %v", dc, err) continue } @@ -98,7 +98,7 @@ func (nm *NodeMapper) GenerateNodeMap(vSphereInstances map[string]*VSphere, node } for _, dc := range datacenters { - e2elog.Logf("Search candidates vc=%s and datacenter=%s", vs.Config.Hostname, dc.Name()) + framework.Logf("Search candidates vc=%s and datacenter=%s", vs.Config.Hostname, dc.Name()) queueChannel = append(queueChannel, &VmSearch{vs: vs, datacenter: dc}) } } @@ -107,20 +107,20 @@ func (nm *NodeMapper) GenerateNodeMap(vSphereInstances map[string]*VSphere, node n := node go func() { nodeUUID := getUUIDFromProviderID(n.Spec.ProviderID) - e2elog.Logf("Searching for node with UUID: %s", nodeUUID) + framework.Logf("Searching for node with UUID: %s", nodeUUID) for _, res := range queueChannel { ctx, cancel := context.WithCancel(context.Background()) defer cancel() vm, err := res.vs.GetVMByUUID(ctx, nodeUUID, res.datacenter) if err != nil { - e2elog.Logf("Error %v while looking for node=%s in vc=%s and datacenter=%s", + framework.Logf("Error %v while looking for node=%s in vc=%s and datacenter=%s", err, n.Name, res.vs.Config.Hostname, res.datacenter.Name()) continue } if vm != nil { hostSystemRef := res.vs.GetHostFromVMReference(ctx, vm.Reference()) zones := retrieveZoneInformationForNode(n.Name, res.vs, hostSystemRef) - e2elog.Logf("Found node %s as vm=%+v placed on host=%+v under zones %s in vc=%s and datacenter=%s", + framework.Logf("Found node %s as vm=%+v placed on host=%+v under zones %s in vc=%s and datacenter=%s", n.Name, vm, hostSystemRef, zones, res.vs.Config.Hostname, res.datacenter.Name()) nodeInfo := &NodeInfo{Name: n.Name, DataCenterRef: res.datacenter.Reference(), VirtualMachineRef: vm.Reference(), HostSystemRef: hostSystemRef, VSphere: res.vs, Zones: zones} nm.SetNodeInfo(n.Name, nodeInfo) @@ -192,10 +192,10 @@ func retrieveZoneInformationForNode(nodeName string, connection *VSphere, hostSy } switch { case category.Name == "k8s-zone": - e2elog.Logf("Found %s associated with %s for %s", tag.Name, ancestor.Name, nodeName) + framework.Logf("Found %s associated with %s for %s", tag.Name, ancestor.Name, nodeName) zonesAttachedToObject = append(zonesAttachedToObject, tag.Name) case category.Name == "k8s-region": - e2elog.Logf("Found %s associated with %s for %s", tag.Name, ancestor.Name, nodeName) + framework.Logf("Found %s associated with %s for %s", tag.Name, ancestor.Name, nodeName) } } // Overwrite zone information if it exists for this object @@ -250,7 +250,7 @@ func (nm *NodeMapper) GenerateZoneToDatastoreMap() error { vcToZoneDatastoresMap[vc][zone] = commonDatastores } } - e2elog.Logf("Zone to datastores map : %+v", vcToZoneDatastoresMap) + framework.Logf("Zone to datastores map : %+v", vcToZoneDatastoresMap) return nil } diff --git a/test/e2e/storage/vsphere/persistent_volumes-vsphere.go b/test/e2e/storage/vsphere/persistent_volumes-vsphere.go index 31eb85dc146..d3a0495f58b 100644 --- a/test/e2e/storage/vsphere/persistent_volumes-vsphere.go +++ b/test/e2e/storage/vsphere/persistent_volumes-vsphere.go @@ -26,7 +26,6 @@ import ( "k8s.io/apimachinery/pkg/labels" clientset "k8s.io/client-go/kubernetes" "k8s.io/kubernetes/test/e2e/framework" - e2elog "k8s.io/kubernetes/test/e2e/framework/log" e2epod "k8s.io/kubernetes/test/e2e/framework/pod" "k8s.io/kubernetes/test/e2e/storage/utils" ) @@ -113,7 +112,7 @@ var _ = utils.SIGDescribe("PersistentVolumes:vsphere", func() { }) ginkgo.AfterEach(func() { - e2elog.Logf("AfterEach: Cleaning up test resources") + framework.Logf("AfterEach: Cleaning up test resources") if c != nil { framework.ExpectNoError(e2epod.DeletePodWithWait(c, clientPod), "AfterEach: failed to delete pod ", clientPod.Name) diff --git a/test/e2e/storage/vsphere/pv_reclaimpolicy.go b/test/e2e/storage/vsphere/pv_reclaimpolicy.go index 31b34573615..1122f5932dc 100644 --- a/test/e2e/storage/vsphere/pv_reclaimpolicy.go +++ b/test/e2e/storage/vsphere/pv_reclaimpolicy.go @@ -27,7 +27,6 @@ import ( metav1 "k8s.io/apimachinery/pkg/apis/meta/v1" clientset "k8s.io/client-go/kubernetes" "k8s.io/kubernetes/test/e2e/framework" - e2elog "k8s.io/kubernetes/test/e2e/framework/log" e2epod "k8s.io/kubernetes/test/e2e/framework/pod" "k8s.io/kubernetes/test/e2e/storage/utils" ) @@ -132,7 +131,7 @@ var _ = utils.SIGDescribe("PersistentVolumes [Feature:ReclaimPolicy]", func() { ginkgo.By("Verify the volume is accessible and available in the pod") verifyVSphereVolumesAccessible(c, pod, []*v1.PersistentVolume{pv}) - e2elog.Logf("Verified that Volume is accessible in the POD after deleting PV claim") + framework.Logf("Verified that Volume is accessible in the POD after deleting PV claim") ginkgo.By("Deleting the Pod") framework.ExpectNoError(e2epod.DeletePodWithWait(c, pod), "Failed to delete pod ", pod.Name) @@ -179,7 +178,7 @@ var _ = utils.SIGDescribe("PersistentVolumes [Feature:ReclaimPolicy]", func() { pvc = nil ginkgo.By("Verify PV is retained") - e2elog.Logf("Waiting for PV %v to become Released", pv.Name) + framework.Logf("Waiting for PV %v to become Released", pv.Name) err = framework.WaitForPersistentVolumePhase(v1.VolumeReleased, c, pv.Name, 3*time.Second, 300*time.Second) framework.ExpectNoError(err) framework.ExpectNoError(framework.DeletePersistentVolume(c, pv.Name), "Failed to delete PV ", pv.Name) diff --git a/test/e2e/storage/vsphere/vsphere.go b/test/e2e/storage/vsphere/vsphere.go index 1f27bcf935c..f853b6cdeed 100644 --- a/test/e2e/storage/vsphere/vsphere.go +++ b/test/e2e/storage/vsphere/vsphere.go @@ -30,7 +30,7 @@ import ( "github.com/vmware/govmomi/vim25/mo" "github.com/vmware/govmomi/vim25/soap" "github.com/vmware/govmomi/vim25/types" - e2elog "k8s.io/kubernetes/test/e2e/framework/log" + "k8s.io/kubernetes/test/e2e/framework" ) const ( @@ -121,7 +121,7 @@ func (vs *VSphere) GetFolderByPath(ctx context.Context, dc object.Reference, fol finder.SetDatacenter(datacenter) vmFolder, err := finder.Folder(ctx, folderPath) if err != nil { - e2elog.Logf("Failed to get the folder reference for %s. err: %+v", folderPath, err) + framework.Logf("Failed to get the folder reference for %s. err: %+v", folderPath, err) return vmFolderMor, err } return vmFolder.Reference(), nil @@ -156,15 +156,15 @@ func (vs *VSphere) CreateVolume(volumeOptions *VolumeOptions, dataCenterRef type soapFault := soap.ToSoapFault(err) if _, ok := soapFault.VimFault().(types.FileAlreadyExists); ok { directoryAlreadyPresent = true - e2elog.Logf("Directory with the path %+q is already present", directoryPath) + framework.Logf("Directory with the path %+q is already present", directoryPath) } } if !directoryAlreadyPresent { - e2elog.Logf("Cannot create dir %#v. err %s", directoryPath, err) + framework.Logf("Cannot create dir %#v. err %s", directoryPath, err) return "", err } } - e2elog.Logf("Created dir with path as %+q", directoryPath) + framework.Logf("Created dir with path as %+q", directoryPath) vmdkPath := directoryPath + volumeOptions.Name + ".vmdk" // Create a virtual disk manager @@ -180,12 +180,12 @@ func (vs *VSphere) CreateVolume(volumeOptions *VolumeOptions, dataCenterRef type // Create virtual disk task, err := vdm.CreateVirtualDisk(ctx, vmdkPath, datacenter, vmDiskSpec) if err != nil { - e2elog.Logf("Failed to create virtual disk: %s. err: %+v", vmdkPath, err) + framework.Logf("Failed to create virtual disk: %s. err: %+v", vmdkPath, err) return "", err } taskInfo, err := task.WaitForResult(ctx, nil) if err != nil { - e2elog.Logf("Failed to complete virtual disk creation: %s. err: %+v", vmdkPath, err) + framework.Logf("Failed to complete virtual disk creation: %s. err: %+v", vmdkPath, err) return "", err } volumePath := taskInfo.Result.(string) @@ -209,12 +209,12 @@ func (vs *VSphere) DeleteVolume(volumePath string, dataCenterRef types.ManagedOb // Delete virtual disk task, err := virtualDiskManager.DeleteVirtualDisk(ctx, diskPath, datacenter) if err != nil { - e2elog.Logf("Failed to delete virtual disk. err: %v", err) + framework.Logf("Failed to delete virtual disk. err: %v", err) return err } err = task.Wait(ctx) if err != nil { - e2elog.Logf("Failed to delete virtual disk. err: %v", err) + framework.Logf("Failed to delete virtual disk. err: %v", err) return err } return nil @@ -233,7 +233,7 @@ func (vs *VSphere) IsVMPresent(vmName string, dataCenterRef types.ManagedObjectR vmFolder := object.NewFolder(vs.Client.Client, folderMor) vmFoldersChildren, err := vmFolder.Children(ctx) if err != nil { - e2elog.Logf("Failed to get children from Folder: %s. err: %+v", vmFolder.InventoryPath, err) + framework.Logf("Failed to get children from Folder: %s. err: %+v", vmFolder.InventoryPath, err) return } for _, vmFoldersChild := range vmFoldersChildren { diff --git a/test/e2e/storage/vsphere/vsphere_statefulsets.go b/test/e2e/storage/vsphere/vsphere_statefulsets.go index 6b0654f417f..253c5885db8 100644 --- a/test/e2e/storage/vsphere/vsphere_statefulsets.go +++ b/test/e2e/storage/vsphere/vsphere_statefulsets.go @@ -25,7 +25,6 @@ import ( metav1 "k8s.io/apimachinery/pkg/apis/meta/v1" clientset "k8s.io/client-go/kubernetes" "k8s.io/kubernetes/test/e2e/framework" - e2elog "k8s.io/kubernetes/test/e2e/framework/log" e2epod "k8s.io/kubernetes/test/e2e/framework/pod" e2esset "k8s.io/kubernetes/test/e2e/framework/statefulset" "k8s.io/kubernetes/test/e2e/storage/utils" @@ -66,7 +65,7 @@ var _ = utils.SIGDescribe("vsphere statefulset", func() { Bootstrap(f) }) ginkgo.AfterEach(func() { - e2elog.Logf("Deleting all statefulset in namespace: %v", namespace) + framework.Logf("Deleting all statefulset in namespace: %v", namespace) e2esset.DeleteAllStatefulSets(client, namespace) }) @@ -117,7 +116,7 @@ var _ = utils.SIGDescribe("vsphere statefulset", func() { for _, volumespec := range sspod.Spec.Volumes { if volumespec.PersistentVolumeClaim != nil { vSpherediskPath := getvSphereVolumePathFromClaim(client, statefulset.Namespace, volumespec.PersistentVolumeClaim.ClaimName) - e2elog.Logf("Waiting for Volume: %q to detach from Node: %q", vSpherediskPath, sspod.Spec.NodeName) + framework.Logf("Waiting for Volume: %q to detach from Node: %q", vSpherediskPath, sspod.Spec.NodeName) framework.ExpectNoError(waitForVSphereDiskToDetach(vSpherediskPath, sspod.Spec.NodeName)) } } @@ -144,7 +143,7 @@ var _ = utils.SIGDescribe("vsphere statefulset", func() { for _, volumespec := range pod.Spec.Volumes { if volumespec.PersistentVolumeClaim != nil { vSpherediskPath := getvSphereVolumePathFromClaim(client, statefulset.Namespace, volumespec.PersistentVolumeClaim.ClaimName) - e2elog.Logf("Verify Volume: %q is attached to the Node: %q", vSpherediskPath, sspod.Spec.NodeName) + framework.Logf("Verify Volume: %q is attached to the Node: %q", vSpherediskPath, sspod.Spec.NodeName) // Verify scale up has re-attached the same volumes and not introduced new volume gomega.Expect(volumesBeforeScaleDown[vSpherediskPath] == "").To(gomega.BeFalse()) isVolumeAttached, verifyDiskAttachedError := diskIsAttached(vSpherediskPath, sspod.Spec.NodeName) diff --git a/test/e2e/storage/vsphere/vsphere_utils.go b/test/e2e/storage/vsphere/vsphere_utils.go index 27c616caa03..168c8df25c8 100644 --- a/test/e2e/storage/vsphere/vsphere_utils.go +++ b/test/e2e/storage/vsphere/vsphere_utils.go @@ -41,7 +41,6 @@ import ( "k8s.io/apimachinery/pkg/util/wait" clientset "k8s.io/client-go/kubernetes" "k8s.io/kubernetes/test/e2e/framework" - e2elog "k8s.io/kubernetes/test/e2e/framework/log" e2essh "k8s.io/kubernetes/test/e2e/framework/ssh" "k8s.io/kubernetes/test/e2e/storage/utils" imageutils "k8s.io/kubernetes/test/utils/image" @@ -81,13 +80,13 @@ func waitForVSphereDisksToDetach(nodeVolumes map[string][]string) error { for nodeName, nodeVolumes := range attachedResult { for volumePath, attached := range nodeVolumes { if attached { - e2elog.Logf("Waiting for volumes %q to detach from %q.", volumePath, string(nodeName)) + framework.Logf("Waiting for volumes %q to detach from %q.", volumePath, string(nodeName)) return false, nil } } } disksAttached = false - e2elog.Logf("Volume are successfully detached from all the nodes: %+v", nodeVolumes) + framework.Logf("Volume are successfully detached from all the nodes: %+v", nodeVolumes) return true, nil }) if err != nil { @@ -127,10 +126,10 @@ func waitForVSphereDiskStatus(volumePath string, nodeName string, expectedState currentState = attachedState[diskAttached] if currentState == expectedState { - e2elog.Logf("Volume %q has successfully %s %q", volumePath, attachedStateMsg[currentState], nodeName) + framework.Logf("Volume %q has successfully %s %q", volumePath, attachedStateMsg[currentState], nodeName) return true, nil } - e2elog.Logf("Waiting for Volume %q to be %s %q.", volumePath, attachedStateMsg[expectedState], nodeName) + framework.Logf("Waiting for Volume %q to be %s %q.", volumePath, attachedStateMsg[expectedState], nodeName) return false, nil }) if err != nil { @@ -203,13 +202,13 @@ func getVSpherePersistentVolumeClaimSpec(namespace string, labels map[string]str // function to write content to the volume backed by given PVC func writeContentToVSpherePV(client clientset.Interface, pvc *v1.PersistentVolumeClaim, expectedContent string) { utils.RunInPodWithVolume(client, pvc.Namespace, pvc.Name, "echo "+expectedContent+" > /mnt/test/data") - e2elog.Logf("Done with writing content to volume") + framework.Logf("Done with writing content to volume") } // function to verify content is matching on the volume backed for given PVC func verifyContentOfVSpherePV(client clientset.Interface, pvc *v1.PersistentVolumeClaim, expectedContent string) { utils.RunInPodWithVolume(client, pvc.Namespace, pvc.Name, "grep '"+expectedContent+"' /mnt/test/data") - e2elog.Logf("Successfully verified content of the volume") + framework.Logf("Successfully verified content of the volume") } func getVSphereStorageClassSpec(name string, scParameters map[string]string, zones []string, volumeBindingMode storagev1.VolumeBindingMode) *storagev1.StorageClass { @@ -476,7 +475,7 @@ func getPathFromVMDiskPath(vmDiskPath string) string { datastorePathObj := new(object.DatastorePath) isSuccess := datastorePathObj.FromString(vmDiskPath) if !isSuccess { - e2elog.Logf("Failed to parse vmDiskPath: %s", vmDiskPath) + framework.Logf("Failed to parse vmDiskPath: %s", vmDiskPath) return "" } return datastorePathObj.Path @@ -487,7 +486,7 @@ func getDatastorePathObjFromVMDiskPath(vmDiskPath string) (*object.DatastorePath datastorePathObj := new(object.DatastorePath) isSuccess := datastorePathObj.FromString(vmDiskPath) if !isSuccess { - e2elog.Logf("Failed to parse volPath: %s", vmDiskPath) + framework.Logf("Failed to parse volPath: %s", vmDiskPath) return nil, fmt.Errorf("Failed to parse volPath: %s", vmDiskPath) } return datastorePathObj, nil @@ -539,7 +538,7 @@ func removeStorageClusterORFolderNameFromVDiskPath(vDiskPath string) string { func getVirtualDeviceByPath(ctx context.Context, vm *object.VirtualMachine, diskPath string) (vim25types.BaseVirtualDevice, error) { vmDevices, err := vm.Device(ctx) if err != nil { - e2elog.Logf("Failed to get the devices for VM: %q. err: %+v", vm.InventoryPath, err) + framework.Logf("Failed to get the devices for VM: %q. err: %+v", vm.InventoryPath, err) return nil, err } @@ -549,10 +548,10 @@ func getVirtualDeviceByPath(ctx context.Context, vm *object.VirtualMachine, disk virtualDevice := device.GetVirtualDevice() if backing, ok := virtualDevice.Backing.(*vim25types.VirtualDiskFlatVer2BackingInfo); ok { if matchVirtualDiskAndVolPath(backing.FileName, diskPath) { - e2elog.Logf("Found VirtualDisk backing with filename %q for diskPath %q", backing.FileName, diskPath) + framework.Logf("Found VirtualDisk backing with filename %q for diskPath %q", backing.FileName, diskPath) return device, nil } else { - e2elog.Logf("VirtualDisk backing filename %q does not match with diskPath %q", backing.FileName, diskPath) + framework.Logf("VirtualDisk backing filename %q does not match with diskPath %q", backing.FileName, diskPath) } } } @@ -576,7 +575,7 @@ func convertVolPathsToDevicePaths(ctx context.Context, nodeVolumes map[string][] for i, volPath := range volPaths { deviceVolPath, err := convertVolPathToDevicePath(ctx, datacenter, volPath) if err != nil { - e2elog.Logf("Failed to convert vsphere volume path %s to device path for volume %s. err: %+v", volPath, deviceVolPath, err) + framework.Logf("Failed to convert vsphere volume path %s to device path for volume %s. err: %+v", volPath, deviceVolPath, err) return nil, err } volPaths[i] = deviceVolPath @@ -592,7 +591,7 @@ func convertVolPathToDevicePath(ctx context.Context, dc *object.Datacenter, volP // Get the canonical volume path for volPath. canonicalVolumePath, err := getCanonicalVolumePath(ctx, dc, volPath) if err != nil { - e2elog.Logf("Failed to get canonical vsphere volume path for volume: %s. err: %+v", volPath, err) + framework.Logf("Failed to get canonical vsphere volume path for volume: %s. err: %+v", volPath, err) return "", err } // Check if the volume path contains .vmdk extension. If not, add the extension and update the nodeVolumes Map @@ -613,7 +612,7 @@ func getVMXFilePath(vmObject *object.VirtualMachine) (vmxPath string) { gomega.Expect(nodeVM.Config).NotTo(gomega.BeNil()) vmxPath = nodeVM.Config.Files.VmPathName - e2elog.Logf("vmx file path is %s", vmxPath) + framework.Logf("vmx file path is %s", vmxPath) return vmxPath } @@ -638,7 +637,7 @@ func poweroffNodeVM(nodeName string, vm *object.VirtualMachine) { ctx, cancel := context.WithCancel(context.Background()) defer cancel() - e2elog.Logf("Powering off node VM %s", nodeName) + framework.Logf("Powering off node VM %s", nodeName) _, err := vm.PowerOff(ctx) framework.ExpectNoError(err) @@ -651,7 +650,7 @@ func poweronNodeVM(nodeName string, vm *object.VirtualMachine) { ctx, cancel := context.WithCancel(context.Background()) defer cancel() - e2elog.Logf("Powering on node VM %s", nodeName) + framework.Logf("Powering on node VM %s", nodeName) vm.PowerOn(ctx) err := vm.WaitForPowerState(ctx, vim25types.VirtualMachinePowerStatePoweredOn) @@ -665,7 +664,7 @@ func unregisterNodeVM(nodeName string, vm *object.VirtualMachine) { poweroffNodeVM(nodeName, vm) - e2elog.Logf("Unregistering node VM %s", nodeName) + framework.Logf("Unregistering node VM %s", nodeName) err := vm.Unregister(ctx) framework.ExpectNoError(err, "Unable to unregister the node") } @@ -675,7 +674,7 @@ func registerNodeVM(nodeName, workingDir, vmxFilePath string, rpool *object.Reso ctx, cancel := context.WithCancel(context.Background()) defer cancel() - e2elog.Logf("Registering node VM %s with vmx file path %s", nodeName, vmxFilePath) + framework.Logf("Registering node VM %s with vmx file path %s", nodeName, vmxFilePath) nodeInfo := TestContext.NodeMapper.GetNodeInfo(nodeName) finder := find.NewFinder(nodeInfo.VSphere.Client.Client, false) @@ -707,7 +706,7 @@ func disksAreAttached(nodeVolumes map[string][]string) (nodeVolumesAttachMap map // Convert VolPaths into canonical form so that it can be compared with the VM device path. vmVolumes, err := convertVolPathsToDevicePaths(ctx, nodeVolumes) if err != nil { - e2elog.Logf("Failed to convert volPaths to devicePaths: %+v. err: %+v", nodeVolumes, err) + framework.Logf("Failed to convert volPaths to devicePaths: %+v. err: %+v", nodeVolumes, err) return nil, err } for vm, volumes := range vmVolumes { @@ -735,7 +734,7 @@ func diskIsAttached(volPath string, nodeName string) (bool, error) { volPath = removeStorageClusterORFolderNameFromVDiskPath(volPath) device, err := getVirtualDeviceByPath(ctx, vm, volPath) if err != nil { - e2elog.Logf("diskIsAttached failed to determine whether disk %q is still attached on node %q", + framework.Logf("diskIsAttached failed to determine whether disk %q is still attached on node %q", volPath, nodeName) return false, err @@ -743,7 +742,7 @@ func diskIsAttached(volPath string, nodeName string) (bool, error) { if device == nil { return false, nil } - e2elog.Logf("diskIsAttached found the disk %q attached on node %q", volPath, nodeName) + framework.Logf("diskIsAttached found the disk %q attached on node %q", volPath, nodeName) return true, nil } @@ -780,7 +779,7 @@ func GetReadySchedulableRandomNodeInfo() *NodeInfo { // via service-control on the given vCenter host over SSH. func invokeVCenterServiceControl(command, service, host string) error { sshCmd := fmt.Sprintf("service-control --%s %s", command, service) - e2elog.Logf("Invoking command %v on vCenter host %v", sshCmd, host) + framework.Logf("Invoking command %v on vCenter host %v", sshCmd, host) result, err := e2essh.SSH(sshCmd, host, framework.TestContext.Provider) if err != nil || result.Code != 0 { e2essh.LogResult(result) diff --git a/test/e2e/storage/vsphere/vsphere_volume_datastore.go b/test/e2e/storage/vsphere/vsphere_volume_datastore.go index 9005427daf4..042a34fd84c 100644 --- a/test/e2e/storage/vsphere/vsphere_volume_datastore.go +++ b/test/e2e/storage/vsphere/vsphere_volume_datastore.go @@ -22,11 +22,10 @@ import ( "time" "github.com/onsi/ginkgo" - "k8s.io/api/core/v1" + v1 "k8s.io/api/core/v1" metav1 "k8s.io/apimachinery/pkg/apis/meta/v1" clientset "k8s.io/client-go/kubernetes" "k8s.io/kubernetes/test/e2e/framework" - e2elog "k8s.io/kubernetes/test/e2e/framework/log" "k8s.io/kubernetes/test/e2e/storage/utils" ) @@ -60,7 +59,7 @@ var _ = utils.SIGDescribe("Volume Provisioning on Datastore [Feature:vsphere]", scParameters = make(map[string]string) nodeList := framework.GetReadySchedulableNodesOrDie(f.ClientSet) if !(len(nodeList.Items) > 0) { - e2elog.Failf("Unable to find ready and schedulable Node") + framework.Failf("Unable to find ready and schedulable Node") } }) diff --git a/test/e2e/storage/vsphere/vsphere_volume_diskformat.go b/test/e2e/storage/vsphere/vsphere_volume_diskformat.go index 50ba273280a..fcf3e9aed9a 100644 --- a/test/e2e/storage/vsphere/vsphere_volume_diskformat.go +++ b/test/e2e/storage/vsphere/vsphere_volume_diskformat.go @@ -30,7 +30,6 @@ import ( "k8s.io/apimachinery/pkg/util/uuid" clientset "k8s.io/client-go/kubernetes" "k8s.io/kubernetes/test/e2e/framework" - e2elog "k8s.io/kubernetes/test/e2e/framework/log" e2epod "k8s.io/kubernetes/test/e2e/framework/pod" "k8s.io/kubernetes/test/e2e/storage/utils" ) @@ -103,7 +102,7 @@ var _ = utils.SIGDescribe("Volume Disk Format [Feature:vsphere]", func() { func invokeTest(f *framework.Framework, client clientset.Interface, namespace string, nodeName string, nodeKeyValueLabel map[string]string, diskFormat string) { - e2elog.Logf("Invoking Test for DiskFomat: %s", diskFormat) + framework.Logf("Invoking Test for DiskFomat: %s", diskFormat) scParameters := make(map[string]string) scParameters["diskformat"] = diskFormat diff --git a/test/e2e/storage/vsphere/vsphere_volume_fstype.go b/test/e2e/storage/vsphere/vsphere_volume_fstype.go index 5b310be3132..b87095b3dce 100644 --- a/test/e2e/storage/vsphere/vsphere_volume_fstype.go +++ b/test/e2e/storage/vsphere/vsphere_volume_fstype.go @@ -26,7 +26,6 @@ import ( metav1 "k8s.io/apimachinery/pkg/apis/meta/v1" clientset "k8s.io/client-go/kubernetes" "k8s.io/kubernetes/test/e2e/framework" - e2elog "k8s.io/kubernetes/test/e2e/framework/log" e2epod "k8s.io/kubernetes/test/e2e/framework/pod" "k8s.io/kubernetes/test/e2e/storage/utils" ) @@ -95,7 +94,7 @@ var _ = utils.SIGDescribe("Volume FStype [Feature:vsphere]", func() { }) func invokeTestForFstype(f *framework.Framework, client clientset.Interface, namespace string, fstype string, expectedContent string) { - e2elog.Logf("Invoking Test for fstype: %s", fstype) + framework.Logf("Invoking Test for fstype: %s", fstype) scParameters := make(map[string]string) scParameters["fstype"] = fstype diff --git a/test/e2e/storage/vsphere/vsphere_volume_node_poweroff.go b/test/e2e/storage/vsphere/vsphere_volume_node_poweroff.go index 30d7c1beea8..02e9ee53aa7 100644 --- a/test/e2e/storage/vsphere/vsphere_volume_node_poweroff.go +++ b/test/e2e/storage/vsphere/vsphere_volume_node_poweroff.go @@ -33,7 +33,6 @@ import ( clientset "k8s.io/client-go/kubernetes" "k8s.io/kubernetes/test/e2e/framework" e2edeploy "k8s.io/kubernetes/test/e2e/framework/deployment" - e2elog "k8s.io/kubernetes/test/e2e/framework/log" "k8s.io/kubernetes/test/e2e/storage/utils" ) @@ -159,19 +158,19 @@ func waitForPodToFailover(client clientset.Interface, deployment *appsv1.Deploym } if newNode != oldNode { - e2elog.Logf("The pod has been failed over from %q to %q", oldNode, newNode) + framework.Logf("The pod has been failed over from %q to %q", oldNode, newNode) return true, nil } - e2elog.Logf("Waiting for pod to be failed over from %q", oldNode) + framework.Logf("Waiting for pod to be failed over from %q", oldNode) return false, nil }) if err != nil { if err == wait.ErrWaitTimeout { - e2elog.Logf("Time out after waiting for %v", timeout) + framework.Logf("Time out after waiting for %v", timeout) } - e2elog.Logf("Pod did not fail over from %q with error: %v", oldNode, err) + framework.Logf("Pod did not fail over from %q with error: %v", oldNode, err) return "", err } diff --git a/test/e2e/storage/vsphere/vsphere_volume_perf.go b/test/e2e/storage/vsphere/vsphere_volume_perf.go index 042eb05f729..d0129a458ef 100644 --- a/test/e2e/storage/vsphere/vsphere_volume_perf.go +++ b/test/e2e/storage/vsphere/vsphere_volume_perf.go @@ -26,7 +26,6 @@ import ( storagev1 "k8s.io/api/storage/v1" clientset "k8s.io/client-go/kubernetes" "k8s.io/kubernetes/test/e2e/framework" - e2elog "k8s.io/kubernetes/test/e2e/framework/log" e2epod "k8s.io/kubernetes/test/e2e/framework/pod" "k8s.io/kubernetes/test/e2e/storage/utils" ) @@ -105,11 +104,11 @@ var _ = utils.SIGDescribe("vcp-performance [Feature:vsphere]", func() { } iterations64 := float64(iterations) - e2elog.Logf("Average latency for below operations") - e2elog.Logf("Creating %d PVCs and waiting for bound phase: %v seconds", volumeCount, sumLatency[CreateOp]/iterations64) - e2elog.Logf("Creating %v Pod: %v seconds", volumeCount/volumesPerPod, sumLatency[AttachOp]/iterations64) - e2elog.Logf("Deleting %v Pod and waiting for disk to be detached: %v seconds", volumeCount/volumesPerPod, sumLatency[DetachOp]/iterations64) - e2elog.Logf("Deleting %v PVCs: %v seconds", volumeCount, sumLatency[DeleteOp]/iterations64) + framework.Logf("Average latency for below operations") + framework.Logf("Creating %d PVCs and waiting for bound phase: %v seconds", volumeCount, sumLatency[CreateOp]/iterations64) + framework.Logf("Creating %v Pod: %v seconds", volumeCount/volumesPerPod, sumLatency[AttachOp]/iterations64) + framework.Logf("Deleting %v Pod and waiting for disk to be detached: %v seconds", volumeCount/volumesPerPod, sumLatency[DetachOp]/iterations64) + framework.Logf("Deleting %v PVCs: %v seconds", volumeCount, sumLatency[DeleteOp]/iterations64) }) }) diff --git a/test/e2e/storage/vsphere/vsphere_volume_vpxd_restart.go b/test/e2e/storage/vsphere/vsphere_volume_vpxd_restart.go index 5fe3ff8ef5f..d6ef1c1a801 100644 --- a/test/e2e/storage/vsphere/vsphere_volume_vpxd_restart.go +++ b/test/e2e/storage/vsphere/vsphere_volume_vpxd_restart.go @@ -29,7 +29,6 @@ import ( "k8s.io/apimachinery/pkg/util/uuid" clientset "k8s.io/client-go/kubernetes" "k8s.io/kubernetes/test/e2e/framework" - e2elog "k8s.io/kubernetes/test/e2e/framework/log" e2epod "k8s.io/kubernetes/test/e2e/framework/pod" "k8s.io/kubernetes/test/e2e/storage/utils" ) @@ -109,7 +108,7 @@ var _ = utils.SIGDescribe("Verify Volume Attach Through vpxd Restart [Feature:vs pods []*v1.Pod ) - e2elog.Logf("Testing for nodes on vCenter host: %s", vcHost) + framework.Logf("Testing for nodes on vCenter host: %s", vcHost) for i, node := range nodes { ginkgo.By(fmt.Sprintf("Creating test vsphere volume %d", i)) diff --git a/test/e2e/storage/vsphere/vsphere_volume_vsan_policy.go b/test/e2e/storage/vsphere/vsphere_volume_vsan_policy.go index 8b27d89440a..73992701aa0 100644 --- a/test/e2e/storage/vsphere/vsphere_volume_vsan_policy.go +++ b/test/e2e/storage/vsphere/vsphere_volume_vsan_policy.go @@ -29,7 +29,6 @@ import ( metav1 "k8s.io/apimachinery/pkg/apis/meta/v1" clientset "k8s.io/client-go/kubernetes" "k8s.io/kubernetes/test/e2e/framework" - e2elog "k8s.io/kubernetes/test/e2e/framework/log" e2enode "k8s.io/kubernetes/test/e2e/framework/node" e2epod "k8s.io/kubernetes/test/e2e/framework/pod" "k8s.io/kubernetes/test/e2e/storage/utils" @@ -106,15 +105,15 @@ var _ = utils.SIGDescribe("Storage Policy Based Volume Provisioning [Feature:vsp namespace = f.Namespace.Name policyName = GetAndExpectStringEnvVar(SPBMPolicyName) tagPolicy = GetAndExpectStringEnvVar(SPBMTagPolicy) - e2elog.Logf("framework: %+v", f) + framework.Logf("framework: %+v", f) scParameters = make(map[string]string) nodeList := framework.GetReadySchedulableNodesOrDie(f.ClientSet) if !(len(nodeList.Items) > 0) { - e2elog.Failf("Unable to find ready and schedulable Node") + framework.Failf("Unable to find ready and schedulable Node") } masternodes, _, err := e2enode.GetMasterAndWorkerNodes(client) if err != nil { - e2elog.Logf("Unexpected error occurred: %v", err) + framework.Logf("Unexpected error occurred: %v", err) } // TODO: write a wrapper for ExpectNoErrorWithOffset() framework.ExpectNoErrorWithOffset(0, err) @@ -127,7 +126,7 @@ var _ = utils.SIGDescribe("Storage Policy Based Volume Provisioning [Feature:vsp ginkgo.By(fmt.Sprintf("Invoking test for VSAN policy hostFailuresToTolerate: %s, cacheReservation: %s", HostFailuresToTolerateCapabilityVal, CacheReservationCapabilityVal)) scParameters[Policy_HostFailuresToTolerate] = HostFailuresToTolerateCapabilityVal scParameters[Policy_CacheReservation] = CacheReservationCapabilityVal - e2elog.Logf("Invoking test for VSAN storage capabilities: %+v", scParameters) + framework.Logf("Invoking test for VSAN storage capabilities: %+v", scParameters) invokeValidPolicyTest(f, client, namespace, scParameters) }) @@ -136,7 +135,7 @@ var _ = utils.SIGDescribe("Storage Policy Based Volume Provisioning [Feature:vsp ginkgo.By(fmt.Sprintf("Invoking test for VSAN policy diskStripes: %s, objectSpaceReservation: %s", DiskStripesCapabilityVal, ObjectSpaceReservationCapabilityVal)) scParameters[Policy_DiskStripes] = "1" scParameters[Policy_ObjectSpaceReservation] = "30" - e2elog.Logf("Invoking test for VSAN storage capabilities: %+v", scParameters) + framework.Logf("Invoking test for VSAN storage capabilities: %+v", scParameters) invokeValidPolicyTest(f, client, namespace, scParameters) }) @@ -146,7 +145,7 @@ var _ = utils.SIGDescribe("Storage Policy Based Volume Provisioning [Feature:vsp scParameters[Policy_DiskStripes] = DiskStripesCapabilityVal scParameters[Policy_ObjectSpaceReservation] = ObjectSpaceReservationCapabilityVal scParameters[Datastore] = VsanDatastore - e2elog.Logf("Invoking test for VSAN storage capabilities: %+v", scParameters) + framework.Logf("Invoking test for VSAN storage capabilities: %+v", scParameters) invokeValidPolicyTest(f, client, namespace, scParameters) }) @@ -155,7 +154,7 @@ var _ = utils.SIGDescribe("Storage Policy Based Volume Provisioning [Feature:vsp ginkgo.By(fmt.Sprintf("Invoking test for VSAN policy objectSpaceReservation: %s, iopsLimit: %s", ObjectSpaceReservationCapabilityVal, IopsLimitCapabilityVal)) scParameters[Policy_ObjectSpaceReservation] = ObjectSpaceReservationCapabilityVal scParameters[Policy_IopsLimit] = IopsLimitCapabilityVal - e2elog.Logf("Invoking test for VSAN storage capabilities: %+v", scParameters) + framework.Logf("Invoking test for VSAN storage capabilities: %+v", scParameters) invokeValidPolicyTest(f, client, namespace, scParameters) }) @@ -164,7 +163,7 @@ var _ = utils.SIGDescribe("Storage Policy Based Volume Provisioning [Feature:vsp ginkgo.By(fmt.Sprintf("Invoking test for VSAN policy objectSpaceReserve: %s, stripeWidth: %s", ObjectSpaceReservationCapabilityVal, StripeWidthCapabilityVal)) scParameters["objectSpaceReserve"] = ObjectSpaceReservationCapabilityVal scParameters[Policy_DiskStripes] = StripeWidthCapabilityVal - e2elog.Logf("Invoking test for VSAN storage capabilities: %+v", scParameters) + framework.Logf("Invoking test for VSAN storage capabilities: %+v", scParameters) err := invokeInvalidPolicyTestNeg(client, namespace, scParameters) framework.ExpectError(err) errorMsg := "invalid option \\\"objectSpaceReserve\\\" for volume plugin kubernetes.io/vsphere-volume" @@ -179,7 +178,7 @@ var _ = utils.SIGDescribe("Storage Policy Based Volume Provisioning [Feature:vsp ginkgo.By(fmt.Sprintf("Invoking test for VSAN policy diskStripes: %s, cacheReservation: %s", DiskStripesCapabilityInvalidVal, CacheReservationCapabilityVal)) scParameters[Policy_DiskStripes] = DiskStripesCapabilityInvalidVal scParameters[Policy_CacheReservation] = CacheReservationCapabilityVal - e2elog.Logf("Invoking test for VSAN storage capabilities: %+v", scParameters) + framework.Logf("Invoking test for VSAN storage capabilities: %+v", scParameters) err := invokeInvalidPolicyTestNeg(client, namespace, scParameters) framework.ExpectError(err) errorMsg := "Invalid value for " + Policy_DiskStripes + "." @@ -193,7 +192,7 @@ var _ = utils.SIGDescribe("Storage Policy Based Volume Provisioning [Feature:vsp ginkgo.It("verify VSAN storage capability with invalid hostFailuresToTolerate value is not honored for dynamically provisioned pvc using storageclass", func() { ginkgo.By(fmt.Sprintf("Invoking test for VSAN policy hostFailuresToTolerate: %s", HostFailuresToTolerateCapabilityInvalidVal)) scParameters[Policy_HostFailuresToTolerate] = HostFailuresToTolerateCapabilityInvalidVal - e2elog.Logf("Invoking test for VSAN storage capabilities: %+v", scParameters) + framework.Logf("Invoking test for VSAN storage capabilities: %+v", scParameters) err := invokeInvalidPolicyTestNeg(client, namespace, scParameters) framework.ExpectError(err) errorMsg := "Invalid value for " + Policy_HostFailuresToTolerate + "." @@ -209,7 +208,7 @@ var _ = utils.SIGDescribe("Storage Policy Based Volume Provisioning [Feature:vsp scParameters[Policy_DiskStripes] = DiskStripesCapabilityVal scParameters[Policy_ObjectSpaceReservation] = ObjectSpaceReservationCapabilityVal scParameters[Datastore] = VmfsDatastore - e2elog.Logf("Invoking test for VSAN storage capabilities: %+v", scParameters) + framework.Logf("Invoking test for VSAN storage capabilities: %+v", scParameters) err := invokeInvalidPolicyTestNeg(client, namespace, scParameters) framework.ExpectError(err) errorMsg := "The specified datastore: \\\"" + VmfsDatastore + "\\\" is not a VSAN datastore. " + @@ -223,7 +222,7 @@ var _ = utils.SIGDescribe("Storage Policy Based Volume Provisioning [Feature:vsp ginkgo.By(fmt.Sprintf("Invoking test for SPBM policy: %s", policyName)) scParameters[SpbmStoragePolicy] = policyName scParameters[DiskFormat] = ThinDisk - e2elog.Logf("Invoking test for SPBM storage policy: %+v", scParameters) + framework.Logf("Invoking test for SPBM storage policy: %+v", scParameters) invokeValidPolicyTest(f, client, namespace, scParameters) }) @@ -231,7 +230,7 @@ var _ = utils.SIGDescribe("Storage Policy Based Volume Provisioning [Feature:vsp scParameters[Policy_DiskStripes] = DiskStripesCapabilityMaxVal scParameters[Policy_ObjectSpaceReservation] = ObjectSpaceReservationCapabilityVal scParameters[Datastore] = VsanDatastore - e2elog.Logf("Invoking test for SPBM storage policy: %+v", scParameters) + framework.Logf("Invoking test for SPBM storage policy: %+v", scParameters) kubernetesClusterName := GetAndExpectStringEnvVar(KubernetesClusterName) invokeStaleDummyVMTestWithStoragePolicy(client, masterNode, namespace, kubernetesClusterName, scParameters) }) @@ -241,7 +240,7 @@ var _ = utils.SIGDescribe("Storage Policy Based Volume Provisioning [Feature:vsp scParameters[SpbmStoragePolicy] = tagPolicy scParameters[Datastore] = VsanDatastore scParameters[DiskFormat] = ThinDisk - e2elog.Logf("Invoking test for SPBM storage policy on a non-compatible datastore: %+v", scParameters) + framework.Logf("Invoking test for SPBM storage policy on a non-compatible datastore: %+v", scParameters) err := invokeInvalidPolicyTestNeg(client, namespace, scParameters) framework.ExpectError(err) errorMsg := "User specified datastore is not compatible with the storagePolicy: \\\"" + tagPolicy + "\\\"" @@ -254,7 +253,7 @@ var _ = utils.SIGDescribe("Storage Policy Based Volume Provisioning [Feature:vsp ginkgo.By(fmt.Sprintf("Invoking test for SPBM policy: %s", BronzeStoragePolicy)) scParameters[SpbmStoragePolicy] = BronzeStoragePolicy scParameters[DiskFormat] = ThinDisk - e2elog.Logf("Invoking test for non-existing SPBM storage policy: %+v", scParameters) + framework.Logf("Invoking test for non-existing SPBM storage policy: %+v", scParameters) err := invokeInvalidPolicyTestNeg(client, namespace, scParameters) framework.ExpectError(err) errorMsg := "no pbm profile found with name: \\\"" + BronzeStoragePolicy + "\\" @@ -269,7 +268,7 @@ var _ = utils.SIGDescribe("Storage Policy Based Volume Provisioning [Feature:vsp gomega.Expect(scParameters[SpbmStoragePolicy]).NotTo(gomega.BeEmpty()) scParameters[Policy_DiskStripes] = DiskStripesCapabilityVal scParameters[DiskFormat] = ThinDisk - e2elog.Logf("Invoking test for SPBM storage policy and VSAN capabilities together: %+v", scParameters) + framework.Logf("Invoking test for SPBM storage policy and VSAN capabilities together: %+v", scParameters) err := invokeInvalidPolicyTestNeg(client, namespace, scParameters) framework.ExpectError(err) errorMsg := "Cannot specify storage policy capabilities along with storage policy name. Please specify only one" diff --git a/test/e2e/storage/vsphere/vsphere_zone_support.go b/test/e2e/storage/vsphere/vsphere_zone_support.go index f5aa994b4cd..1baf2481f30 100644 --- a/test/e2e/storage/vsphere/vsphere_zone_support.go +++ b/test/e2e/storage/vsphere/vsphere_zone_support.go @@ -30,7 +30,6 @@ import ( clientset "k8s.io/client-go/kubernetes" volumeevents "k8s.io/kubernetes/pkg/controller/volume/events" "k8s.io/kubernetes/test/e2e/framework" - e2elog "k8s.io/kubernetes/test/e2e/framework/log" e2epod "k8s.io/kubernetes/test/e2e/framework/pod" "k8s.io/kubernetes/test/e2e/storage/utils" ) @@ -114,7 +113,7 @@ var _ = utils.SIGDescribe("Zone Support", func() { zones = make([]string, 0) nodeList := framework.GetReadySchedulableNodesOrDie(f.ClientSet) if !(len(nodeList.Items) > 0) { - e2elog.Failf("Unable to find ready and schedulable Node") + framework.Failf("Unable to find ready and schedulable Node") } }) @@ -475,7 +474,7 @@ func verifyPVCCreationFails(client clientset.Interface, namespace string, scPara framework.ExpectError(err) eventList, err := client.CoreV1().Events(pvclaim.Namespace).List(metav1.ListOptions{}) - e2elog.Logf("Failure message : %+q", eventList.Items[0].Message) + framework.Logf("Failure message : %+q", eventList.Items[0].Message) return fmt.Errorf("Failure message: %+q", eventList.Items[0].Message) }