mirror of
				https://github.com/optim-enterprises-bv/kubernetes.git
				synced 2025-10-31 18:28:13 +00:00 
			
		
		
		
	
		
			
				
	
	
		
			6514 lines
		
	
	
		
			276 KiB
		
	
	
	
		
			Go
		
	
	
	
	
	
			
		
		
	
	
			6514 lines
		
	
	
		
			276 KiB
		
	
	
	
		
			Go
		
	
	
	
	
	
| /*
 | |
| Copyright 2014 The Kubernetes Authors.
 | |
| 
 | |
| Licensed under the Apache License, Version 2.0 (the "License");
 | |
| you may not use this file except in compliance with the License.
 | |
| You may obtain a copy of the License at
 | |
| 
 | |
|     http://www.apache.org/licenses/LICENSE-2.0
 | |
| 
 | |
| Unless required by applicable law or agreed to in writing, software
 | |
| distributed under the License is distributed on an "AS IS" BASIS,
 | |
| WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
 | |
| See the License for the specific language governing permissions and
 | |
| limitations under the License.
 | |
| */
 | |
| 
 | |
| package validation
 | |
| 
 | |
| import (
 | |
| 	"encoding/json"
 | |
| 	"fmt"
 | |
| 	"math"
 | |
| 	"net"
 | |
| 	"path"
 | |
| 	"path/filepath"
 | |
| 	"reflect"
 | |
| 	"regexp"
 | |
| 	"strings"
 | |
| 	"unicode"
 | |
| 	"unicode/utf8"
 | |
| 
 | |
| 	v1 "k8s.io/api/core/v1"
 | |
| 	apiequality "k8s.io/apimachinery/pkg/api/equality"
 | |
| 	"k8s.io/apimachinery/pkg/api/resource"
 | |
| 	apimachineryvalidation "k8s.io/apimachinery/pkg/api/validation"
 | |
| 	metav1 "k8s.io/apimachinery/pkg/apis/meta/v1"
 | |
| 	unversionedvalidation "k8s.io/apimachinery/pkg/apis/meta/v1/validation"
 | |
| 	"k8s.io/apimachinery/pkg/labels"
 | |
| 	"k8s.io/apimachinery/pkg/util/diff"
 | |
| 	"k8s.io/apimachinery/pkg/util/intstr"
 | |
| 	"k8s.io/apimachinery/pkg/util/sets"
 | |
| 	"k8s.io/apimachinery/pkg/util/validation"
 | |
| 	"k8s.io/apimachinery/pkg/util/validation/field"
 | |
| 	utilfeature "k8s.io/apiserver/pkg/util/feature"
 | |
| 	schedulinghelper "k8s.io/component-helpers/scheduling/corev1"
 | |
| 	apiservice "k8s.io/kubernetes/pkg/api/service"
 | |
| 	"k8s.io/kubernetes/pkg/apis/core"
 | |
| 	"k8s.io/kubernetes/pkg/apis/core/helper"
 | |
| 	podshelper "k8s.io/kubernetes/pkg/apis/core/pods"
 | |
| 	corev1 "k8s.io/kubernetes/pkg/apis/core/v1"
 | |
| 	v1helper "k8s.io/kubernetes/pkg/apis/core/v1/helper"
 | |
| 	"k8s.io/kubernetes/pkg/capabilities"
 | |
| 	"k8s.io/kubernetes/pkg/cluster/ports"
 | |
| 	"k8s.io/kubernetes/pkg/features"
 | |
| 	"k8s.io/kubernetes/pkg/fieldpath"
 | |
| 	"k8s.io/kubernetes/pkg/security/apparmor"
 | |
| 	netutils "k8s.io/utils/net"
 | |
| )
 | |
| 
 | |
| const isNegativeErrorMsg string = apimachineryvalidation.IsNegativeErrorMsg
 | |
| const isInvalidQuotaResource string = `must be a standard resource for quota`
 | |
| const fieldImmutableErrorMsg string = apimachineryvalidation.FieldImmutableErrorMsg
 | |
| const isNotIntegerErrorMsg string = `must be an integer`
 | |
| const isNotPositiveErrorMsg string = `must be greater than zero`
 | |
| 
 | |
| var pdPartitionErrorMsg string = validation.InclusiveRangeError(1, 255)
 | |
| var fileModeErrorMsg = "must be a number between 0 and 0777 (octal), both inclusive"
 | |
| 
 | |
| // BannedOwners is a black list of object that are not allowed to be owners.
 | |
| var BannedOwners = apimachineryvalidation.BannedOwners
 | |
| 
 | |
| var iscsiInitiatorIqnRegex = regexp.MustCompile(`iqn\.\d{4}-\d{2}\.([[:alnum:]-.]+)(:[^,;*&$|\s]+)$`)
 | |
| var iscsiInitiatorEuiRegex = regexp.MustCompile(`^eui.[[:alnum:]]{16}$`)
 | |
| var iscsiInitiatorNaaRegex = regexp.MustCompile(`^naa.[[:alnum:]]{32}$`)
 | |
| 
 | |
| var allowedEphemeralContainerFields = map[string]bool{
 | |
| 	"Name":                     true,
 | |
| 	"Image":                    true,
 | |
| 	"Command":                  true,
 | |
| 	"Args":                     true,
 | |
| 	"WorkingDir":               true,
 | |
| 	"EnvFrom":                  true,
 | |
| 	"Env":                      true,
 | |
| 	"VolumeMounts":             true,
 | |
| 	"TerminationMessagePath":   true,
 | |
| 	"TerminationMessagePolicy": true,
 | |
| 	"ImagePullPolicy":          true,
 | |
| 	"Stdin":                    true,
 | |
| 	"StdinOnce":                true,
 | |
| 	"TTY":                      true,
 | |
| }
 | |
| 
 | |
| // ValidateHasLabel requires that metav1.ObjectMeta has a Label with key and expectedValue
 | |
| func ValidateHasLabel(meta metav1.ObjectMeta, fldPath *field.Path, key, expectedValue string) field.ErrorList {
 | |
| 	allErrs := field.ErrorList{}
 | |
| 	actualValue, found := meta.Labels[key]
 | |
| 	if !found {
 | |
| 		allErrs = append(allErrs, field.Required(fldPath.Child("labels").Key(key),
 | |
| 			fmt.Sprintf("must be '%s'", expectedValue)))
 | |
| 		return allErrs
 | |
| 	}
 | |
| 	if actualValue != expectedValue {
 | |
| 		allErrs = append(allErrs, field.Invalid(fldPath.Child("labels").Key(key), meta.Labels,
 | |
| 			fmt.Sprintf("must be '%s'", expectedValue)))
 | |
| 	}
 | |
| 	return allErrs
 | |
| }
 | |
| 
 | |
| // ValidateAnnotations validates that a set of annotations are correctly defined.
 | |
| func ValidateAnnotations(annotations map[string]string, fldPath *field.Path) field.ErrorList {
 | |
| 	return apimachineryvalidation.ValidateAnnotations(annotations, fldPath)
 | |
| }
 | |
| 
 | |
| func ValidateDNS1123Label(value string, fldPath *field.Path) field.ErrorList {
 | |
| 	allErrs := field.ErrorList{}
 | |
| 	for _, msg := range validation.IsDNS1123Label(value) {
 | |
| 		allErrs = append(allErrs, field.Invalid(fldPath, value, msg))
 | |
| 	}
 | |
| 	return allErrs
 | |
| }
 | |
| 
 | |
| // ValidateQualifiedName validates if name is what Kubernetes calls a "qualified name".
 | |
| func ValidateQualifiedName(value string, fldPath *field.Path) field.ErrorList {
 | |
| 	allErrs := field.ErrorList{}
 | |
| 	for _, msg := range validation.IsQualifiedName(value) {
 | |
| 		allErrs = append(allErrs, field.Invalid(fldPath, value, msg))
 | |
| 	}
 | |
| 	return allErrs
 | |
| }
 | |
| 
 | |
| // ValidateDNS1123Subdomain validates that a name is a proper DNS subdomain.
 | |
| func ValidateDNS1123Subdomain(value string, fldPath *field.Path) field.ErrorList {
 | |
| 	allErrs := field.ErrorList{}
 | |
| 	for _, msg := range validation.IsDNS1123Subdomain(value) {
 | |
| 		allErrs = append(allErrs, field.Invalid(fldPath, value, msg))
 | |
| 	}
 | |
| 	return allErrs
 | |
| }
 | |
| 
 | |
| func ValidatePodSpecificAnnotations(annotations map[string]string, spec *core.PodSpec, fldPath *field.Path, opts PodValidationOptions) field.ErrorList {
 | |
| 	allErrs := field.ErrorList{}
 | |
| 
 | |
| 	if value, isMirror := annotations[core.MirrorPodAnnotationKey]; isMirror {
 | |
| 		if len(spec.NodeName) == 0 {
 | |
| 			allErrs = append(allErrs, field.Invalid(fldPath.Key(core.MirrorPodAnnotationKey), value, "must set spec.nodeName if mirror pod annotation is set"))
 | |
| 		}
 | |
| 	}
 | |
| 
 | |
| 	if annotations[core.TolerationsAnnotationKey] != "" {
 | |
| 		allErrs = append(allErrs, ValidateTolerationsInPodAnnotations(annotations, fldPath)...)
 | |
| 	}
 | |
| 
 | |
| 	if !opts.AllowInvalidPodDeletionCost {
 | |
| 		if _, err := helper.GetDeletionCostFromPodAnnotations(annotations); err != nil {
 | |
| 			allErrs = append(allErrs, field.Invalid(fldPath.Key(core.PodDeletionCost), annotations[core.PodDeletionCost], "must be a 32bit integer"))
 | |
| 		}
 | |
| 	}
 | |
| 
 | |
| 	allErrs = append(allErrs, ValidateSeccompPodAnnotations(annotations, fldPath)...)
 | |
| 	allErrs = append(allErrs, ValidateAppArmorPodAnnotations(annotations, spec, fldPath)...)
 | |
| 
 | |
| 	return allErrs
 | |
| }
 | |
| 
 | |
| // ValidateTolerationsInPodAnnotations tests that the serialized tolerations in Pod.Annotations has valid data
 | |
| func ValidateTolerationsInPodAnnotations(annotations map[string]string, fldPath *field.Path) field.ErrorList {
 | |
| 	allErrs := field.ErrorList{}
 | |
| 
 | |
| 	tolerations, err := helper.GetTolerationsFromPodAnnotations(annotations)
 | |
| 	if err != nil {
 | |
| 		allErrs = append(allErrs, field.Invalid(fldPath, core.TolerationsAnnotationKey, err.Error()))
 | |
| 		return allErrs
 | |
| 	}
 | |
| 
 | |
| 	if len(tolerations) > 0 {
 | |
| 		allErrs = append(allErrs, ValidateTolerations(tolerations, fldPath.Child(core.TolerationsAnnotationKey))...)
 | |
| 	}
 | |
| 
 | |
| 	return allErrs
 | |
| }
 | |
| 
 | |
| func ValidatePodSpecificAnnotationUpdates(newPod, oldPod *core.Pod, fldPath *field.Path, opts PodValidationOptions) field.ErrorList {
 | |
| 	allErrs := field.ErrorList{}
 | |
| 	newAnnotations := newPod.Annotations
 | |
| 	oldAnnotations := oldPod.Annotations
 | |
| 	for k, oldVal := range oldAnnotations {
 | |
| 		if newVal, exists := newAnnotations[k]; exists && newVal == oldVal {
 | |
| 			continue // No change.
 | |
| 		}
 | |
| 		if strings.HasPrefix(k, v1.AppArmorBetaContainerAnnotationKeyPrefix) {
 | |
| 			allErrs = append(allErrs, field.Forbidden(fldPath.Key(k), "may not remove or update AppArmor annotations"))
 | |
| 		}
 | |
| 		if k == core.MirrorPodAnnotationKey {
 | |
| 			allErrs = append(allErrs, field.Forbidden(fldPath.Key(k), "may not remove or update mirror pod annotation"))
 | |
| 		}
 | |
| 	}
 | |
| 	// Check for additions
 | |
| 	for k := range newAnnotations {
 | |
| 		if _, ok := oldAnnotations[k]; ok {
 | |
| 			continue // No change.
 | |
| 		}
 | |
| 		if strings.HasPrefix(k, v1.AppArmorBetaContainerAnnotationKeyPrefix) {
 | |
| 			allErrs = append(allErrs, field.Forbidden(fldPath.Key(k), "may not add AppArmor annotations"))
 | |
| 		}
 | |
| 		if k == core.MirrorPodAnnotationKey {
 | |
| 			allErrs = append(allErrs, field.Forbidden(fldPath.Key(k), "may not add mirror pod annotation"))
 | |
| 		}
 | |
| 	}
 | |
| 	allErrs = append(allErrs, ValidatePodSpecificAnnotations(newAnnotations, &newPod.Spec, fldPath, opts)...)
 | |
| 	return allErrs
 | |
| }
 | |
| 
 | |
| func ValidateEndpointsSpecificAnnotations(annotations map[string]string, fldPath *field.Path) field.ErrorList {
 | |
| 	allErrs := field.ErrorList{}
 | |
| 	return allErrs
 | |
| }
 | |
| 
 | |
| // ValidateNameFunc validates that the provided name is valid for a given resource type.
 | |
| // Not all resources have the same validation rules for names. Prefix is true
 | |
| // if the name will have a value appended to it.  If the name is not valid,
 | |
| // this returns a list of descriptions of individual characteristics of the
 | |
| // value that were not valid.  Otherwise this returns an empty list or nil.
 | |
| type ValidateNameFunc apimachineryvalidation.ValidateNameFunc
 | |
| 
 | |
| // ValidatePodName can be used to check whether the given pod name is valid.
 | |
| // Prefix indicates this name will be used as part of generation, in which case
 | |
| // trailing dashes are allowed.
 | |
| var ValidatePodName = apimachineryvalidation.NameIsDNSSubdomain
 | |
| 
 | |
| // ValidateReplicationControllerName can be used to check whether the given replication
 | |
| // controller name is valid.
 | |
| // Prefix indicates this name will be used as part of generation, in which case
 | |
| // trailing dashes are allowed.
 | |
| var ValidateReplicationControllerName = apimachineryvalidation.NameIsDNSSubdomain
 | |
| 
 | |
| // ValidateServiceName can be used to check whether the given service name is valid.
 | |
| // Prefix indicates this name will be used as part of generation, in which case
 | |
| // trailing dashes are allowed.
 | |
| var ValidateServiceName = apimachineryvalidation.NameIsDNS1035Label
 | |
| 
 | |
| // ValidateNodeName can be used to check whether the given node name is valid.
 | |
| // Prefix indicates this name will be used as part of generation, in which case
 | |
| // trailing dashes are allowed.
 | |
| var ValidateNodeName = apimachineryvalidation.NameIsDNSSubdomain
 | |
| 
 | |
| // ValidateNamespaceName can be used to check whether the given namespace name is valid.
 | |
| // Prefix indicates this name will be used as part of generation, in which case
 | |
| // trailing dashes are allowed.
 | |
| var ValidateNamespaceName = apimachineryvalidation.ValidateNamespaceName
 | |
| 
 | |
| // ValidateLimitRangeName can be used to check whether the given limit range name is valid.
 | |
| // Prefix indicates this name will be used as part of generation, in which case
 | |
| // trailing dashes are allowed.
 | |
| var ValidateLimitRangeName = apimachineryvalidation.NameIsDNSSubdomain
 | |
| 
 | |
| // ValidateResourceQuotaName can be used to check whether the given
 | |
| // resource quota name is valid.
 | |
| // Prefix indicates this name will be used as part of generation, in which case
 | |
| // trailing dashes are allowed.
 | |
| var ValidateResourceQuotaName = apimachineryvalidation.NameIsDNSSubdomain
 | |
| 
 | |
| // ValidateSecretName can be used to check whether the given secret name is valid.
 | |
| // Prefix indicates this name will be used as part of generation, in which case
 | |
| // trailing dashes are allowed.
 | |
| var ValidateSecretName = apimachineryvalidation.NameIsDNSSubdomain
 | |
| 
 | |
| // ValidateServiceAccountName can be used to check whether the given service account name is valid.
 | |
| // Prefix indicates this name will be used as part of generation, in which case
 | |
| // trailing dashes are allowed.
 | |
| var ValidateServiceAccountName = apimachineryvalidation.ValidateServiceAccountName
 | |
| 
 | |
| // ValidateEndpointsName can be used to check whether the given endpoints name is valid.
 | |
| // Prefix indicates this name will be used as part of generation, in which case
 | |
| // trailing dashes are allowed.
 | |
| var ValidateEndpointsName = apimachineryvalidation.NameIsDNSSubdomain
 | |
| 
 | |
| // ValidateClusterName can be used to check whether the given cluster name is valid.
 | |
| var ValidateClusterName = apimachineryvalidation.ValidateClusterName
 | |
| 
 | |
| // ValidateClassName can be used to check whether the given class name is valid.
 | |
| // It is defined here to avoid import cycle between pkg/apis/storage/validation
 | |
| // (where it should be) and this file.
 | |
| var ValidateClassName = apimachineryvalidation.NameIsDNSSubdomain
 | |
| 
 | |
| // ValidatePriorityClassName can be used to check whether the given priority
 | |
| // class name is valid.
 | |
| var ValidatePriorityClassName = apimachineryvalidation.NameIsDNSSubdomain
 | |
| 
 | |
| // ValidateRuntimeClassName can be used to check whether the given RuntimeClass name is valid.
 | |
| // Prefix indicates this name will be used as part of generation, in which case
 | |
| // trailing dashes are allowed.
 | |
| func ValidateRuntimeClassName(name string, fldPath *field.Path) field.ErrorList {
 | |
| 	var allErrs field.ErrorList
 | |
| 	for _, msg := range apimachineryvalidation.NameIsDNSSubdomain(name, false) {
 | |
| 		allErrs = append(allErrs, field.Invalid(fldPath, name, msg))
 | |
| 	}
 | |
| 	return allErrs
 | |
| }
 | |
| 
 | |
| // validateOverhead can be used to check whether the given Overhead is valid.
 | |
| func validateOverhead(overhead core.ResourceList, fldPath *field.Path, opts PodValidationOptions) field.ErrorList {
 | |
| 	// reuse the ResourceRequirements validation logic
 | |
| 	return ValidateResourceRequirements(&core.ResourceRequirements{Limits: overhead}, fldPath, opts)
 | |
| }
 | |
| 
 | |
| // Validates that given value is not negative.
 | |
| func ValidateNonnegativeField(value int64, fldPath *field.Path) field.ErrorList {
 | |
| 	return apimachineryvalidation.ValidateNonnegativeField(value, fldPath)
 | |
| }
 | |
| 
 | |
| // Validates that a Quantity is not negative
 | |
| func ValidateNonnegativeQuantity(value resource.Quantity, fldPath *field.Path) field.ErrorList {
 | |
| 	allErrs := field.ErrorList{}
 | |
| 	if value.Cmp(resource.Quantity{}) < 0 {
 | |
| 		allErrs = append(allErrs, field.Invalid(fldPath, value.String(), isNegativeErrorMsg))
 | |
| 	}
 | |
| 	return allErrs
 | |
| }
 | |
| 
 | |
| // Validates that a Quantity is positive
 | |
| func ValidatePositiveQuantityValue(value resource.Quantity, fldPath *field.Path) field.ErrorList {
 | |
| 	allErrs := field.ErrorList{}
 | |
| 	if value.Cmp(resource.Quantity{}) <= 0 {
 | |
| 		allErrs = append(allErrs, field.Invalid(fldPath, value.String(), isNotPositiveErrorMsg))
 | |
| 	}
 | |
| 	return allErrs
 | |
| }
 | |
| 
 | |
| func ValidateImmutableField(newVal, oldVal interface{}, fldPath *field.Path) field.ErrorList {
 | |
| 	return apimachineryvalidation.ValidateImmutableField(newVal, oldVal, fldPath)
 | |
| }
 | |
| 
 | |
| func ValidateImmutableAnnotation(newVal string, oldVal string, annotation string, fldPath *field.Path) field.ErrorList {
 | |
| 	allErrs := field.ErrorList{}
 | |
| 
 | |
| 	if oldVal != newVal {
 | |
| 		allErrs = append(allErrs, field.Invalid(fldPath.Child("annotations", annotation), newVal, fieldImmutableErrorMsg))
 | |
| 	}
 | |
| 	return allErrs
 | |
| }
 | |
| 
 | |
| // ValidateObjectMeta validates an object's metadata on creation. It expects that name generation has already
 | |
| // been performed.
 | |
| // It doesn't return an error for rootscoped resources with namespace, because namespace should already be cleared before.
 | |
| // TODO: Remove calls to this method scattered in validations of specific resources, e.g., ValidatePodUpdate.
 | |
| func ValidateObjectMeta(meta *metav1.ObjectMeta, requiresNamespace bool, nameFn ValidateNameFunc, fldPath *field.Path) field.ErrorList {
 | |
| 	allErrs := apimachineryvalidation.ValidateObjectMeta(meta, requiresNamespace, apimachineryvalidation.ValidateNameFunc(nameFn), fldPath)
 | |
| 	// run additional checks for the finalizer name
 | |
| 	for i := range meta.Finalizers {
 | |
| 		allErrs = append(allErrs, validateKubeFinalizerName(string(meta.Finalizers[i]), fldPath.Child("finalizers").Index(i))...)
 | |
| 	}
 | |
| 	return allErrs
 | |
| }
 | |
| 
 | |
| // ValidateObjectMetaUpdate validates an object's metadata when updated
 | |
| func ValidateObjectMetaUpdate(newMeta, oldMeta *metav1.ObjectMeta, fldPath *field.Path) field.ErrorList {
 | |
| 	allErrs := apimachineryvalidation.ValidateObjectMetaUpdate(newMeta, oldMeta, fldPath)
 | |
| 	// run additional checks for the finalizer name
 | |
| 	for i := range newMeta.Finalizers {
 | |
| 		allErrs = append(allErrs, validateKubeFinalizerName(string(newMeta.Finalizers[i]), fldPath.Child("finalizers").Index(i))...)
 | |
| 	}
 | |
| 
 | |
| 	return allErrs
 | |
| }
 | |
| 
 | |
| func ValidateVolumes(volumes []core.Volume, podMeta *metav1.ObjectMeta, fldPath *field.Path, opts PodValidationOptions) (map[string]core.VolumeSource, field.ErrorList) {
 | |
| 	allErrs := field.ErrorList{}
 | |
| 
 | |
| 	allNames := sets.String{}
 | |
| 	allCreatedPVCs := sets.String{}
 | |
| 	// Determine which PVCs will be created for this pod. We need
 | |
| 	// the exact name of the pod for this. Without it, this sanity
 | |
| 	// check has to be skipped.
 | |
| 	if podMeta != nil && podMeta.Name != "" {
 | |
| 		for _, vol := range volumes {
 | |
| 			if vol.VolumeSource.Ephemeral != nil {
 | |
| 				allCreatedPVCs.Insert(podMeta.Name + "-" + vol.Name)
 | |
| 			}
 | |
| 		}
 | |
| 	}
 | |
| 	vols := make(map[string]core.VolumeSource)
 | |
| 	for i, vol := range volumes {
 | |
| 		idxPath := fldPath.Index(i)
 | |
| 		namePath := idxPath.Child("name")
 | |
| 		el := validateVolumeSource(&vol.VolumeSource, idxPath, vol.Name, podMeta, opts)
 | |
| 		if len(vol.Name) == 0 {
 | |
| 			el = append(el, field.Required(namePath, ""))
 | |
| 		} else {
 | |
| 			el = append(el, ValidateDNS1123Label(vol.Name, namePath)...)
 | |
| 		}
 | |
| 		if allNames.Has(vol.Name) {
 | |
| 			el = append(el, field.Duplicate(namePath, vol.Name))
 | |
| 		}
 | |
| 		if len(el) == 0 {
 | |
| 			allNames.Insert(vol.Name)
 | |
| 			vols[vol.Name] = vol.VolumeSource
 | |
| 		} else {
 | |
| 			allErrs = append(allErrs, el...)
 | |
| 		}
 | |
| 		// A PersistentVolumeClaimSource should not reference a created PVC. That doesn't
 | |
| 		// make sense.
 | |
| 		if vol.PersistentVolumeClaim != nil && allCreatedPVCs.Has(vol.PersistentVolumeClaim.ClaimName) {
 | |
| 			allErrs = append(allErrs, field.Invalid(idxPath.Child("persistentVolumeClaim").Child("claimName"), vol.PersistentVolumeClaim.ClaimName,
 | |
| 				"must not reference a PVC that gets created for an ephemeral volume"))
 | |
| 		}
 | |
| 	}
 | |
| 
 | |
| 	return vols, allErrs
 | |
| }
 | |
| 
 | |
| func IsMatchedVolume(name string, volumes map[string]core.VolumeSource) bool {
 | |
| 	if _, ok := volumes[name]; ok {
 | |
| 		return true
 | |
| 	}
 | |
| 	return false
 | |
| }
 | |
| 
 | |
| func isMatchedDevice(name string, volumes map[string]core.VolumeSource) (bool, bool) {
 | |
| 	if source, ok := volumes[name]; ok {
 | |
| 		if source.PersistentVolumeClaim != nil {
 | |
| 			return true, true
 | |
| 		}
 | |
| 		return true, false
 | |
| 	}
 | |
| 	return false, false
 | |
| }
 | |
| 
 | |
| func mountNameAlreadyExists(name string, devices map[string]string) bool {
 | |
| 	if _, ok := devices[name]; ok {
 | |
| 		return true
 | |
| 	}
 | |
| 	return false
 | |
| }
 | |
| 
 | |
| func mountPathAlreadyExists(mountPath string, devices map[string]string) bool {
 | |
| 	for _, devPath := range devices {
 | |
| 		if mountPath == devPath {
 | |
| 			return true
 | |
| 		}
 | |
| 	}
 | |
| 
 | |
| 	return false
 | |
| }
 | |
| 
 | |
| func deviceNameAlreadyExists(name string, mounts map[string]string) bool {
 | |
| 	if _, ok := mounts[name]; ok {
 | |
| 		return true
 | |
| 	}
 | |
| 	return false
 | |
| }
 | |
| 
 | |
| func devicePathAlreadyExists(devicePath string, mounts map[string]string) bool {
 | |
| 	for _, mountPath := range mounts {
 | |
| 		if mountPath == devicePath {
 | |
| 			return true
 | |
| 		}
 | |
| 	}
 | |
| 
 | |
| 	return false
 | |
| }
 | |
| 
 | |
| func validateVolumeSource(source *core.VolumeSource, fldPath *field.Path, volName string, podMeta *metav1.ObjectMeta, opts PodValidationOptions) field.ErrorList {
 | |
| 	numVolumes := 0
 | |
| 	allErrs := field.ErrorList{}
 | |
| 	if source.EmptyDir != nil {
 | |
| 		numVolumes++
 | |
| 		if source.EmptyDir.SizeLimit != nil && source.EmptyDir.SizeLimit.Cmp(resource.Quantity{}) < 0 {
 | |
| 			allErrs = append(allErrs, field.Forbidden(fldPath.Child("emptyDir").Child("sizeLimit"), "SizeLimit field must be a valid resource quantity"))
 | |
| 		}
 | |
| 	}
 | |
| 	if source.HostPath != nil {
 | |
| 		if numVolumes > 0 {
 | |
| 			allErrs = append(allErrs, field.Forbidden(fldPath.Child("hostPath"), "may not specify more than 1 volume type"))
 | |
| 		} else {
 | |
| 			numVolumes++
 | |
| 			allErrs = append(allErrs, validateHostPathVolumeSource(source.HostPath, fldPath.Child("hostPath"))...)
 | |
| 		}
 | |
| 	}
 | |
| 	if source.GitRepo != nil {
 | |
| 		if numVolumes > 0 {
 | |
| 			allErrs = append(allErrs, field.Forbidden(fldPath.Child("gitRepo"), "may not specify more than 1 volume type"))
 | |
| 		} else {
 | |
| 			numVolumes++
 | |
| 			allErrs = append(allErrs, validateGitRepoVolumeSource(source.GitRepo, fldPath.Child("gitRepo"))...)
 | |
| 		}
 | |
| 	}
 | |
| 	if source.GCEPersistentDisk != nil {
 | |
| 		if numVolumes > 0 {
 | |
| 			allErrs = append(allErrs, field.Forbidden(fldPath.Child("gcePersistentDisk"), "may not specify more than 1 volume type"))
 | |
| 		} else {
 | |
| 			numVolumes++
 | |
| 			allErrs = append(allErrs, validateGCEPersistentDiskVolumeSource(source.GCEPersistentDisk, fldPath.Child("persistentDisk"))...)
 | |
| 		}
 | |
| 	}
 | |
| 	if source.AWSElasticBlockStore != nil {
 | |
| 		if numVolumes > 0 {
 | |
| 			allErrs = append(allErrs, field.Forbidden(fldPath.Child("awsElasticBlockStore"), "may not specify more than 1 volume type"))
 | |
| 		} else {
 | |
| 			numVolumes++
 | |
| 			allErrs = append(allErrs, validateAWSElasticBlockStoreVolumeSource(source.AWSElasticBlockStore, fldPath.Child("awsElasticBlockStore"))...)
 | |
| 		}
 | |
| 	}
 | |
| 	if source.Secret != nil {
 | |
| 		if numVolumes > 0 {
 | |
| 			allErrs = append(allErrs, field.Forbidden(fldPath.Child("secret"), "may not specify more than 1 volume type"))
 | |
| 		} else {
 | |
| 			numVolumes++
 | |
| 			allErrs = append(allErrs, validateSecretVolumeSource(source.Secret, fldPath.Child("secret"))...)
 | |
| 		}
 | |
| 	}
 | |
| 	if source.NFS != nil {
 | |
| 		if numVolumes > 0 {
 | |
| 			allErrs = append(allErrs, field.Forbidden(fldPath.Child("nfs"), "may not specify more than 1 volume type"))
 | |
| 		} else {
 | |
| 			numVolumes++
 | |
| 			allErrs = append(allErrs, validateNFSVolumeSource(source.NFS, fldPath.Child("nfs"))...)
 | |
| 		}
 | |
| 	}
 | |
| 	if source.ISCSI != nil {
 | |
| 		if numVolumes > 0 {
 | |
| 			allErrs = append(allErrs, field.Forbidden(fldPath.Child("iscsi"), "may not specify more than 1 volume type"))
 | |
| 		} else {
 | |
| 			numVolumes++
 | |
| 			allErrs = append(allErrs, validateISCSIVolumeSource(source.ISCSI, fldPath.Child("iscsi"))...)
 | |
| 		}
 | |
| 		if source.ISCSI.InitiatorName != nil && len(volName+":"+source.ISCSI.TargetPortal) > 64 {
 | |
| 			tooLongErr := "Total length of <volume name>:<iscsi.targetPortal> must be under 64 characters if iscsi.initiatorName is specified."
 | |
| 			allErrs = append(allErrs, field.Invalid(fldPath.Child("name"), volName, tooLongErr))
 | |
| 		}
 | |
| 	}
 | |
| 	if source.Glusterfs != nil {
 | |
| 		if numVolumes > 0 {
 | |
| 			allErrs = append(allErrs, field.Forbidden(fldPath.Child("glusterfs"), "may not specify more than 1 volume type"))
 | |
| 		} else {
 | |
| 			numVolumes++
 | |
| 			allErrs = append(allErrs, validateGlusterfsVolumeSource(source.Glusterfs, fldPath.Child("glusterfs"))...)
 | |
| 		}
 | |
| 	}
 | |
| 	if source.Flocker != nil {
 | |
| 		if numVolumes > 0 {
 | |
| 			allErrs = append(allErrs, field.Forbidden(fldPath.Child("flocker"), "may not specify more than 1 volume type"))
 | |
| 		} else {
 | |
| 			numVolumes++
 | |
| 			allErrs = append(allErrs, validateFlockerVolumeSource(source.Flocker, fldPath.Child("flocker"))...)
 | |
| 		}
 | |
| 	}
 | |
| 	if source.PersistentVolumeClaim != nil {
 | |
| 		if numVolumes > 0 {
 | |
| 			allErrs = append(allErrs, field.Forbidden(fldPath.Child("persistentVolumeClaim"), "may not specify more than 1 volume type"))
 | |
| 		} else {
 | |
| 			numVolumes++
 | |
| 			allErrs = append(allErrs, validatePersistentClaimVolumeSource(source.PersistentVolumeClaim, fldPath.Child("persistentVolumeClaim"))...)
 | |
| 		}
 | |
| 	}
 | |
| 	if source.RBD != nil {
 | |
| 		if numVolumes > 0 {
 | |
| 			allErrs = append(allErrs, field.Forbidden(fldPath.Child("rbd"), "may not specify more than 1 volume type"))
 | |
| 		} else {
 | |
| 			numVolumes++
 | |
| 			allErrs = append(allErrs, validateRBDVolumeSource(source.RBD, fldPath.Child("rbd"))...)
 | |
| 		}
 | |
| 	}
 | |
| 	if source.Cinder != nil {
 | |
| 		if numVolumes > 0 {
 | |
| 			allErrs = append(allErrs, field.Forbidden(fldPath.Child("cinder"), "may not specify more than 1 volume type"))
 | |
| 		} else {
 | |
| 			numVolumes++
 | |
| 			allErrs = append(allErrs, validateCinderVolumeSource(source.Cinder, fldPath.Child("cinder"))...)
 | |
| 		}
 | |
| 	}
 | |
| 	if source.CephFS != nil {
 | |
| 		if numVolumes > 0 {
 | |
| 			allErrs = append(allErrs, field.Forbidden(fldPath.Child("cephFS"), "may not specify more than 1 volume type"))
 | |
| 		} else {
 | |
| 			numVolumes++
 | |
| 			allErrs = append(allErrs, validateCephFSVolumeSource(source.CephFS, fldPath.Child("cephfs"))...)
 | |
| 		}
 | |
| 	}
 | |
| 	if source.Quobyte != nil {
 | |
| 		if numVolumes > 0 {
 | |
| 			allErrs = append(allErrs, field.Forbidden(fldPath.Child("quobyte"), "may not specify more than 1 volume type"))
 | |
| 		} else {
 | |
| 			numVolumes++
 | |
| 			allErrs = append(allErrs, validateQuobyteVolumeSource(source.Quobyte, fldPath.Child("quobyte"))...)
 | |
| 		}
 | |
| 	}
 | |
| 	if source.DownwardAPI != nil {
 | |
| 		if numVolumes > 0 {
 | |
| 			allErrs = append(allErrs, field.Forbidden(fldPath.Child("downwarAPI"), "may not specify more than 1 volume type"))
 | |
| 		} else {
 | |
| 			numVolumes++
 | |
| 			allErrs = append(allErrs, validateDownwardAPIVolumeSource(source.DownwardAPI, fldPath.Child("downwardAPI"), opts)...)
 | |
| 		}
 | |
| 	}
 | |
| 	if source.FC != nil {
 | |
| 		if numVolumes > 0 {
 | |
| 			allErrs = append(allErrs, field.Forbidden(fldPath.Child("fc"), "may not specify more than 1 volume type"))
 | |
| 		} else {
 | |
| 			numVolumes++
 | |
| 			allErrs = append(allErrs, validateFCVolumeSource(source.FC, fldPath.Child("fc"))...)
 | |
| 		}
 | |
| 	}
 | |
| 	if source.FlexVolume != nil {
 | |
| 		if numVolumes > 0 {
 | |
| 			allErrs = append(allErrs, field.Forbidden(fldPath.Child("flexVolume"), "may not specify more than 1 volume type"))
 | |
| 		} else {
 | |
| 			numVolumes++
 | |
| 			allErrs = append(allErrs, validateFlexVolumeSource(source.FlexVolume, fldPath.Child("flexVolume"))...)
 | |
| 		}
 | |
| 	}
 | |
| 	if source.ConfigMap != nil {
 | |
| 		if numVolumes > 0 {
 | |
| 			allErrs = append(allErrs, field.Forbidden(fldPath.Child("configMap"), "may not specify more than 1 volume type"))
 | |
| 		} else {
 | |
| 			numVolumes++
 | |
| 			allErrs = append(allErrs, validateConfigMapVolumeSource(source.ConfigMap, fldPath.Child("configMap"))...)
 | |
| 		}
 | |
| 	}
 | |
| 
 | |
| 	if source.AzureFile != nil {
 | |
| 		if numVolumes > 0 {
 | |
| 			allErrs = append(allErrs, field.Forbidden(fldPath.Child("azureFile"), "may not specify more than 1 volume type"))
 | |
| 		} else {
 | |
| 			numVolumes++
 | |
| 			allErrs = append(allErrs, validateAzureFile(source.AzureFile, fldPath.Child("azureFile"))...)
 | |
| 		}
 | |
| 	}
 | |
| 
 | |
| 	if source.VsphereVolume != nil {
 | |
| 		if numVolumes > 0 {
 | |
| 			allErrs = append(allErrs, field.Forbidden(fldPath.Child("vsphereVolume"), "may not specify more than 1 volume type"))
 | |
| 		} else {
 | |
| 			numVolumes++
 | |
| 			allErrs = append(allErrs, validateVsphereVolumeSource(source.VsphereVolume, fldPath.Child("vsphereVolume"))...)
 | |
| 		}
 | |
| 	}
 | |
| 	if source.PhotonPersistentDisk != nil {
 | |
| 		if numVolumes > 0 {
 | |
| 			allErrs = append(allErrs, field.Forbidden(fldPath.Child("photonPersistentDisk"), "may not specify more than 1 volume type"))
 | |
| 		} else {
 | |
| 			numVolumes++
 | |
| 			allErrs = append(allErrs, validatePhotonPersistentDiskVolumeSource(source.PhotonPersistentDisk, fldPath.Child("photonPersistentDisk"))...)
 | |
| 		}
 | |
| 	}
 | |
| 	if source.PortworxVolume != nil {
 | |
| 		if numVolumes > 0 {
 | |
| 			allErrs = append(allErrs, field.Forbidden(fldPath.Child("portworxVolume"), "may not specify more than 1 volume type"))
 | |
| 		} else {
 | |
| 			numVolumes++
 | |
| 			allErrs = append(allErrs, validatePortworxVolumeSource(source.PortworxVolume, fldPath.Child("portworxVolume"))...)
 | |
| 		}
 | |
| 	}
 | |
| 	if source.AzureDisk != nil {
 | |
| 		if numVolumes > 0 {
 | |
| 			allErrs = append(allErrs, field.Forbidden(fldPath.Child("azureDisk"), "may not specify more than 1 volume type"))
 | |
| 		} else {
 | |
| 			numVolumes++
 | |
| 			allErrs = append(allErrs, validateAzureDisk(source.AzureDisk, fldPath.Child("azureDisk"))...)
 | |
| 		}
 | |
| 	}
 | |
| 	if source.StorageOS != nil {
 | |
| 		if numVolumes > 0 {
 | |
| 			allErrs = append(allErrs, field.Forbidden(fldPath.Child("storageos"), "may not specify more than 1 volume type"))
 | |
| 		} else {
 | |
| 			numVolumes++
 | |
| 			allErrs = append(allErrs, validateStorageOSVolumeSource(source.StorageOS, fldPath.Child("storageos"))...)
 | |
| 		}
 | |
| 	}
 | |
| 	if source.Projected != nil {
 | |
| 		if numVolumes > 0 {
 | |
| 			allErrs = append(allErrs, field.Forbidden(fldPath.Child("projected"), "may not specify more than 1 volume type"))
 | |
| 		} else {
 | |
| 			numVolumes++
 | |
| 			allErrs = append(allErrs, validateProjectedVolumeSource(source.Projected, fldPath.Child("projected"), opts)...)
 | |
| 		}
 | |
| 	}
 | |
| 	if source.ScaleIO != nil {
 | |
| 		if numVolumes > 0 {
 | |
| 			allErrs = append(allErrs, field.Forbidden(fldPath.Child("scaleIO"), "may not specify more than 1 volume type"))
 | |
| 		} else {
 | |
| 			numVolumes++
 | |
| 			allErrs = append(allErrs, validateScaleIOVolumeSource(source.ScaleIO, fldPath.Child("scaleIO"))...)
 | |
| 		}
 | |
| 	}
 | |
| 	if source.CSI != nil {
 | |
| 		if numVolumes > 0 {
 | |
| 			allErrs = append(allErrs, field.Forbidden(fldPath.Child("csi"), "may not specify more than 1 volume type"))
 | |
| 		} else {
 | |
| 			numVolumes++
 | |
| 			allErrs = append(allErrs, validateCSIVolumeSource(source.CSI, fldPath.Child("csi"))...)
 | |
| 		}
 | |
| 	}
 | |
| 	if source.Ephemeral != nil {
 | |
| 		if numVolumes > 0 {
 | |
| 			allErrs = append(allErrs, field.Forbidden(fldPath.Child("ephemeral"), "may not specify more than 1 volume type"))
 | |
| 		} else {
 | |
| 			numVolumes++
 | |
| 			allErrs = append(allErrs, validateEphemeralVolumeSource(source.Ephemeral, fldPath.Child("ephemeral"))...)
 | |
| 			// Check the expected name for the PVC. This gets skipped if information is missing,
 | |
| 			// because that already gets flagged as a problem elsewhere. For example,
 | |
| 			// ValidateObjectMeta as called by validatePodMetadataAndSpec checks that the name is set.
 | |
| 			if podMeta != nil && podMeta.Name != "" && volName != "" {
 | |
| 				pvcName := podMeta.Name + "-" + volName
 | |
| 				for _, msg := range ValidatePersistentVolumeName(pvcName, false) {
 | |
| 					allErrs = append(allErrs, field.Invalid(fldPath.Child("name"), volName, fmt.Sprintf("PVC name %q: %v", pvcName, msg)))
 | |
| 				}
 | |
| 			}
 | |
| 		}
 | |
| 	}
 | |
| 
 | |
| 	if numVolumes == 0 {
 | |
| 		allErrs = append(allErrs, field.Required(fldPath, "must specify a volume type"))
 | |
| 	}
 | |
| 
 | |
| 	return allErrs
 | |
| }
 | |
| 
 | |
| func validateHostPathVolumeSource(hostPath *core.HostPathVolumeSource, fldPath *field.Path) field.ErrorList {
 | |
| 	allErrs := field.ErrorList{}
 | |
| 	if len(hostPath.Path) == 0 {
 | |
| 		allErrs = append(allErrs, field.Required(fldPath.Child("path"), ""))
 | |
| 		return allErrs
 | |
| 	}
 | |
| 
 | |
| 	allErrs = append(allErrs, validatePathNoBacksteps(hostPath.Path, fldPath.Child("path"))...)
 | |
| 	allErrs = append(allErrs, validateHostPathType(hostPath.Type, fldPath.Child("type"))...)
 | |
| 	return allErrs
 | |
| }
 | |
| 
 | |
| func validateGitRepoVolumeSource(gitRepo *core.GitRepoVolumeSource, fldPath *field.Path) field.ErrorList {
 | |
| 	allErrs := field.ErrorList{}
 | |
| 	if len(gitRepo.Repository) == 0 {
 | |
| 		allErrs = append(allErrs, field.Required(fldPath.Child("repository"), ""))
 | |
| 	}
 | |
| 
 | |
| 	pathErrs := validateLocalDescendingPath(gitRepo.Directory, fldPath.Child("directory"))
 | |
| 	allErrs = append(allErrs, pathErrs...)
 | |
| 	return allErrs
 | |
| }
 | |
| 
 | |
| func validateISCSIVolumeSource(iscsi *core.ISCSIVolumeSource, fldPath *field.Path) field.ErrorList {
 | |
| 	allErrs := field.ErrorList{}
 | |
| 	if len(iscsi.TargetPortal) == 0 {
 | |
| 		allErrs = append(allErrs, field.Required(fldPath.Child("targetPortal"), ""))
 | |
| 	}
 | |
| 	if len(iscsi.IQN) == 0 {
 | |
| 		allErrs = append(allErrs, field.Required(fldPath.Child("iqn"), ""))
 | |
| 	} else {
 | |
| 		if !strings.HasPrefix(iscsi.IQN, "iqn") && !strings.HasPrefix(iscsi.IQN, "eui") && !strings.HasPrefix(iscsi.IQN, "naa") {
 | |
| 			allErrs = append(allErrs, field.Invalid(fldPath.Child("iqn"), iscsi.IQN, "must be valid format starting with iqn, eui, or naa"))
 | |
| 		} else if strings.HasPrefix(iscsi.IQN, "iqn") && !iscsiInitiatorIqnRegex.MatchString(iscsi.IQN) {
 | |
| 			allErrs = append(allErrs, field.Invalid(fldPath.Child("iqn"), iscsi.IQN, "must be valid format"))
 | |
| 		} else if strings.HasPrefix(iscsi.IQN, "eui") && !iscsiInitiatorEuiRegex.MatchString(iscsi.IQN) {
 | |
| 			allErrs = append(allErrs, field.Invalid(fldPath.Child("iqn"), iscsi.IQN, "must be valid format"))
 | |
| 		} else if strings.HasPrefix(iscsi.IQN, "naa") && !iscsiInitiatorNaaRegex.MatchString(iscsi.IQN) {
 | |
| 			allErrs = append(allErrs, field.Invalid(fldPath.Child("iqn"), iscsi.IQN, "must be valid format"))
 | |
| 		}
 | |
| 	}
 | |
| 	if iscsi.Lun < 0 || iscsi.Lun > 255 {
 | |
| 		allErrs = append(allErrs, field.Invalid(fldPath.Child("lun"), iscsi.Lun, validation.InclusiveRangeError(0, 255)))
 | |
| 	}
 | |
| 	if (iscsi.DiscoveryCHAPAuth || iscsi.SessionCHAPAuth) && iscsi.SecretRef == nil {
 | |
| 		allErrs = append(allErrs, field.Required(fldPath.Child("secretRef"), ""))
 | |
| 	}
 | |
| 	if iscsi.InitiatorName != nil {
 | |
| 		initiator := *iscsi.InitiatorName
 | |
| 		if !strings.HasPrefix(initiator, "iqn") && !strings.HasPrefix(initiator, "eui") && !strings.HasPrefix(initiator, "naa") {
 | |
| 			allErrs = append(allErrs, field.Invalid(fldPath.Child("initiatorname"), initiator, "must be valid format starting with iqn, eui, or naa"))
 | |
| 		}
 | |
| 		if strings.HasPrefix(initiator, "iqn") && !iscsiInitiatorIqnRegex.MatchString(initiator) {
 | |
| 			allErrs = append(allErrs, field.Invalid(fldPath.Child("initiatorname"), initiator, "must be valid format"))
 | |
| 		} else if strings.HasPrefix(initiator, "eui") && !iscsiInitiatorEuiRegex.MatchString(initiator) {
 | |
| 			allErrs = append(allErrs, field.Invalid(fldPath.Child("initiatorname"), initiator, "must be valid format"))
 | |
| 		} else if strings.HasPrefix(initiator, "naa") && !iscsiInitiatorNaaRegex.MatchString(initiator) {
 | |
| 			allErrs = append(allErrs, field.Invalid(fldPath.Child("initiatorname"), initiator, "must be valid format"))
 | |
| 		}
 | |
| 	}
 | |
| 	return allErrs
 | |
| }
 | |
| 
 | |
| func validateISCSIPersistentVolumeSource(iscsi *core.ISCSIPersistentVolumeSource, pvName string, fldPath *field.Path) field.ErrorList {
 | |
| 	allErrs := field.ErrorList{}
 | |
| 	if len(iscsi.TargetPortal) == 0 {
 | |
| 		allErrs = append(allErrs, field.Required(fldPath.Child("targetPortal"), ""))
 | |
| 	}
 | |
| 	if iscsi.InitiatorName != nil && len(pvName+":"+iscsi.TargetPortal) > 64 {
 | |
| 		tooLongErr := "Total length of <volume name>:<iscsi.targetPortal> must be under 64 characters if iscsi.initiatorName is specified."
 | |
| 		allErrs = append(allErrs, field.Invalid(fldPath.Child("targetportal"), iscsi.TargetPortal, tooLongErr))
 | |
| 	}
 | |
| 	if len(iscsi.IQN) == 0 {
 | |
| 		allErrs = append(allErrs, field.Required(fldPath.Child("iqn"), ""))
 | |
| 	} else {
 | |
| 		if !strings.HasPrefix(iscsi.IQN, "iqn") && !strings.HasPrefix(iscsi.IQN, "eui") && !strings.HasPrefix(iscsi.IQN, "naa") {
 | |
| 			allErrs = append(allErrs, field.Invalid(fldPath.Child("iqn"), iscsi.IQN, "must be valid format"))
 | |
| 		} else if strings.HasPrefix(iscsi.IQN, "iqn") && !iscsiInitiatorIqnRegex.MatchString(iscsi.IQN) {
 | |
| 			allErrs = append(allErrs, field.Invalid(fldPath.Child("iqn"), iscsi.IQN, "must be valid format"))
 | |
| 		} else if strings.HasPrefix(iscsi.IQN, "eui") && !iscsiInitiatorEuiRegex.MatchString(iscsi.IQN) {
 | |
| 			allErrs = append(allErrs, field.Invalid(fldPath.Child("iqn"), iscsi.IQN, "must be valid format"))
 | |
| 		} else if strings.HasPrefix(iscsi.IQN, "naa") && !iscsiInitiatorNaaRegex.MatchString(iscsi.IQN) {
 | |
| 			allErrs = append(allErrs, field.Invalid(fldPath.Child("iqn"), iscsi.IQN, "must be valid format"))
 | |
| 		}
 | |
| 	}
 | |
| 	if iscsi.Lun < 0 || iscsi.Lun > 255 {
 | |
| 		allErrs = append(allErrs, field.Invalid(fldPath.Child("lun"), iscsi.Lun, validation.InclusiveRangeError(0, 255)))
 | |
| 	}
 | |
| 	if (iscsi.DiscoveryCHAPAuth || iscsi.SessionCHAPAuth) && iscsi.SecretRef == nil {
 | |
| 		allErrs = append(allErrs, field.Required(fldPath.Child("secretRef"), ""))
 | |
| 	}
 | |
| 	if iscsi.SecretRef != nil {
 | |
| 		if len(iscsi.SecretRef.Name) == 0 {
 | |
| 			allErrs = append(allErrs, field.Required(fldPath.Child("secretRef", "name"), ""))
 | |
| 		}
 | |
| 	}
 | |
| 	if iscsi.InitiatorName != nil {
 | |
| 		initiator := *iscsi.InitiatorName
 | |
| 		if !strings.HasPrefix(initiator, "iqn") && !strings.HasPrefix(initiator, "eui") && !strings.HasPrefix(initiator, "naa") {
 | |
| 			allErrs = append(allErrs, field.Invalid(fldPath.Child("initiatorname"), initiator, "must be valid format"))
 | |
| 		}
 | |
| 		if strings.HasPrefix(initiator, "iqn") && !iscsiInitiatorIqnRegex.MatchString(initiator) {
 | |
| 			allErrs = append(allErrs, field.Invalid(fldPath.Child("initiatorname"), initiator, "must be valid format"))
 | |
| 		} else if strings.HasPrefix(initiator, "eui") && !iscsiInitiatorEuiRegex.MatchString(initiator) {
 | |
| 			allErrs = append(allErrs, field.Invalid(fldPath.Child("initiatorname"), initiator, "must be valid format"))
 | |
| 		} else if strings.HasPrefix(initiator, "naa") && !iscsiInitiatorNaaRegex.MatchString(initiator) {
 | |
| 			allErrs = append(allErrs, field.Invalid(fldPath.Child("initiatorname"), initiator, "must be valid format"))
 | |
| 		}
 | |
| 	}
 | |
| 	return allErrs
 | |
| }
 | |
| 
 | |
| func validateFCVolumeSource(fc *core.FCVolumeSource, fldPath *field.Path) field.ErrorList {
 | |
| 	allErrs := field.ErrorList{}
 | |
| 	if len(fc.TargetWWNs) < 1 && len(fc.WWIDs) < 1 {
 | |
| 		allErrs = append(allErrs, field.Required(fldPath.Child("targetWWNs"), "must specify either targetWWNs or wwids, but not both"))
 | |
| 	}
 | |
| 
 | |
| 	if len(fc.TargetWWNs) != 0 && len(fc.WWIDs) != 0 {
 | |
| 		allErrs = append(allErrs, field.Invalid(fldPath.Child("targetWWNs"), fc.TargetWWNs, "targetWWNs and wwids can not be specified simultaneously"))
 | |
| 	}
 | |
| 
 | |
| 	if len(fc.TargetWWNs) != 0 {
 | |
| 		if fc.Lun == nil {
 | |
| 			allErrs = append(allErrs, field.Required(fldPath.Child("lun"), "lun is required if targetWWNs is specified"))
 | |
| 		} else {
 | |
| 			if *fc.Lun < 0 || *fc.Lun > 255 {
 | |
| 				allErrs = append(allErrs, field.Invalid(fldPath.Child("lun"), fc.Lun, validation.InclusiveRangeError(0, 255)))
 | |
| 			}
 | |
| 		}
 | |
| 	}
 | |
| 	return allErrs
 | |
| }
 | |
| 
 | |
| func validateGCEPersistentDiskVolumeSource(pd *core.GCEPersistentDiskVolumeSource, fldPath *field.Path) field.ErrorList {
 | |
| 	allErrs := field.ErrorList{}
 | |
| 	if len(pd.PDName) == 0 {
 | |
| 		allErrs = append(allErrs, field.Required(fldPath.Child("pdName"), ""))
 | |
| 	}
 | |
| 	if pd.Partition < 0 || pd.Partition > 255 {
 | |
| 		allErrs = append(allErrs, field.Invalid(fldPath.Child("partition"), pd.Partition, pdPartitionErrorMsg))
 | |
| 	}
 | |
| 	return allErrs
 | |
| }
 | |
| 
 | |
| func validateAWSElasticBlockStoreVolumeSource(PD *core.AWSElasticBlockStoreVolumeSource, fldPath *field.Path) field.ErrorList {
 | |
| 	allErrs := field.ErrorList{}
 | |
| 	if len(PD.VolumeID) == 0 {
 | |
| 		allErrs = append(allErrs, field.Required(fldPath.Child("volumeID"), ""))
 | |
| 	}
 | |
| 	if PD.Partition < 0 || PD.Partition > 255 {
 | |
| 		allErrs = append(allErrs, field.Invalid(fldPath.Child("partition"), PD.Partition, pdPartitionErrorMsg))
 | |
| 	}
 | |
| 	return allErrs
 | |
| }
 | |
| 
 | |
| func validateSecretVolumeSource(secretSource *core.SecretVolumeSource, fldPath *field.Path) field.ErrorList {
 | |
| 	allErrs := field.ErrorList{}
 | |
| 	if len(secretSource.SecretName) == 0 {
 | |
| 		allErrs = append(allErrs, field.Required(fldPath.Child("secretName"), ""))
 | |
| 	}
 | |
| 
 | |
| 	secretMode := secretSource.DefaultMode
 | |
| 	if secretMode != nil && (*secretMode > 0777 || *secretMode < 0) {
 | |
| 		allErrs = append(allErrs, field.Invalid(fldPath.Child("defaultMode"), *secretMode, fileModeErrorMsg))
 | |
| 	}
 | |
| 
 | |
| 	itemsPath := fldPath.Child("items")
 | |
| 	for i, kp := range secretSource.Items {
 | |
| 		itemPath := itemsPath.Index(i)
 | |
| 		allErrs = append(allErrs, validateKeyToPath(&kp, itemPath)...)
 | |
| 	}
 | |
| 	return allErrs
 | |
| }
 | |
| 
 | |
| func validateConfigMapVolumeSource(configMapSource *core.ConfigMapVolumeSource, fldPath *field.Path) field.ErrorList {
 | |
| 	allErrs := field.ErrorList{}
 | |
| 	if len(configMapSource.Name) == 0 {
 | |
| 		allErrs = append(allErrs, field.Required(fldPath.Child("name"), ""))
 | |
| 	}
 | |
| 
 | |
| 	configMapMode := configMapSource.DefaultMode
 | |
| 	if configMapMode != nil && (*configMapMode > 0777 || *configMapMode < 0) {
 | |
| 		allErrs = append(allErrs, field.Invalid(fldPath.Child("defaultMode"), *configMapMode, fileModeErrorMsg))
 | |
| 	}
 | |
| 
 | |
| 	itemsPath := fldPath.Child("items")
 | |
| 	for i, kp := range configMapSource.Items {
 | |
| 		itemPath := itemsPath.Index(i)
 | |
| 		allErrs = append(allErrs, validateKeyToPath(&kp, itemPath)...)
 | |
| 	}
 | |
| 	return allErrs
 | |
| }
 | |
| 
 | |
| func validateKeyToPath(kp *core.KeyToPath, fldPath *field.Path) field.ErrorList {
 | |
| 	allErrs := field.ErrorList{}
 | |
| 	if len(kp.Key) == 0 {
 | |
| 		allErrs = append(allErrs, field.Required(fldPath.Child("key"), ""))
 | |
| 	}
 | |
| 	if len(kp.Path) == 0 {
 | |
| 		allErrs = append(allErrs, field.Required(fldPath.Child("path"), ""))
 | |
| 	}
 | |
| 	allErrs = append(allErrs, validateLocalNonReservedPath(kp.Path, fldPath.Child("path"))...)
 | |
| 	if kp.Mode != nil && (*kp.Mode > 0777 || *kp.Mode < 0) {
 | |
| 		allErrs = append(allErrs, field.Invalid(fldPath.Child("mode"), *kp.Mode, fileModeErrorMsg))
 | |
| 	}
 | |
| 
 | |
| 	return allErrs
 | |
| }
 | |
| 
 | |
| func validatePersistentClaimVolumeSource(claim *core.PersistentVolumeClaimVolumeSource, fldPath *field.Path) field.ErrorList {
 | |
| 	allErrs := field.ErrorList{}
 | |
| 	if len(claim.ClaimName) == 0 {
 | |
| 		allErrs = append(allErrs, field.Required(fldPath.Child("claimName"), ""))
 | |
| 	}
 | |
| 	return allErrs
 | |
| }
 | |
| 
 | |
| func validateNFSVolumeSource(nfs *core.NFSVolumeSource, fldPath *field.Path) field.ErrorList {
 | |
| 	allErrs := field.ErrorList{}
 | |
| 	if len(nfs.Server) == 0 {
 | |
| 		allErrs = append(allErrs, field.Required(fldPath.Child("server"), ""))
 | |
| 	}
 | |
| 	if len(nfs.Path) == 0 {
 | |
| 		allErrs = append(allErrs, field.Required(fldPath.Child("path"), ""))
 | |
| 	}
 | |
| 	if !path.IsAbs(nfs.Path) {
 | |
| 		allErrs = append(allErrs, field.Invalid(fldPath.Child("path"), nfs.Path, "must be an absolute path"))
 | |
| 	}
 | |
| 	return allErrs
 | |
| }
 | |
| 
 | |
| func validateQuobyteVolumeSource(quobyte *core.QuobyteVolumeSource, fldPath *field.Path) field.ErrorList {
 | |
| 	allErrs := field.ErrorList{}
 | |
| 	if len(quobyte.Registry) == 0 {
 | |
| 		allErrs = append(allErrs, field.Required(fldPath.Child("registry"), "must be a host:port pair or multiple pairs separated by commas"))
 | |
| 	} else if len(quobyte.Tenant) >= 65 {
 | |
| 		allErrs = append(allErrs, field.Required(fldPath.Child("tenant"), "must be a UUID and may not exceed a length of 64 characters"))
 | |
| 	} else {
 | |
| 		for _, hostPortPair := range strings.Split(quobyte.Registry, ",") {
 | |
| 			if _, _, err := net.SplitHostPort(hostPortPair); err != nil {
 | |
| 				allErrs = append(allErrs, field.Invalid(fldPath.Child("registry"), quobyte.Registry, "must be a host:port pair or multiple pairs separated by commas"))
 | |
| 			}
 | |
| 		}
 | |
| 	}
 | |
| 
 | |
| 	if len(quobyte.Volume) == 0 {
 | |
| 		allErrs = append(allErrs, field.Required(fldPath.Child("volume"), ""))
 | |
| 	}
 | |
| 	return allErrs
 | |
| }
 | |
| 
 | |
| func validateGlusterfsVolumeSource(glusterfs *core.GlusterfsVolumeSource, fldPath *field.Path) field.ErrorList {
 | |
| 	allErrs := field.ErrorList{}
 | |
| 	if len(glusterfs.EndpointsName) == 0 {
 | |
| 		allErrs = append(allErrs, field.Required(fldPath.Child("endpoints"), ""))
 | |
| 	}
 | |
| 	if len(glusterfs.Path) == 0 {
 | |
| 		allErrs = append(allErrs, field.Required(fldPath.Child("path"), ""))
 | |
| 	}
 | |
| 	return allErrs
 | |
| }
 | |
| func validateGlusterfsPersistentVolumeSource(glusterfs *core.GlusterfsPersistentVolumeSource, fldPath *field.Path) field.ErrorList {
 | |
| 	allErrs := field.ErrorList{}
 | |
| 	if len(glusterfs.EndpointsName) == 0 {
 | |
| 		allErrs = append(allErrs, field.Required(fldPath.Child("endpoints"), ""))
 | |
| 	}
 | |
| 	if len(glusterfs.Path) == 0 {
 | |
| 		allErrs = append(allErrs, field.Required(fldPath.Child("path"), ""))
 | |
| 	}
 | |
| 	if glusterfs.EndpointsNamespace != nil {
 | |
| 		endpointNs := glusterfs.EndpointsNamespace
 | |
| 		if *endpointNs == "" {
 | |
| 			allErrs = append(allErrs, field.Invalid(fldPath.Child("endpointsNamespace"), *endpointNs, "if the endpointnamespace is set, it must be a valid namespace name"))
 | |
| 		} else {
 | |
| 			for _, msg := range ValidateNamespaceName(*endpointNs, false) {
 | |
| 				allErrs = append(allErrs, field.Invalid(fldPath.Child("endpointsNamespace"), *endpointNs, msg))
 | |
| 			}
 | |
| 		}
 | |
| 	}
 | |
| 	return allErrs
 | |
| }
 | |
| 
 | |
| func validateFlockerVolumeSource(flocker *core.FlockerVolumeSource, fldPath *field.Path) field.ErrorList {
 | |
| 	allErrs := field.ErrorList{}
 | |
| 	if len(flocker.DatasetName) == 0 && len(flocker.DatasetUUID) == 0 {
 | |
| 		//TODO: consider adding a RequiredOneOf() error for this and similar cases
 | |
| 		allErrs = append(allErrs, field.Required(fldPath, "one of datasetName and datasetUUID is required"))
 | |
| 	}
 | |
| 	if len(flocker.DatasetName) != 0 && len(flocker.DatasetUUID) != 0 {
 | |
| 		allErrs = append(allErrs, field.Invalid(fldPath, "resource", "datasetName and datasetUUID can not be specified simultaneously"))
 | |
| 	}
 | |
| 	if strings.Contains(flocker.DatasetName, "/") {
 | |
| 		allErrs = append(allErrs, field.Invalid(fldPath.Child("datasetName"), flocker.DatasetName, "must not contain '/'"))
 | |
| 	}
 | |
| 	return allErrs
 | |
| }
 | |
| 
 | |
| var validVolumeDownwardAPIFieldPathExpressions = sets.NewString(
 | |
| 	"metadata.name",
 | |
| 	"metadata.namespace",
 | |
| 	"metadata.labels",
 | |
| 	"metadata.annotations",
 | |
| 	"metadata.uid")
 | |
| 
 | |
| func validateDownwardAPIVolumeFile(file *core.DownwardAPIVolumeFile, fldPath *field.Path, opts PodValidationOptions) field.ErrorList {
 | |
| 	allErrs := field.ErrorList{}
 | |
| 	if len(file.Path) == 0 {
 | |
| 		allErrs = append(allErrs, field.Required(fldPath.Child("path"), ""))
 | |
| 	}
 | |
| 	allErrs = append(allErrs, validateLocalNonReservedPath(file.Path, fldPath.Child("path"))...)
 | |
| 	if file.FieldRef != nil {
 | |
| 		allErrs = append(allErrs, validateObjectFieldSelector(file.FieldRef, &validVolumeDownwardAPIFieldPathExpressions, fldPath.Child("fieldRef"))...)
 | |
| 		if file.ResourceFieldRef != nil {
 | |
| 			allErrs = append(allErrs, field.Invalid(fldPath, "resource", "fieldRef and resourceFieldRef can not be specified simultaneously"))
 | |
| 		}
 | |
| 	} else if file.ResourceFieldRef != nil {
 | |
| 		localValidContainerResourceFieldPathPrefixes := validContainerResourceFieldPathPrefixes
 | |
| 		if opts.AllowDownwardAPIHugePages {
 | |
| 			localValidContainerResourceFieldPathPrefixes = validContainerResourceFieldPathPrefixesWithDownwardAPIHugePages
 | |
| 		}
 | |
| 		allErrs = append(allErrs, validateContainerResourceFieldSelector(file.ResourceFieldRef, &validContainerResourceFieldPathExpressions, &localValidContainerResourceFieldPathPrefixes, fldPath.Child("resourceFieldRef"), true)...)
 | |
| 	} else {
 | |
| 		allErrs = append(allErrs, field.Required(fldPath, "one of fieldRef and resourceFieldRef is required"))
 | |
| 	}
 | |
| 	if file.Mode != nil && (*file.Mode > 0777 || *file.Mode < 0) {
 | |
| 		allErrs = append(allErrs, field.Invalid(fldPath.Child("mode"), *file.Mode, fileModeErrorMsg))
 | |
| 	}
 | |
| 
 | |
| 	return allErrs
 | |
| }
 | |
| 
 | |
| func validateDownwardAPIVolumeSource(downwardAPIVolume *core.DownwardAPIVolumeSource, fldPath *field.Path, opts PodValidationOptions) field.ErrorList {
 | |
| 	allErrs := field.ErrorList{}
 | |
| 
 | |
| 	downwardAPIMode := downwardAPIVolume.DefaultMode
 | |
| 	if downwardAPIMode != nil && (*downwardAPIMode > 0777 || *downwardAPIMode < 0) {
 | |
| 		allErrs = append(allErrs, field.Invalid(fldPath.Child("defaultMode"), *downwardAPIMode, fileModeErrorMsg))
 | |
| 	}
 | |
| 
 | |
| 	for _, file := range downwardAPIVolume.Items {
 | |
| 		allErrs = append(allErrs, validateDownwardAPIVolumeFile(&file, fldPath, opts)...)
 | |
| 	}
 | |
| 	return allErrs
 | |
| }
 | |
| 
 | |
| func validateProjectionSources(projection *core.ProjectedVolumeSource, projectionMode *int32, fldPath *field.Path, opts PodValidationOptions) field.ErrorList {
 | |
| 	allErrs := field.ErrorList{}
 | |
| 	allPaths := sets.String{}
 | |
| 
 | |
| 	for i, source := range projection.Sources {
 | |
| 		numSources := 0
 | |
| 		srcPath := fldPath.Child("sources").Index(i)
 | |
| 		if projPath := srcPath.Child("secret"); source.Secret != nil {
 | |
| 			numSources++
 | |
| 			if len(source.Secret.Name) == 0 {
 | |
| 				allErrs = append(allErrs, field.Required(projPath.Child("name"), ""))
 | |
| 			}
 | |
| 			itemsPath := projPath.Child("items")
 | |
| 			for i, kp := range source.Secret.Items {
 | |
| 				itemPath := itemsPath.Index(i)
 | |
| 				allErrs = append(allErrs, validateKeyToPath(&kp, itemPath)...)
 | |
| 				if len(kp.Path) > 0 {
 | |
| 					curPath := kp.Path
 | |
| 					if !allPaths.Has(curPath) {
 | |
| 						allPaths.Insert(curPath)
 | |
| 					} else {
 | |
| 						allErrs = append(allErrs, field.Invalid(fldPath, source.Secret.Name, "conflicting duplicate paths"))
 | |
| 					}
 | |
| 				}
 | |
| 			}
 | |
| 		}
 | |
| 		if projPath := srcPath.Child("configMap"); source.ConfigMap != nil {
 | |
| 			numSources++
 | |
| 			if len(source.ConfigMap.Name) == 0 {
 | |
| 				allErrs = append(allErrs, field.Required(projPath.Child("name"), ""))
 | |
| 			}
 | |
| 			itemsPath := projPath.Child("items")
 | |
| 			for i, kp := range source.ConfigMap.Items {
 | |
| 				itemPath := itemsPath.Index(i)
 | |
| 				allErrs = append(allErrs, validateKeyToPath(&kp, itemPath)...)
 | |
| 				if len(kp.Path) > 0 {
 | |
| 					curPath := kp.Path
 | |
| 					if !allPaths.Has(curPath) {
 | |
| 						allPaths.Insert(curPath)
 | |
| 					} else {
 | |
| 						allErrs = append(allErrs, field.Invalid(fldPath, source.ConfigMap.Name, "conflicting duplicate paths"))
 | |
| 					}
 | |
| 				}
 | |
| 			}
 | |
| 		}
 | |
| 		if projPath := srcPath.Child("downwardAPI"); source.DownwardAPI != nil {
 | |
| 			numSources++
 | |
| 			for _, file := range source.DownwardAPI.Items {
 | |
| 				allErrs = append(allErrs, validateDownwardAPIVolumeFile(&file, projPath, opts)...)
 | |
| 				if len(file.Path) > 0 {
 | |
| 					curPath := file.Path
 | |
| 					if !allPaths.Has(curPath) {
 | |
| 						allPaths.Insert(curPath)
 | |
| 					} else {
 | |
| 						allErrs = append(allErrs, field.Invalid(fldPath, curPath, "conflicting duplicate paths"))
 | |
| 					}
 | |
| 				}
 | |
| 			}
 | |
| 		}
 | |
| 		if projPath := srcPath.Child("serviceAccountToken"); source.ServiceAccountToken != nil {
 | |
| 			numSources++
 | |
| 			if source.ServiceAccountToken.ExpirationSeconds < 10*60 {
 | |
| 				allErrs = append(allErrs, field.Invalid(projPath.Child("expirationSeconds"), source.ServiceAccountToken.ExpirationSeconds, "may not specify a duration less than 10 minutes"))
 | |
| 			}
 | |
| 			if source.ServiceAccountToken.ExpirationSeconds > 1<<32 {
 | |
| 				allErrs = append(allErrs, field.Invalid(projPath.Child("expirationSeconds"), source.ServiceAccountToken.ExpirationSeconds, "may not specify a duration larger than 2^32 seconds"))
 | |
| 			}
 | |
| 			if source.ServiceAccountToken.Path == "" {
 | |
| 				allErrs = append(allErrs, field.Required(fldPath.Child("path"), ""))
 | |
| 			}
 | |
| 		}
 | |
| 		if numSources > 1 {
 | |
| 			allErrs = append(allErrs, field.Forbidden(srcPath, "may not specify more than 1 volume type"))
 | |
| 		}
 | |
| 	}
 | |
| 	return allErrs
 | |
| }
 | |
| 
 | |
| func validateProjectedVolumeSource(projection *core.ProjectedVolumeSource, fldPath *field.Path, opts PodValidationOptions) field.ErrorList {
 | |
| 	allErrs := field.ErrorList{}
 | |
| 
 | |
| 	projectionMode := projection.DefaultMode
 | |
| 	if projectionMode != nil && (*projectionMode > 0777 || *projectionMode < 0) {
 | |
| 		allErrs = append(allErrs, field.Invalid(fldPath.Child("defaultMode"), *projectionMode, fileModeErrorMsg))
 | |
| 	}
 | |
| 
 | |
| 	allErrs = append(allErrs, validateProjectionSources(projection, projectionMode, fldPath, opts)...)
 | |
| 	return allErrs
 | |
| }
 | |
| 
 | |
| var supportedHostPathTypes = sets.NewString(
 | |
| 	string(core.HostPathUnset),
 | |
| 	string(core.HostPathDirectoryOrCreate),
 | |
| 	string(core.HostPathDirectory),
 | |
| 	string(core.HostPathFileOrCreate),
 | |
| 	string(core.HostPathFile),
 | |
| 	string(core.HostPathSocket),
 | |
| 	string(core.HostPathCharDev),
 | |
| 	string(core.HostPathBlockDev))
 | |
| 
 | |
| func validateHostPathType(hostPathType *core.HostPathType, fldPath *field.Path) field.ErrorList {
 | |
| 	allErrs := field.ErrorList{}
 | |
| 
 | |
| 	if hostPathType != nil && !supportedHostPathTypes.Has(string(*hostPathType)) {
 | |
| 		allErrs = append(allErrs, field.NotSupported(fldPath, hostPathType, supportedHostPathTypes.List()))
 | |
| 	}
 | |
| 
 | |
| 	return allErrs
 | |
| }
 | |
| 
 | |
| // This validate will make sure targetPath:
 | |
| // 1. is not abs path
 | |
| // 2. does not have any element which is ".."
 | |
| func validateLocalDescendingPath(targetPath string, fldPath *field.Path) field.ErrorList {
 | |
| 	allErrs := field.ErrorList{}
 | |
| 	if path.IsAbs(targetPath) {
 | |
| 		allErrs = append(allErrs, field.Invalid(fldPath, targetPath, "must be a relative path"))
 | |
| 	}
 | |
| 
 | |
| 	allErrs = append(allErrs, validatePathNoBacksteps(targetPath, fldPath)...)
 | |
| 
 | |
| 	return allErrs
 | |
| }
 | |
| 
 | |
| // validatePathNoBacksteps makes sure the targetPath does not have any `..` path elements when split
 | |
| //
 | |
| // This assumes the OS of the apiserver and the nodes are the same. The same check should be done
 | |
| // on the node to ensure there are no backsteps.
 | |
| func validatePathNoBacksteps(targetPath string, fldPath *field.Path) field.ErrorList {
 | |
| 	allErrs := field.ErrorList{}
 | |
| 	parts := strings.Split(filepath.ToSlash(targetPath), "/")
 | |
| 	for _, item := range parts {
 | |
| 		if item == ".." {
 | |
| 			allErrs = append(allErrs, field.Invalid(fldPath, targetPath, "must not contain '..'"))
 | |
| 			break // even for `../../..`, one error is sufficient to make the point
 | |
| 		}
 | |
| 	}
 | |
| 	return allErrs
 | |
| }
 | |
| 
 | |
| // validateMountPropagation verifies that MountPropagation field is valid and
 | |
| // allowed for given container.
 | |
| func validateMountPropagation(mountPropagation *core.MountPropagationMode, container *core.Container, fldPath *field.Path) field.ErrorList {
 | |
| 	allErrs := field.ErrorList{}
 | |
| 
 | |
| 	if mountPropagation == nil {
 | |
| 		return allErrs
 | |
| 	}
 | |
| 
 | |
| 	supportedMountPropagations := sets.NewString(string(core.MountPropagationBidirectional), string(core.MountPropagationHostToContainer), string(core.MountPropagationNone))
 | |
| 	if !supportedMountPropagations.Has(string(*mountPropagation)) {
 | |
| 		allErrs = append(allErrs, field.NotSupported(fldPath, *mountPropagation, supportedMountPropagations.List()))
 | |
| 	}
 | |
| 
 | |
| 	if container == nil {
 | |
| 		// The container is not available yet.
 | |
| 		// Stop validation now, Pod validation will refuse final
 | |
| 		// Pods with Bidirectional propagation in non-privileged containers.
 | |
| 		return allErrs
 | |
| 	}
 | |
| 
 | |
| 	privileged := container.SecurityContext != nil && container.SecurityContext.Privileged != nil && *container.SecurityContext.Privileged
 | |
| 	if *mountPropagation == core.MountPropagationBidirectional && !privileged {
 | |
| 		allErrs = append(allErrs, field.Forbidden(fldPath, "Bidirectional mount propagation is available only to privileged containers"))
 | |
| 	}
 | |
| 	return allErrs
 | |
| }
 | |
| 
 | |
| // This validate will make sure targetPath:
 | |
| // 1. is not abs path
 | |
| // 2. does not contain any '..' elements
 | |
| // 3. does not start with '..'
 | |
| func validateLocalNonReservedPath(targetPath string, fldPath *field.Path) field.ErrorList {
 | |
| 	allErrs := field.ErrorList{}
 | |
| 	allErrs = append(allErrs, validateLocalDescendingPath(targetPath, fldPath)...)
 | |
| 	// Don't report this error if the check for .. elements already caught it.
 | |
| 	if strings.HasPrefix(targetPath, "..") && !strings.HasPrefix(targetPath, "../") {
 | |
| 		allErrs = append(allErrs, field.Invalid(fldPath, targetPath, "must not start with '..'"))
 | |
| 	}
 | |
| 	return allErrs
 | |
| }
 | |
| 
 | |
| func validateRBDVolumeSource(rbd *core.RBDVolumeSource, fldPath *field.Path) field.ErrorList {
 | |
| 	allErrs := field.ErrorList{}
 | |
| 	if len(rbd.CephMonitors) == 0 {
 | |
| 		allErrs = append(allErrs, field.Required(fldPath.Child("monitors"), ""))
 | |
| 	}
 | |
| 	if len(rbd.RBDImage) == 0 {
 | |
| 		allErrs = append(allErrs, field.Required(fldPath.Child("image"), ""))
 | |
| 	}
 | |
| 	return allErrs
 | |
| }
 | |
| 
 | |
| func validateRBDPersistentVolumeSource(rbd *core.RBDPersistentVolumeSource, fldPath *field.Path) field.ErrorList {
 | |
| 	allErrs := field.ErrorList{}
 | |
| 	if len(rbd.CephMonitors) == 0 {
 | |
| 		allErrs = append(allErrs, field.Required(fldPath.Child("monitors"), ""))
 | |
| 	}
 | |
| 	if len(rbd.RBDImage) == 0 {
 | |
| 		allErrs = append(allErrs, field.Required(fldPath.Child("image"), ""))
 | |
| 	}
 | |
| 	return allErrs
 | |
| }
 | |
| 
 | |
| func validateCinderVolumeSource(cd *core.CinderVolumeSource, fldPath *field.Path) field.ErrorList {
 | |
| 	allErrs := field.ErrorList{}
 | |
| 	if len(cd.VolumeID) == 0 {
 | |
| 		allErrs = append(allErrs, field.Required(fldPath.Child("volumeID"), ""))
 | |
| 	}
 | |
| 	if cd.SecretRef != nil {
 | |
| 		if len(cd.SecretRef.Name) == 0 {
 | |
| 			allErrs = append(allErrs, field.Required(fldPath.Child("secretRef", "name"), ""))
 | |
| 		}
 | |
| 	}
 | |
| 	return allErrs
 | |
| }
 | |
| 
 | |
| func validateCinderPersistentVolumeSource(cd *core.CinderPersistentVolumeSource, fldPath *field.Path) field.ErrorList {
 | |
| 	allErrs := field.ErrorList{}
 | |
| 	if len(cd.VolumeID) == 0 {
 | |
| 		allErrs = append(allErrs, field.Required(fldPath.Child("volumeID"), ""))
 | |
| 	}
 | |
| 	if cd.SecretRef != nil {
 | |
| 		if len(cd.SecretRef.Name) == 0 {
 | |
| 			allErrs = append(allErrs, field.Required(fldPath.Child("secretRef", "name"), ""))
 | |
| 		}
 | |
| 		if len(cd.SecretRef.Namespace) == 0 {
 | |
| 			allErrs = append(allErrs, field.Required(fldPath.Child("secretRef", "namespace"), ""))
 | |
| 		}
 | |
| 	}
 | |
| 	return allErrs
 | |
| }
 | |
| 
 | |
| func validateCephFSVolumeSource(cephfs *core.CephFSVolumeSource, fldPath *field.Path) field.ErrorList {
 | |
| 	allErrs := field.ErrorList{}
 | |
| 	if len(cephfs.Monitors) == 0 {
 | |
| 		allErrs = append(allErrs, field.Required(fldPath.Child("monitors"), ""))
 | |
| 	}
 | |
| 	return allErrs
 | |
| }
 | |
| 
 | |
| func validateCephFSPersistentVolumeSource(cephfs *core.CephFSPersistentVolumeSource, fldPath *field.Path) field.ErrorList {
 | |
| 	allErrs := field.ErrorList{}
 | |
| 	if len(cephfs.Monitors) == 0 {
 | |
| 		allErrs = append(allErrs, field.Required(fldPath.Child("monitors"), ""))
 | |
| 	}
 | |
| 	return allErrs
 | |
| }
 | |
| 
 | |
| func validateFlexVolumeSource(fv *core.FlexVolumeSource, fldPath *field.Path) field.ErrorList {
 | |
| 	allErrs := field.ErrorList{}
 | |
| 	if len(fv.Driver) == 0 {
 | |
| 		allErrs = append(allErrs, field.Required(fldPath.Child("driver"), ""))
 | |
| 	}
 | |
| 
 | |
| 	// Make sure user-specified options don't use kubernetes namespaces
 | |
| 	for k := range fv.Options {
 | |
| 		namespace := k
 | |
| 		if parts := strings.SplitN(k, "/", 2); len(parts) == 2 {
 | |
| 			namespace = parts[0]
 | |
| 		}
 | |
| 		normalized := "." + strings.ToLower(namespace)
 | |
| 		if strings.HasSuffix(normalized, ".kubernetes.io") || strings.HasSuffix(normalized, ".k8s.io") {
 | |
| 			allErrs = append(allErrs, field.Invalid(fldPath.Child("options").Key(k), k, "kubernetes.io and k8s.io namespaces are reserved"))
 | |
| 		}
 | |
| 	}
 | |
| 
 | |
| 	return allErrs
 | |
| }
 | |
| 
 | |
| func validateFlexPersistentVolumeSource(fv *core.FlexPersistentVolumeSource, fldPath *field.Path) field.ErrorList {
 | |
| 	allErrs := field.ErrorList{}
 | |
| 	if len(fv.Driver) == 0 {
 | |
| 		allErrs = append(allErrs, field.Required(fldPath.Child("driver"), ""))
 | |
| 	}
 | |
| 
 | |
| 	// Make sure user-specified options don't use kubernetes namespaces
 | |
| 	for k := range fv.Options {
 | |
| 		namespace := k
 | |
| 		if parts := strings.SplitN(k, "/", 2); len(parts) == 2 {
 | |
| 			namespace = parts[0]
 | |
| 		}
 | |
| 		normalized := "." + strings.ToLower(namespace)
 | |
| 		if strings.HasSuffix(normalized, ".kubernetes.io") || strings.HasSuffix(normalized, ".k8s.io") {
 | |
| 			allErrs = append(allErrs, field.Invalid(fldPath.Child("options").Key(k), k, "kubernetes.io and k8s.io namespaces are reserved"))
 | |
| 		}
 | |
| 	}
 | |
| 
 | |
| 	return allErrs
 | |
| }
 | |
| 
 | |
| func validateAzureFile(azure *core.AzureFileVolumeSource, fldPath *field.Path) field.ErrorList {
 | |
| 	allErrs := field.ErrorList{}
 | |
| 	if azure.SecretName == "" {
 | |
| 		allErrs = append(allErrs, field.Required(fldPath.Child("secretName"), ""))
 | |
| 	}
 | |
| 	if azure.ShareName == "" {
 | |
| 		allErrs = append(allErrs, field.Required(fldPath.Child("shareName"), ""))
 | |
| 	}
 | |
| 	return allErrs
 | |
| }
 | |
| 
 | |
| func validateAzureFilePV(azure *core.AzureFilePersistentVolumeSource, fldPath *field.Path) field.ErrorList {
 | |
| 	allErrs := field.ErrorList{}
 | |
| 	if azure.SecretName == "" {
 | |
| 		allErrs = append(allErrs, field.Required(fldPath.Child("secretName"), ""))
 | |
| 	}
 | |
| 	if azure.ShareName == "" {
 | |
| 		allErrs = append(allErrs, field.Required(fldPath.Child("shareName"), ""))
 | |
| 	}
 | |
| 	if azure.SecretNamespace != nil {
 | |
| 		if len(*azure.SecretNamespace) == 0 {
 | |
| 			allErrs = append(allErrs, field.Required(fldPath.Child("secretNamespace"), ""))
 | |
| 		}
 | |
| 	}
 | |
| 	return allErrs
 | |
| }
 | |
| 
 | |
| func validateAzureDisk(azure *core.AzureDiskVolumeSource, fldPath *field.Path) field.ErrorList {
 | |
| 	var supportedCachingModes = sets.NewString(string(core.AzureDataDiskCachingNone), string(core.AzureDataDiskCachingReadOnly), string(core.AzureDataDiskCachingReadWrite))
 | |
| 	var supportedDiskKinds = sets.NewString(string(core.AzureSharedBlobDisk), string(core.AzureDedicatedBlobDisk), string(core.AzureManagedDisk))
 | |
| 
 | |
| 	diskURISupportedManaged := []string{"/subscriptions/{sub-id}/resourcegroups/{group-name}/providers/microsoft.compute/disks/{disk-id}"}
 | |
| 	diskURISupportedblob := []string{"https://{account-name}.blob.core.windows.net/{container-name}/{disk-name}.vhd"}
 | |
| 
 | |
| 	allErrs := field.ErrorList{}
 | |
| 	if azure.DiskName == "" {
 | |
| 		allErrs = append(allErrs, field.Required(fldPath.Child("diskName"), ""))
 | |
| 	}
 | |
| 
 | |
| 	if azure.DataDiskURI == "" {
 | |
| 		allErrs = append(allErrs, field.Required(fldPath.Child("diskURI"), ""))
 | |
| 	}
 | |
| 
 | |
| 	if azure.CachingMode != nil && !supportedCachingModes.Has(string(*azure.CachingMode)) {
 | |
| 		allErrs = append(allErrs, field.NotSupported(fldPath.Child("cachingMode"), *azure.CachingMode, supportedCachingModes.List()))
 | |
| 	}
 | |
| 
 | |
| 	if azure.Kind != nil && !supportedDiskKinds.Has(string(*azure.Kind)) {
 | |
| 		allErrs = append(allErrs, field.NotSupported(fldPath.Child("kind"), *azure.Kind, supportedDiskKinds.List()))
 | |
| 	}
 | |
| 
 | |
| 	// validate that DiskUri is the correct format
 | |
| 	if azure.Kind != nil && *azure.Kind == core.AzureManagedDisk && strings.Index(azure.DataDiskURI, "/subscriptions/") != 0 {
 | |
| 		allErrs = append(allErrs, field.NotSupported(fldPath.Child("diskURI"), azure.DataDiskURI, diskURISupportedManaged))
 | |
| 	}
 | |
| 
 | |
| 	if azure.Kind != nil && *azure.Kind != core.AzureManagedDisk && strings.Index(azure.DataDiskURI, "https://") != 0 {
 | |
| 		allErrs = append(allErrs, field.NotSupported(fldPath.Child("diskURI"), azure.DataDiskURI, diskURISupportedblob))
 | |
| 	}
 | |
| 
 | |
| 	return allErrs
 | |
| }
 | |
| 
 | |
| func validateVsphereVolumeSource(cd *core.VsphereVirtualDiskVolumeSource, fldPath *field.Path) field.ErrorList {
 | |
| 	allErrs := field.ErrorList{}
 | |
| 	if len(cd.VolumePath) == 0 {
 | |
| 		allErrs = append(allErrs, field.Required(fldPath.Child("volumePath"), ""))
 | |
| 	}
 | |
| 	return allErrs
 | |
| }
 | |
| 
 | |
| func validatePhotonPersistentDiskVolumeSource(cd *core.PhotonPersistentDiskVolumeSource, fldPath *field.Path) field.ErrorList {
 | |
| 	allErrs := field.ErrorList{}
 | |
| 	if len(cd.PdID) == 0 {
 | |
| 		allErrs = append(allErrs, field.Required(fldPath.Child("pdID"), ""))
 | |
| 	}
 | |
| 	return allErrs
 | |
| }
 | |
| 
 | |
| func validatePortworxVolumeSource(pwx *core.PortworxVolumeSource, fldPath *field.Path) field.ErrorList {
 | |
| 	allErrs := field.ErrorList{}
 | |
| 	if len(pwx.VolumeID) == 0 {
 | |
| 		allErrs = append(allErrs, field.Required(fldPath.Child("volumeID"), ""))
 | |
| 	}
 | |
| 	return allErrs
 | |
| }
 | |
| 
 | |
| func validateScaleIOVolumeSource(sio *core.ScaleIOVolumeSource, fldPath *field.Path) field.ErrorList {
 | |
| 	allErrs := field.ErrorList{}
 | |
| 	if sio.Gateway == "" {
 | |
| 		allErrs = append(allErrs, field.Required(fldPath.Child("gateway"), ""))
 | |
| 	}
 | |
| 	if sio.System == "" {
 | |
| 		allErrs = append(allErrs, field.Required(fldPath.Child("system"), ""))
 | |
| 	}
 | |
| 	if sio.VolumeName == "" {
 | |
| 		allErrs = append(allErrs, field.Required(fldPath.Child("volumeName"), ""))
 | |
| 	}
 | |
| 	return allErrs
 | |
| }
 | |
| 
 | |
| func validateScaleIOPersistentVolumeSource(sio *core.ScaleIOPersistentVolumeSource, fldPath *field.Path) field.ErrorList {
 | |
| 	allErrs := field.ErrorList{}
 | |
| 	if sio.Gateway == "" {
 | |
| 		allErrs = append(allErrs, field.Required(fldPath.Child("gateway"), ""))
 | |
| 	}
 | |
| 	if sio.System == "" {
 | |
| 		allErrs = append(allErrs, field.Required(fldPath.Child("system"), ""))
 | |
| 	}
 | |
| 	if sio.VolumeName == "" {
 | |
| 		allErrs = append(allErrs, field.Required(fldPath.Child("volumeName"), ""))
 | |
| 	}
 | |
| 	return allErrs
 | |
| }
 | |
| 
 | |
| func validateLocalVolumeSource(ls *core.LocalVolumeSource, fldPath *field.Path) field.ErrorList {
 | |
| 	allErrs := field.ErrorList{}
 | |
| 	if ls.Path == "" {
 | |
| 		allErrs = append(allErrs, field.Required(fldPath.Child("path"), ""))
 | |
| 		return allErrs
 | |
| 	}
 | |
| 
 | |
| 	allErrs = append(allErrs, validatePathNoBacksteps(ls.Path, fldPath.Child("path"))...)
 | |
| 	return allErrs
 | |
| }
 | |
| 
 | |
| func validateStorageOSVolumeSource(storageos *core.StorageOSVolumeSource, fldPath *field.Path) field.ErrorList {
 | |
| 	allErrs := field.ErrorList{}
 | |
| 	if len(storageos.VolumeName) == 0 {
 | |
| 		allErrs = append(allErrs, field.Required(fldPath.Child("volumeName"), ""))
 | |
| 	} else {
 | |
| 		allErrs = append(allErrs, ValidateDNS1123Label(storageos.VolumeName, fldPath.Child("volumeName"))...)
 | |
| 	}
 | |
| 	if len(storageos.VolumeNamespace) > 0 {
 | |
| 		allErrs = append(allErrs, ValidateDNS1123Label(storageos.VolumeNamespace, fldPath.Child("volumeNamespace"))...)
 | |
| 	}
 | |
| 	if storageos.SecretRef != nil {
 | |
| 		if len(storageos.SecretRef.Name) == 0 {
 | |
| 			allErrs = append(allErrs, field.Required(fldPath.Child("secretRef", "name"), ""))
 | |
| 		}
 | |
| 	}
 | |
| 	return allErrs
 | |
| }
 | |
| 
 | |
| func validateStorageOSPersistentVolumeSource(storageos *core.StorageOSPersistentVolumeSource, fldPath *field.Path) field.ErrorList {
 | |
| 	allErrs := field.ErrorList{}
 | |
| 	if len(storageos.VolumeName) == 0 {
 | |
| 		allErrs = append(allErrs, field.Required(fldPath.Child("volumeName"), ""))
 | |
| 	} else {
 | |
| 		allErrs = append(allErrs, ValidateDNS1123Label(storageos.VolumeName, fldPath.Child("volumeName"))...)
 | |
| 	}
 | |
| 	if len(storageos.VolumeNamespace) > 0 {
 | |
| 		allErrs = append(allErrs, ValidateDNS1123Label(storageos.VolumeNamespace, fldPath.Child("volumeNamespace"))...)
 | |
| 	}
 | |
| 	if storageos.SecretRef != nil {
 | |
| 		if len(storageos.SecretRef.Name) == 0 {
 | |
| 			allErrs = append(allErrs, field.Required(fldPath.Child("secretRef", "name"), ""))
 | |
| 		}
 | |
| 		if len(storageos.SecretRef.Namespace) == 0 {
 | |
| 			allErrs = append(allErrs, field.Required(fldPath.Child("secretRef", "namespace"), ""))
 | |
| 		}
 | |
| 	}
 | |
| 	return allErrs
 | |
| }
 | |
| 
 | |
| func ValidateCSIDriverName(driverName string, fldPath *field.Path) field.ErrorList {
 | |
| 	allErrs := field.ErrorList{}
 | |
| 
 | |
| 	if len(driverName) == 0 {
 | |
| 		allErrs = append(allErrs, field.Required(fldPath, ""))
 | |
| 	}
 | |
| 
 | |
| 	if len(driverName) > 63 {
 | |
| 		allErrs = append(allErrs, field.TooLong(fldPath, driverName, 63))
 | |
| 	}
 | |
| 
 | |
| 	for _, msg := range validation.IsDNS1123Subdomain(strings.ToLower(driverName)) {
 | |
| 		allErrs = append(allErrs, field.Invalid(fldPath, driverName, msg))
 | |
| 	}
 | |
| 
 | |
| 	return allErrs
 | |
| }
 | |
| 
 | |
| func validateCSIPersistentVolumeSource(csi *core.CSIPersistentVolumeSource, fldPath *field.Path) field.ErrorList {
 | |
| 	allErrs := field.ErrorList{}
 | |
| 
 | |
| 	allErrs = append(allErrs, ValidateCSIDriverName(csi.Driver, fldPath.Child("driver"))...)
 | |
| 
 | |
| 	if len(csi.VolumeHandle) == 0 {
 | |
| 		allErrs = append(allErrs, field.Required(fldPath.Child("volumeHandle"), ""))
 | |
| 	}
 | |
| 
 | |
| 	if csi.ControllerPublishSecretRef != nil {
 | |
| 		if len(csi.ControllerPublishSecretRef.Name) == 0 {
 | |
| 			allErrs = append(allErrs, field.Required(fldPath.Child("controllerPublishSecretRef", "name"), ""))
 | |
| 		} else {
 | |
| 			allErrs = append(allErrs, ValidateDNS1123Label(csi.ControllerPublishSecretRef.Name, fldPath.Child("name"))...)
 | |
| 		}
 | |
| 		if len(csi.ControllerPublishSecretRef.Namespace) == 0 {
 | |
| 			allErrs = append(allErrs, field.Required(fldPath.Child("controllerPublishSecretRef", "namespace"), ""))
 | |
| 		} else {
 | |
| 			allErrs = append(allErrs, ValidateDNS1123Label(csi.ControllerPublishSecretRef.Namespace, fldPath.Child("namespace"))...)
 | |
| 		}
 | |
| 	}
 | |
| 
 | |
| 	if csi.ControllerExpandSecretRef != nil {
 | |
| 		if len(csi.ControllerExpandSecretRef.Name) == 0 {
 | |
| 			allErrs = append(allErrs, field.Required(fldPath.Child("controllerExpandSecretRef", "name"), ""))
 | |
| 		} else {
 | |
| 			allErrs = append(allErrs, ValidateDNS1123Label(csi.ControllerExpandSecretRef.Name, fldPath.Child("name"))...)
 | |
| 		}
 | |
| 		if len(csi.ControllerExpandSecretRef.Namespace) == 0 {
 | |
| 			allErrs = append(allErrs, field.Required(fldPath.Child("controllerExpandSecretRef", "namespace"), ""))
 | |
| 		} else {
 | |
| 			allErrs = append(allErrs, ValidateDNS1123Label(csi.ControllerExpandSecretRef.Namespace, fldPath.Child("namespace"))...)
 | |
| 		}
 | |
| 	}
 | |
| 
 | |
| 	if csi.NodePublishSecretRef != nil {
 | |
| 		if len(csi.NodePublishSecretRef.Name) == 0 {
 | |
| 			allErrs = append(allErrs, field.Required(fldPath.Child("nodePublishSecretRef ", "name"), ""))
 | |
| 		} else {
 | |
| 			allErrs = append(allErrs, ValidateDNS1123Label(csi.NodePublishSecretRef.Name, fldPath.Child("name"))...)
 | |
| 		}
 | |
| 		if len(csi.NodePublishSecretRef.Namespace) == 0 {
 | |
| 			allErrs = append(allErrs, field.Required(fldPath.Child("nodePublishSecretRef ", "namespace"), ""))
 | |
| 		} else {
 | |
| 			allErrs = append(allErrs, ValidateDNS1123Label(csi.NodePublishSecretRef.Namespace, fldPath.Child("namespace"))...)
 | |
| 		}
 | |
| 	}
 | |
| 
 | |
| 	return allErrs
 | |
| }
 | |
| 
 | |
| func validateCSIVolumeSource(csi *core.CSIVolumeSource, fldPath *field.Path) field.ErrorList {
 | |
| 	allErrs := field.ErrorList{}
 | |
| 	allErrs = append(allErrs, ValidateCSIDriverName(csi.Driver, fldPath.Child("driver"))...)
 | |
| 
 | |
| 	if csi.NodePublishSecretRef != nil {
 | |
| 		if len(csi.NodePublishSecretRef.Name) == 0 {
 | |
| 			allErrs = append(allErrs, field.Required(fldPath.Child("nodePublishSecretRef ", "name"), ""))
 | |
| 		} else {
 | |
| 			for _, msg := range ValidateSecretName(csi.NodePublishSecretRef.Name, false) {
 | |
| 				allErrs = append(allErrs, field.Invalid(fldPath.Child("name"), csi.NodePublishSecretRef.Name, msg))
 | |
| 			}
 | |
| 		}
 | |
| 	}
 | |
| 
 | |
| 	return allErrs
 | |
| }
 | |
| 
 | |
| func validateEphemeralVolumeSource(ephemeral *core.EphemeralVolumeSource, fldPath *field.Path) field.ErrorList {
 | |
| 	allErrs := field.ErrorList{}
 | |
| 	if ephemeral.VolumeClaimTemplate == nil {
 | |
| 		allErrs = append(allErrs, field.Required(fldPath.Child("volumeClaimTemplate"), ""))
 | |
| 	} else {
 | |
| 		allErrs = append(allErrs, ValidatePersistentVolumeClaimTemplate(ephemeral.VolumeClaimTemplate, fldPath.Child("volumeClaimTemplate"))...)
 | |
| 	}
 | |
| 	return allErrs
 | |
| }
 | |
| 
 | |
| // ValidatePersistentVolumeClaimTemplate verifies that the embedded object meta and spec are valid.
 | |
| // Checking of the object data is very minimal because only labels and annotations are used.
 | |
| func ValidatePersistentVolumeClaimTemplate(claimTemplate *core.PersistentVolumeClaimTemplate, fldPath *field.Path) field.ErrorList {
 | |
| 	allErrs := validatePersistentVolumeClaimTemplateObjectMeta(&claimTemplate.ObjectMeta, fldPath.Child("metadata"))
 | |
| 	allErrs = append(allErrs, ValidatePersistentVolumeClaimSpec(&claimTemplate.Spec, fldPath.Child("spec"))...)
 | |
| 	return allErrs
 | |
| }
 | |
| 
 | |
| func validatePersistentVolumeClaimTemplateObjectMeta(objMeta *metav1.ObjectMeta, fldPath *field.Path) field.ErrorList {
 | |
| 	allErrs := apimachineryvalidation.ValidateAnnotations(objMeta.Annotations, fldPath.Child("annotations"))
 | |
| 	allErrs = append(allErrs, unversionedvalidation.ValidateLabels(objMeta.Labels, fldPath.Child("labels"))...)
 | |
| 	// All other fields are not supported and thus must not be set
 | |
| 	// to avoid confusion.  We could reject individual fields,
 | |
| 	// but then adding a new one to ObjectMeta wouldn't be checked
 | |
| 	// unless this code gets updated. Instead, we ensure that
 | |
| 	// only allowed fields are set via reflection.
 | |
| 	allErrs = append(allErrs, validateFieldAllowList(*objMeta, allowedPVCTemplateObjectMetaFields, "cannot be set for an ephemeral volume", fldPath)...)
 | |
| 	return allErrs
 | |
| }
 | |
| 
 | |
| var allowedPVCTemplateObjectMetaFields = map[string]bool{
 | |
| 	"Annotations": true,
 | |
| 	"Labels":      true,
 | |
| }
 | |
| 
 | |
| // ValidatePersistentVolumeName checks that a name is appropriate for a
 | |
| // PersistentVolumeName object.
 | |
| var ValidatePersistentVolumeName = apimachineryvalidation.NameIsDNSSubdomain
 | |
| 
 | |
| var supportedAccessModes = sets.NewString(string(core.ReadWriteOnce), string(core.ReadOnlyMany), string(core.ReadWriteMany))
 | |
| 
 | |
| var supportedReclaimPolicy = sets.NewString(string(core.PersistentVolumeReclaimDelete), string(core.PersistentVolumeReclaimRecycle), string(core.PersistentVolumeReclaimRetain))
 | |
| 
 | |
| var supportedVolumeModes = sets.NewString(string(core.PersistentVolumeBlock), string(core.PersistentVolumeFilesystem))
 | |
| 
 | |
| func ValidatePersistentVolumeSpec(pvSpec *core.PersistentVolumeSpec, pvName string, validateInlinePersistentVolumeSpec bool, fldPath *field.Path) field.ErrorList {
 | |
| 	allErrs := field.ErrorList{}
 | |
| 
 | |
| 	if validateInlinePersistentVolumeSpec {
 | |
| 		if pvSpec.ClaimRef != nil {
 | |
| 			allErrs = append(allErrs, field.Forbidden(fldPath.Child("claimRef"), "may not be specified in the context of inline volumes"))
 | |
| 		}
 | |
| 		if len(pvSpec.Capacity) != 0 {
 | |
| 			allErrs = append(allErrs, field.Forbidden(fldPath.Child("capacity"), "may not be specified in the context of inline volumes"))
 | |
| 		}
 | |
| 		if pvSpec.CSI == nil {
 | |
| 			allErrs = append(allErrs, field.Required(fldPath.Child("csi"), "has to be specified in the context of inline volumes"))
 | |
| 		}
 | |
| 	}
 | |
| 
 | |
| 	if len(pvSpec.AccessModes) == 0 {
 | |
| 		allErrs = append(allErrs, field.Required(fldPath.Child("accessModes"), ""))
 | |
| 	}
 | |
| 	for _, mode := range pvSpec.AccessModes {
 | |
| 		if !supportedAccessModes.Has(string(mode)) {
 | |
| 			allErrs = append(allErrs, field.NotSupported(fldPath.Child("accessModes"), mode, supportedAccessModes.List()))
 | |
| 		}
 | |
| 	}
 | |
| 
 | |
| 	if !validateInlinePersistentVolumeSpec {
 | |
| 		if len(pvSpec.Capacity) == 0 {
 | |
| 			allErrs = append(allErrs, field.Required(fldPath.Child("capacity"), ""))
 | |
| 		}
 | |
| 
 | |
| 		if _, ok := pvSpec.Capacity[core.ResourceStorage]; !ok || len(pvSpec.Capacity) > 1 {
 | |
| 			allErrs = append(allErrs, field.NotSupported(fldPath.Child("capacity"), pvSpec.Capacity, []string{string(core.ResourceStorage)}))
 | |
| 		}
 | |
| 		capPath := fldPath.Child("capacity")
 | |
| 		for r, qty := range pvSpec.Capacity {
 | |
| 			allErrs = append(allErrs, validateBasicResource(qty, capPath.Key(string(r)))...)
 | |
| 			allErrs = append(allErrs, ValidatePositiveQuantityValue(qty, capPath.Key(string(r)))...)
 | |
| 		}
 | |
| 	}
 | |
| 
 | |
| 	if len(string(pvSpec.PersistentVolumeReclaimPolicy)) > 0 {
 | |
| 		if validateInlinePersistentVolumeSpec {
 | |
| 			if pvSpec.PersistentVolumeReclaimPolicy != core.PersistentVolumeReclaimRetain {
 | |
| 				allErrs = append(allErrs, field.Forbidden(fldPath.Child("persistentVolumeReclaimPolicy"), "may only be "+string(core.PersistentVolumeReclaimRetain)+" in the context of inline volumes"))
 | |
| 			}
 | |
| 		} else {
 | |
| 			if !supportedReclaimPolicy.Has(string(pvSpec.PersistentVolumeReclaimPolicy)) {
 | |
| 				allErrs = append(allErrs, field.NotSupported(fldPath.Child("persistentVolumeReclaimPolicy"), pvSpec.PersistentVolumeReclaimPolicy, supportedReclaimPolicy.List()))
 | |
| 			}
 | |
| 		}
 | |
| 	}
 | |
| 
 | |
| 	var nodeAffinitySpecified bool
 | |
| 	var errs field.ErrorList
 | |
| 	if pvSpec.NodeAffinity != nil {
 | |
| 		if validateInlinePersistentVolumeSpec {
 | |
| 			allErrs = append(allErrs, field.Forbidden(fldPath.Child("nodeAffinity"), "may not be specified in the context of inline volumes"))
 | |
| 		} else {
 | |
| 			nodeAffinitySpecified, errs = validateVolumeNodeAffinity(pvSpec.NodeAffinity, fldPath.Child("nodeAffinity"))
 | |
| 			allErrs = append(allErrs, errs...)
 | |
| 		}
 | |
| 	}
 | |
| 	numVolumes := 0
 | |
| 	if pvSpec.HostPath != nil {
 | |
| 		if numVolumes > 0 {
 | |
| 			allErrs = append(allErrs, field.Forbidden(fldPath.Child("hostPath"), "may not specify more than 1 volume type"))
 | |
| 		} else {
 | |
| 			numVolumes++
 | |
| 			allErrs = append(allErrs, validateHostPathVolumeSource(pvSpec.HostPath, fldPath.Child("hostPath"))...)
 | |
| 		}
 | |
| 	}
 | |
| 	if pvSpec.GCEPersistentDisk != nil {
 | |
| 		if numVolumes > 0 {
 | |
| 			allErrs = append(allErrs, field.Forbidden(fldPath.Child("gcePersistentDisk"), "may not specify more than 1 volume type"))
 | |
| 		} else {
 | |
| 			numVolumes++
 | |
| 			allErrs = append(allErrs, validateGCEPersistentDiskVolumeSource(pvSpec.GCEPersistentDisk, fldPath.Child("persistentDisk"))...)
 | |
| 		}
 | |
| 	}
 | |
| 	if pvSpec.AWSElasticBlockStore != nil {
 | |
| 		if numVolumes > 0 {
 | |
| 			allErrs = append(allErrs, field.Forbidden(fldPath.Child("awsElasticBlockStore"), "may not specify more than 1 volume type"))
 | |
| 		} else {
 | |
| 			numVolumes++
 | |
| 			allErrs = append(allErrs, validateAWSElasticBlockStoreVolumeSource(pvSpec.AWSElasticBlockStore, fldPath.Child("awsElasticBlockStore"))...)
 | |
| 		}
 | |
| 	}
 | |
| 	if pvSpec.Glusterfs != nil {
 | |
| 		if numVolumes > 0 {
 | |
| 			allErrs = append(allErrs, field.Forbidden(fldPath.Child("glusterfs"), "may not specify more than 1 volume type"))
 | |
| 		} else {
 | |
| 			numVolumes++
 | |
| 			allErrs = append(allErrs, validateGlusterfsPersistentVolumeSource(pvSpec.Glusterfs, fldPath.Child("glusterfs"))...)
 | |
| 		}
 | |
| 	}
 | |
| 	if pvSpec.Flocker != nil {
 | |
| 		if numVolumes > 0 {
 | |
| 			allErrs = append(allErrs, field.Forbidden(fldPath.Child("flocker"), "may not specify more than 1 volume type"))
 | |
| 		} else {
 | |
| 			numVolumes++
 | |
| 			allErrs = append(allErrs, validateFlockerVolumeSource(pvSpec.Flocker, fldPath.Child("flocker"))...)
 | |
| 		}
 | |
| 	}
 | |
| 	if pvSpec.NFS != nil {
 | |
| 		if numVolumes > 0 {
 | |
| 			allErrs = append(allErrs, field.Forbidden(fldPath.Child("nfs"), "may not specify more than 1 volume type"))
 | |
| 		} else {
 | |
| 			numVolumes++
 | |
| 			allErrs = append(allErrs, validateNFSVolumeSource(pvSpec.NFS, fldPath.Child("nfs"))...)
 | |
| 		}
 | |
| 	}
 | |
| 	if pvSpec.RBD != nil {
 | |
| 		if numVolumes > 0 {
 | |
| 			allErrs = append(allErrs, field.Forbidden(fldPath.Child("rbd"), "may not specify more than 1 volume type"))
 | |
| 		} else {
 | |
| 			numVolumes++
 | |
| 			allErrs = append(allErrs, validateRBDPersistentVolumeSource(pvSpec.RBD, fldPath.Child("rbd"))...)
 | |
| 		}
 | |
| 	}
 | |
| 	if pvSpec.Quobyte != nil {
 | |
| 		if numVolumes > 0 {
 | |
| 			allErrs = append(allErrs, field.Forbidden(fldPath.Child("quobyte"), "may not specify more than 1 volume type"))
 | |
| 		} else {
 | |
| 			numVolumes++
 | |
| 			allErrs = append(allErrs, validateQuobyteVolumeSource(pvSpec.Quobyte, fldPath.Child("quobyte"))...)
 | |
| 		}
 | |
| 	}
 | |
| 	if pvSpec.CephFS != nil {
 | |
| 		if numVolumes > 0 {
 | |
| 			allErrs = append(allErrs, field.Forbidden(fldPath.Child("cephFS"), "may not specify more than 1 volume type"))
 | |
| 		} else {
 | |
| 			numVolumes++
 | |
| 			allErrs = append(allErrs, validateCephFSPersistentVolumeSource(pvSpec.CephFS, fldPath.Child("cephfs"))...)
 | |
| 		}
 | |
| 	}
 | |
| 	if pvSpec.ISCSI != nil {
 | |
| 		if numVolumes > 0 {
 | |
| 			allErrs = append(allErrs, field.Forbidden(fldPath.Child("iscsi"), "may not specify more than 1 volume type"))
 | |
| 		} else {
 | |
| 			numVolumes++
 | |
| 			allErrs = append(allErrs, validateISCSIPersistentVolumeSource(pvSpec.ISCSI, pvName, fldPath.Child("iscsi"))...)
 | |
| 		}
 | |
| 	}
 | |
| 	if pvSpec.Cinder != nil {
 | |
| 		if numVolumes > 0 {
 | |
| 			allErrs = append(allErrs, field.Forbidden(fldPath.Child("cinder"), "may not specify more than 1 volume type"))
 | |
| 		} else {
 | |
| 			numVolumes++
 | |
| 			allErrs = append(allErrs, validateCinderPersistentVolumeSource(pvSpec.Cinder, fldPath.Child("cinder"))...)
 | |
| 		}
 | |
| 	}
 | |
| 	if pvSpec.FC != nil {
 | |
| 		if numVolumes > 0 {
 | |
| 			allErrs = append(allErrs, field.Forbidden(fldPath.Child("fc"), "may not specify more than 1 volume type"))
 | |
| 		} else {
 | |
| 			numVolumes++
 | |
| 			allErrs = append(allErrs, validateFCVolumeSource(pvSpec.FC, fldPath.Child("fc"))...)
 | |
| 		}
 | |
| 	}
 | |
| 	if pvSpec.FlexVolume != nil {
 | |
| 		numVolumes++
 | |
| 		allErrs = append(allErrs, validateFlexPersistentVolumeSource(pvSpec.FlexVolume, fldPath.Child("flexVolume"))...)
 | |
| 	}
 | |
| 	if pvSpec.AzureFile != nil {
 | |
| 		if numVolumes > 0 {
 | |
| 			allErrs = append(allErrs, field.Forbidden(fldPath.Child("azureFile"), "may not specify more than 1 volume type"))
 | |
| 
 | |
| 		} else {
 | |
| 			numVolumes++
 | |
| 			allErrs = append(allErrs, validateAzureFilePV(pvSpec.AzureFile, fldPath.Child("azureFile"))...)
 | |
| 		}
 | |
| 	}
 | |
| 
 | |
| 	if pvSpec.VsphereVolume != nil {
 | |
| 		if numVolumes > 0 {
 | |
| 			allErrs = append(allErrs, field.Forbidden(fldPath.Child("vsphereVolume"), "may not specify more than 1 volume type"))
 | |
| 		} else {
 | |
| 			numVolumes++
 | |
| 			allErrs = append(allErrs, validateVsphereVolumeSource(pvSpec.VsphereVolume, fldPath.Child("vsphereVolume"))...)
 | |
| 		}
 | |
| 	}
 | |
| 	if pvSpec.PhotonPersistentDisk != nil {
 | |
| 		if numVolumes > 0 {
 | |
| 			allErrs = append(allErrs, field.Forbidden(fldPath.Child("photonPersistentDisk"), "may not specify more than 1 volume type"))
 | |
| 		} else {
 | |
| 			numVolumes++
 | |
| 			allErrs = append(allErrs, validatePhotonPersistentDiskVolumeSource(pvSpec.PhotonPersistentDisk, fldPath.Child("photonPersistentDisk"))...)
 | |
| 		}
 | |
| 	}
 | |
| 	if pvSpec.PortworxVolume != nil {
 | |
| 		if numVolumes > 0 {
 | |
| 			allErrs = append(allErrs, field.Forbidden(fldPath.Child("portworxVolume"), "may not specify more than 1 volume type"))
 | |
| 		} else {
 | |
| 			numVolumes++
 | |
| 			allErrs = append(allErrs, validatePortworxVolumeSource(pvSpec.PortworxVolume, fldPath.Child("portworxVolume"))...)
 | |
| 		}
 | |
| 	}
 | |
| 	if pvSpec.AzureDisk != nil {
 | |
| 		if numVolumes > 0 {
 | |
| 			allErrs = append(allErrs, field.Forbidden(fldPath.Child("azureDisk"), "may not specify more than 1 volume type"))
 | |
| 		} else {
 | |
| 			numVolumes++
 | |
| 			allErrs = append(allErrs, validateAzureDisk(pvSpec.AzureDisk, fldPath.Child("azureDisk"))...)
 | |
| 		}
 | |
| 	}
 | |
| 	if pvSpec.ScaleIO != nil {
 | |
| 		if numVolumes > 0 {
 | |
| 			allErrs = append(allErrs, field.Forbidden(fldPath.Child("scaleIO"), "may not specify more than 1 volume type"))
 | |
| 		} else {
 | |
| 			numVolumes++
 | |
| 			allErrs = append(allErrs, validateScaleIOPersistentVolumeSource(pvSpec.ScaleIO, fldPath.Child("scaleIO"))...)
 | |
| 		}
 | |
| 	}
 | |
| 	if pvSpec.Local != nil {
 | |
| 		if numVolumes > 0 {
 | |
| 			allErrs = append(allErrs, field.Forbidden(fldPath.Child("local"), "may not specify more than 1 volume type"))
 | |
| 		} else {
 | |
| 			numVolumes++
 | |
| 			allErrs = append(allErrs, validateLocalVolumeSource(pvSpec.Local, fldPath.Child("local"))...)
 | |
| 			// NodeAffinity is required
 | |
| 			if !nodeAffinitySpecified {
 | |
| 				allErrs = append(allErrs, field.Required(fldPath.Child("nodeAffinity"), "Local volume requires node affinity"))
 | |
| 			}
 | |
| 		}
 | |
| 	}
 | |
| 	if pvSpec.StorageOS != nil {
 | |
| 		if numVolumes > 0 {
 | |
| 			allErrs = append(allErrs, field.Forbidden(fldPath.Child("storageos"), "may not specify more than 1 volume type"))
 | |
| 		} else {
 | |
| 			numVolumes++
 | |
| 			allErrs = append(allErrs, validateStorageOSPersistentVolumeSource(pvSpec.StorageOS, fldPath.Child("storageos"))...)
 | |
| 		}
 | |
| 	}
 | |
| 
 | |
| 	if pvSpec.CSI != nil {
 | |
| 		if numVolumes > 0 {
 | |
| 			allErrs = append(allErrs, field.Forbidden(fldPath.Child("csi"), "may not specify more than 1 volume type"))
 | |
| 		} else {
 | |
| 			numVolumes++
 | |
| 			allErrs = append(allErrs, validateCSIPersistentVolumeSource(pvSpec.CSI, fldPath.Child("csi"))...)
 | |
| 		}
 | |
| 	}
 | |
| 
 | |
| 	if numVolumes == 0 {
 | |
| 		allErrs = append(allErrs, field.Required(fldPath, "must specify a volume type"))
 | |
| 	}
 | |
| 
 | |
| 	// do not allow hostPath mounts of '/' to have a 'recycle' reclaim policy
 | |
| 	if pvSpec.HostPath != nil && path.Clean(pvSpec.HostPath.Path) == "/" && pvSpec.PersistentVolumeReclaimPolicy == core.PersistentVolumeReclaimRecycle {
 | |
| 		allErrs = append(allErrs, field.Forbidden(fldPath.Child("persistentVolumeReclaimPolicy"), "may not be 'recycle' for a hostPath mount of '/'"))
 | |
| 	}
 | |
| 
 | |
| 	if len(pvSpec.StorageClassName) > 0 {
 | |
| 		if validateInlinePersistentVolumeSpec {
 | |
| 			allErrs = append(allErrs, field.Forbidden(fldPath.Child("storageClassName"), "may not be specified in the context of inline volumes"))
 | |
| 		} else {
 | |
| 			for _, msg := range ValidateClassName(pvSpec.StorageClassName, false) {
 | |
| 				allErrs = append(allErrs, field.Invalid(fldPath.Child("storageClassName"), pvSpec.StorageClassName, msg))
 | |
| 			}
 | |
| 		}
 | |
| 	}
 | |
| 	if pvSpec.VolumeMode != nil {
 | |
| 		if validateInlinePersistentVolumeSpec {
 | |
| 			if *pvSpec.VolumeMode != core.PersistentVolumeFilesystem {
 | |
| 				allErrs = append(allErrs, field.Forbidden(fldPath.Child("volumeMode"), "may not specify volumeMode other than "+string(core.PersistentVolumeFilesystem)+" in the context of inline volumes"))
 | |
| 			}
 | |
| 		} else {
 | |
| 			if !supportedVolumeModes.Has(string(*pvSpec.VolumeMode)) {
 | |
| 				allErrs = append(allErrs, field.NotSupported(fldPath.Child("volumeMode"), *pvSpec.VolumeMode, supportedVolumeModes.List()))
 | |
| 			}
 | |
| 		}
 | |
| 	}
 | |
| 	return allErrs
 | |
| }
 | |
| 
 | |
| func ValidatePersistentVolume(pv *core.PersistentVolume) field.ErrorList {
 | |
| 	metaPath := field.NewPath("metadata")
 | |
| 	allErrs := ValidateObjectMeta(&pv.ObjectMeta, false, ValidatePersistentVolumeName, metaPath)
 | |
| 	allErrs = append(allErrs, ValidatePersistentVolumeSpec(&pv.Spec, pv.ObjectMeta.Name, false, field.NewPath("spec"))...)
 | |
| 	return allErrs
 | |
| }
 | |
| 
 | |
| // ValidatePersistentVolumeUpdate tests to see if the update is legal for an end user to make.
 | |
| // newPv is updated with fields that cannot be changed.
 | |
| func ValidatePersistentVolumeUpdate(newPv, oldPv *core.PersistentVolume) field.ErrorList {
 | |
| 	allErrs := ValidatePersistentVolume(newPv)
 | |
| 
 | |
| 	// if oldPV does not have ControllerExpandSecretRef then allow it to be set
 | |
| 	if (oldPv.Spec.CSI != nil && oldPv.Spec.CSI.ControllerExpandSecretRef == nil) &&
 | |
| 		(newPv.Spec.CSI != nil && newPv.Spec.CSI.ControllerExpandSecretRef != nil) {
 | |
| 		newPv = newPv.DeepCopy()
 | |
| 		newPv.Spec.CSI.ControllerExpandSecretRef = nil
 | |
| 	}
 | |
| 
 | |
| 	// PersistentVolumeSource should be immutable after creation.
 | |
| 	if !apiequality.Semantic.DeepEqual(newPv.Spec.PersistentVolumeSource, oldPv.Spec.PersistentVolumeSource) {
 | |
| 		pvcSourceDiff := diff.ObjectDiff(newPv.Spec.PersistentVolumeSource, oldPv.Spec.PersistentVolumeSource)
 | |
| 		allErrs = append(allErrs, field.Forbidden(field.NewPath("spec", "persistentvolumesource"), fmt.Sprintf("spec.persistentvolumesource is immutable after creation\n%v", pvcSourceDiff)))
 | |
| 	}
 | |
| 	allErrs = append(allErrs, ValidateImmutableField(newPv.Spec.VolumeMode, oldPv.Spec.VolumeMode, field.NewPath("volumeMode"))...)
 | |
| 
 | |
| 	// Allow setting NodeAffinity if oldPv NodeAffinity was not set
 | |
| 	if oldPv.Spec.NodeAffinity != nil {
 | |
| 		allErrs = append(allErrs, ValidateImmutableField(newPv.Spec.NodeAffinity, oldPv.Spec.NodeAffinity, field.NewPath("nodeAffinity"))...)
 | |
| 	}
 | |
| 
 | |
| 	return allErrs
 | |
| }
 | |
| 
 | |
| // ValidatePersistentVolumeStatusUpdate tests to see if the status update is legal for an end user to make.
 | |
| func ValidatePersistentVolumeStatusUpdate(newPv, oldPv *core.PersistentVolume) field.ErrorList {
 | |
| 	allErrs := ValidateObjectMetaUpdate(&newPv.ObjectMeta, &oldPv.ObjectMeta, field.NewPath("metadata"))
 | |
| 	if len(newPv.ResourceVersion) == 0 {
 | |
| 		allErrs = append(allErrs, field.Required(field.NewPath("resourceVersion"), ""))
 | |
| 	}
 | |
| 	return allErrs
 | |
| }
 | |
| 
 | |
| // ValidatePersistentVolumeClaim validates a PersistentVolumeClaim
 | |
| func ValidatePersistentVolumeClaim(pvc *core.PersistentVolumeClaim) field.ErrorList {
 | |
| 	allErrs := ValidateObjectMeta(&pvc.ObjectMeta, true, ValidatePersistentVolumeName, field.NewPath("metadata"))
 | |
| 	allErrs = append(allErrs, ValidatePersistentVolumeClaimSpec(&pvc.Spec, field.NewPath("spec"))...)
 | |
| 	return allErrs
 | |
| }
 | |
| 
 | |
| // ValidatePersistentVolumeClaimSpec validates a PersistentVolumeClaimSpec
 | |
| func ValidatePersistentVolumeClaimSpec(spec *core.PersistentVolumeClaimSpec, fldPath *field.Path) field.ErrorList {
 | |
| 	allErrs := field.ErrorList{}
 | |
| 	if len(spec.AccessModes) == 0 {
 | |
| 		allErrs = append(allErrs, field.Required(fldPath.Child("accessModes"), "at least 1 access mode is required"))
 | |
| 	}
 | |
| 	if spec.Selector != nil {
 | |
| 		allErrs = append(allErrs, unversionedvalidation.ValidateLabelSelector(spec.Selector, fldPath.Child("selector"))...)
 | |
| 	}
 | |
| 	for _, mode := range spec.AccessModes {
 | |
| 		if mode != core.ReadWriteOnce && mode != core.ReadOnlyMany && mode != core.ReadWriteMany {
 | |
| 			allErrs = append(allErrs, field.NotSupported(fldPath.Child("accessModes"), mode, supportedAccessModes.List()))
 | |
| 		}
 | |
| 	}
 | |
| 	storageValue, ok := spec.Resources.Requests[core.ResourceStorage]
 | |
| 	if !ok {
 | |
| 		allErrs = append(allErrs, field.Required(fldPath.Child("resources").Key(string(core.ResourceStorage)), ""))
 | |
| 	} else if errs := ValidatePositiveQuantityValue(storageValue, fldPath.Child("resources").Key(string(core.ResourceStorage))); len(errs) > 0 {
 | |
| 		allErrs = append(allErrs, errs...)
 | |
| 	} else {
 | |
| 		allErrs = append(allErrs, ValidateResourceQuantityValue(string(core.ResourceStorage), storageValue, fldPath.Child("resources").Key(string(core.ResourceStorage)))...)
 | |
| 	}
 | |
| 
 | |
| 	if spec.StorageClassName != nil && len(*spec.StorageClassName) > 0 {
 | |
| 		for _, msg := range ValidateClassName(*spec.StorageClassName, false) {
 | |
| 			allErrs = append(allErrs, field.Invalid(fldPath.Child("storageClassName"), *spec.StorageClassName, msg))
 | |
| 		}
 | |
| 	}
 | |
| 	if spec.VolumeMode != nil && !supportedVolumeModes.Has(string(*spec.VolumeMode)) {
 | |
| 		allErrs = append(allErrs, field.NotSupported(fldPath.Child("volumeMode"), *spec.VolumeMode, supportedVolumeModes.List()))
 | |
| 	}
 | |
| 
 | |
| 	if spec.DataSource != nil {
 | |
| 		if len(spec.DataSource.Name) == 0 {
 | |
| 			allErrs = append(allErrs, field.Required(fldPath.Child("dataSource", "name"), ""))
 | |
| 		}
 | |
| 		if len(spec.DataSource.Kind) == 0 {
 | |
| 			allErrs = append(allErrs, field.Required(fldPath.Child("dataSource", "kind"), ""))
 | |
| 		}
 | |
| 		apiGroup := ""
 | |
| 		if spec.DataSource.APIGroup != nil {
 | |
| 			apiGroup = *spec.DataSource.APIGroup
 | |
| 		}
 | |
| 		if len(apiGroup) == 0 && spec.DataSource.Kind != "PersistentVolumeClaim" {
 | |
| 			allErrs = append(allErrs, field.Invalid(fldPath.Child("dataSource"), spec.DataSource.Kind, ""))
 | |
| 		}
 | |
| 	}
 | |
| 
 | |
| 	return allErrs
 | |
| }
 | |
| 
 | |
| // ValidatePersistentVolumeClaimUpdate validates an update to a PersistentVolumeClaim
 | |
| func ValidatePersistentVolumeClaimUpdate(newPvc, oldPvc *core.PersistentVolumeClaim) field.ErrorList {
 | |
| 	allErrs := ValidateObjectMetaUpdate(&newPvc.ObjectMeta, &oldPvc.ObjectMeta, field.NewPath("metadata"))
 | |
| 	allErrs = append(allErrs, ValidatePersistentVolumeClaim(newPvc)...)
 | |
| 	newPvcClone := newPvc.DeepCopy()
 | |
| 	oldPvcClone := oldPvc.DeepCopy()
 | |
| 
 | |
| 	// PVController needs to update PVC.Spec w/ VolumeName.
 | |
| 	// Claims are immutable in order to enforce quota, range limits, etc. without gaming the system.
 | |
| 	if len(oldPvc.Spec.VolumeName) == 0 {
 | |
| 		// volumeName changes are allowed once.
 | |
| 		oldPvcClone.Spec.VolumeName = newPvcClone.Spec.VolumeName // +k8s:verify-mutation:reason=clone
 | |
| 	}
 | |
| 
 | |
| 	if validateStorageClassUpgrade(oldPvcClone.Annotations, newPvcClone.Annotations,
 | |
| 		oldPvcClone.Spec.StorageClassName, newPvcClone.Spec.StorageClassName) {
 | |
| 		newPvcClone.Spec.StorageClassName = nil
 | |
| 		metav1.SetMetaDataAnnotation(&newPvcClone.ObjectMeta, core.BetaStorageClassAnnotation, oldPvcClone.Annotations[core.BetaStorageClassAnnotation])
 | |
| 	} else {
 | |
| 		// storageclass annotation should be immutable after creation
 | |
| 		// TODO: remove Beta when no longer needed
 | |
| 		allErrs = append(allErrs, ValidateImmutableAnnotation(newPvc.ObjectMeta.Annotations[v1.BetaStorageClassAnnotation], oldPvc.ObjectMeta.Annotations[v1.BetaStorageClassAnnotation], v1.BetaStorageClassAnnotation, field.NewPath("metadata"))...)
 | |
| 	}
 | |
| 
 | |
| 	if utilfeature.DefaultFeatureGate.Enabled(features.ExpandPersistentVolumes) {
 | |
| 		// lets make sure storage values are same.
 | |
| 		if newPvc.Status.Phase == core.ClaimBound && newPvcClone.Spec.Resources.Requests != nil {
 | |
| 			newPvcClone.Spec.Resources.Requests["storage"] = oldPvc.Spec.Resources.Requests["storage"] // +k8s:verify-mutation:reason=clone
 | |
| 		}
 | |
| 
 | |
| 		oldSize := oldPvc.Spec.Resources.Requests["storage"]
 | |
| 		newSize := newPvc.Spec.Resources.Requests["storage"]
 | |
| 
 | |
| 		if !apiequality.Semantic.DeepEqual(newPvcClone.Spec, oldPvcClone.Spec) {
 | |
| 			specDiff := diff.ObjectDiff(newPvcClone.Spec, oldPvcClone.Spec)
 | |
| 			allErrs = append(allErrs, field.Forbidden(field.NewPath("spec"), fmt.Sprintf("spec is immutable after creation except resources.requests for bound claims\n%v", specDiff)))
 | |
| 		}
 | |
| 		if newSize.Cmp(oldSize) < 0 {
 | |
| 			allErrs = append(allErrs, field.Forbidden(field.NewPath("spec", "resources", "requests", "storage"), "field can not be less than previous value"))
 | |
| 		}
 | |
| 
 | |
| 	} else {
 | |
| 		// changes to Spec are not allowed, but updates to label/and some annotations are OK.
 | |
| 		// no-op updates pass validation.
 | |
| 		if !apiequality.Semantic.DeepEqual(newPvcClone.Spec, oldPvcClone.Spec) {
 | |
| 			specDiff := diff.ObjectDiff(newPvcClone.Spec, oldPvcClone.Spec)
 | |
| 			allErrs = append(allErrs, field.Forbidden(field.NewPath("spec"), fmt.Sprintf("field is immutable after creation\n%v", specDiff)))
 | |
| 		}
 | |
| 	}
 | |
| 
 | |
| 	allErrs = append(allErrs, ValidateImmutableField(newPvc.Spec.VolumeMode, oldPvc.Spec.VolumeMode, field.NewPath("volumeMode"))...)
 | |
| 
 | |
| 	return allErrs
 | |
| }
 | |
| 
 | |
| // Provide an upgrade path from PVC with storage class specified in beta
 | |
| // annotation to storage class specified in attribute. We allow update of
 | |
| // StorageClassName only if following four conditions are met at the same time:
 | |
| // 1. The old pvc's StorageClassAnnotation is set
 | |
| // 2. The old pvc's StorageClassName is not set
 | |
| // 3. The new pvc's StorageClassName is set and equal to the old value in annotation
 | |
| // 4. If the new pvc's StorageClassAnnotation is set,it must be equal to the old pv/pvc's StorageClassAnnotation
 | |
| func validateStorageClassUpgrade(oldAnnotations, newAnnotations map[string]string, oldScName, newScName *string) bool {
 | |
| 	oldSc, oldAnnotationExist := oldAnnotations[core.BetaStorageClassAnnotation]
 | |
| 	newScInAnnotation, newAnnotationExist := newAnnotations[core.BetaStorageClassAnnotation]
 | |
| 	return oldAnnotationExist /* condition 1 */ &&
 | |
| 		oldScName == nil /* condition 2*/ &&
 | |
| 		(newScName != nil && *newScName == oldSc) /* condition 3 */ &&
 | |
| 		(!newAnnotationExist || newScInAnnotation == oldSc) /* condition 4 */
 | |
| }
 | |
| 
 | |
| // ValidatePersistentVolumeClaimStatusUpdate validates an update to status of a PersistentVolumeClaim
 | |
| func ValidatePersistentVolumeClaimStatusUpdate(newPvc, oldPvc *core.PersistentVolumeClaim) field.ErrorList {
 | |
| 	allErrs := ValidateObjectMetaUpdate(&newPvc.ObjectMeta, &oldPvc.ObjectMeta, field.NewPath("metadata"))
 | |
| 	if len(newPvc.ResourceVersion) == 0 {
 | |
| 		allErrs = append(allErrs, field.Required(field.NewPath("resourceVersion"), ""))
 | |
| 	}
 | |
| 	if len(newPvc.Spec.AccessModes) == 0 {
 | |
| 		allErrs = append(allErrs, field.Required(field.NewPath("Spec", "accessModes"), ""))
 | |
| 	}
 | |
| 	capPath := field.NewPath("status", "capacity")
 | |
| 	for r, qty := range newPvc.Status.Capacity {
 | |
| 		allErrs = append(allErrs, validateBasicResource(qty, capPath.Key(string(r)))...)
 | |
| 	}
 | |
| 	return allErrs
 | |
| }
 | |
| 
 | |
| var supportedPortProtocols = sets.NewString(string(core.ProtocolTCP), string(core.ProtocolUDP), string(core.ProtocolSCTP))
 | |
| 
 | |
| func validateContainerPorts(ports []core.ContainerPort, fldPath *field.Path) field.ErrorList {
 | |
| 	allErrs := field.ErrorList{}
 | |
| 
 | |
| 	allNames := sets.String{}
 | |
| 	for i, port := range ports {
 | |
| 		idxPath := fldPath.Index(i)
 | |
| 		if len(port.Name) > 0 {
 | |
| 			if msgs := validation.IsValidPortName(port.Name); len(msgs) != 0 {
 | |
| 				for i = range msgs {
 | |
| 					allErrs = append(allErrs, field.Invalid(idxPath.Child("name"), port.Name, msgs[i]))
 | |
| 				}
 | |
| 			} else if allNames.Has(port.Name) {
 | |
| 				allErrs = append(allErrs, field.Duplicate(idxPath.Child("name"), port.Name))
 | |
| 			} else {
 | |
| 				allNames.Insert(port.Name)
 | |
| 			}
 | |
| 		}
 | |
| 		if port.ContainerPort == 0 {
 | |
| 			allErrs = append(allErrs, field.Required(idxPath.Child("containerPort"), ""))
 | |
| 		} else {
 | |
| 			for _, msg := range validation.IsValidPortNum(int(port.ContainerPort)) {
 | |
| 				allErrs = append(allErrs, field.Invalid(idxPath.Child("containerPort"), port.ContainerPort, msg))
 | |
| 			}
 | |
| 		}
 | |
| 		if port.HostPort != 0 {
 | |
| 			for _, msg := range validation.IsValidPortNum(int(port.HostPort)) {
 | |
| 				allErrs = append(allErrs, field.Invalid(idxPath.Child("hostPort"), port.HostPort, msg))
 | |
| 			}
 | |
| 		}
 | |
| 		if len(port.Protocol) == 0 {
 | |
| 			allErrs = append(allErrs, field.Required(idxPath.Child("protocol"), ""))
 | |
| 		} else if !supportedPortProtocols.Has(string(port.Protocol)) {
 | |
| 			allErrs = append(allErrs, field.NotSupported(idxPath.Child("protocol"), port.Protocol, supportedPortProtocols.List()))
 | |
| 		}
 | |
| 	}
 | |
| 	return allErrs
 | |
| }
 | |
| 
 | |
| // ValidateEnv validates env vars
 | |
| func ValidateEnv(vars []core.EnvVar, fldPath *field.Path, opts PodValidationOptions) field.ErrorList {
 | |
| 	allErrs := field.ErrorList{}
 | |
| 
 | |
| 	for i, ev := range vars {
 | |
| 		idxPath := fldPath.Index(i)
 | |
| 		if len(ev.Name) == 0 {
 | |
| 			allErrs = append(allErrs, field.Required(idxPath.Child("name"), ""))
 | |
| 		} else {
 | |
| 			for _, msg := range validation.IsEnvVarName(ev.Name) {
 | |
| 				allErrs = append(allErrs, field.Invalid(idxPath.Child("name"), ev.Name, msg))
 | |
| 			}
 | |
| 		}
 | |
| 		allErrs = append(allErrs, validateEnvVarValueFrom(ev, idxPath.Child("valueFrom"), opts)...)
 | |
| 	}
 | |
| 	return allErrs
 | |
| }
 | |
| 
 | |
| var validEnvDownwardAPIFieldPathExpressions = sets.NewString(
 | |
| 	"metadata.name",
 | |
| 	"metadata.namespace",
 | |
| 	"metadata.uid",
 | |
| 	"spec.nodeName",
 | |
| 	"spec.serviceAccountName",
 | |
| 	"status.hostIP",
 | |
| 	"status.podIP",
 | |
| 	// status.podIPs is populated even if IPv6DualStack feature gate
 | |
| 	// is not enabled. This will work for single stack and dual stack.
 | |
| 	"status.podIPs")
 | |
| 
 | |
| var validContainerResourceFieldPathExpressions = sets.NewString("limits.cpu", "limits.memory", "limits.ephemeral-storage", "requests.cpu", "requests.memory", "requests.ephemeral-storage")
 | |
| 
 | |
| // NOTE: this is only valid with DownwardAPIHugePages enabled
 | |
| var validContainerResourceFieldPathPrefixes = sets.NewString()
 | |
| var validContainerResourceFieldPathPrefixesWithDownwardAPIHugePages = sets.NewString(hugepagesRequestsPrefixDownwardAPI, hugepagesLimitsPrefixDownwardAPI)
 | |
| 
 | |
| const hugepagesRequestsPrefixDownwardAPI string = `requests.hugepages-`
 | |
| const hugepagesLimitsPrefixDownwardAPI string = `limits.hugepages-`
 | |
| 
 | |
| func validateEnvVarValueFrom(ev core.EnvVar, fldPath *field.Path, opts PodValidationOptions) field.ErrorList {
 | |
| 	allErrs := field.ErrorList{}
 | |
| 
 | |
| 	if ev.ValueFrom == nil {
 | |
| 		return allErrs
 | |
| 	}
 | |
| 
 | |
| 	numSources := 0
 | |
| 
 | |
| 	if ev.ValueFrom.FieldRef != nil {
 | |
| 		numSources++
 | |
| 		allErrs = append(allErrs, validateObjectFieldSelector(ev.ValueFrom.FieldRef, &validEnvDownwardAPIFieldPathExpressions, fldPath.Child("fieldRef"))...)
 | |
| 	}
 | |
| 	if ev.ValueFrom.ResourceFieldRef != nil {
 | |
| 		numSources++
 | |
| 		localValidContainerResourceFieldPathPrefixes := validContainerResourceFieldPathPrefixes
 | |
| 		if opts.AllowDownwardAPIHugePages {
 | |
| 			localValidContainerResourceFieldPathPrefixes = validContainerResourceFieldPathPrefixesWithDownwardAPIHugePages
 | |
| 		}
 | |
| 		allErrs = append(allErrs, validateContainerResourceFieldSelector(ev.ValueFrom.ResourceFieldRef, &validContainerResourceFieldPathExpressions, &localValidContainerResourceFieldPathPrefixes, fldPath.Child("resourceFieldRef"), false)...)
 | |
| 	}
 | |
| 	if ev.ValueFrom.ConfigMapKeyRef != nil {
 | |
| 		numSources++
 | |
| 		allErrs = append(allErrs, validateConfigMapKeySelector(ev.ValueFrom.ConfigMapKeyRef, fldPath.Child("configMapKeyRef"))...)
 | |
| 	}
 | |
| 	if ev.ValueFrom.SecretKeyRef != nil {
 | |
| 		numSources++
 | |
| 		allErrs = append(allErrs, validateSecretKeySelector(ev.ValueFrom.SecretKeyRef, fldPath.Child("secretKeyRef"))...)
 | |
| 	}
 | |
| 
 | |
| 	if numSources == 0 {
 | |
| 		allErrs = append(allErrs, field.Invalid(fldPath, "", "must specify one of: `fieldRef`, `resourceFieldRef`, `configMapKeyRef` or `secretKeyRef`"))
 | |
| 	} else if len(ev.Value) != 0 {
 | |
| 		if numSources != 0 {
 | |
| 			allErrs = append(allErrs, field.Invalid(fldPath, "", "may not be specified when `value` is not empty"))
 | |
| 		}
 | |
| 	} else if numSources > 1 {
 | |
| 		allErrs = append(allErrs, field.Invalid(fldPath, "", "may not have more than one field specified at a time"))
 | |
| 	}
 | |
| 
 | |
| 	return allErrs
 | |
| }
 | |
| 
 | |
| func validateObjectFieldSelector(fs *core.ObjectFieldSelector, expressions *sets.String, fldPath *field.Path) field.ErrorList {
 | |
| 	allErrs := field.ErrorList{}
 | |
| 
 | |
| 	if len(fs.APIVersion) == 0 {
 | |
| 		allErrs = append(allErrs, field.Required(fldPath.Child("apiVersion"), ""))
 | |
| 		return allErrs
 | |
| 	}
 | |
| 	if len(fs.FieldPath) == 0 {
 | |
| 		allErrs = append(allErrs, field.Required(fldPath.Child("fieldPath"), ""))
 | |
| 		return allErrs
 | |
| 	}
 | |
| 
 | |
| 	internalFieldPath, _, err := podshelper.ConvertDownwardAPIFieldLabel(fs.APIVersion, fs.FieldPath, "")
 | |
| 	if err != nil {
 | |
| 		allErrs = append(allErrs, field.Invalid(fldPath.Child("fieldPath"), fs.FieldPath, fmt.Sprintf("error converting fieldPath: %v", err)))
 | |
| 		return allErrs
 | |
| 	}
 | |
| 
 | |
| 	if path, subscript, ok := fieldpath.SplitMaybeSubscriptedPath(internalFieldPath); ok {
 | |
| 		switch path {
 | |
| 		case "metadata.annotations":
 | |
| 			for _, msg := range validation.IsQualifiedName(strings.ToLower(subscript)) {
 | |
| 				allErrs = append(allErrs, field.Invalid(fldPath, subscript, msg))
 | |
| 			}
 | |
| 		case "metadata.labels":
 | |
| 			for _, msg := range validation.IsQualifiedName(subscript) {
 | |
| 				allErrs = append(allErrs, field.Invalid(fldPath, subscript, msg))
 | |
| 			}
 | |
| 		default:
 | |
| 			allErrs = append(allErrs, field.Invalid(fldPath, path, "does not support subscript"))
 | |
| 		}
 | |
| 	} else if !expressions.Has(path) {
 | |
| 		allErrs = append(allErrs, field.NotSupported(fldPath.Child("fieldPath"), path, expressions.List()))
 | |
| 		return allErrs
 | |
| 	}
 | |
| 
 | |
| 	return allErrs
 | |
| }
 | |
| 
 | |
| func validateContainerResourceFieldSelector(fs *core.ResourceFieldSelector, expressions *sets.String, prefixes *sets.String, fldPath *field.Path, volume bool) field.ErrorList {
 | |
| 	allErrs := field.ErrorList{}
 | |
| 
 | |
| 	if volume && len(fs.ContainerName) == 0 {
 | |
| 		allErrs = append(allErrs, field.Required(fldPath.Child("containerName"), ""))
 | |
| 	} else if len(fs.Resource) == 0 {
 | |
| 		allErrs = append(allErrs, field.Required(fldPath.Child("resource"), ""))
 | |
| 	} else if !expressions.Has(fs.Resource) {
 | |
| 		// check if the prefix is present
 | |
| 		foundPrefix := false
 | |
| 		if prefixes != nil {
 | |
| 			for _, prefix := range prefixes.List() {
 | |
| 				if strings.HasPrefix(fs.Resource, prefix) {
 | |
| 					foundPrefix = true
 | |
| 				}
 | |
| 			}
 | |
| 		}
 | |
| 		if !foundPrefix {
 | |
| 			allErrs = append(allErrs, field.NotSupported(fldPath.Child("resource"), fs.Resource, expressions.List()))
 | |
| 		}
 | |
| 	}
 | |
| 	allErrs = append(allErrs, validateContainerResourceDivisor(fs.Resource, fs.Divisor, fldPath)...)
 | |
| 	return allErrs
 | |
| }
 | |
| 
 | |
| func ValidateEnvFrom(vars []core.EnvFromSource, fldPath *field.Path) field.ErrorList {
 | |
| 	allErrs := field.ErrorList{}
 | |
| 	for i, ev := range vars {
 | |
| 		idxPath := fldPath.Index(i)
 | |
| 		if len(ev.Prefix) > 0 {
 | |
| 			for _, msg := range validation.IsEnvVarName(ev.Prefix) {
 | |
| 				allErrs = append(allErrs, field.Invalid(idxPath.Child("prefix"), ev.Prefix, msg))
 | |
| 			}
 | |
| 		}
 | |
| 
 | |
| 		numSources := 0
 | |
| 		if ev.ConfigMapRef != nil {
 | |
| 			numSources++
 | |
| 			allErrs = append(allErrs, validateConfigMapEnvSource(ev.ConfigMapRef, idxPath.Child("configMapRef"))...)
 | |
| 		}
 | |
| 		if ev.SecretRef != nil {
 | |
| 			numSources++
 | |
| 			allErrs = append(allErrs, validateSecretEnvSource(ev.SecretRef, idxPath.Child("secretRef"))...)
 | |
| 		}
 | |
| 
 | |
| 		if numSources == 0 {
 | |
| 			allErrs = append(allErrs, field.Invalid(fldPath, "", "must specify one of: `configMapRef` or `secretRef`"))
 | |
| 		} else if numSources > 1 {
 | |
| 			allErrs = append(allErrs, field.Invalid(fldPath, "", "may not have more than one field specified at a time"))
 | |
| 		}
 | |
| 	}
 | |
| 	return allErrs
 | |
| }
 | |
| 
 | |
| func validateConfigMapEnvSource(configMapSource *core.ConfigMapEnvSource, fldPath *field.Path) field.ErrorList {
 | |
| 	allErrs := field.ErrorList{}
 | |
| 	if len(configMapSource.Name) == 0 {
 | |
| 		allErrs = append(allErrs, field.Required(fldPath.Child("name"), ""))
 | |
| 	} else {
 | |
| 		for _, msg := range ValidateConfigMapName(configMapSource.Name, true) {
 | |
| 			allErrs = append(allErrs, field.Invalid(fldPath.Child("name"), configMapSource.Name, msg))
 | |
| 		}
 | |
| 	}
 | |
| 	return allErrs
 | |
| }
 | |
| 
 | |
| func validateSecretEnvSource(secretSource *core.SecretEnvSource, fldPath *field.Path) field.ErrorList {
 | |
| 	allErrs := field.ErrorList{}
 | |
| 	if len(secretSource.Name) == 0 {
 | |
| 		allErrs = append(allErrs, field.Required(fldPath.Child("name"), ""))
 | |
| 	} else {
 | |
| 		for _, msg := range ValidateSecretName(secretSource.Name, true) {
 | |
| 			allErrs = append(allErrs, field.Invalid(fldPath.Child("name"), secretSource.Name, msg))
 | |
| 		}
 | |
| 	}
 | |
| 	return allErrs
 | |
| }
 | |
| 
 | |
| var validContainerResourceDivisorForCPU = sets.NewString("1m", "1")
 | |
| var validContainerResourceDivisorForMemory = sets.NewString("1", "1k", "1M", "1G", "1T", "1P", "1E", "1Ki", "1Mi", "1Gi", "1Ti", "1Pi", "1Ei")
 | |
| var validContainerResourceDivisorForHugePages = sets.NewString("1", "1k", "1M", "1G", "1T", "1P", "1E", "1Ki", "1Mi", "1Gi", "1Ti", "1Pi", "1Ei")
 | |
| var validContainerResourceDivisorForEphemeralStorage = sets.NewString("1", "1k", "1M", "1G", "1T", "1P", "1E", "1Ki", "1Mi", "1Gi", "1Ti", "1Pi", "1Ei")
 | |
| 
 | |
| func validateContainerResourceDivisor(rName string, divisor resource.Quantity, fldPath *field.Path) field.ErrorList {
 | |
| 	allErrs := field.ErrorList{}
 | |
| 	unsetDivisor := resource.Quantity{}
 | |
| 	if unsetDivisor.Cmp(divisor) == 0 {
 | |
| 		return allErrs
 | |
| 	}
 | |
| 	switch rName {
 | |
| 	case "limits.cpu", "requests.cpu":
 | |
| 		if !validContainerResourceDivisorForCPU.Has(divisor.String()) {
 | |
| 			allErrs = append(allErrs, field.Invalid(fldPath.Child("divisor"), rName, "only divisor's values 1m and 1 are supported with the cpu resource"))
 | |
| 		}
 | |
| 	case "limits.memory", "requests.memory":
 | |
| 		if !validContainerResourceDivisorForMemory.Has(divisor.String()) {
 | |
| 			allErrs = append(allErrs, field.Invalid(fldPath.Child("divisor"), rName, "only divisor's values 1, 1k, 1M, 1G, 1T, 1P, 1E, 1Ki, 1Mi, 1Gi, 1Ti, 1Pi, 1Ei are supported with the memory resource"))
 | |
| 		}
 | |
| 	case "limits.ephemeral-storage", "requests.ephemeral-storage":
 | |
| 		if !validContainerResourceDivisorForEphemeralStorage.Has(divisor.String()) {
 | |
| 			allErrs = append(allErrs, field.Invalid(fldPath.Child("divisor"), rName, "only divisor's values 1, 1k, 1M, 1G, 1T, 1P, 1E, 1Ki, 1Mi, 1Gi, 1Ti, 1Pi, 1Ei are supported with the local ephemeral storage resource"))
 | |
| 		}
 | |
| 	}
 | |
| 	if strings.HasPrefix(rName, hugepagesRequestsPrefixDownwardAPI) || strings.HasPrefix(rName, hugepagesLimitsPrefixDownwardAPI) {
 | |
| 		if !validContainerResourceDivisorForHugePages.Has(divisor.String()) {
 | |
| 			allErrs = append(allErrs, field.Invalid(fldPath.Child("divisor"), rName, "only divisor's values 1, 1k, 1M, 1G, 1T, 1P, 1E, 1Ki, 1Mi, 1Gi, 1Ti, 1Pi, 1Ei are supported with the hugepages resource"))
 | |
| 		}
 | |
| 	}
 | |
| 	return allErrs
 | |
| }
 | |
| 
 | |
| func validateConfigMapKeySelector(s *core.ConfigMapKeySelector, fldPath *field.Path) field.ErrorList {
 | |
| 	allErrs := field.ErrorList{}
 | |
| 
 | |
| 	nameFn := ValidateNameFunc(ValidateSecretName)
 | |
| 	for _, msg := range nameFn(s.Name, false) {
 | |
| 		allErrs = append(allErrs, field.Invalid(fldPath.Child("name"), s.Name, msg))
 | |
| 	}
 | |
| 	if len(s.Key) == 0 {
 | |
| 		allErrs = append(allErrs, field.Required(fldPath.Child("key"), ""))
 | |
| 	} else {
 | |
| 		for _, msg := range validation.IsConfigMapKey(s.Key) {
 | |
| 			allErrs = append(allErrs, field.Invalid(fldPath.Child("key"), s.Key, msg))
 | |
| 		}
 | |
| 	}
 | |
| 
 | |
| 	return allErrs
 | |
| }
 | |
| 
 | |
| func validateSecretKeySelector(s *core.SecretKeySelector, fldPath *field.Path) field.ErrorList {
 | |
| 	allErrs := field.ErrorList{}
 | |
| 
 | |
| 	nameFn := ValidateNameFunc(ValidateSecretName)
 | |
| 	for _, msg := range nameFn(s.Name, false) {
 | |
| 		allErrs = append(allErrs, field.Invalid(fldPath.Child("name"), s.Name, msg))
 | |
| 	}
 | |
| 	if len(s.Key) == 0 {
 | |
| 		allErrs = append(allErrs, field.Required(fldPath.Child("key"), ""))
 | |
| 	} else {
 | |
| 		for _, msg := range validation.IsConfigMapKey(s.Key) {
 | |
| 			allErrs = append(allErrs, field.Invalid(fldPath.Child("key"), s.Key, msg))
 | |
| 		}
 | |
| 	}
 | |
| 
 | |
| 	return allErrs
 | |
| }
 | |
| 
 | |
| func GetVolumeMountMap(mounts []core.VolumeMount) map[string]string {
 | |
| 	volmounts := make(map[string]string)
 | |
| 
 | |
| 	for _, mnt := range mounts {
 | |
| 		volmounts[mnt.Name] = mnt.MountPath
 | |
| 	}
 | |
| 
 | |
| 	return volmounts
 | |
| }
 | |
| 
 | |
| func GetVolumeDeviceMap(devices []core.VolumeDevice) map[string]string {
 | |
| 	volDevices := make(map[string]string)
 | |
| 
 | |
| 	for _, dev := range devices {
 | |
| 		volDevices[dev.Name] = dev.DevicePath
 | |
| 	}
 | |
| 
 | |
| 	return volDevices
 | |
| }
 | |
| 
 | |
| func ValidateVolumeMounts(mounts []core.VolumeMount, voldevices map[string]string, volumes map[string]core.VolumeSource, container *core.Container, fldPath *field.Path) field.ErrorList {
 | |
| 	allErrs := field.ErrorList{}
 | |
| 	mountpoints := sets.NewString()
 | |
| 
 | |
| 	for i, mnt := range mounts {
 | |
| 		idxPath := fldPath.Index(i)
 | |
| 		if len(mnt.Name) == 0 {
 | |
| 			allErrs = append(allErrs, field.Required(idxPath.Child("name"), ""))
 | |
| 		}
 | |
| 		if !IsMatchedVolume(mnt.Name, volumes) {
 | |
| 			allErrs = append(allErrs, field.NotFound(idxPath.Child("name"), mnt.Name))
 | |
| 		}
 | |
| 		if len(mnt.MountPath) == 0 {
 | |
| 			allErrs = append(allErrs, field.Required(idxPath.Child("mountPath"), ""))
 | |
| 		}
 | |
| 		if mountpoints.Has(mnt.MountPath) {
 | |
| 			allErrs = append(allErrs, field.Invalid(idxPath.Child("mountPath"), mnt.MountPath, "must be unique"))
 | |
| 		}
 | |
| 		mountpoints.Insert(mnt.MountPath)
 | |
| 
 | |
| 		// check for overlap with VolumeDevice
 | |
| 		if mountNameAlreadyExists(mnt.Name, voldevices) {
 | |
| 			allErrs = append(allErrs, field.Invalid(idxPath.Child("name"), mnt.Name, "must not already exist in volumeDevices"))
 | |
| 		}
 | |
| 		if mountPathAlreadyExists(mnt.MountPath, voldevices) {
 | |
| 			allErrs = append(allErrs, field.Invalid(idxPath.Child("mountPath"), mnt.MountPath, "must not already exist as a path in volumeDevices"))
 | |
| 		}
 | |
| 
 | |
| 		if len(mnt.SubPath) > 0 {
 | |
| 			allErrs = append(allErrs, validateLocalDescendingPath(mnt.SubPath, fldPath.Child("subPath"))...)
 | |
| 		}
 | |
| 
 | |
| 		if len(mnt.SubPathExpr) > 0 {
 | |
| 			if len(mnt.SubPath) > 0 {
 | |
| 				allErrs = append(allErrs, field.Invalid(idxPath.Child("subPathExpr"), mnt.SubPathExpr, "subPathExpr and subPath are mutually exclusive"))
 | |
| 			}
 | |
| 
 | |
| 			allErrs = append(allErrs, validateLocalDescendingPath(mnt.SubPathExpr, fldPath.Child("subPathExpr"))...)
 | |
| 		}
 | |
| 
 | |
| 		if mnt.MountPropagation != nil {
 | |
| 			allErrs = append(allErrs, validateMountPropagation(mnt.MountPropagation, container, fldPath.Child("mountPropagation"))...)
 | |
| 		}
 | |
| 	}
 | |
| 	return allErrs
 | |
| }
 | |
| 
 | |
| func ValidateVolumeDevices(devices []core.VolumeDevice, volmounts map[string]string, volumes map[string]core.VolumeSource, fldPath *field.Path) field.ErrorList {
 | |
| 	allErrs := field.ErrorList{}
 | |
| 	devicepath := sets.NewString()
 | |
| 	devicename := sets.NewString()
 | |
| 
 | |
| 	for i, dev := range devices {
 | |
| 		idxPath := fldPath.Index(i)
 | |
| 		devName := dev.Name
 | |
| 		devPath := dev.DevicePath
 | |
| 		didMatch, isPVC := isMatchedDevice(devName, volumes)
 | |
| 		if len(devName) == 0 {
 | |
| 			allErrs = append(allErrs, field.Required(idxPath.Child("name"), ""))
 | |
| 		}
 | |
| 		if devicename.Has(devName) {
 | |
| 			allErrs = append(allErrs, field.Invalid(idxPath.Child("name"), devName, "must be unique"))
 | |
| 		}
 | |
| 		// Must be PersistentVolumeClaim volume source
 | |
| 		if didMatch && !isPVC {
 | |
| 			allErrs = append(allErrs, field.Invalid(idxPath.Child("name"), devName, "can only use volume source type of PersistentVolumeClaim for block mode"))
 | |
| 		}
 | |
| 		if !didMatch {
 | |
| 			allErrs = append(allErrs, field.NotFound(idxPath.Child("name"), devName))
 | |
| 		}
 | |
| 		if len(devPath) == 0 {
 | |
| 			allErrs = append(allErrs, field.Required(idxPath.Child("devicePath"), ""))
 | |
| 		}
 | |
| 		if devicepath.Has(devPath) {
 | |
| 			allErrs = append(allErrs, field.Invalid(idxPath.Child("devicePath"), devPath, "must be unique"))
 | |
| 		}
 | |
| 		if len(devPath) > 0 && len(validatePathNoBacksteps(devPath, fldPath.Child("devicePath"))) > 0 {
 | |
| 			allErrs = append(allErrs, field.Invalid(idxPath.Child("devicePath"), devPath, "can not contain backsteps ('..')"))
 | |
| 		} else {
 | |
| 			devicepath.Insert(devPath)
 | |
| 		}
 | |
| 		// check for overlap with VolumeMount
 | |
| 		if deviceNameAlreadyExists(devName, volmounts) {
 | |
| 			allErrs = append(allErrs, field.Invalid(idxPath.Child("name"), devName, "must not already exist in volumeMounts"))
 | |
| 		}
 | |
| 		if devicePathAlreadyExists(devPath, volmounts) {
 | |
| 			allErrs = append(allErrs, field.Invalid(idxPath.Child("devicePath"), devPath, "must not already exist as a path in volumeMounts"))
 | |
| 		}
 | |
| 		if len(devName) > 0 {
 | |
| 			devicename.Insert(devName)
 | |
| 		}
 | |
| 	}
 | |
| 	return allErrs
 | |
| }
 | |
| 
 | |
| func validateProbe(probe *core.Probe, fldPath *field.Path) field.ErrorList {
 | |
| 	allErrs := field.ErrorList{}
 | |
| 
 | |
| 	if probe == nil {
 | |
| 		return allErrs
 | |
| 	}
 | |
| 	allErrs = append(allErrs, validateHandler(&probe.Handler, fldPath)...)
 | |
| 
 | |
| 	allErrs = append(allErrs, ValidateNonnegativeField(int64(probe.InitialDelaySeconds), fldPath.Child("initialDelaySeconds"))...)
 | |
| 	allErrs = append(allErrs, ValidateNonnegativeField(int64(probe.TimeoutSeconds), fldPath.Child("timeoutSeconds"))...)
 | |
| 	allErrs = append(allErrs, ValidateNonnegativeField(int64(probe.PeriodSeconds), fldPath.Child("periodSeconds"))...)
 | |
| 	allErrs = append(allErrs, ValidateNonnegativeField(int64(probe.SuccessThreshold), fldPath.Child("successThreshold"))...)
 | |
| 	allErrs = append(allErrs, ValidateNonnegativeField(int64(probe.FailureThreshold), fldPath.Child("failureThreshold"))...)
 | |
| 	return allErrs
 | |
| }
 | |
| 
 | |
| func validateClientIPAffinityConfig(config *core.SessionAffinityConfig, fldPath *field.Path) field.ErrorList {
 | |
| 	allErrs := field.ErrorList{}
 | |
| 	if config == nil {
 | |
| 		allErrs = append(allErrs, field.Required(fldPath, fmt.Sprintf("when session affinity type is %s", core.ServiceAffinityClientIP)))
 | |
| 		return allErrs
 | |
| 	}
 | |
| 	if config.ClientIP == nil {
 | |
| 		allErrs = append(allErrs, field.Required(fldPath.Child("clientIP"), fmt.Sprintf("when session affinity type is %s", core.ServiceAffinityClientIP)))
 | |
| 		return allErrs
 | |
| 	}
 | |
| 	if config.ClientIP.TimeoutSeconds == nil {
 | |
| 		allErrs = append(allErrs, field.Required(fldPath.Child("clientIP").Child("timeoutSeconds"), fmt.Sprintf("when session affinity type is %s", core.ServiceAffinityClientIP)))
 | |
| 		return allErrs
 | |
| 	}
 | |
| 	allErrs = append(allErrs, validateAffinityTimeout(config.ClientIP.TimeoutSeconds, fldPath.Child("clientIP").Child("timeoutSeconds"))...)
 | |
| 
 | |
| 	return allErrs
 | |
| }
 | |
| 
 | |
| func validateAffinityTimeout(timeout *int32, fldPath *field.Path) field.ErrorList {
 | |
| 	allErrs := field.ErrorList{}
 | |
| 	if *timeout <= 0 || *timeout > core.MaxClientIPServiceAffinitySeconds {
 | |
| 		allErrs = append(allErrs, field.Invalid(fldPath, timeout, fmt.Sprintf("must be greater than 0 and less than %d", core.MaxClientIPServiceAffinitySeconds)))
 | |
| 	}
 | |
| 	return allErrs
 | |
| }
 | |
| 
 | |
| // AccumulateUniqueHostPorts extracts each HostPort of each Container,
 | |
| // accumulating the results and returning an error if any ports conflict.
 | |
| func AccumulateUniqueHostPorts(containers []core.Container, accumulator *sets.String, fldPath *field.Path) field.ErrorList {
 | |
| 	allErrs := field.ErrorList{}
 | |
| 
 | |
| 	for ci, ctr := range containers {
 | |
| 		idxPath := fldPath.Index(ci)
 | |
| 		portsPath := idxPath.Child("ports")
 | |
| 		for pi := range ctr.Ports {
 | |
| 			idxPath := portsPath.Index(pi)
 | |
| 			port := ctr.Ports[pi].HostPort
 | |
| 			if port == 0 {
 | |
| 				continue
 | |
| 			}
 | |
| 			str := fmt.Sprintf("%s/%s/%d", ctr.Ports[pi].Protocol, ctr.Ports[pi].HostIP, port)
 | |
| 			if accumulator.Has(str) {
 | |
| 				allErrs = append(allErrs, field.Duplicate(idxPath.Child("hostPort"), str))
 | |
| 			} else {
 | |
| 				accumulator.Insert(str)
 | |
| 			}
 | |
| 		}
 | |
| 	}
 | |
| 	return allErrs
 | |
| }
 | |
| 
 | |
| // checkHostPortConflicts checks for colliding Port.HostPort values across
 | |
| // a slice of containers.
 | |
| func checkHostPortConflicts(containers []core.Container, fldPath *field.Path) field.ErrorList {
 | |
| 	allPorts := sets.String{}
 | |
| 	return AccumulateUniqueHostPorts(containers, &allPorts, fldPath)
 | |
| }
 | |
| 
 | |
| func validateExecAction(exec *core.ExecAction, fldPath *field.Path) field.ErrorList {
 | |
| 	allErrors := field.ErrorList{}
 | |
| 	if len(exec.Command) == 0 {
 | |
| 		allErrors = append(allErrors, field.Required(fldPath.Child("command"), ""))
 | |
| 	}
 | |
| 	return allErrors
 | |
| }
 | |
| 
 | |
| var supportedHTTPSchemes = sets.NewString(string(core.URISchemeHTTP), string(core.URISchemeHTTPS))
 | |
| 
 | |
| func validateHTTPGetAction(http *core.HTTPGetAction, fldPath *field.Path) field.ErrorList {
 | |
| 	allErrors := field.ErrorList{}
 | |
| 	if len(http.Path) == 0 {
 | |
| 		allErrors = append(allErrors, field.Required(fldPath.Child("path"), ""))
 | |
| 	}
 | |
| 	allErrors = append(allErrors, ValidatePortNumOrName(http.Port, fldPath.Child("port"))...)
 | |
| 	if !supportedHTTPSchemes.Has(string(http.Scheme)) {
 | |
| 		allErrors = append(allErrors, field.NotSupported(fldPath.Child("scheme"), http.Scheme, supportedHTTPSchemes.List()))
 | |
| 	}
 | |
| 	for _, header := range http.HTTPHeaders {
 | |
| 		for _, msg := range validation.IsHTTPHeaderName(header.Name) {
 | |
| 			allErrors = append(allErrors, field.Invalid(fldPath.Child("httpHeaders"), header.Name, msg))
 | |
| 		}
 | |
| 	}
 | |
| 	return allErrors
 | |
| }
 | |
| 
 | |
| func ValidatePortNumOrName(port intstr.IntOrString, fldPath *field.Path) field.ErrorList {
 | |
| 	allErrs := field.ErrorList{}
 | |
| 	if port.Type == intstr.Int {
 | |
| 		for _, msg := range validation.IsValidPortNum(port.IntValue()) {
 | |
| 			allErrs = append(allErrs, field.Invalid(fldPath, port.IntValue(), msg))
 | |
| 		}
 | |
| 	} else if port.Type == intstr.String {
 | |
| 		for _, msg := range validation.IsValidPortName(port.StrVal) {
 | |
| 			allErrs = append(allErrs, field.Invalid(fldPath, port.StrVal, msg))
 | |
| 		}
 | |
| 	} else {
 | |
| 		allErrs = append(allErrs, field.InternalError(fldPath, fmt.Errorf("unknown type: %v", port.Type)))
 | |
| 	}
 | |
| 	return allErrs
 | |
| }
 | |
| 
 | |
| func validateTCPSocketAction(tcp *core.TCPSocketAction, fldPath *field.Path) field.ErrorList {
 | |
| 	return ValidatePortNumOrName(tcp.Port, fldPath.Child("port"))
 | |
| }
 | |
| 
 | |
| func validateHandler(handler *core.Handler, fldPath *field.Path) field.ErrorList {
 | |
| 	numHandlers := 0
 | |
| 	allErrors := field.ErrorList{}
 | |
| 	if handler.Exec != nil {
 | |
| 		if numHandlers > 0 {
 | |
| 			allErrors = append(allErrors, field.Forbidden(fldPath.Child("exec"), "may not specify more than 1 handler type"))
 | |
| 		} else {
 | |
| 			numHandlers++
 | |
| 			allErrors = append(allErrors, validateExecAction(handler.Exec, fldPath.Child("exec"))...)
 | |
| 		}
 | |
| 	}
 | |
| 	if handler.HTTPGet != nil {
 | |
| 		if numHandlers > 0 {
 | |
| 			allErrors = append(allErrors, field.Forbidden(fldPath.Child("httpGet"), "may not specify more than 1 handler type"))
 | |
| 		} else {
 | |
| 			numHandlers++
 | |
| 			allErrors = append(allErrors, validateHTTPGetAction(handler.HTTPGet, fldPath.Child("httpGet"))...)
 | |
| 		}
 | |
| 	}
 | |
| 	if handler.TCPSocket != nil {
 | |
| 		if numHandlers > 0 {
 | |
| 			allErrors = append(allErrors, field.Forbidden(fldPath.Child("tcpSocket"), "may not specify more than 1 handler type"))
 | |
| 		} else {
 | |
| 			numHandlers++
 | |
| 			allErrors = append(allErrors, validateTCPSocketAction(handler.TCPSocket, fldPath.Child("tcpSocket"))...)
 | |
| 		}
 | |
| 	}
 | |
| 	if numHandlers == 0 {
 | |
| 		allErrors = append(allErrors, field.Required(fldPath, "must specify a handler type"))
 | |
| 	}
 | |
| 	return allErrors
 | |
| }
 | |
| 
 | |
| func validateLifecycle(lifecycle *core.Lifecycle, fldPath *field.Path) field.ErrorList {
 | |
| 	allErrs := field.ErrorList{}
 | |
| 	if lifecycle.PostStart != nil {
 | |
| 		allErrs = append(allErrs, validateHandler(lifecycle.PostStart, fldPath.Child("postStart"))...)
 | |
| 	}
 | |
| 	if lifecycle.PreStop != nil {
 | |
| 		allErrs = append(allErrs, validateHandler(lifecycle.PreStop, fldPath.Child("preStop"))...)
 | |
| 	}
 | |
| 	return allErrs
 | |
| }
 | |
| 
 | |
| var supportedPullPolicies = sets.NewString(string(core.PullAlways), string(core.PullIfNotPresent), string(core.PullNever))
 | |
| 
 | |
| func validatePullPolicy(policy core.PullPolicy, fldPath *field.Path) field.ErrorList {
 | |
| 	allErrors := field.ErrorList{}
 | |
| 
 | |
| 	switch policy {
 | |
| 	case core.PullAlways, core.PullIfNotPresent, core.PullNever:
 | |
| 		break
 | |
| 	case "":
 | |
| 		allErrors = append(allErrors, field.Required(fldPath, ""))
 | |
| 	default:
 | |
| 		allErrors = append(allErrors, field.NotSupported(fldPath, policy, supportedPullPolicies.List()))
 | |
| 	}
 | |
| 
 | |
| 	return allErrors
 | |
| }
 | |
| 
 | |
| func validateEphemeralContainers(ephemeralContainers []core.EphemeralContainer, containers, initContainers []core.Container, volumes map[string]core.VolumeSource, fldPath *field.Path, opts PodValidationOptions) field.ErrorList {
 | |
| 	allErrs := field.ErrorList{}
 | |
| 
 | |
| 	if len(ephemeralContainers) == 0 {
 | |
| 		return allErrs
 | |
| 	}
 | |
| 
 | |
| 	allNames := sets.String{}
 | |
| 	for _, c := range containers {
 | |
| 		allNames.Insert(c.Name)
 | |
| 	}
 | |
| 	for _, c := range initContainers {
 | |
| 		allNames.Insert(c.Name)
 | |
| 	}
 | |
| 
 | |
| 	for i, ec := range ephemeralContainers {
 | |
| 		idxPath := fldPath.Index(i)
 | |
| 
 | |
| 		if ec.TargetContainerName != "" && !allNames.Has(ec.TargetContainerName) {
 | |
| 			allErrs = append(allErrs, field.NotFound(idxPath.Child("targetContainerName"), ec.TargetContainerName))
 | |
| 		}
 | |
| 
 | |
| 		if ec.Name == "" {
 | |
| 			allErrs = append(allErrs, field.Required(idxPath, "ephemeralContainer requires a name"))
 | |
| 			continue
 | |
| 		}
 | |
| 
 | |
| 		// Using validateContainers() here isn't ideal because it adds an index to the error message that
 | |
| 		// doesn't really exist for EphemeralContainers (i.e. ephemeralContainers[0].spec[0].name instead
 | |
| 		// of ephemeralContainers[0].spec.name)
 | |
| 		// TODO(verb): factor a validateContainer() out of validateContainers() to be used here
 | |
| 		c := core.Container(ec.EphemeralContainerCommon)
 | |
| 		allErrs = append(allErrs, validateContainers([]core.Container{c}, false, volumes, idxPath, opts)...)
 | |
| 		// EphemeralContainers don't require the backwards-compatibility distinction between pod/podTemplate validation
 | |
| 		allErrs = append(allErrs, validateContainersOnlyForPod([]core.Container{c}, idxPath)...)
 | |
| 
 | |
| 		if allNames.Has(ec.Name) {
 | |
| 			allErrs = append(allErrs, field.Duplicate(idxPath.Child("name"), ec.Name))
 | |
| 		} else {
 | |
| 			allNames.Insert(ec.Name)
 | |
| 		}
 | |
| 
 | |
| 		// Ephemeral Containers should not be relied upon for fundamental pod services, so fields such as
 | |
| 		// Lifecycle, probes, resources and ports should be disallowed. This is implemented as a list
 | |
| 		// of allowed fields so that new fields will be given consideration prior to inclusion in Ephemeral Containers.
 | |
| 		allErrs = append(allErrs, validateFieldAllowList(ec.EphemeralContainerCommon, allowedEphemeralContainerFields, "cannot be set for an Ephemeral Container", idxPath)...)
 | |
| 	}
 | |
| 
 | |
| 	return allErrs
 | |
| }
 | |
| 
 | |
| // validateFieldAcceptList checks that only allowed fields are set.
 | |
| // The value must be a struct (not a pointer to a struct!).
 | |
| func validateFieldAllowList(value interface{}, allowedFields map[string]bool, errorText string, fldPath *field.Path) field.ErrorList {
 | |
| 	var allErrs field.ErrorList
 | |
| 
 | |
| 	reflectType, reflectValue := reflect.TypeOf(value), reflect.ValueOf(value)
 | |
| 	for i := 0; i < reflectType.NumField(); i++ {
 | |
| 		f := reflectType.Field(i)
 | |
| 		if allowedFields[f.Name] {
 | |
| 			continue
 | |
| 		}
 | |
| 
 | |
| 		// Compare the value of this field to its zero value to determine if it has been set
 | |
| 		if !reflect.DeepEqual(reflectValue.Field(i).Interface(), reflect.Zero(f.Type).Interface()) {
 | |
| 			r, n := utf8.DecodeRuneInString(f.Name)
 | |
| 			lcName := string(unicode.ToLower(r)) + f.Name[n:]
 | |
| 			allErrs = append(allErrs, field.Forbidden(fldPath.Child(lcName), errorText))
 | |
| 		}
 | |
| 	}
 | |
| 
 | |
| 	return allErrs
 | |
| }
 | |
| 
 | |
| func validateInitContainers(containers, otherContainers []core.Container, deviceVolumes map[string]core.VolumeSource, fldPath *field.Path, opts PodValidationOptions) field.ErrorList {
 | |
| 	var allErrs field.ErrorList
 | |
| 	if len(containers) > 0 {
 | |
| 		allErrs = append(allErrs, validateContainers(containers, true, deviceVolumes, fldPath, opts)...)
 | |
| 	}
 | |
| 
 | |
| 	allNames := sets.String{}
 | |
| 	for _, ctr := range otherContainers {
 | |
| 		allNames.Insert(ctr.Name)
 | |
| 	}
 | |
| 	for i, ctr := range containers {
 | |
| 		idxPath := fldPath.Index(i)
 | |
| 		if allNames.Has(ctr.Name) {
 | |
| 			allErrs = append(allErrs, field.Duplicate(idxPath.Child("name"), ctr.Name))
 | |
| 		}
 | |
| 		if len(ctr.Name) > 0 {
 | |
| 			allNames.Insert(ctr.Name)
 | |
| 		}
 | |
| 		if ctr.Lifecycle != nil {
 | |
| 			allErrs = append(allErrs, field.Invalid(idxPath.Child("lifecycle"), ctr.Lifecycle, "must not be set for init containers"))
 | |
| 		}
 | |
| 		if ctr.LivenessProbe != nil {
 | |
| 			allErrs = append(allErrs, field.Invalid(idxPath.Child("livenessProbe"), ctr.LivenessProbe, "must not be set for init containers"))
 | |
| 		}
 | |
| 		if ctr.ReadinessProbe != nil {
 | |
| 			allErrs = append(allErrs, field.Invalid(idxPath.Child("readinessProbe"), ctr.ReadinessProbe, "must not be set for init containers"))
 | |
| 		}
 | |
| 		if ctr.StartupProbe != nil {
 | |
| 			allErrs = append(allErrs, field.Invalid(idxPath.Child("startupProbe"), ctr.StartupProbe, "must not be set for init containers"))
 | |
| 		}
 | |
| 	}
 | |
| 	return allErrs
 | |
| }
 | |
| 
 | |
| func validateContainers(containers []core.Container, isInitContainers bool, volumes map[string]core.VolumeSource, fldPath *field.Path, opts PodValidationOptions) field.ErrorList {
 | |
| 	allErrs := field.ErrorList{}
 | |
| 
 | |
| 	if len(containers) == 0 {
 | |
| 		return append(allErrs, field.Required(fldPath, ""))
 | |
| 	}
 | |
| 
 | |
| 	allNames := sets.String{}
 | |
| 	for i, ctr := range containers {
 | |
| 		idxPath := fldPath.Index(i)
 | |
| 		namePath := idxPath.Child("name")
 | |
| 		volMounts := GetVolumeMountMap(ctr.VolumeMounts)
 | |
| 		volDevices := GetVolumeDeviceMap(ctr.VolumeDevices)
 | |
| 
 | |
| 		if len(ctr.Name) == 0 {
 | |
| 			allErrs = append(allErrs, field.Required(namePath, ""))
 | |
| 		} else {
 | |
| 			allErrs = append(allErrs, ValidateDNS1123Label(ctr.Name, namePath)...)
 | |
| 		}
 | |
| 		if allNames.Has(ctr.Name) {
 | |
| 			allErrs = append(allErrs, field.Duplicate(namePath, ctr.Name))
 | |
| 		} else {
 | |
| 			allNames.Insert(ctr.Name)
 | |
| 		}
 | |
| 		// TODO: do not validate leading and trailing whitespace to preserve backward compatibility.
 | |
| 		// for example: https://github.com/openshift/origin/issues/14659 image = " " is special token in pod template
 | |
| 		// others may have done similar
 | |
| 		if len(ctr.Image) == 0 {
 | |
| 			allErrs = append(allErrs, field.Required(idxPath.Child("image"), ""))
 | |
| 		}
 | |
| 		if ctr.Lifecycle != nil {
 | |
| 			allErrs = append(allErrs, validateLifecycle(ctr.Lifecycle, idxPath.Child("lifecycle"))...)
 | |
| 		}
 | |
| 		allErrs = append(allErrs, validateProbe(ctr.LivenessProbe, idxPath.Child("livenessProbe"))...)
 | |
| 		// Readiness-specific validation
 | |
| 		if ctr.ReadinessProbe != nil && ctr.ReadinessProbe.TerminationGracePeriodSeconds != nil {
 | |
| 			allErrs = append(allErrs, field.Invalid(idxPath.Child("readinessProbe", "terminationGracePeriodSeconds"), ctr.ReadinessProbe.TerminationGracePeriodSeconds, "must not be set for readinessProbes"))
 | |
| 		}
 | |
| 		allErrs = append(allErrs, validateProbe(ctr.StartupProbe, idxPath.Child("startupProbe"))...)
 | |
| 		// Liveness-specific validation
 | |
| 		if ctr.LivenessProbe != nil && ctr.LivenessProbe.SuccessThreshold != 1 {
 | |
| 			allErrs = append(allErrs, field.Invalid(idxPath.Child("livenessProbe", "successThreshold"), ctr.LivenessProbe.SuccessThreshold, "must be 1"))
 | |
| 		}
 | |
| 		allErrs = append(allErrs, validateProbe(ctr.StartupProbe, idxPath.Child("startupProbe"))...)
 | |
| 		// Startup-specific validation
 | |
| 		if ctr.StartupProbe != nil && ctr.StartupProbe.SuccessThreshold != 1 {
 | |
| 			allErrs = append(allErrs, field.Invalid(idxPath.Child("startupProbe", "successThreshold"), ctr.StartupProbe.SuccessThreshold, "must be 1"))
 | |
| 		}
 | |
| 
 | |
| 		switch ctr.TerminationMessagePolicy {
 | |
| 		case core.TerminationMessageReadFile, core.TerminationMessageFallbackToLogsOnError:
 | |
| 		case "":
 | |
| 			allErrs = append(allErrs, field.Required(idxPath.Child("terminationMessagePolicy"), "must be 'File' or 'FallbackToLogsOnError'"))
 | |
| 		default:
 | |
| 			allErrs = append(allErrs, field.Invalid(idxPath.Child("terminationMessagePolicy"), ctr.TerminationMessagePolicy, "must be 'File' or 'FallbackToLogsOnError'"))
 | |
| 		}
 | |
| 
 | |
| 		allErrs = append(allErrs, validateProbe(ctr.ReadinessProbe, idxPath.Child("readinessProbe"))...)
 | |
| 		allErrs = append(allErrs, validateContainerPorts(ctr.Ports, idxPath.Child("ports"))...)
 | |
| 		allErrs = append(allErrs, ValidateEnv(ctr.Env, idxPath.Child("env"), opts)...)
 | |
| 		allErrs = append(allErrs, ValidateEnvFrom(ctr.EnvFrom, idxPath.Child("envFrom"))...)
 | |
| 		allErrs = append(allErrs, ValidateVolumeMounts(ctr.VolumeMounts, volDevices, volumes, &ctr, idxPath.Child("volumeMounts"))...)
 | |
| 		allErrs = append(allErrs, ValidateVolumeDevices(ctr.VolumeDevices, volMounts, volumes, idxPath.Child("volumeDevices"))...)
 | |
| 		allErrs = append(allErrs, validatePullPolicy(ctr.ImagePullPolicy, idxPath.Child("imagePullPolicy"))...)
 | |
| 		allErrs = append(allErrs, ValidateResourceRequirements(&ctr.Resources, idxPath.Child("resources"), opts)...)
 | |
| 		allErrs = append(allErrs, ValidateSecurityContext(ctr.SecurityContext, idxPath.Child("securityContext"))...)
 | |
| 	}
 | |
| 
 | |
| 	if isInitContainers {
 | |
| 		// check initContainers one by one since they are running in sequential order.
 | |
| 		for _, initContainer := range containers {
 | |
| 			allErrs = append(allErrs, checkHostPortConflicts([]core.Container{initContainer}, fldPath)...)
 | |
| 		}
 | |
| 	} else {
 | |
| 		// Check for colliding ports across all containers.
 | |
| 		allErrs = append(allErrs, checkHostPortConflicts(containers, fldPath)...)
 | |
| 	}
 | |
| 
 | |
| 	return allErrs
 | |
| }
 | |
| 
 | |
| func validateRestartPolicy(restartPolicy *core.RestartPolicy, fldPath *field.Path) field.ErrorList {
 | |
| 	allErrors := field.ErrorList{}
 | |
| 	switch *restartPolicy {
 | |
| 	case core.RestartPolicyAlways, core.RestartPolicyOnFailure, core.RestartPolicyNever:
 | |
| 		break
 | |
| 	case "":
 | |
| 		allErrors = append(allErrors, field.Required(fldPath, ""))
 | |
| 	default:
 | |
| 		validValues := []string{string(core.RestartPolicyAlways), string(core.RestartPolicyOnFailure), string(core.RestartPolicyNever)}
 | |
| 		allErrors = append(allErrors, field.NotSupported(fldPath, *restartPolicy, validValues))
 | |
| 	}
 | |
| 
 | |
| 	return allErrors
 | |
| }
 | |
| 
 | |
| func ValidatePreemptionPolicy(preemptionPolicy *core.PreemptionPolicy, fldPath *field.Path) field.ErrorList {
 | |
| 	allErrors := field.ErrorList{}
 | |
| 	switch *preemptionPolicy {
 | |
| 	case core.PreemptLowerPriority, core.PreemptNever:
 | |
| 	case "":
 | |
| 		allErrors = append(allErrors, field.Required(fldPath, ""))
 | |
| 	default:
 | |
| 		validValues := []string{string(core.PreemptLowerPriority), string(core.PreemptNever)}
 | |
| 		allErrors = append(allErrors, field.NotSupported(fldPath, preemptionPolicy, validValues))
 | |
| 	}
 | |
| 	return allErrors
 | |
| }
 | |
| 
 | |
| func validateDNSPolicy(dnsPolicy *core.DNSPolicy, fldPath *field.Path) field.ErrorList {
 | |
| 	allErrors := field.ErrorList{}
 | |
| 	switch *dnsPolicy {
 | |
| 	case core.DNSClusterFirstWithHostNet, core.DNSClusterFirst, core.DNSDefault, core.DNSNone:
 | |
| 	case "":
 | |
| 		allErrors = append(allErrors, field.Required(fldPath, ""))
 | |
| 	default:
 | |
| 		validValues := []string{string(core.DNSClusterFirstWithHostNet), string(core.DNSClusterFirst), string(core.DNSDefault), string(core.DNSNone)}
 | |
| 		allErrors = append(allErrors, field.NotSupported(fldPath, dnsPolicy, validValues))
 | |
| 	}
 | |
| 	return allErrors
 | |
| }
 | |
| 
 | |
| var validFSGroupChangePolicies = sets.NewString(string(core.FSGroupChangeOnRootMismatch), string(core.FSGroupChangeAlways))
 | |
| 
 | |
| func validateFSGroupChangePolicy(fsGroupPolicy *core.PodFSGroupChangePolicy, fldPath *field.Path) field.ErrorList {
 | |
| 	allErrors := field.ErrorList{}
 | |
| 	if !validFSGroupChangePolicies.Has(string(*fsGroupPolicy)) {
 | |
| 		allErrors = append(allErrors, field.NotSupported(fldPath, fsGroupPolicy, validFSGroupChangePolicies.List()))
 | |
| 	}
 | |
| 	return allErrors
 | |
| }
 | |
| 
 | |
| const (
 | |
| 	// Limits on various DNS parameters. These are derived from
 | |
| 	// restrictions in Linux libc name resolution handling.
 | |
| 	// Max number of DNS name servers.
 | |
| 	MaxDNSNameservers = 3
 | |
| 	// Max number of domains in search path.
 | |
| 	MaxDNSSearchPaths = 6
 | |
| 	// Max number of characters in search path.
 | |
| 	MaxDNSSearchListChars = 256
 | |
| )
 | |
| 
 | |
| func validateReadinessGates(readinessGates []core.PodReadinessGate, fldPath *field.Path) field.ErrorList {
 | |
| 	allErrs := field.ErrorList{}
 | |
| 	for i, value := range readinessGates {
 | |
| 		for _, msg := range validation.IsQualifiedName(string(value.ConditionType)) {
 | |
| 			allErrs = append(allErrs, field.Invalid(fldPath.Index(i).Child("conditionType"), string(value.ConditionType), msg))
 | |
| 		}
 | |
| 	}
 | |
| 	return allErrs
 | |
| }
 | |
| 
 | |
| func validatePodDNSConfig(dnsConfig *core.PodDNSConfig, dnsPolicy *core.DNSPolicy, fldPath *field.Path) field.ErrorList {
 | |
| 	allErrs := field.ErrorList{}
 | |
| 
 | |
| 	// Validate DNSNone case. Must provide at least one DNS name server.
 | |
| 	if dnsPolicy != nil && *dnsPolicy == core.DNSNone {
 | |
| 		if dnsConfig == nil {
 | |
| 			return append(allErrs, field.Required(fldPath, fmt.Sprintf("must provide `dnsConfig` when `dnsPolicy` is %s", core.DNSNone)))
 | |
| 		}
 | |
| 		if len(dnsConfig.Nameservers) == 0 {
 | |
| 			return append(allErrs, field.Required(fldPath.Child("nameservers"), fmt.Sprintf("must provide at least one DNS nameserver when `dnsPolicy` is %s", core.DNSNone)))
 | |
| 		}
 | |
| 	}
 | |
| 
 | |
| 	if dnsConfig != nil {
 | |
| 		// Validate nameservers.
 | |
| 		if len(dnsConfig.Nameservers) > MaxDNSNameservers {
 | |
| 			allErrs = append(allErrs, field.Invalid(fldPath.Child("nameservers"), dnsConfig.Nameservers, fmt.Sprintf("must not have more than %v nameservers", MaxDNSNameservers)))
 | |
| 		}
 | |
| 		for i, ns := range dnsConfig.Nameservers {
 | |
| 			if ip := net.ParseIP(ns); ip == nil {
 | |
| 				allErrs = append(allErrs, field.Invalid(fldPath.Child("nameservers").Index(i), ns, "must be valid IP address"))
 | |
| 			}
 | |
| 		}
 | |
| 		// Validate searches.
 | |
| 		if len(dnsConfig.Searches) > MaxDNSSearchPaths {
 | |
| 			allErrs = append(allErrs, field.Invalid(fldPath.Child("searches"), dnsConfig.Searches, fmt.Sprintf("must not have more than %v search paths", MaxDNSSearchPaths)))
 | |
| 		}
 | |
| 		// Include the space between search paths.
 | |
| 		if len(strings.Join(dnsConfig.Searches, " ")) > MaxDNSSearchListChars {
 | |
| 			allErrs = append(allErrs, field.Invalid(fldPath.Child("searches"), dnsConfig.Searches, "must not have more than 256 characters (including spaces) in the search list"))
 | |
| 		}
 | |
| 		for i, search := range dnsConfig.Searches {
 | |
| 			// it is fine to have a trailing dot
 | |
| 			if strings.HasSuffix(search, ".") {
 | |
| 				search = search[0 : len(search)-1]
 | |
| 			}
 | |
| 			allErrs = append(allErrs, ValidateDNS1123Subdomain(search, fldPath.Child("searches").Index(i))...)
 | |
| 		}
 | |
| 		// Validate options.
 | |
| 		for i, option := range dnsConfig.Options {
 | |
| 			if len(option.Name) == 0 {
 | |
| 				allErrs = append(allErrs, field.Required(fldPath.Child("options").Index(i), "must not be empty"))
 | |
| 			}
 | |
| 		}
 | |
| 	}
 | |
| 	return allErrs
 | |
| }
 | |
| 
 | |
| func validateHostNetwork(hostNetwork bool, containers []core.Container, fldPath *field.Path) field.ErrorList {
 | |
| 	allErrors := field.ErrorList{}
 | |
| 	if hostNetwork {
 | |
| 		for i, container := range containers {
 | |
| 			portsPath := fldPath.Index(i).Child("ports")
 | |
| 			for i, port := range container.Ports {
 | |
| 				idxPath := portsPath.Index(i)
 | |
| 				if port.HostPort != port.ContainerPort {
 | |
| 					allErrors = append(allErrors, field.Invalid(idxPath.Child("containerPort"), port.ContainerPort, "must match `hostPort` when `hostNetwork` is true"))
 | |
| 				}
 | |
| 			}
 | |
| 		}
 | |
| 	}
 | |
| 	return allErrors
 | |
| }
 | |
| 
 | |
| // validateImagePullSecrets checks to make sure the pull secrets are well
 | |
| // formed.  Right now, we only expect name to be set (it's the only field).  If
 | |
| // this ever changes and someone decides to set those fields, we'd like to
 | |
| // know.
 | |
| func validateImagePullSecrets(imagePullSecrets []core.LocalObjectReference, fldPath *field.Path) field.ErrorList {
 | |
| 	allErrors := field.ErrorList{}
 | |
| 	for i, currPullSecret := range imagePullSecrets {
 | |
| 		idxPath := fldPath.Index(i)
 | |
| 		strippedRef := core.LocalObjectReference{Name: currPullSecret.Name}
 | |
| 		if !reflect.DeepEqual(strippedRef, currPullSecret) {
 | |
| 			allErrors = append(allErrors, field.Invalid(idxPath, currPullSecret, "only name may be set"))
 | |
| 		}
 | |
| 	}
 | |
| 	return allErrors
 | |
| }
 | |
| 
 | |
| // validateAffinity checks if given affinities are valid
 | |
| func validateAffinity(affinity *core.Affinity, fldPath *field.Path) field.ErrorList {
 | |
| 	allErrs := field.ErrorList{}
 | |
| 
 | |
| 	if affinity != nil {
 | |
| 		if affinity.NodeAffinity != nil {
 | |
| 			allErrs = append(allErrs, validateNodeAffinity(affinity.NodeAffinity, fldPath.Child("nodeAffinity"))...)
 | |
| 		}
 | |
| 		if affinity.PodAffinity != nil {
 | |
| 			allErrs = append(allErrs, validatePodAffinity(affinity.PodAffinity, fldPath.Child("podAffinity"))...)
 | |
| 		}
 | |
| 		if affinity.PodAntiAffinity != nil {
 | |
| 			allErrs = append(allErrs, validatePodAntiAffinity(affinity.PodAntiAffinity, fldPath.Child("podAntiAffinity"))...)
 | |
| 		}
 | |
| 	}
 | |
| 
 | |
| 	return allErrs
 | |
| }
 | |
| 
 | |
| func validateTaintEffect(effect *core.TaintEffect, allowEmpty bool, fldPath *field.Path) field.ErrorList {
 | |
| 	if !allowEmpty && len(*effect) == 0 {
 | |
| 		return field.ErrorList{field.Required(fldPath, "")}
 | |
| 	}
 | |
| 
 | |
| 	allErrors := field.ErrorList{}
 | |
| 	switch *effect {
 | |
| 	// TODO: Replace next line with subsequent commented-out line when implement TaintEffectNoScheduleNoAdmit.
 | |
| 	case core.TaintEffectNoSchedule, core.TaintEffectPreferNoSchedule, core.TaintEffectNoExecute:
 | |
| 		// case core.TaintEffectNoSchedule, core.TaintEffectPreferNoSchedule, core.TaintEffectNoScheduleNoAdmit, core.TaintEffectNoExecute:
 | |
| 	default:
 | |
| 		validValues := []string{
 | |
| 			string(core.TaintEffectNoSchedule),
 | |
| 			string(core.TaintEffectPreferNoSchedule),
 | |
| 			string(core.TaintEffectNoExecute),
 | |
| 			// TODO: Uncomment this block when implement TaintEffectNoScheduleNoAdmit.
 | |
| 			// string(core.TaintEffectNoScheduleNoAdmit),
 | |
| 		}
 | |
| 		allErrors = append(allErrors, field.NotSupported(fldPath, *effect, validValues))
 | |
| 	}
 | |
| 	return allErrors
 | |
| }
 | |
| 
 | |
| // validateOnlyAddedTolerations validates updated pod tolerations.
 | |
| func validateOnlyAddedTolerations(newTolerations []core.Toleration, oldTolerations []core.Toleration, fldPath *field.Path) field.ErrorList {
 | |
| 	allErrs := field.ErrorList{}
 | |
| 	for _, old := range oldTolerations {
 | |
| 		found := false
 | |
| 		oldTolerationClone := old.DeepCopy()
 | |
| 		for _, newToleration := range newTolerations {
 | |
| 			// assign to our clone before doing a deep equal so we can allow tolerationseconds to change.
 | |
| 			oldTolerationClone.TolerationSeconds = newToleration.TolerationSeconds // +k8s:verify-mutation:reason=clone
 | |
| 			if reflect.DeepEqual(*oldTolerationClone, newToleration) {
 | |
| 				found = true
 | |
| 				break
 | |
| 			}
 | |
| 		}
 | |
| 		if !found {
 | |
| 			allErrs = append(allErrs, field.Forbidden(fldPath, "existing toleration can not be modified except its tolerationSeconds"))
 | |
| 			return allErrs
 | |
| 		}
 | |
| 	}
 | |
| 
 | |
| 	allErrs = append(allErrs, ValidateTolerations(newTolerations, fldPath)...)
 | |
| 	return allErrs
 | |
| }
 | |
| 
 | |
| func ValidateHostAliases(hostAliases []core.HostAlias, fldPath *field.Path) field.ErrorList {
 | |
| 	allErrs := field.ErrorList{}
 | |
| 	for _, hostAlias := range hostAliases {
 | |
| 		if ip := net.ParseIP(hostAlias.IP); ip == nil {
 | |
| 			allErrs = append(allErrs, field.Invalid(fldPath.Child("ip"), hostAlias.IP, "must be valid IP address"))
 | |
| 		}
 | |
| 		for _, hostname := range hostAlias.Hostnames {
 | |
| 			allErrs = append(allErrs, ValidateDNS1123Subdomain(hostname, fldPath.Child("hostnames"))...)
 | |
| 		}
 | |
| 	}
 | |
| 	return allErrs
 | |
| }
 | |
| 
 | |
| // ValidateTolerations tests if given tolerations have valid data.
 | |
| func ValidateTolerations(tolerations []core.Toleration, fldPath *field.Path) field.ErrorList {
 | |
| 	allErrors := field.ErrorList{}
 | |
| 	for i, toleration := range tolerations {
 | |
| 		idxPath := fldPath.Index(i)
 | |
| 		// validate the toleration key
 | |
| 		if len(toleration.Key) > 0 {
 | |
| 			allErrors = append(allErrors, unversionedvalidation.ValidateLabelName(toleration.Key, idxPath.Child("key"))...)
 | |
| 		}
 | |
| 
 | |
| 		// empty toleration key with Exists operator and empty value means match all taints
 | |
| 		if len(toleration.Key) == 0 && toleration.Operator != core.TolerationOpExists {
 | |
| 			allErrors = append(allErrors, field.Invalid(idxPath.Child("operator"), toleration.Operator,
 | |
| 				"operator must be Exists when `key` is empty, which means \"match all values and all keys\""))
 | |
| 		}
 | |
| 
 | |
| 		if toleration.TolerationSeconds != nil && toleration.Effect != core.TaintEffectNoExecute {
 | |
| 			allErrors = append(allErrors, field.Invalid(idxPath.Child("effect"), toleration.Effect,
 | |
| 				"effect must be 'NoExecute' when `tolerationSeconds` is set"))
 | |
| 		}
 | |
| 
 | |
| 		// validate toleration operator and value
 | |
| 		switch toleration.Operator {
 | |
| 		// empty operator means Equal
 | |
| 		case core.TolerationOpEqual, "":
 | |
| 			if errs := validation.IsValidLabelValue(toleration.Value); len(errs) != 0 {
 | |
| 				allErrors = append(allErrors, field.Invalid(idxPath.Child("operator"), toleration.Value, strings.Join(errs, ";")))
 | |
| 			}
 | |
| 		case core.TolerationOpExists:
 | |
| 			if len(toleration.Value) > 0 {
 | |
| 				allErrors = append(allErrors, field.Invalid(idxPath.Child("operator"), toleration, "value must be empty when `operator` is 'Exists'"))
 | |
| 			}
 | |
| 		default:
 | |
| 			validValues := []string{string(core.TolerationOpEqual), string(core.TolerationOpExists)}
 | |
| 			allErrors = append(allErrors, field.NotSupported(idxPath.Child("operator"), toleration.Operator, validValues))
 | |
| 		}
 | |
| 
 | |
| 		// validate toleration effect, empty toleration effect means match all taint effects
 | |
| 		if len(toleration.Effect) > 0 {
 | |
| 			allErrors = append(allErrors, validateTaintEffect(&toleration.Effect, true, idxPath.Child("effect"))...)
 | |
| 		}
 | |
| 	}
 | |
| 	return allErrors
 | |
| }
 | |
| 
 | |
| // validateContainersOnlyForPod does additional validation for containers on a pod versus a pod template
 | |
| // it only does additive validation of fields not covered in validateContainers
 | |
| func validateContainersOnlyForPod(containers []core.Container, fldPath *field.Path) field.ErrorList {
 | |
| 	allErrs := field.ErrorList{}
 | |
| 	for i, ctr := range containers {
 | |
| 		idxPath := fldPath.Index(i)
 | |
| 		if len(ctr.Image) != len(strings.TrimSpace(ctr.Image)) {
 | |
| 			allErrs = append(allErrs, field.Invalid(idxPath.Child("image"), ctr.Image, "must not have leading or trailing whitespace"))
 | |
| 		}
 | |
| 	}
 | |
| 	return allErrs
 | |
| }
 | |
| 
 | |
| // PodValidationOptions contains the different settings for pod validation
 | |
| type PodValidationOptions struct {
 | |
| 	// Allow pod spec to have more than one huge page resource (with different sizes)
 | |
| 	AllowMultipleHugePageResources bool
 | |
| 	// Allow pod spec to use hugepages in downward API
 | |
| 	AllowDownwardAPIHugePages bool
 | |
| 	// Allow invalid pod-deletion-cost annotation value for backward compatibility.
 | |
| 	AllowInvalidPodDeletionCost bool
 | |
| 	// Allow pod spec to use non-integer multiple of huge page unit size
 | |
| 	AllowIndivisibleHugePagesValues bool
 | |
| 	// Allow hostProcess field to be set in windows security context
 | |
| 	AllowWindowsHostProcessField bool
 | |
| }
 | |
| 
 | |
| // ValidatePodSingleHugePageResources checks if there are multiple huge
 | |
| // pages resources in the pod object.
 | |
| func ValidatePodSingleHugePageResources(pod *core.Pod, specPath *field.Path) field.ErrorList {
 | |
| 	allErrs := field.ErrorList{}
 | |
| 	resourceSet := helper.ToPodResourcesSet(&pod.Spec)
 | |
| 	hugePageResources := sets.NewString()
 | |
| 	for resourceStr := range resourceSet {
 | |
| 		if v1helper.IsHugePageResourceName(v1.ResourceName(resourceStr)) {
 | |
| 			hugePageResources.Insert(resourceStr)
 | |
| 		}
 | |
| 	}
 | |
| 	if len(hugePageResources) > 1 {
 | |
| 		allErrs = append(allErrs, field.Invalid(specPath, hugePageResources.List(), "must use a single hugepage size in a pod spec"))
 | |
| 	}
 | |
| 	return allErrs
 | |
| }
 | |
| 
 | |
| // validatePodMetadataAndSpec tests if required fields in the pod.metadata and pod.spec are set,
 | |
| // and is called by ValidatePodCreate and ValidatePodUpdate.
 | |
| func validatePodMetadataAndSpec(pod *core.Pod, opts PodValidationOptions) field.ErrorList {
 | |
| 	fldPath := field.NewPath("metadata")
 | |
| 	allErrs := ValidateObjectMeta(&pod.ObjectMeta, true, ValidatePodName, fldPath)
 | |
| 	allErrs = append(allErrs, ValidatePodSpecificAnnotations(pod.ObjectMeta.Annotations, &pod.Spec, fldPath.Child("annotations"), opts)...)
 | |
| 	allErrs = append(allErrs, ValidatePodSpec(&pod.Spec, &pod.ObjectMeta, field.NewPath("spec"), opts)...)
 | |
| 
 | |
| 	// we do additional validation only pertinent for pods and not pod templates
 | |
| 	// this was done to preserve backwards compatibility
 | |
| 	specPath := field.NewPath("spec")
 | |
| 
 | |
| 	if pod.Spec.ServiceAccountName == "" {
 | |
| 		for vi, volume := range pod.Spec.Volumes {
 | |
| 			path := specPath.Child("volumes").Index(vi).Child("projected")
 | |
| 			if volume.Projected != nil {
 | |
| 				for si, source := range volume.Projected.Sources {
 | |
| 					saPath := path.Child("sources").Index(si).Child("serviceAccountToken")
 | |
| 					if source.ServiceAccountToken != nil {
 | |
| 						allErrs = append(allErrs, field.Forbidden(saPath, "must not be specified when serviceAccountName is not set"))
 | |
| 					}
 | |
| 				}
 | |
| 			}
 | |
| 		}
 | |
| 	}
 | |
| 
 | |
| 	allErrs = append(allErrs, validateContainersOnlyForPod(pod.Spec.Containers, specPath.Child("containers"))...)
 | |
| 	allErrs = append(allErrs, validateContainersOnlyForPod(pod.Spec.InitContainers, specPath.Child("initContainers"))...)
 | |
| 
 | |
| 	if !opts.AllowMultipleHugePageResources {
 | |
| 		allErrs = append(allErrs, ValidatePodSingleHugePageResources(pod, specPath)...)
 | |
| 	}
 | |
| 
 | |
| 	return allErrs
 | |
| }
 | |
| 
 | |
| // validatePodIPs validates IPs in pod status
 | |
| func validatePodIPs(pod *core.Pod) field.ErrorList {
 | |
| 	allErrs := field.ErrorList{}
 | |
| 
 | |
| 	podIPsField := field.NewPath("status", "podIPs")
 | |
| 
 | |
| 	// all PodIPs must be valid IPs
 | |
| 	for i, podIP := range pod.Status.PodIPs {
 | |
| 		for _, msg := range validation.IsValidIP(podIP.IP) {
 | |
| 			allErrs = append(allErrs, field.Invalid(podIPsField.Index(i), podIP.IP, msg))
 | |
| 		}
 | |
| 	}
 | |
| 
 | |
| 	// if we have more than one Pod.PodIP then
 | |
| 	// - validate for dual stack
 | |
| 	// - validate for duplication
 | |
| 	if len(pod.Status.PodIPs) > 1 {
 | |
| 		podIPs := make([]string, 0, len(pod.Status.PodIPs))
 | |
| 		for _, podIP := range pod.Status.PodIPs {
 | |
| 			podIPs = append(podIPs, podIP.IP)
 | |
| 		}
 | |
| 
 | |
| 		dualStack, err := netutils.IsDualStackIPStrings(podIPs)
 | |
| 		if err != nil {
 | |
| 			allErrs = append(allErrs, field.InternalError(podIPsField, fmt.Errorf("failed to check for dual stack with error:%v", err)))
 | |
| 		}
 | |
| 
 | |
| 		// We only support one from each IP family (i.e. max two IPs in this list).
 | |
| 		if !dualStack || len(podIPs) > 2 {
 | |
| 			allErrs = append(allErrs, field.Invalid(podIPsField, pod.Status.PodIPs, "may specify no more than one IP for each IP family"))
 | |
| 		}
 | |
| 
 | |
| 		// There should be no duplicates in list of Pod.PodIPs
 | |
| 		seen := sets.String{} //:= make(map[string]int)
 | |
| 		for i, podIP := range pod.Status.PodIPs {
 | |
| 			if seen.Has(podIP.IP) {
 | |
| 				allErrs = append(allErrs, field.Duplicate(podIPsField.Index(i), podIP))
 | |
| 			}
 | |
| 			seen.Insert(podIP.IP)
 | |
| 		}
 | |
| 	}
 | |
| 
 | |
| 	return allErrs
 | |
| }
 | |
| 
 | |
| // ValidatePodSpec tests that the specified PodSpec has valid data.
 | |
| // This includes checking formatting and uniqueness.  It also canonicalizes the
 | |
| // structure by setting default values and implementing any backwards-compatibility
 | |
| // tricks.
 | |
| // The pod metadata is needed to validate generic ephemeral volumes. It is optional
 | |
| // and should be left empty unless the spec is from a real pod object.
 | |
| func ValidatePodSpec(spec *core.PodSpec, podMeta *metav1.ObjectMeta, fldPath *field.Path, opts PodValidationOptions) field.ErrorList {
 | |
| 	allErrs := field.ErrorList{}
 | |
| 
 | |
| 	vols, vErrs := ValidateVolumes(spec.Volumes, podMeta, fldPath.Child("volumes"), opts)
 | |
| 	allErrs = append(allErrs, vErrs...)
 | |
| 	allErrs = append(allErrs, validateContainers(spec.Containers, false, vols, fldPath.Child("containers"), opts)...)
 | |
| 	allErrs = append(allErrs, validateInitContainers(spec.InitContainers, spec.Containers, vols, fldPath.Child("initContainers"), opts)...)
 | |
| 	allErrs = append(allErrs, validateEphemeralContainers(spec.EphemeralContainers, spec.Containers, spec.InitContainers, vols, fldPath.Child("ephemeralContainers"), opts)...)
 | |
| 	allErrs = append(allErrs, validateRestartPolicy(&spec.RestartPolicy, fldPath.Child("restartPolicy"))...)
 | |
| 	allErrs = append(allErrs, validateDNSPolicy(&spec.DNSPolicy, fldPath.Child("dnsPolicy"))...)
 | |
| 	allErrs = append(allErrs, unversionedvalidation.ValidateLabels(spec.NodeSelector, fldPath.Child("nodeSelector"))...)
 | |
| 	allErrs = append(allErrs, ValidatePodSecurityContext(spec.SecurityContext, spec, fldPath, fldPath.Child("securityContext"))...)
 | |
| 	allErrs = append(allErrs, validateImagePullSecrets(spec.ImagePullSecrets, fldPath.Child("imagePullSecrets"))...)
 | |
| 	allErrs = append(allErrs, validateAffinity(spec.Affinity, fldPath.Child("affinity"))...)
 | |
| 	allErrs = append(allErrs, validatePodDNSConfig(spec.DNSConfig, &spec.DNSPolicy, fldPath.Child("dnsConfig"))...)
 | |
| 	allErrs = append(allErrs, validateReadinessGates(spec.ReadinessGates, fldPath.Child("readinessGates"))...)
 | |
| 	allErrs = append(allErrs, validateTopologySpreadConstraints(spec.TopologySpreadConstraints, fldPath.Child("topologySpreadConstraints"))...)
 | |
| 	allErrs = append(allErrs, validateWindowsHostProcessPod(spec, fldPath, opts)...)
 | |
| 	if len(spec.ServiceAccountName) > 0 {
 | |
| 		for _, msg := range ValidateServiceAccountName(spec.ServiceAccountName, false) {
 | |
| 			allErrs = append(allErrs, field.Invalid(fldPath.Child("serviceAccountName"), spec.ServiceAccountName, msg))
 | |
| 		}
 | |
| 	}
 | |
| 
 | |
| 	if len(spec.NodeName) > 0 {
 | |
| 		for _, msg := range ValidateNodeName(spec.NodeName, false) {
 | |
| 			allErrs = append(allErrs, field.Invalid(fldPath.Child("nodeName"), spec.NodeName, msg))
 | |
| 		}
 | |
| 	}
 | |
| 
 | |
| 	if spec.ActiveDeadlineSeconds != nil {
 | |
| 		value := *spec.ActiveDeadlineSeconds
 | |
| 		if value < 1 || value > math.MaxInt32 {
 | |
| 			allErrs = append(allErrs, field.Invalid(fldPath.Child("activeDeadlineSeconds"), value, validation.InclusiveRangeError(1, math.MaxInt32)))
 | |
| 		}
 | |
| 	}
 | |
| 
 | |
| 	if len(spec.Hostname) > 0 {
 | |
| 		allErrs = append(allErrs, ValidateDNS1123Label(spec.Hostname, fldPath.Child("hostname"))...)
 | |
| 	}
 | |
| 
 | |
| 	if len(spec.Subdomain) > 0 {
 | |
| 		allErrs = append(allErrs, ValidateDNS1123Label(spec.Subdomain, fldPath.Child("subdomain"))...)
 | |
| 	}
 | |
| 
 | |
| 	if len(spec.Tolerations) > 0 {
 | |
| 		allErrs = append(allErrs, ValidateTolerations(spec.Tolerations, fldPath.Child("tolerations"))...)
 | |
| 	}
 | |
| 
 | |
| 	if len(spec.HostAliases) > 0 {
 | |
| 		allErrs = append(allErrs, ValidateHostAliases(spec.HostAliases, fldPath.Child("hostAliases"))...)
 | |
| 	}
 | |
| 
 | |
| 	if len(spec.PriorityClassName) > 0 {
 | |
| 		for _, msg := range ValidatePriorityClassName(spec.PriorityClassName, false) {
 | |
| 			allErrs = append(allErrs, field.Invalid(fldPath.Child("priorityClassName"), spec.PriorityClassName, msg))
 | |
| 		}
 | |
| 	}
 | |
| 
 | |
| 	if spec.RuntimeClassName != nil {
 | |
| 		allErrs = append(allErrs, ValidateRuntimeClassName(*spec.RuntimeClassName, fldPath.Child("runtimeClassName"))...)
 | |
| 	}
 | |
| 
 | |
| 	if spec.PreemptionPolicy != nil {
 | |
| 		allErrs = append(allErrs, ValidatePreemptionPolicy(spec.PreemptionPolicy, fldPath.Child("preemptionPolicy"))...)
 | |
| 	}
 | |
| 
 | |
| 	if spec.Overhead != nil {
 | |
| 		allErrs = append(allErrs, validateOverhead(spec.Overhead, fldPath.Child("overhead"), opts)...)
 | |
| 	}
 | |
| 
 | |
| 	return allErrs
 | |
| }
 | |
| 
 | |
| // ValidateNodeSelectorRequirement tests that the specified NodeSelectorRequirement fields has valid data
 | |
| func ValidateNodeSelectorRequirement(rq core.NodeSelectorRequirement, fldPath *field.Path) field.ErrorList {
 | |
| 	allErrs := field.ErrorList{}
 | |
| 	switch rq.Operator {
 | |
| 	case core.NodeSelectorOpIn, core.NodeSelectorOpNotIn:
 | |
| 		if len(rq.Values) == 0 {
 | |
| 			allErrs = append(allErrs, field.Required(fldPath.Child("values"), "must be specified when `operator` is 'In' or 'NotIn'"))
 | |
| 		}
 | |
| 	case core.NodeSelectorOpExists, core.NodeSelectorOpDoesNotExist:
 | |
| 		if len(rq.Values) > 0 {
 | |
| 			allErrs = append(allErrs, field.Forbidden(fldPath.Child("values"), "may not be specified when `operator` is 'Exists' or 'DoesNotExist'"))
 | |
| 		}
 | |
| 
 | |
| 	case core.NodeSelectorOpGt, core.NodeSelectorOpLt:
 | |
| 		if len(rq.Values) != 1 {
 | |
| 			allErrs = append(allErrs, field.Required(fldPath.Child("values"), "must be specified single value when `operator` is 'Lt' or 'Gt'"))
 | |
| 		}
 | |
| 	default:
 | |
| 		allErrs = append(allErrs, field.Invalid(fldPath.Child("operator"), rq.Operator, "not a valid selector operator"))
 | |
| 	}
 | |
| 
 | |
| 	allErrs = append(allErrs, unversionedvalidation.ValidateLabelName(rq.Key, fldPath.Child("key"))...)
 | |
| 
 | |
| 	return allErrs
 | |
| }
 | |
| 
 | |
| var nodeFieldSelectorValidators = map[string]func(string, bool) []string{
 | |
| 	metav1.ObjectNameField: ValidateNodeName,
 | |
| }
 | |
| 
 | |
| // ValidateNodeFieldSelectorRequirement tests that the specified NodeSelectorRequirement fields has valid data
 | |
| func ValidateNodeFieldSelectorRequirement(req core.NodeSelectorRequirement, fldPath *field.Path) field.ErrorList {
 | |
| 	allErrs := field.ErrorList{}
 | |
| 
 | |
| 	switch req.Operator {
 | |
| 	case core.NodeSelectorOpIn, core.NodeSelectorOpNotIn:
 | |
| 		if len(req.Values) != 1 {
 | |
| 			allErrs = append(allErrs, field.Required(fldPath.Child("values"),
 | |
| 				"must be only one value when `operator` is 'In' or 'NotIn' for node field selector"))
 | |
| 		}
 | |
| 	default:
 | |
| 		allErrs = append(allErrs, field.Invalid(fldPath.Child("operator"), req.Operator, "not a valid selector operator"))
 | |
| 	}
 | |
| 
 | |
| 	if vf, found := nodeFieldSelectorValidators[req.Key]; !found {
 | |
| 		allErrs = append(allErrs, field.Invalid(fldPath.Child("key"), req.Key, "not a valid field selector key"))
 | |
| 	} else {
 | |
| 		for i, v := range req.Values {
 | |
| 			for _, msg := range vf(v, false) {
 | |
| 				allErrs = append(allErrs, field.Invalid(fldPath.Child("values").Index(i), v, msg))
 | |
| 			}
 | |
| 		}
 | |
| 	}
 | |
| 
 | |
| 	return allErrs
 | |
| }
 | |
| 
 | |
| // ValidateNodeSelectorTerm tests that the specified node selector term has valid data
 | |
| func ValidateNodeSelectorTerm(term core.NodeSelectorTerm, fldPath *field.Path) field.ErrorList {
 | |
| 	allErrs := field.ErrorList{}
 | |
| 
 | |
| 	for j, req := range term.MatchExpressions {
 | |
| 		allErrs = append(allErrs, ValidateNodeSelectorRequirement(req, fldPath.Child("matchExpressions").Index(j))...)
 | |
| 	}
 | |
| 
 | |
| 	for j, req := range term.MatchFields {
 | |
| 		allErrs = append(allErrs, ValidateNodeFieldSelectorRequirement(req, fldPath.Child("matchFields").Index(j))...)
 | |
| 	}
 | |
| 
 | |
| 	return allErrs
 | |
| }
 | |
| 
 | |
| // ValidateNodeSelector tests that the specified nodeSelector fields has valid data
 | |
| func ValidateNodeSelector(nodeSelector *core.NodeSelector, fldPath *field.Path) field.ErrorList {
 | |
| 	allErrs := field.ErrorList{}
 | |
| 
 | |
| 	termFldPath := fldPath.Child("nodeSelectorTerms")
 | |
| 	if len(nodeSelector.NodeSelectorTerms) == 0 {
 | |
| 		return append(allErrs, field.Required(termFldPath, "must have at least one node selector term"))
 | |
| 	}
 | |
| 
 | |
| 	for i, term := range nodeSelector.NodeSelectorTerms {
 | |
| 		allErrs = append(allErrs, ValidateNodeSelectorTerm(term, termFldPath.Index(i))...)
 | |
| 	}
 | |
| 
 | |
| 	return allErrs
 | |
| }
 | |
| 
 | |
| // validateTopologySelectorLabelRequirement tests that the specified TopologySelectorLabelRequirement fields has valid data,
 | |
| // and constructs a set containing all of its Values.
 | |
| func validateTopologySelectorLabelRequirement(rq core.TopologySelectorLabelRequirement, fldPath *field.Path) (sets.String, field.ErrorList) {
 | |
| 	allErrs := field.ErrorList{}
 | |
| 	valueSet := make(sets.String)
 | |
| 	valuesPath := fldPath.Child("values")
 | |
| 	if len(rq.Values) == 0 {
 | |
| 		allErrs = append(allErrs, field.Required(valuesPath, ""))
 | |
| 	}
 | |
| 
 | |
| 	// Validate set property of Values field
 | |
| 	for i, value := range rq.Values {
 | |
| 		if valueSet.Has(value) {
 | |
| 			allErrs = append(allErrs, field.Duplicate(valuesPath.Index(i), value))
 | |
| 		}
 | |
| 		valueSet.Insert(value)
 | |
| 	}
 | |
| 
 | |
| 	allErrs = append(allErrs, unversionedvalidation.ValidateLabelName(rq.Key, fldPath.Child("key"))...)
 | |
| 
 | |
| 	return valueSet, allErrs
 | |
| }
 | |
| 
 | |
| // ValidateTopologySelectorTerm tests that the specified topology selector term has valid data,
 | |
| // and constructs a map representing the term in raw form.
 | |
| func ValidateTopologySelectorTerm(term core.TopologySelectorTerm, fldPath *field.Path) (map[string]sets.String, field.ErrorList) {
 | |
| 	allErrs := field.ErrorList{}
 | |
| 	exprMap := make(map[string]sets.String)
 | |
| 	exprPath := fldPath.Child("matchLabelExpressions")
 | |
| 
 | |
| 	// Allow empty MatchLabelExpressions, in case this field becomes optional in the future.
 | |
| 	for i, req := range term.MatchLabelExpressions {
 | |
| 		idxPath := exprPath.Index(i)
 | |
| 		valueSet, exprErrs := validateTopologySelectorLabelRequirement(req, idxPath)
 | |
| 		allErrs = append(allErrs, exprErrs...)
 | |
| 
 | |
| 		// Validate no duplicate keys exist.
 | |
| 		if _, exists := exprMap[req.Key]; exists {
 | |
| 			allErrs = append(allErrs, field.Duplicate(idxPath.Child("key"), req.Key))
 | |
| 		}
 | |
| 		exprMap[req.Key] = valueSet
 | |
| 	}
 | |
| 
 | |
| 	return exprMap, allErrs
 | |
| }
 | |
| 
 | |
| // ValidateAvoidPodsInNodeAnnotations tests that the serialized AvoidPods in Node.Annotations has valid data
 | |
| func ValidateAvoidPodsInNodeAnnotations(annotations map[string]string, fldPath *field.Path) field.ErrorList {
 | |
| 	allErrs := field.ErrorList{}
 | |
| 
 | |
| 	v1Avoids, err := schedulinghelper.GetAvoidPodsFromNodeAnnotations(annotations)
 | |
| 	if err != nil {
 | |
| 		allErrs = append(allErrs, field.Invalid(fldPath.Child("AvoidPods"), core.PreferAvoidPodsAnnotationKey, err.Error()))
 | |
| 		return allErrs
 | |
| 	}
 | |
| 	var avoids core.AvoidPods
 | |
| 	if err := corev1.Convert_v1_AvoidPods_To_core_AvoidPods(&v1Avoids, &avoids, nil); err != nil {
 | |
| 		allErrs = append(allErrs, field.Invalid(fldPath.Child("AvoidPods"), core.PreferAvoidPodsAnnotationKey, err.Error()))
 | |
| 		return allErrs
 | |
| 	}
 | |
| 
 | |
| 	if len(avoids.PreferAvoidPods) != 0 {
 | |
| 		for i, pa := range avoids.PreferAvoidPods {
 | |
| 			idxPath := fldPath.Child(core.PreferAvoidPodsAnnotationKey).Index(i)
 | |
| 			allErrs = append(allErrs, validatePreferAvoidPodsEntry(pa, idxPath)...)
 | |
| 		}
 | |
| 	}
 | |
| 
 | |
| 	return allErrs
 | |
| }
 | |
| 
 | |
| // validatePreferAvoidPodsEntry tests if given PreferAvoidPodsEntry has valid data.
 | |
| func validatePreferAvoidPodsEntry(avoidPodEntry core.PreferAvoidPodsEntry, fldPath *field.Path) field.ErrorList {
 | |
| 	allErrors := field.ErrorList{}
 | |
| 	if avoidPodEntry.PodSignature.PodController == nil {
 | |
| 		allErrors = append(allErrors, field.Required(fldPath.Child("PodSignature"), ""))
 | |
| 	} else {
 | |
| 		if !*(avoidPodEntry.PodSignature.PodController.Controller) {
 | |
| 			allErrors = append(allErrors,
 | |
| 				field.Invalid(fldPath.Child("PodSignature").Child("PodController").Child("Controller"),
 | |
| 					*(avoidPodEntry.PodSignature.PodController.Controller), "must point to a controller"))
 | |
| 		}
 | |
| 	}
 | |
| 	return allErrors
 | |
| }
 | |
| 
 | |
| // ValidatePreferredSchedulingTerms tests that the specified SoftNodeAffinity fields has valid data
 | |
| func ValidatePreferredSchedulingTerms(terms []core.PreferredSchedulingTerm, fldPath *field.Path) field.ErrorList {
 | |
| 	allErrs := field.ErrorList{}
 | |
| 
 | |
| 	for i, term := range terms {
 | |
| 		if term.Weight <= 0 || term.Weight > 100 {
 | |
| 			allErrs = append(allErrs, field.Invalid(fldPath.Index(i).Child("weight"), term.Weight, "must be in the range 1-100"))
 | |
| 		}
 | |
| 
 | |
| 		allErrs = append(allErrs, ValidateNodeSelectorTerm(term.Preference, fldPath.Index(i).Child("preference"))...)
 | |
| 	}
 | |
| 	return allErrs
 | |
| }
 | |
| 
 | |
| // validatePodAffinityTerm tests that the specified podAffinityTerm fields have valid data
 | |
| func validatePodAffinityTerm(podAffinityTerm core.PodAffinityTerm, fldPath *field.Path) field.ErrorList {
 | |
| 	allErrs := field.ErrorList{}
 | |
| 
 | |
| 	allErrs = append(allErrs, unversionedvalidation.ValidateLabelSelector(podAffinityTerm.LabelSelector, fldPath.Child("labelSelector"))...)
 | |
| 	allErrs = append(allErrs, unversionedvalidation.ValidateLabelSelector(podAffinityTerm.NamespaceSelector, fldPath.Child("namespaceSelector"))...)
 | |
| 
 | |
| 	for _, name := range podAffinityTerm.Namespaces {
 | |
| 		for _, msg := range ValidateNamespaceName(name, false) {
 | |
| 			allErrs = append(allErrs, field.Invalid(fldPath.Child("namespace"), name, msg))
 | |
| 		}
 | |
| 	}
 | |
| 	if len(podAffinityTerm.TopologyKey) == 0 {
 | |
| 		allErrs = append(allErrs, field.Required(fldPath.Child("topologyKey"), "can not be empty"))
 | |
| 	}
 | |
| 	return append(allErrs, unversionedvalidation.ValidateLabelName(podAffinityTerm.TopologyKey, fldPath.Child("topologyKey"))...)
 | |
| }
 | |
| 
 | |
| // validatePodAffinityTerms tests that the specified podAffinityTerms fields have valid data
 | |
| func validatePodAffinityTerms(podAffinityTerms []core.PodAffinityTerm, fldPath *field.Path) field.ErrorList {
 | |
| 	allErrs := field.ErrorList{}
 | |
| 	for i, podAffinityTerm := range podAffinityTerms {
 | |
| 		allErrs = append(allErrs, validatePodAffinityTerm(podAffinityTerm, fldPath.Index(i))...)
 | |
| 	}
 | |
| 	return allErrs
 | |
| }
 | |
| 
 | |
| // validateWeightedPodAffinityTerms tests that the specified weightedPodAffinityTerms fields have valid data
 | |
| func validateWeightedPodAffinityTerms(weightedPodAffinityTerms []core.WeightedPodAffinityTerm, fldPath *field.Path) field.ErrorList {
 | |
| 	allErrs := field.ErrorList{}
 | |
| 	for j, weightedTerm := range weightedPodAffinityTerms {
 | |
| 		if weightedTerm.Weight <= 0 || weightedTerm.Weight > 100 {
 | |
| 			allErrs = append(allErrs, field.Invalid(fldPath.Index(j).Child("weight"), weightedTerm.Weight, "must be in the range 1-100"))
 | |
| 		}
 | |
| 		allErrs = append(allErrs, validatePodAffinityTerm(weightedTerm.PodAffinityTerm, fldPath.Index(j).Child("podAffinityTerm"))...)
 | |
| 	}
 | |
| 	return allErrs
 | |
| }
 | |
| 
 | |
| // validatePodAntiAffinity tests that the specified podAntiAffinity fields have valid data
 | |
| func validatePodAntiAffinity(podAntiAffinity *core.PodAntiAffinity, fldPath *field.Path) field.ErrorList {
 | |
| 	allErrs := field.ErrorList{}
 | |
| 	// TODO:Uncomment below code once RequiredDuringSchedulingRequiredDuringExecution is implemented.
 | |
| 	// if podAntiAffinity.RequiredDuringSchedulingRequiredDuringExecution != nil {
 | |
| 	//	allErrs = append(allErrs, validatePodAffinityTerms(podAntiAffinity.RequiredDuringSchedulingRequiredDuringExecution, false,
 | |
| 	//		fldPath.Child("requiredDuringSchedulingRequiredDuringExecution"))...)
 | |
| 	//}
 | |
| 	if podAntiAffinity.RequiredDuringSchedulingIgnoredDuringExecution != nil {
 | |
| 		allErrs = append(allErrs, validatePodAffinityTerms(podAntiAffinity.RequiredDuringSchedulingIgnoredDuringExecution,
 | |
| 			fldPath.Child("requiredDuringSchedulingIgnoredDuringExecution"))...)
 | |
| 	}
 | |
| 	if podAntiAffinity.PreferredDuringSchedulingIgnoredDuringExecution != nil {
 | |
| 		allErrs = append(allErrs, validateWeightedPodAffinityTerms(podAntiAffinity.PreferredDuringSchedulingIgnoredDuringExecution,
 | |
| 			fldPath.Child("preferredDuringSchedulingIgnoredDuringExecution"))...)
 | |
| 	}
 | |
| 	return allErrs
 | |
| }
 | |
| 
 | |
| // validateNodeAffinity tests that the specified nodeAffinity fields have valid data
 | |
| func validateNodeAffinity(na *core.NodeAffinity, fldPath *field.Path) field.ErrorList {
 | |
| 	allErrs := field.ErrorList{}
 | |
| 	// TODO: Uncomment the next three lines once RequiredDuringSchedulingRequiredDuringExecution is implemented.
 | |
| 	// if na.RequiredDuringSchedulingRequiredDuringExecution != nil {
 | |
| 	//	allErrs = append(allErrs, ValidateNodeSelector(na.RequiredDuringSchedulingRequiredDuringExecution, fldPath.Child("requiredDuringSchedulingRequiredDuringExecution"))...)
 | |
| 	// }
 | |
| 	if na.RequiredDuringSchedulingIgnoredDuringExecution != nil {
 | |
| 		allErrs = append(allErrs, ValidateNodeSelector(na.RequiredDuringSchedulingIgnoredDuringExecution, fldPath.Child("requiredDuringSchedulingIgnoredDuringExecution"))...)
 | |
| 	}
 | |
| 	if len(na.PreferredDuringSchedulingIgnoredDuringExecution) > 0 {
 | |
| 		allErrs = append(allErrs, ValidatePreferredSchedulingTerms(na.PreferredDuringSchedulingIgnoredDuringExecution, fldPath.Child("preferredDuringSchedulingIgnoredDuringExecution"))...)
 | |
| 	}
 | |
| 	return allErrs
 | |
| }
 | |
| 
 | |
| // validatePodAffinity tests that the specified podAffinity fields have valid data
 | |
| func validatePodAffinity(podAffinity *core.PodAffinity, fldPath *field.Path) field.ErrorList {
 | |
| 	allErrs := field.ErrorList{}
 | |
| 	// TODO:Uncomment below code once RequiredDuringSchedulingRequiredDuringExecution is implemented.
 | |
| 	// if podAffinity.RequiredDuringSchedulingRequiredDuringExecution != nil {
 | |
| 	//	allErrs = append(allErrs, validatePodAffinityTerms(podAffinity.RequiredDuringSchedulingRequiredDuringExecution, false,
 | |
| 	//		fldPath.Child("requiredDuringSchedulingRequiredDuringExecution"))...)
 | |
| 	//}
 | |
| 	if podAffinity.RequiredDuringSchedulingIgnoredDuringExecution != nil {
 | |
| 		allErrs = append(allErrs, validatePodAffinityTerms(podAffinity.RequiredDuringSchedulingIgnoredDuringExecution,
 | |
| 			fldPath.Child("requiredDuringSchedulingIgnoredDuringExecution"))...)
 | |
| 	}
 | |
| 	if podAffinity.PreferredDuringSchedulingIgnoredDuringExecution != nil {
 | |
| 		allErrs = append(allErrs, validateWeightedPodAffinityTerms(podAffinity.PreferredDuringSchedulingIgnoredDuringExecution,
 | |
| 			fldPath.Child("preferredDuringSchedulingIgnoredDuringExecution"))...)
 | |
| 	}
 | |
| 	return allErrs
 | |
| }
 | |
| 
 | |
| func validateSeccompProfileField(sp *core.SeccompProfile, fldPath *field.Path) field.ErrorList {
 | |
| 	allErrs := field.ErrorList{}
 | |
| 	if sp == nil {
 | |
| 		return allErrs
 | |
| 	}
 | |
| 
 | |
| 	if err := validateSeccompProfileType(fldPath.Child("type"), sp.Type); err != nil {
 | |
| 		allErrs = append(allErrs, err)
 | |
| 	}
 | |
| 
 | |
| 	if sp.Type == core.SeccompProfileTypeLocalhost {
 | |
| 		if sp.LocalhostProfile == nil {
 | |
| 			allErrs = append(allErrs, field.Required(fldPath.Child("localhostProfile"), "must be set when seccomp type is Localhost"))
 | |
| 		} else {
 | |
| 			allErrs = append(allErrs, validateLocalDescendingPath(*sp.LocalhostProfile, fldPath.Child("localhostProfile"))...)
 | |
| 		}
 | |
| 	} else {
 | |
| 		if sp.LocalhostProfile != nil {
 | |
| 			allErrs = append(allErrs, field.Invalid(fldPath.Child("localhostProfile"), sp, "can only be set when seccomp type is Localhost"))
 | |
| 		}
 | |
| 	}
 | |
| 
 | |
| 	return allErrs
 | |
| }
 | |
| 
 | |
| func ValidateSeccompProfile(p string, fldPath *field.Path) field.ErrorList {
 | |
| 	if p == core.SeccompProfileRuntimeDefault || p == core.DeprecatedSeccompProfileDockerDefault {
 | |
| 		return nil
 | |
| 	}
 | |
| 	if p == v1.SeccompProfileNameUnconfined {
 | |
| 		return nil
 | |
| 	}
 | |
| 	if strings.HasPrefix(p, v1.SeccompLocalhostProfileNamePrefix) {
 | |
| 		return validateLocalDescendingPath(strings.TrimPrefix(p, v1.SeccompLocalhostProfileNamePrefix), fldPath)
 | |
| 	}
 | |
| 	return field.ErrorList{field.Invalid(fldPath, p, "must be a valid seccomp profile")}
 | |
| }
 | |
| 
 | |
| func ValidateSeccompPodAnnotations(annotations map[string]string, fldPath *field.Path) field.ErrorList {
 | |
| 	allErrs := field.ErrorList{}
 | |
| 	if p, exists := annotations[core.SeccompPodAnnotationKey]; exists {
 | |
| 		allErrs = append(allErrs, ValidateSeccompProfile(p, fldPath.Child(core.SeccompPodAnnotationKey))...)
 | |
| 	}
 | |
| 	for k, p := range annotations {
 | |
| 		if strings.HasPrefix(k, core.SeccompContainerAnnotationKeyPrefix) {
 | |
| 			allErrs = append(allErrs, ValidateSeccompProfile(p, fldPath.Child(k))...)
 | |
| 		}
 | |
| 	}
 | |
| 
 | |
| 	return allErrs
 | |
| }
 | |
| 
 | |
| // ValidateSeccompProfileType tests that the argument is a valid SeccompProfileType.
 | |
| func validateSeccompProfileType(fldPath *field.Path, seccompProfileType core.SeccompProfileType) *field.Error {
 | |
| 	switch seccompProfileType {
 | |
| 	case core.SeccompProfileTypeLocalhost, core.SeccompProfileTypeRuntimeDefault, core.SeccompProfileTypeUnconfined:
 | |
| 		return nil
 | |
| 	case "":
 | |
| 		return field.Required(fldPath, "type is required when seccompProfile is set")
 | |
| 	default:
 | |
| 		return field.NotSupported(fldPath, seccompProfileType, []string{string(core.SeccompProfileTypeLocalhost), string(core.SeccompProfileTypeRuntimeDefault), string(core.SeccompProfileTypeUnconfined)})
 | |
| 	}
 | |
| }
 | |
| 
 | |
| func ValidateAppArmorPodAnnotations(annotations map[string]string, spec *core.PodSpec, fldPath *field.Path) field.ErrorList {
 | |
| 	allErrs := field.ErrorList{}
 | |
| 	for k, p := range annotations {
 | |
| 		if !strings.HasPrefix(k, v1.AppArmorBetaContainerAnnotationKeyPrefix) {
 | |
| 			continue
 | |
| 		}
 | |
| 		containerName := strings.TrimPrefix(k, v1.AppArmorBetaContainerAnnotationKeyPrefix)
 | |
| 		if !podSpecHasContainer(spec, containerName) {
 | |
| 			allErrs = append(allErrs, field.Invalid(fldPath.Key(k), containerName, "container not found"))
 | |
| 		}
 | |
| 
 | |
| 		if err := apparmor.ValidateProfileFormat(p); err != nil {
 | |
| 			allErrs = append(allErrs, field.Invalid(fldPath.Key(k), p, err.Error()))
 | |
| 		}
 | |
| 	}
 | |
| 
 | |
| 	return allErrs
 | |
| }
 | |
| 
 | |
| func podSpecHasContainer(spec *core.PodSpec, containerName string) bool {
 | |
| 	var hasContainer bool
 | |
| 	podshelper.VisitContainersWithPath(spec, field.NewPath("spec"), func(c *core.Container, _ *field.Path) bool {
 | |
| 		if c.Name == containerName {
 | |
| 			hasContainer = true
 | |
| 			return false
 | |
| 		}
 | |
| 		return true
 | |
| 	})
 | |
| 	return hasContainer
 | |
| }
 | |
| 
 | |
| const (
 | |
| 	// a sysctl segment regex, concatenated with dots to form a sysctl name
 | |
| 	SysctlSegmentFmt string = "[a-z0-9]([-_a-z0-9]*[a-z0-9])?"
 | |
| 
 | |
| 	// a sysctl name regex
 | |
| 	SysctlFmt string = "(" + SysctlSegmentFmt + "\\.)*" + SysctlSegmentFmt
 | |
| 
 | |
| 	// the maximal length of a sysctl name
 | |
| 	SysctlMaxLength int = 253
 | |
| )
 | |
| 
 | |
| var sysctlRegexp = regexp.MustCompile("^" + SysctlFmt + "$")
 | |
| 
 | |
| // IsValidSysctlName checks that the given string is a valid sysctl name,
 | |
| // i.e. matches SysctlFmt.
 | |
| func IsValidSysctlName(name string) bool {
 | |
| 	if len(name) > SysctlMaxLength {
 | |
| 		return false
 | |
| 	}
 | |
| 	return sysctlRegexp.MatchString(name)
 | |
| }
 | |
| 
 | |
| func validateSysctls(sysctls []core.Sysctl, fldPath *field.Path) field.ErrorList {
 | |
| 	allErrs := field.ErrorList{}
 | |
| 	names := make(map[string]struct{})
 | |
| 	for i, s := range sysctls {
 | |
| 		if len(s.Name) == 0 {
 | |
| 			allErrs = append(allErrs, field.Required(fldPath.Index(i).Child("name"), ""))
 | |
| 		} else if !IsValidSysctlName(s.Name) {
 | |
| 			allErrs = append(allErrs, field.Invalid(fldPath.Index(i).Child("name"), s.Name, fmt.Sprintf("must have at most %d characters and match regex %s", SysctlMaxLength, SysctlFmt)))
 | |
| 		} else if _, ok := names[s.Name]; ok {
 | |
| 			allErrs = append(allErrs, field.Duplicate(fldPath.Index(i).Child("name"), s.Name))
 | |
| 		}
 | |
| 		names[s.Name] = struct{}{}
 | |
| 	}
 | |
| 	return allErrs
 | |
| }
 | |
| 
 | |
| // ValidatePodSecurityContext test that the specified PodSecurityContext has valid data.
 | |
| func ValidatePodSecurityContext(securityContext *core.PodSecurityContext, spec *core.PodSpec, specPath, fldPath *field.Path) field.ErrorList {
 | |
| 	allErrs := field.ErrorList{}
 | |
| 
 | |
| 	if securityContext != nil {
 | |
| 		allErrs = append(allErrs, validateHostNetwork(securityContext.HostNetwork, spec.Containers, specPath.Child("containers"))...)
 | |
| 		if securityContext.FSGroup != nil {
 | |
| 			for _, msg := range validation.IsValidGroupID(*securityContext.FSGroup) {
 | |
| 				allErrs = append(allErrs, field.Invalid(fldPath.Child("fsGroup"), *(securityContext.FSGroup), msg))
 | |
| 			}
 | |
| 		}
 | |
| 		if securityContext.RunAsUser != nil {
 | |
| 			for _, msg := range validation.IsValidUserID(*securityContext.RunAsUser) {
 | |
| 				allErrs = append(allErrs, field.Invalid(fldPath.Child("runAsUser"), *(securityContext.RunAsUser), msg))
 | |
| 			}
 | |
| 		}
 | |
| 		if securityContext.RunAsGroup != nil {
 | |
| 			for _, msg := range validation.IsValidGroupID(*securityContext.RunAsGroup) {
 | |
| 				allErrs = append(allErrs, field.Invalid(fldPath.Child("runAsGroup"), *(securityContext.RunAsGroup), msg))
 | |
| 			}
 | |
| 		}
 | |
| 		for g, gid := range securityContext.SupplementalGroups {
 | |
| 			for _, msg := range validation.IsValidGroupID(gid) {
 | |
| 				allErrs = append(allErrs, field.Invalid(fldPath.Child("supplementalGroups").Index(g), gid, msg))
 | |
| 			}
 | |
| 		}
 | |
| 		if securityContext.ShareProcessNamespace != nil && securityContext.HostPID && *securityContext.ShareProcessNamespace {
 | |
| 			allErrs = append(allErrs, field.Invalid(fldPath.Child("shareProcessNamespace"), *securityContext.ShareProcessNamespace, "ShareProcessNamespace and HostPID cannot both be enabled"))
 | |
| 		}
 | |
| 
 | |
| 		if len(securityContext.Sysctls) != 0 {
 | |
| 			allErrs = append(allErrs, validateSysctls(securityContext.Sysctls, fldPath.Child("sysctls"))...)
 | |
| 		}
 | |
| 
 | |
| 		if securityContext.FSGroupChangePolicy != nil {
 | |
| 			allErrs = append(allErrs, validateFSGroupChangePolicy(securityContext.FSGroupChangePolicy, fldPath.Child("fsGroupChangePolicy"))...)
 | |
| 		}
 | |
| 
 | |
| 		allErrs = append(allErrs, validateSeccompProfileField(securityContext.SeccompProfile, fldPath.Child("seccompProfile"))...)
 | |
| 		allErrs = append(allErrs, validateWindowsSecurityContextOptions(securityContext.WindowsOptions, fldPath.Child("windowsOptions"))...)
 | |
| 	}
 | |
| 
 | |
| 	return allErrs
 | |
| }
 | |
| 
 | |
| func ValidateContainerUpdates(newContainers, oldContainers []core.Container, fldPath *field.Path) (allErrs field.ErrorList, stop bool) {
 | |
| 	allErrs = field.ErrorList{}
 | |
| 	if len(newContainers) != len(oldContainers) {
 | |
| 		//TODO: Pinpoint the specific container that causes the invalid error after we have strategic merge diff
 | |
| 		allErrs = append(allErrs, field.Forbidden(fldPath, "pod updates may not add or remove containers"))
 | |
| 		return allErrs, true
 | |
| 	}
 | |
| 
 | |
| 	// validate updated container images
 | |
| 	for i, ctr := range newContainers {
 | |
| 		if len(ctr.Image) == 0 {
 | |
| 			allErrs = append(allErrs, field.Required(fldPath.Index(i).Child("image"), ""))
 | |
| 		}
 | |
| 		// this is only called from ValidatePodUpdate so its safe to check leading/trailing whitespace.
 | |
| 		if len(strings.TrimSpace(ctr.Image)) != len(ctr.Image) {
 | |
| 			allErrs = append(allErrs, field.Invalid(fldPath.Index(i).Child("image"), ctr.Image, "must not have leading or trailing whitespace"))
 | |
| 		}
 | |
| 	}
 | |
| 	return allErrs, false
 | |
| }
 | |
| 
 | |
| // ValidatePodCreate validates a pod in the context of its initial create
 | |
| func ValidatePodCreate(pod *core.Pod, opts PodValidationOptions) field.ErrorList {
 | |
| 	allErrs := validatePodMetadataAndSpec(pod, opts)
 | |
| 
 | |
| 	fldPath := field.NewPath("spec")
 | |
| 	// EphemeralContainers can only be set on update using the ephemeralcontainers subresource
 | |
| 	if len(pod.Spec.EphemeralContainers) > 0 {
 | |
| 		allErrs = append(allErrs, field.Forbidden(fldPath.Child("ephemeralContainers"), "cannot be set on create"))
 | |
| 	}
 | |
| 	allErrs = append(allErrs, validateSeccompAnnotationsAndFields(pod.ObjectMeta, &pod.Spec, fldPath)...)
 | |
| 
 | |
| 	return allErrs
 | |
| }
 | |
| 
 | |
| // ValidateSeccompAnnotationsAndFields iterates through all containers and ensure that when both seccompProfile and seccomp annotations exist they match.
 | |
| func validateSeccompAnnotationsAndFields(objectMeta metav1.ObjectMeta, podSpec *core.PodSpec, specPath *field.Path) field.ErrorList {
 | |
| 	allErrs := field.ErrorList{}
 | |
| 
 | |
| 	if podSpec.SecurityContext != nil && podSpec.SecurityContext.SeccompProfile != nil {
 | |
| 		// If both seccomp annotations and fields are specified, the values must match.
 | |
| 		if annotation, found := objectMeta.Annotations[v1.SeccompPodAnnotationKey]; found {
 | |
| 			seccompPath := specPath.Child("securityContext").Child("seccompProfile")
 | |
| 			err := validateSeccompAnnotationsAndFieldsMatch(annotation, podSpec.SecurityContext.SeccompProfile, seccompPath)
 | |
| 			if err != nil {
 | |
| 				allErrs = append(allErrs, err)
 | |
| 			}
 | |
| 		}
 | |
| 	}
 | |
| 
 | |
| 	podshelper.VisitContainersWithPath(podSpec, specPath, func(c *core.Container, cFldPath *field.Path) bool {
 | |
| 		var field *core.SeccompProfile
 | |
| 		if c.SecurityContext != nil {
 | |
| 			field = c.SecurityContext.SeccompProfile
 | |
| 		}
 | |
| 
 | |
| 		if field == nil {
 | |
| 			return true
 | |
| 		}
 | |
| 
 | |
| 		key := v1.SeccompContainerAnnotationKeyPrefix + c.Name
 | |
| 		if annotation, found := objectMeta.Annotations[key]; found {
 | |
| 			seccompPath := cFldPath.Child("securityContext").Child("seccompProfile")
 | |
| 			err := validateSeccompAnnotationsAndFieldsMatch(annotation, field, seccompPath)
 | |
| 			if err != nil {
 | |
| 				allErrs = append(allErrs, err)
 | |
| 			}
 | |
| 		}
 | |
| 		return true
 | |
| 	})
 | |
| 
 | |
| 	return allErrs
 | |
| }
 | |
| 
 | |
| func validateSeccompAnnotationsAndFieldsMatch(annotationValue string, seccompField *core.SeccompProfile, fldPath *field.Path) *field.Error {
 | |
| 	if seccompField == nil {
 | |
| 		return nil
 | |
| 	}
 | |
| 
 | |
| 	switch seccompField.Type {
 | |
| 	case core.SeccompProfileTypeUnconfined:
 | |
| 		if annotationValue != v1.SeccompProfileNameUnconfined {
 | |
| 			return field.Forbidden(fldPath.Child("type"), "seccomp type in annotation and field must match")
 | |
| 		}
 | |
| 
 | |
| 	case core.SeccompProfileTypeRuntimeDefault:
 | |
| 		if annotationValue != v1.SeccompProfileRuntimeDefault && annotationValue != v1.DeprecatedSeccompProfileDockerDefault {
 | |
| 			return field.Forbidden(fldPath.Child("type"), "seccomp type in annotation and field must match")
 | |
| 		}
 | |
| 
 | |
| 	case core.SeccompProfileTypeLocalhost:
 | |
| 		if !strings.HasPrefix(annotationValue, v1.SeccompLocalhostProfileNamePrefix) {
 | |
| 			return field.Forbidden(fldPath.Child("type"), "seccomp type in annotation and field must match")
 | |
| 		} else if seccompField.LocalhostProfile == nil || strings.TrimPrefix(annotationValue, v1.SeccompLocalhostProfileNamePrefix) != *seccompField.LocalhostProfile {
 | |
| 			return field.Forbidden(fldPath.Child("localhostProfile"), "seccomp profile in annotation and field must match")
 | |
| 		}
 | |
| 	}
 | |
| 
 | |
| 	return nil
 | |
| }
 | |
| 
 | |
| // ValidatePodUpdate tests to see if the update is legal for an end user to make. newPod is updated with fields
 | |
| // that cannot be changed.
 | |
| func ValidatePodUpdate(newPod, oldPod *core.Pod, opts PodValidationOptions) field.ErrorList {
 | |
| 	fldPath := field.NewPath("metadata")
 | |
| 	allErrs := ValidateObjectMetaUpdate(&newPod.ObjectMeta, &oldPod.ObjectMeta, fldPath)
 | |
| 	allErrs = append(allErrs, validatePodMetadataAndSpec(newPod, opts)...)
 | |
| 	allErrs = append(allErrs, ValidatePodSpecificAnnotationUpdates(newPod, oldPod, fldPath.Child("annotations"), opts)...)
 | |
| 	specPath := field.NewPath("spec")
 | |
| 
 | |
| 	if !opts.AllowMultipleHugePageResources {
 | |
| 		allErrs = append(allErrs, ValidatePodSingleHugePageResources(newPod, specPath)...)
 | |
| 	}
 | |
| 
 | |
| 	// validate updateable fields:
 | |
| 	// 1.  spec.containers[*].image
 | |
| 	// 2.  spec.initContainers[*].image
 | |
| 	// 3.  spec.activeDeadlineSeconds
 | |
| 
 | |
| 	containerErrs, stop := ValidateContainerUpdates(newPod.Spec.Containers, oldPod.Spec.Containers, specPath.Child("containers"))
 | |
| 	allErrs = append(allErrs, containerErrs...)
 | |
| 	if stop {
 | |
| 		return allErrs
 | |
| 	}
 | |
| 	containerErrs, stop = ValidateContainerUpdates(newPod.Spec.InitContainers, oldPod.Spec.InitContainers, specPath.Child("initContainers"))
 | |
| 	allErrs = append(allErrs, containerErrs...)
 | |
| 	if stop {
 | |
| 		return allErrs
 | |
| 	}
 | |
| 
 | |
| 	// validate updated spec.activeDeadlineSeconds.  two types of updates are allowed:
 | |
| 	// 1.  from nil to a positive value
 | |
| 	// 2.  from a positive value to a lesser, non-negative value
 | |
| 	if newPod.Spec.ActiveDeadlineSeconds != nil {
 | |
| 		newActiveDeadlineSeconds := *newPod.Spec.ActiveDeadlineSeconds
 | |
| 		if newActiveDeadlineSeconds < 0 || newActiveDeadlineSeconds > math.MaxInt32 {
 | |
| 			allErrs = append(allErrs, field.Invalid(specPath.Child("activeDeadlineSeconds"), newActiveDeadlineSeconds, validation.InclusiveRangeError(0, math.MaxInt32)))
 | |
| 			return allErrs
 | |
| 		}
 | |
| 		if oldPod.Spec.ActiveDeadlineSeconds != nil {
 | |
| 			oldActiveDeadlineSeconds := *oldPod.Spec.ActiveDeadlineSeconds
 | |
| 			if oldActiveDeadlineSeconds < newActiveDeadlineSeconds {
 | |
| 				allErrs = append(allErrs, field.Invalid(specPath.Child("activeDeadlineSeconds"), newActiveDeadlineSeconds, "must be less than or equal to previous value"))
 | |
| 				return allErrs
 | |
| 			}
 | |
| 		}
 | |
| 	} else if oldPod.Spec.ActiveDeadlineSeconds != nil {
 | |
| 		allErrs = append(allErrs, field.Invalid(specPath.Child("activeDeadlineSeconds"), newPod.Spec.ActiveDeadlineSeconds, "must not update from a positive integer to nil value"))
 | |
| 	}
 | |
| 
 | |
| 	// Allow only additions to tolerations updates.
 | |
| 	allErrs = append(allErrs, validateOnlyAddedTolerations(newPod.Spec.Tolerations, oldPod.Spec.Tolerations, specPath.Child("tolerations"))...)
 | |
| 
 | |
| 	// the last thing to check is pod spec equality.  If the pod specs are equal, then we can simply return the errors we have
 | |
| 	// so far and save the cost of a deep copy.
 | |
| 	if apiequality.Semantic.DeepEqual(newPod.Spec, oldPod.Spec) {
 | |
| 		return allErrs
 | |
| 	}
 | |
| 
 | |
| 	// handle updateable fields by munging those fields prior to deep equal comparison.
 | |
| 	mungedPodSpec := *newPod.Spec.DeepCopy()
 | |
| 	// munge spec.containers[*].image
 | |
| 	var newContainers []core.Container
 | |
| 	for ix, container := range mungedPodSpec.Containers {
 | |
| 		container.Image = oldPod.Spec.Containers[ix].Image // +k8s:verify-mutation:reason=clone
 | |
| 		newContainers = append(newContainers, container)
 | |
| 	}
 | |
| 	mungedPodSpec.Containers = newContainers
 | |
| 	// munge spec.initContainers[*].image
 | |
| 	var newInitContainers []core.Container
 | |
| 	for ix, container := range mungedPodSpec.InitContainers {
 | |
| 		container.Image = oldPod.Spec.InitContainers[ix].Image // +k8s:verify-mutation:reason=clone
 | |
| 		newInitContainers = append(newInitContainers, container)
 | |
| 	}
 | |
| 	mungedPodSpec.InitContainers = newInitContainers
 | |
| 	// munge spec.activeDeadlineSeconds
 | |
| 	mungedPodSpec.ActiveDeadlineSeconds = nil
 | |
| 	if oldPod.Spec.ActiveDeadlineSeconds != nil {
 | |
| 		activeDeadlineSeconds := *oldPod.Spec.ActiveDeadlineSeconds
 | |
| 		mungedPodSpec.ActiveDeadlineSeconds = &activeDeadlineSeconds
 | |
| 	}
 | |
| 	// tolerations are checked before the deep copy, so munge those too
 | |
| 	mungedPodSpec.Tolerations = oldPod.Spec.Tolerations // +k8s:verify-mutation:reason=clone
 | |
| 
 | |
| 	if !apiequality.Semantic.DeepEqual(mungedPodSpec, oldPod.Spec) {
 | |
| 		// This diff isn't perfect, but it's a helluva lot better an "I'm not going to tell you what the difference is".
 | |
| 		//TODO: Pinpoint the specific field that causes the invalid error after we have strategic merge diff
 | |
| 		specDiff := diff.ObjectDiff(mungedPodSpec, oldPod.Spec)
 | |
| 		allErrs = append(allErrs, field.Forbidden(specPath, fmt.Sprintf("pod updates may not change fields other than `spec.containers[*].image`, `spec.initContainers[*].image`, `spec.activeDeadlineSeconds` or `spec.tolerations` (only additions to existing tolerations)\n%v", specDiff)))
 | |
| 	}
 | |
| 
 | |
| 	return allErrs
 | |
| }
 | |
| 
 | |
| // ValidateContainerStateTransition test to if any illegal container state transitions are being attempted
 | |
| func ValidateContainerStateTransition(newStatuses, oldStatuses []core.ContainerStatus, fldpath *field.Path, restartPolicy core.RestartPolicy) field.ErrorList {
 | |
| 	allErrs := field.ErrorList{}
 | |
| 	// If we should always restart, containers are allowed to leave the terminated state
 | |
| 	if restartPolicy == core.RestartPolicyAlways {
 | |
| 		return allErrs
 | |
| 	}
 | |
| 	for i, oldStatus := range oldStatuses {
 | |
| 		// Skip any container that is not terminated
 | |
| 		if oldStatus.State.Terminated == nil {
 | |
| 			continue
 | |
| 		}
 | |
| 		// Skip any container that failed but is allowed to restart
 | |
| 		if oldStatus.State.Terminated.ExitCode != 0 && restartPolicy == core.RestartPolicyOnFailure {
 | |
| 			continue
 | |
| 		}
 | |
| 		for _, newStatus := range newStatuses {
 | |
| 			if oldStatus.Name == newStatus.Name && newStatus.State.Terminated == nil {
 | |
| 				allErrs = append(allErrs, field.Forbidden(fldpath.Index(i).Child("state"), "may not be transitioned to non-terminated state"))
 | |
| 			}
 | |
| 		}
 | |
| 	}
 | |
| 	return allErrs
 | |
| }
 | |
| 
 | |
| // ValidatePodStatusUpdate tests to see if the update is legal for an end user to make.
 | |
| func ValidatePodStatusUpdate(newPod, oldPod *core.Pod, opts PodValidationOptions) field.ErrorList {
 | |
| 	fldPath := field.NewPath("metadata")
 | |
| 	allErrs := ValidateObjectMetaUpdate(&newPod.ObjectMeta, &oldPod.ObjectMeta, fldPath)
 | |
| 	allErrs = append(allErrs, ValidatePodSpecificAnnotationUpdates(newPod, oldPod, fldPath.Child("annotations"), opts)...)
 | |
| 	allErrs = append(allErrs, validatePodConditions(newPod.Status.Conditions, fldPath.Child("conditions"))...)
 | |
| 
 | |
| 	fldPath = field.NewPath("status")
 | |
| 	if newPod.Spec.NodeName != oldPod.Spec.NodeName {
 | |
| 		allErrs = append(allErrs, field.Forbidden(fldPath.Child("nodeName"), "may not be changed directly"))
 | |
| 	}
 | |
| 
 | |
| 	if newPod.Status.NominatedNodeName != oldPod.Status.NominatedNodeName && len(newPod.Status.NominatedNodeName) > 0 {
 | |
| 		for _, msg := range ValidateNodeName(newPod.Status.NominatedNodeName, false) {
 | |
| 			allErrs = append(allErrs, field.Invalid(fldPath.Child("nominatedNodeName"), newPod.Status.NominatedNodeName, msg))
 | |
| 		}
 | |
| 	}
 | |
| 
 | |
| 	// If pod should not restart, make sure the status update does not transition
 | |
| 	// any terminated containers to a non-terminated state.
 | |
| 	allErrs = append(allErrs, ValidateContainerStateTransition(newPod.Status.ContainerStatuses, oldPod.Status.ContainerStatuses, fldPath.Child("containerStatuses"), oldPod.Spec.RestartPolicy)...)
 | |
| 	allErrs = append(allErrs, ValidateContainerStateTransition(newPod.Status.InitContainerStatuses, oldPod.Status.InitContainerStatuses, fldPath.Child("initContainerStatuses"), oldPod.Spec.RestartPolicy)...)
 | |
| 
 | |
| 	if newIPErrs := validatePodIPs(newPod); len(newIPErrs) > 0 {
 | |
| 		allErrs = append(allErrs, newIPErrs...)
 | |
| 	}
 | |
| 
 | |
| 	return allErrs
 | |
| }
 | |
| 
 | |
| // validatePodConditions tests if the custom pod conditions are valid.
 | |
| func validatePodConditions(conditions []core.PodCondition, fldPath *field.Path) field.ErrorList {
 | |
| 	allErrs := field.ErrorList{}
 | |
| 	systemConditions := sets.NewString(string(core.PodScheduled), string(core.PodReady), string(core.PodInitialized))
 | |
| 	for i, condition := range conditions {
 | |
| 		if systemConditions.Has(string(condition.Type)) {
 | |
| 			continue
 | |
| 		}
 | |
| 		for _, msg := range validation.IsQualifiedName(string(condition.Type)) {
 | |
| 			allErrs = append(allErrs, field.Invalid(fldPath.Index(i).Child("Type"), string(condition.Type), msg))
 | |
| 		}
 | |
| 	}
 | |
| 	return allErrs
 | |
| }
 | |
| 
 | |
| // ValidatePodEphemeralContainersUpdate tests that a user update to EphemeralContainers is valid.
 | |
| // newPod and oldPod must only differ in their EphemeralContainers.
 | |
| func ValidatePodEphemeralContainersUpdate(newPod, oldPod *core.Pod, opts PodValidationOptions) field.ErrorList {
 | |
| 	spec := newPod.Spec
 | |
| 	specPath := field.NewPath("spec").Child("ephemeralContainers")
 | |
| 
 | |
| 	vols := make(map[string]core.VolumeSource)
 | |
| 	for _, vol := range spec.Volumes {
 | |
| 		vols[vol.Name] = vol.VolumeSource
 | |
| 	}
 | |
| 	allErrs := validateEphemeralContainers(spec.EphemeralContainers, spec.Containers, spec.InitContainers, vols, specPath, opts)
 | |
| 
 | |
| 	// Existing EphemeralContainers may not be changed. Order isn't preserved by patch, so check each individually.
 | |
| 	newContainerIndex := make(map[string]*core.EphemeralContainer)
 | |
| 	for i := range newPod.Spec.EphemeralContainers {
 | |
| 		newContainerIndex[newPod.Spec.EphemeralContainers[i].Name] = &newPod.Spec.EphemeralContainers[i]
 | |
| 	}
 | |
| 	for _, old := range oldPod.Spec.EphemeralContainers {
 | |
| 		if new, ok := newContainerIndex[old.Name]; !ok {
 | |
| 			allErrs = append(allErrs, field.Forbidden(specPath, fmt.Sprintf("existing ephemeral containers %q may not be removed\n", old.Name)))
 | |
| 		} else if !apiequality.Semantic.DeepEqual(old, *new) {
 | |
| 			specDiff := diff.ObjectDiff(old, *new)
 | |
| 			allErrs = append(allErrs, field.Forbidden(specPath, fmt.Sprintf("existing ephemeral containers %q may not be changed\n%v", old.Name, specDiff)))
 | |
| 		}
 | |
| 	}
 | |
| 
 | |
| 	return allErrs
 | |
| }
 | |
| 
 | |
| // ValidatePodBinding tests if required fields in the pod binding are legal.
 | |
| func ValidatePodBinding(binding *core.Binding) field.ErrorList {
 | |
| 	allErrs := field.ErrorList{}
 | |
| 
 | |
| 	if len(binding.Target.Kind) != 0 && binding.Target.Kind != "Node" {
 | |
| 		// TODO: When validation becomes versioned, this gets more complicated.
 | |
| 		allErrs = append(allErrs, field.NotSupported(field.NewPath("target", "kind"), binding.Target.Kind, []string{"Node", "<empty>"}))
 | |
| 	}
 | |
| 	if len(binding.Target.Name) == 0 {
 | |
| 		// TODO: When validation becomes versioned, this gets more complicated.
 | |
| 		allErrs = append(allErrs, field.Required(field.NewPath("target", "name"), ""))
 | |
| 	}
 | |
| 
 | |
| 	return allErrs
 | |
| }
 | |
| 
 | |
| // ValidatePodTemplate tests if required fields in the pod template are set.
 | |
| func ValidatePodTemplate(pod *core.PodTemplate, opts PodValidationOptions) field.ErrorList {
 | |
| 	allErrs := ValidateObjectMeta(&pod.ObjectMeta, true, ValidatePodName, field.NewPath("metadata"))
 | |
| 	allErrs = append(allErrs, ValidatePodTemplateSpec(&pod.Template, field.NewPath("template"), opts)...)
 | |
| 	return allErrs
 | |
| }
 | |
| 
 | |
| // ValidatePodTemplateUpdate tests to see if the update is legal for an end user to make. newPod is updated with fields
 | |
| // that cannot be changed.
 | |
| func ValidatePodTemplateUpdate(newPod, oldPod *core.PodTemplate, opts PodValidationOptions) field.ErrorList {
 | |
| 	allErrs := ValidateObjectMetaUpdate(&newPod.ObjectMeta, &oldPod.ObjectMeta, field.NewPath("metadata"))
 | |
| 	allErrs = append(allErrs, ValidatePodTemplateSpec(&newPod.Template, field.NewPath("template"), opts)...)
 | |
| 	return allErrs
 | |
| }
 | |
| 
 | |
| var supportedSessionAffinityType = sets.NewString(string(core.ServiceAffinityClientIP), string(core.ServiceAffinityNone))
 | |
| var supportedServiceType = sets.NewString(string(core.ServiceTypeClusterIP), string(core.ServiceTypeNodePort),
 | |
| 	string(core.ServiceTypeLoadBalancer), string(core.ServiceTypeExternalName))
 | |
| 
 | |
| var supportedServiceInternalTrafficPolicy = sets.NewString(string(core.ServiceInternalTrafficPolicyCluster), string(core.ServiceExternalTrafficPolicyTypeLocal))
 | |
| 
 | |
| var supportedServiceIPFamily = sets.NewString(string(core.IPv4Protocol), string(core.IPv6Protocol))
 | |
| var supportedServiceIPFamilyPolicy = sets.NewString(string(core.IPFamilyPolicySingleStack), string(core.IPFamilyPolicyPreferDualStack), string(core.IPFamilyPolicyRequireDualStack))
 | |
| 
 | |
| // ValidateService tests if required fields/annotations of a Service are valid.
 | |
| func ValidateService(service *core.Service) field.ErrorList {
 | |
| 	allErrs := ValidateObjectMeta(&service.ObjectMeta, true, ValidateServiceName, field.NewPath("metadata"))
 | |
| 
 | |
| 	specPath := field.NewPath("spec")
 | |
| 
 | |
| 	if len(service.Spec.Ports) == 0 && !isHeadlessService(service) && service.Spec.Type != core.ServiceTypeExternalName {
 | |
| 		allErrs = append(allErrs, field.Required(specPath.Child("ports"), ""))
 | |
| 	}
 | |
| 	switch service.Spec.Type {
 | |
| 	case core.ServiceTypeLoadBalancer:
 | |
| 		for ix := range service.Spec.Ports {
 | |
| 			port := &service.Spec.Ports[ix]
 | |
| 			// This is a workaround for broken cloud environments that
 | |
| 			// over-open firewalls.  Hopefully it can go away when more clouds
 | |
| 			// understand containers better.
 | |
| 			if port.Port == ports.KubeletPort {
 | |
| 				portPath := specPath.Child("ports").Index(ix)
 | |
| 				allErrs = append(allErrs, field.Invalid(portPath, port.Port, fmt.Sprintf("may not expose port %v externally since it is used by kubelet", ports.KubeletPort)))
 | |
| 			}
 | |
| 		}
 | |
| 		if isHeadlessService(service) {
 | |
| 			allErrs = append(allErrs, field.Invalid(specPath.Child("clusterIPs").Index(0), service.Spec.ClusterIPs[0], "may not be set to 'None' for LoadBalancer services"))
 | |
| 		}
 | |
| 	case core.ServiceTypeNodePort:
 | |
| 		if isHeadlessService(service) {
 | |
| 			allErrs = append(allErrs, field.Invalid(specPath.Child("clusterIPs").Index(0), service.Spec.ClusterIPs[0], "may not be set to 'None' for NodePort services"))
 | |
| 		}
 | |
| 	case core.ServiceTypeExternalName:
 | |
| 		// must have  len(.spec.ClusterIPs) == 0 // note: strategy sets ClusterIPs based on ClusterIP
 | |
| 		if len(service.Spec.ClusterIPs) > 0 {
 | |
| 			allErrs = append(allErrs, field.Forbidden(specPath.Child("clusterIPs"), "may not be set for ExternalName services"))
 | |
| 		}
 | |
| 
 | |
| 		// must have nil families and nil policy
 | |
| 		if len(service.Spec.IPFamilies) > 0 {
 | |
| 			allErrs = append(allErrs, field.Forbidden(specPath.Child("ipFamilies"), "may not be set for ExternalName services"))
 | |
| 		}
 | |
| 		if service.Spec.IPFamilyPolicy != nil {
 | |
| 			allErrs = append(allErrs, field.Forbidden(specPath.Child("ipFamilyPolicy"), "may not be set for ExternalName services"))
 | |
| 		}
 | |
| 
 | |
| 		// The value (a CNAME) may have a trailing dot to denote it as fully qualified
 | |
| 		cname := strings.TrimSuffix(service.Spec.ExternalName, ".")
 | |
| 		if len(cname) > 0 {
 | |
| 			allErrs = append(allErrs, ValidateDNS1123Subdomain(cname, specPath.Child("externalName"))...)
 | |
| 		} else {
 | |
| 			allErrs = append(allErrs, field.Required(specPath.Child("externalName"), ""))
 | |
| 		}
 | |
| 	}
 | |
| 
 | |
| 	allPortNames := sets.String{}
 | |
| 	portsPath := specPath.Child("ports")
 | |
| 	for i := range service.Spec.Ports {
 | |
| 		portPath := portsPath.Index(i)
 | |
| 		allErrs = append(allErrs, validateServicePort(&service.Spec.Ports[i], len(service.Spec.Ports) > 1, isHeadlessService(service), &allPortNames, portPath)...)
 | |
| 	}
 | |
| 
 | |
| 	if service.Spec.Selector != nil {
 | |
| 		allErrs = append(allErrs, unversionedvalidation.ValidateLabels(service.Spec.Selector, specPath.Child("selector"))...)
 | |
| 	}
 | |
| 
 | |
| 	if len(service.Spec.SessionAffinity) == 0 {
 | |
| 		allErrs = append(allErrs, field.Required(specPath.Child("sessionAffinity"), ""))
 | |
| 	} else if !supportedSessionAffinityType.Has(string(service.Spec.SessionAffinity)) {
 | |
| 		allErrs = append(allErrs, field.NotSupported(specPath.Child("sessionAffinity"), service.Spec.SessionAffinity, supportedSessionAffinityType.List()))
 | |
| 	}
 | |
| 
 | |
| 	if service.Spec.SessionAffinity == core.ServiceAffinityClientIP {
 | |
| 		allErrs = append(allErrs, validateClientIPAffinityConfig(service.Spec.SessionAffinityConfig, specPath.Child("sessionAffinityConfig"))...)
 | |
| 	} else if service.Spec.SessionAffinity == core.ServiceAffinityNone {
 | |
| 		if service.Spec.SessionAffinityConfig != nil {
 | |
| 			allErrs = append(allErrs, field.Forbidden(specPath.Child("sessionAffinityConfig"), fmt.Sprintf("must not be set when session affinity is %s", string(core.ServiceAffinityNone))))
 | |
| 		}
 | |
| 	}
 | |
| 
 | |
| 	// dualstack <-> ClusterIPs <-> ipfamilies
 | |
| 	allErrs = append(allErrs, validateServiceClusterIPsRelatedFields(service)...)
 | |
| 
 | |
| 	ipPath := specPath.Child("externalIPs")
 | |
| 	for i, ip := range service.Spec.ExternalIPs {
 | |
| 		idxPath := ipPath.Index(i)
 | |
| 		if msgs := validation.IsValidIP(ip); len(msgs) != 0 {
 | |
| 			for i := range msgs {
 | |
| 				allErrs = append(allErrs, field.Invalid(idxPath, ip, msgs[i]))
 | |
| 			}
 | |
| 		} else {
 | |
| 			allErrs = append(allErrs, ValidateNonSpecialIP(ip, idxPath)...)
 | |
| 		}
 | |
| 	}
 | |
| 
 | |
| 	if len(service.Spec.Type) == 0 {
 | |
| 		allErrs = append(allErrs, field.Required(specPath.Child("type"), ""))
 | |
| 	} else if !supportedServiceType.Has(string(service.Spec.Type)) {
 | |
| 		allErrs = append(allErrs, field.NotSupported(specPath.Child("type"), service.Spec.Type, supportedServiceType.List()))
 | |
| 	}
 | |
| 
 | |
| 	if service.Spec.Type == core.ServiceTypeClusterIP {
 | |
| 		portsPath := specPath.Child("ports")
 | |
| 		for i := range service.Spec.Ports {
 | |
| 			portPath := portsPath.Index(i)
 | |
| 			if service.Spec.Ports[i].NodePort != 0 {
 | |
| 				allErrs = append(allErrs, field.Forbidden(portPath.Child("nodePort"), "may not be used when `type` is 'ClusterIP'"))
 | |
| 			}
 | |
| 		}
 | |
| 	}
 | |
| 
 | |
| 	// Check for duplicate NodePorts, considering (protocol,port) pairs
 | |
| 	portsPath = specPath.Child("ports")
 | |
| 	nodePorts := make(map[core.ServicePort]bool)
 | |
| 	for i := range service.Spec.Ports {
 | |
| 		port := &service.Spec.Ports[i]
 | |
| 		if port.NodePort == 0 {
 | |
| 			continue
 | |
| 		}
 | |
| 		portPath := portsPath.Index(i)
 | |
| 		var key core.ServicePort
 | |
| 		key.Protocol = port.Protocol
 | |
| 		key.NodePort = port.NodePort
 | |
| 		_, found := nodePorts[key]
 | |
| 		if found {
 | |
| 			allErrs = append(allErrs, field.Duplicate(portPath.Child("nodePort"), port.NodePort))
 | |
| 		}
 | |
| 		nodePorts[key] = true
 | |
| 	}
 | |
| 
 | |
| 	// Check for duplicate Ports, considering (protocol,port) pairs
 | |
| 	portsPath = specPath.Child("ports")
 | |
| 	ports := make(map[core.ServicePort]bool)
 | |
| 	for i, port := range service.Spec.Ports {
 | |
| 		portPath := portsPath.Index(i)
 | |
| 		key := core.ServicePort{Protocol: port.Protocol, Port: port.Port}
 | |
| 		_, found := ports[key]
 | |
| 		if found {
 | |
| 			allErrs = append(allErrs, field.Duplicate(portPath, key))
 | |
| 		}
 | |
| 		ports[key] = true
 | |
| 	}
 | |
| 
 | |
| 	// Validate TopologyKeys
 | |
| 	if len(service.Spec.TopologyKeys) > 0 {
 | |
| 		topoPath := specPath.Child("topologyKeys")
 | |
| 		// topologyKeys is mutually exclusive with 'externalTrafficPolicy=Local'
 | |
| 		if service.Spec.ExternalTrafficPolicy == core.ServiceExternalTrafficPolicyTypeLocal {
 | |
| 			allErrs = append(allErrs, field.Forbidden(topoPath, "may not be specified when `externalTrafficPolicy=Local`"))
 | |
| 		}
 | |
| 		if len(service.Spec.TopologyKeys) > core.MaxServiceTopologyKeys {
 | |
| 			allErrs = append(allErrs, field.TooMany(topoPath, len(service.Spec.TopologyKeys), core.MaxServiceTopologyKeys))
 | |
| 		}
 | |
| 		topoKeys := sets.NewString()
 | |
| 		for i, key := range service.Spec.TopologyKeys {
 | |
| 			keyPath := topoPath.Index(i)
 | |
| 			if topoKeys.Has(key) {
 | |
| 				allErrs = append(allErrs, field.Duplicate(keyPath, key))
 | |
| 			}
 | |
| 			topoKeys.Insert(key)
 | |
| 			// "Any" must be the last value specified
 | |
| 			if key == v1.TopologyKeyAny && i != len(service.Spec.TopologyKeys)-1 {
 | |
| 				allErrs = append(allErrs, field.Invalid(keyPath, key, `"*" must be the last value specified`))
 | |
| 			}
 | |
| 			if key != v1.TopologyKeyAny {
 | |
| 				for _, msg := range validation.IsQualifiedName(key) {
 | |
| 					allErrs = append(allErrs, field.Invalid(keyPath, service.Spec.TopologyKeys, msg))
 | |
| 				}
 | |
| 			}
 | |
| 		}
 | |
| 	}
 | |
| 
 | |
| 	// Validate SourceRange field and annotation
 | |
| 	_, ok := service.Annotations[core.AnnotationLoadBalancerSourceRangesKey]
 | |
| 	if len(service.Spec.LoadBalancerSourceRanges) > 0 || ok {
 | |
| 		var fieldPath *field.Path
 | |
| 		var val string
 | |
| 		if len(service.Spec.LoadBalancerSourceRanges) > 0 {
 | |
| 			fieldPath = specPath.Child("LoadBalancerSourceRanges")
 | |
| 			val = fmt.Sprintf("%v", service.Spec.LoadBalancerSourceRanges)
 | |
| 		} else {
 | |
| 			fieldPath = field.NewPath("metadata", "annotations").Key(core.AnnotationLoadBalancerSourceRangesKey)
 | |
| 			val = service.Annotations[core.AnnotationLoadBalancerSourceRangesKey]
 | |
| 		}
 | |
| 		if service.Spec.Type != core.ServiceTypeLoadBalancer {
 | |
| 			allErrs = append(allErrs, field.Forbidden(fieldPath, "may only be used when `type` is 'LoadBalancer'"))
 | |
| 		}
 | |
| 		_, err := apiservice.GetLoadBalancerSourceRanges(service)
 | |
| 		if err != nil {
 | |
| 			allErrs = append(allErrs, field.Invalid(fieldPath, val, "must be a list of IP ranges. For example, 10.240.0.0/24,10.250.0.0/24 "))
 | |
| 		}
 | |
| 	}
 | |
| 
 | |
| 	if service.Spec.AllocateLoadBalancerNodePorts != nil && service.Spec.Type != core.ServiceTypeLoadBalancer {
 | |
| 		allErrs = append(allErrs, field.Forbidden(specPath.Child("allocateLoadBalancerNodePorts"), "may only be used when `type` is 'LoadBalancer'"))
 | |
| 	}
 | |
| 
 | |
| 	if utilfeature.DefaultFeatureGate.Enabled(features.ServiceLBNodePortControl) {
 | |
| 		if service.Spec.Type == core.ServiceTypeLoadBalancer && service.Spec.AllocateLoadBalancerNodePorts == nil {
 | |
| 			allErrs = append(allErrs, field.Required(field.NewPath("allocateLoadBalancerNodePorts"), ""))
 | |
| 		}
 | |
| 	}
 | |
| 
 | |
| 	// validate LoadBalancerClass field
 | |
| 	allErrs = append(allErrs, validateLoadBalancerClassField(nil, service)...)
 | |
| 
 | |
| 	// external traffic fields
 | |
| 	allErrs = append(allErrs, validateServiceExternalTrafficFieldsValue(service)...)
 | |
| 
 | |
| 	// internal traffic policy field
 | |
| 	allErrs = append(allErrs, validateServiceInternalTrafficFieldsValue(service)...)
 | |
| 
 | |
| 	return allErrs
 | |
| }
 | |
| 
 | |
| func validateServicePort(sp *core.ServicePort, requireName, isHeadlessService bool, allNames *sets.String, fldPath *field.Path) field.ErrorList {
 | |
| 	allErrs := field.ErrorList{}
 | |
| 
 | |
| 	if requireName && len(sp.Name) == 0 {
 | |
| 		allErrs = append(allErrs, field.Required(fldPath.Child("name"), ""))
 | |
| 	} else if len(sp.Name) != 0 {
 | |
| 		allErrs = append(allErrs, ValidateDNS1123Label(sp.Name, fldPath.Child("name"))...)
 | |
| 		if allNames.Has(sp.Name) {
 | |
| 			allErrs = append(allErrs, field.Duplicate(fldPath.Child("name"), sp.Name))
 | |
| 		} else {
 | |
| 			allNames.Insert(sp.Name)
 | |
| 		}
 | |
| 	}
 | |
| 
 | |
| 	for _, msg := range validation.IsValidPortNum(int(sp.Port)) {
 | |
| 		allErrs = append(allErrs, field.Invalid(fldPath.Child("port"), sp.Port, msg))
 | |
| 	}
 | |
| 
 | |
| 	if len(sp.Protocol) == 0 {
 | |
| 		allErrs = append(allErrs, field.Required(fldPath.Child("protocol"), ""))
 | |
| 	} else if !supportedPortProtocols.Has(string(sp.Protocol)) {
 | |
| 		allErrs = append(allErrs, field.NotSupported(fldPath.Child("protocol"), sp.Protocol, supportedPortProtocols.List()))
 | |
| 	}
 | |
| 
 | |
| 	allErrs = append(allErrs, ValidatePortNumOrName(sp.TargetPort, fldPath.Child("targetPort"))...)
 | |
| 
 | |
| 	if sp.AppProtocol != nil {
 | |
| 		for _, msg := range validation.IsQualifiedName(*sp.AppProtocol) {
 | |
| 			allErrs = append(allErrs, field.Invalid(fldPath.Child("appProtocol"), sp.AppProtocol, msg))
 | |
| 		}
 | |
| 	}
 | |
| 
 | |
| 	// in the v1 API, targetPorts on headless services were tolerated.
 | |
| 	// once we have version-specific validation, we can reject this on newer API versions, but until then, we have to tolerate it for compatibility.
 | |
| 	//
 | |
| 	// if isHeadlessService {
 | |
| 	// 	if sp.TargetPort.Type == intstr.String || (sp.TargetPort.Type == intstr.Int && sp.Port != sp.TargetPort.IntValue()) {
 | |
| 	// 		allErrs = append(allErrs, field.Invalid(fldPath.Child("targetPort"), sp.TargetPort, "must be equal to the value of 'port' when clusterIP = None"))
 | |
| 	// 	}
 | |
| 	// }
 | |
| 
 | |
| 	return allErrs
 | |
| }
 | |
| 
 | |
| // validateServiceExternalTrafficFieldsValue validates ExternalTraffic related annotations
 | |
| // have legal value.
 | |
| func validateServiceExternalTrafficFieldsValue(service *core.Service) field.ErrorList {
 | |
| 	allErrs := field.ErrorList{}
 | |
| 
 | |
| 	// Check first class fields.
 | |
| 	if service.Spec.ExternalTrafficPolicy != "" &&
 | |
| 		service.Spec.ExternalTrafficPolicy != core.ServiceExternalTrafficPolicyTypeCluster &&
 | |
| 		service.Spec.ExternalTrafficPolicy != core.ServiceExternalTrafficPolicyTypeLocal {
 | |
| 		allErrs = append(allErrs, field.Invalid(field.NewPath("spec").Child("externalTrafficPolicy"), service.Spec.ExternalTrafficPolicy,
 | |
| 			fmt.Sprintf("ExternalTrafficPolicy must be empty, %v or %v", core.ServiceExternalTrafficPolicyTypeCluster, core.ServiceExternalTrafficPolicyTypeLocal)))
 | |
| 	}
 | |
| 
 | |
| 	if service.Spec.HealthCheckNodePort < 0 {
 | |
| 		allErrs = append(allErrs, field.Invalid(field.NewPath("spec").Child("healthCheckNodePort"), service.Spec.HealthCheckNodePort,
 | |
| 			"HealthCheckNodePort must be not less than 0"))
 | |
| 	}
 | |
| 
 | |
| 	return allErrs
 | |
| }
 | |
| 
 | |
| // validateServiceInternalTrafficFieldsValue validates InternalTraffic related
 | |
| // spec have legal value.
 | |
| func validateServiceInternalTrafficFieldsValue(service *core.Service) field.ErrorList {
 | |
| 	allErrs := field.ErrorList{}
 | |
| 
 | |
| 	if utilfeature.DefaultFeatureGate.Enabled(features.ServiceInternalTrafficPolicy) {
 | |
| 		if service.Spec.InternalTrafficPolicy == nil {
 | |
| 			allErrs = append(allErrs, field.Required(field.NewPath("spec").Child("internalTrafficPolicy"), ""))
 | |
| 		}
 | |
| 	}
 | |
| 
 | |
| 	if service.Spec.InternalTrafficPolicy != nil && !supportedServiceInternalTrafficPolicy.Has(string(*service.Spec.InternalTrafficPolicy)) {
 | |
| 		allErrs = append(allErrs, field.NotSupported(field.NewPath("spec").Child("internalTrafficPolicy"), *service.Spec.InternalTrafficPolicy, supportedServiceInternalTrafficPolicy.List()))
 | |
| 	}
 | |
| 
 | |
| 	return allErrs
 | |
| }
 | |
| 
 | |
| // ValidateServiceExternalTrafficFieldsCombination validates if ExternalTrafficPolicy,
 | |
| // HealthCheckNodePort and Type combination are legal. For update, it should be called
 | |
| // after clearing externalTraffic related fields for the ease of transitioning between
 | |
| // different service types.
 | |
| func ValidateServiceExternalTrafficFieldsCombination(service *core.Service) field.ErrorList {
 | |
| 	allErrs := field.ErrorList{}
 | |
| 
 | |
| 	if service.Spec.Type != core.ServiceTypeLoadBalancer &&
 | |
| 		service.Spec.Type != core.ServiceTypeNodePort &&
 | |
| 		service.Spec.ExternalTrafficPolicy != "" {
 | |
| 		allErrs = append(allErrs, field.Invalid(field.NewPath("spec", "externalTrafficPolicy"), service.Spec.ExternalTrafficPolicy,
 | |
| 			"ExternalTrafficPolicy can only be set on NodePort and LoadBalancer service"))
 | |
| 	}
 | |
| 
 | |
| 	if !apiservice.NeedsHealthCheck(service) &&
 | |
| 		service.Spec.HealthCheckNodePort != 0 {
 | |
| 		allErrs = append(allErrs, field.Invalid(field.NewPath("spec", "healthCheckNodePort"), service.Spec.HealthCheckNodePort,
 | |
| 			"HealthCheckNodePort can only be set on LoadBalancer service with ExternalTrafficPolicy=Local"))
 | |
| 	}
 | |
| 
 | |
| 	return allErrs
 | |
| }
 | |
| 
 | |
| // ValidateServiceCreate validates Services as they are created.
 | |
| func ValidateServiceCreate(service *core.Service) field.ErrorList {
 | |
| 	return ValidateService(service)
 | |
| }
 | |
| 
 | |
| // ValidateServiceUpdate tests if required fields in the service are set during an update
 | |
| func ValidateServiceUpdate(service, oldService *core.Service) field.ErrorList {
 | |
| 	allErrs := ValidateObjectMetaUpdate(&service.ObjectMeta, &oldService.ObjectMeta, field.NewPath("metadata"))
 | |
| 
 | |
| 	// User can upgrade (add another clusterIP or ipFamily)
 | |
| 	//      can downgrade (remove secondary clusterIP or ipFamily)
 | |
| 	// but *CAN NOT* change primary/secondary clusterIP || ipFamily *UNLESS*
 | |
| 	// they are changing from/to/ON ExternalName
 | |
| 
 | |
| 	upgradeDowngradeClusterIPsErrs := validateUpgradeDowngradeClusterIPs(oldService, service)
 | |
| 	allErrs = append(allErrs, upgradeDowngradeClusterIPsErrs...)
 | |
| 
 | |
| 	upgradeDowngradeIPFamiliesErrs := validateUpgradeDowngradeIPFamilies(oldService, service)
 | |
| 	allErrs = append(allErrs, upgradeDowngradeIPFamiliesErrs...)
 | |
| 
 | |
| 	upgradeDowngradeLoadBalancerClassErrs := validateLoadBalancerClassField(oldService, service)
 | |
| 	allErrs = append(allErrs, upgradeDowngradeLoadBalancerClassErrs...)
 | |
| 
 | |
| 	return append(allErrs, ValidateService(service)...)
 | |
| }
 | |
| 
 | |
| // ValidateServiceStatusUpdate tests if required fields in the Service are set when updating status.
 | |
| func ValidateServiceStatusUpdate(service, oldService *core.Service) field.ErrorList {
 | |
| 	allErrs := ValidateObjectMetaUpdate(&service.ObjectMeta, &oldService.ObjectMeta, field.NewPath("metadata"))
 | |
| 	allErrs = append(allErrs, ValidateLoadBalancerStatus(&service.Status.LoadBalancer, field.NewPath("status", "loadBalancer"))...)
 | |
| 	return allErrs
 | |
| }
 | |
| 
 | |
| // ValidateReplicationController tests if required fields in the replication controller are set.
 | |
| func ValidateReplicationController(controller *core.ReplicationController, opts PodValidationOptions) field.ErrorList {
 | |
| 	allErrs := ValidateObjectMeta(&controller.ObjectMeta, true, ValidateReplicationControllerName, field.NewPath("metadata"))
 | |
| 	allErrs = append(allErrs, ValidateReplicationControllerSpec(&controller.Spec, field.NewPath("spec"), opts)...)
 | |
| 	return allErrs
 | |
| }
 | |
| 
 | |
| // ValidateReplicationControllerUpdate tests if required fields in the replication controller are set.
 | |
| func ValidateReplicationControllerUpdate(controller, oldController *core.ReplicationController, opts PodValidationOptions) field.ErrorList {
 | |
| 	allErrs := ValidateObjectMetaUpdate(&controller.ObjectMeta, &oldController.ObjectMeta, field.NewPath("metadata"))
 | |
| 	allErrs = append(allErrs, ValidateReplicationControllerSpec(&controller.Spec, field.NewPath("spec"), opts)...)
 | |
| 	return allErrs
 | |
| }
 | |
| 
 | |
| // ValidateReplicationControllerStatusUpdate tests if required fields in the replication controller are set.
 | |
| func ValidateReplicationControllerStatusUpdate(controller, oldController *core.ReplicationController) field.ErrorList {
 | |
| 	allErrs := ValidateObjectMetaUpdate(&controller.ObjectMeta, &oldController.ObjectMeta, field.NewPath("metadata"))
 | |
| 	allErrs = append(allErrs, ValidateReplicationControllerStatus(controller.Status, field.NewPath("status"))...)
 | |
| 	return allErrs
 | |
| }
 | |
| 
 | |
| func ValidateReplicationControllerStatus(status core.ReplicationControllerStatus, statusPath *field.Path) field.ErrorList {
 | |
| 	allErrs := field.ErrorList{}
 | |
| 	allErrs = append(allErrs, ValidateNonnegativeField(int64(status.Replicas), statusPath.Child("replicas"))...)
 | |
| 	allErrs = append(allErrs, ValidateNonnegativeField(int64(status.FullyLabeledReplicas), statusPath.Child("fullyLabeledReplicas"))...)
 | |
| 	allErrs = append(allErrs, ValidateNonnegativeField(int64(status.ReadyReplicas), statusPath.Child("readyReplicas"))...)
 | |
| 	allErrs = append(allErrs, ValidateNonnegativeField(int64(status.AvailableReplicas), statusPath.Child("availableReplicas"))...)
 | |
| 	allErrs = append(allErrs, ValidateNonnegativeField(int64(status.ObservedGeneration), statusPath.Child("observedGeneration"))...)
 | |
| 	msg := "cannot be greater than status.replicas"
 | |
| 	if status.FullyLabeledReplicas > status.Replicas {
 | |
| 		allErrs = append(allErrs, field.Invalid(statusPath.Child("fullyLabeledReplicas"), status.FullyLabeledReplicas, msg))
 | |
| 	}
 | |
| 	if status.ReadyReplicas > status.Replicas {
 | |
| 		allErrs = append(allErrs, field.Invalid(statusPath.Child("readyReplicas"), status.ReadyReplicas, msg))
 | |
| 	}
 | |
| 	if status.AvailableReplicas > status.Replicas {
 | |
| 		allErrs = append(allErrs, field.Invalid(statusPath.Child("availableReplicas"), status.AvailableReplicas, msg))
 | |
| 	}
 | |
| 	if status.AvailableReplicas > status.ReadyReplicas {
 | |
| 		allErrs = append(allErrs, field.Invalid(statusPath.Child("availableReplicas"), status.AvailableReplicas, "cannot be greater than readyReplicas"))
 | |
| 	}
 | |
| 	return allErrs
 | |
| }
 | |
| 
 | |
| // Validates that the given selector is non-empty.
 | |
| func ValidateNonEmptySelector(selectorMap map[string]string, fldPath *field.Path) field.ErrorList {
 | |
| 	allErrs := field.ErrorList{}
 | |
| 	selector := labels.Set(selectorMap).AsSelector()
 | |
| 	if selector.Empty() {
 | |
| 		allErrs = append(allErrs, field.Required(fldPath, ""))
 | |
| 	}
 | |
| 	return allErrs
 | |
| }
 | |
| 
 | |
| // Validates the given template and ensures that it is in accordance with the desired selector and replicas.
 | |
| func ValidatePodTemplateSpecForRC(template *core.PodTemplateSpec, selectorMap map[string]string, replicas int32, fldPath *field.Path, opts PodValidationOptions) field.ErrorList {
 | |
| 	allErrs := field.ErrorList{}
 | |
| 	if template == nil {
 | |
| 		allErrs = append(allErrs, field.Required(fldPath, ""))
 | |
| 	} else {
 | |
| 		selector := labels.Set(selectorMap).AsSelector()
 | |
| 		if !selector.Empty() {
 | |
| 			// Verify that the RC selector matches the labels in template.
 | |
| 			labels := labels.Set(template.Labels)
 | |
| 			if !selector.Matches(labels) {
 | |
| 				allErrs = append(allErrs, field.Invalid(fldPath.Child("metadata", "labels"), template.Labels, "`selector` does not match template `labels`"))
 | |
| 			}
 | |
| 		}
 | |
| 		allErrs = append(allErrs, ValidatePodTemplateSpec(template, fldPath, opts)...)
 | |
| 		if replicas > 1 {
 | |
| 			allErrs = append(allErrs, ValidateReadOnlyPersistentDisks(template.Spec.Volumes, fldPath.Child("spec", "volumes"))...)
 | |
| 		}
 | |
| 		// RestartPolicy has already been first-order validated as per ValidatePodTemplateSpec().
 | |
| 		if template.Spec.RestartPolicy != core.RestartPolicyAlways {
 | |
| 			allErrs = append(allErrs, field.NotSupported(fldPath.Child("spec", "restartPolicy"), template.Spec.RestartPolicy, []string{string(core.RestartPolicyAlways)}))
 | |
| 		}
 | |
| 		if template.Spec.ActiveDeadlineSeconds != nil {
 | |
| 			allErrs = append(allErrs, field.Forbidden(fldPath.Child("spec", "activeDeadlineSeconds"), "activeDeadlineSeconds in ReplicationController is not Supported"))
 | |
| 		}
 | |
| 	}
 | |
| 	return allErrs
 | |
| }
 | |
| 
 | |
| // ValidateReplicationControllerSpec tests if required fields in the replication controller spec are set.
 | |
| func ValidateReplicationControllerSpec(spec *core.ReplicationControllerSpec, fldPath *field.Path, opts PodValidationOptions) field.ErrorList {
 | |
| 	allErrs := field.ErrorList{}
 | |
| 	allErrs = append(allErrs, ValidateNonnegativeField(int64(spec.MinReadySeconds), fldPath.Child("minReadySeconds"))...)
 | |
| 	allErrs = append(allErrs, ValidateNonEmptySelector(spec.Selector, fldPath.Child("selector"))...)
 | |
| 	allErrs = append(allErrs, ValidateNonnegativeField(int64(spec.Replicas), fldPath.Child("replicas"))...)
 | |
| 	allErrs = append(allErrs, ValidatePodTemplateSpecForRC(spec.Template, spec.Selector, spec.Replicas, fldPath.Child("template"), opts)...)
 | |
| 	return allErrs
 | |
| }
 | |
| 
 | |
| // ValidatePodTemplateSpec validates the spec of a pod template
 | |
| func ValidatePodTemplateSpec(spec *core.PodTemplateSpec, fldPath *field.Path, opts PodValidationOptions) field.ErrorList {
 | |
| 	allErrs := field.ErrorList{}
 | |
| 	allErrs = append(allErrs, unversionedvalidation.ValidateLabels(spec.Labels, fldPath.Child("labels"))...)
 | |
| 	allErrs = append(allErrs, ValidateAnnotations(spec.Annotations, fldPath.Child("annotations"))...)
 | |
| 	allErrs = append(allErrs, ValidatePodSpecificAnnotations(spec.Annotations, &spec.Spec, fldPath.Child("annotations"), opts)...)
 | |
| 	allErrs = append(allErrs, ValidatePodSpec(&spec.Spec, nil, fldPath.Child("spec"), opts)...)
 | |
| 	allErrs = append(allErrs, validateSeccompAnnotationsAndFields(spec.ObjectMeta, &spec.Spec, fldPath.Child("spec"))...)
 | |
| 
 | |
| 	if len(spec.Spec.EphemeralContainers) > 0 {
 | |
| 		allErrs = append(allErrs, field.Forbidden(fldPath.Child("spec", "ephemeralContainers"), "ephemeral containers not allowed in pod template"))
 | |
| 	}
 | |
| 
 | |
| 	return allErrs
 | |
| }
 | |
| 
 | |
| func ValidateReadOnlyPersistentDisks(volumes []core.Volume, fldPath *field.Path) field.ErrorList {
 | |
| 	allErrs := field.ErrorList{}
 | |
| 	for i := range volumes {
 | |
| 		vol := &volumes[i]
 | |
| 		idxPath := fldPath.Index(i)
 | |
| 		if vol.GCEPersistentDisk != nil {
 | |
| 			if !vol.GCEPersistentDisk.ReadOnly {
 | |
| 				allErrs = append(allErrs, field.Invalid(idxPath.Child("gcePersistentDisk", "readOnly"), false, "must be true for replicated pods > 1; GCE PD can only be mounted on multiple machines if it is read-only"))
 | |
| 			}
 | |
| 		}
 | |
| 		// TODO: What to do for AWS?  It doesn't support replicas
 | |
| 	}
 | |
| 	return allErrs
 | |
| }
 | |
| 
 | |
| // ValidateTaintsInNodeAnnotations tests that the serialized taints in Node.Annotations has valid data
 | |
| func ValidateTaintsInNodeAnnotations(annotations map[string]string, fldPath *field.Path) field.ErrorList {
 | |
| 	allErrs := field.ErrorList{}
 | |
| 
 | |
| 	taints, err := helper.GetTaintsFromNodeAnnotations(annotations)
 | |
| 	if err != nil {
 | |
| 		allErrs = append(allErrs, field.Invalid(fldPath, core.TaintsAnnotationKey, err.Error()))
 | |
| 		return allErrs
 | |
| 	}
 | |
| 
 | |
| 	if len(taints) > 0 {
 | |
| 		allErrs = append(allErrs, validateNodeTaints(taints, fldPath.Child(core.TaintsAnnotationKey))...)
 | |
| 	}
 | |
| 
 | |
| 	return allErrs
 | |
| }
 | |
| 
 | |
| // validateNodeTaints tests if given taints have valid data.
 | |
| func validateNodeTaints(taints []core.Taint, fldPath *field.Path) field.ErrorList {
 | |
| 	allErrors := field.ErrorList{}
 | |
| 
 | |
| 	uniqueTaints := map[core.TaintEffect]sets.String{}
 | |
| 
 | |
| 	for i, currTaint := range taints {
 | |
| 		idxPath := fldPath.Index(i)
 | |
| 		// validate the taint key
 | |
| 		allErrors = append(allErrors, unversionedvalidation.ValidateLabelName(currTaint.Key, idxPath.Child("key"))...)
 | |
| 		// validate the taint value
 | |
| 		if errs := validation.IsValidLabelValue(currTaint.Value); len(errs) != 0 {
 | |
| 			allErrors = append(allErrors, field.Invalid(idxPath.Child("value"), currTaint.Value, strings.Join(errs, ";")))
 | |
| 		}
 | |
| 		// validate the taint effect
 | |
| 		allErrors = append(allErrors, validateTaintEffect(&currTaint.Effect, false, idxPath.Child("effect"))...)
 | |
| 
 | |
| 		// validate if taint is unique by <key, effect>
 | |
| 		if len(uniqueTaints[currTaint.Effect]) > 0 && uniqueTaints[currTaint.Effect].Has(currTaint.Key) {
 | |
| 			duplicatedError := field.Duplicate(idxPath, currTaint)
 | |
| 			duplicatedError.Detail = "taints must be unique by key and effect pair"
 | |
| 			allErrors = append(allErrors, duplicatedError)
 | |
| 			continue
 | |
| 		}
 | |
| 
 | |
| 		// add taint to existingTaints for uniqueness check
 | |
| 		if len(uniqueTaints[currTaint.Effect]) == 0 {
 | |
| 			uniqueTaints[currTaint.Effect] = sets.String{}
 | |
| 		}
 | |
| 		uniqueTaints[currTaint.Effect].Insert(currTaint.Key)
 | |
| 	}
 | |
| 	return allErrors
 | |
| }
 | |
| 
 | |
| func ValidateNodeSpecificAnnotations(annotations map[string]string, fldPath *field.Path) field.ErrorList {
 | |
| 	allErrs := field.ErrorList{}
 | |
| 
 | |
| 	if annotations[core.TaintsAnnotationKey] != "" {
 | |
| 		allErrs = append(allErrs, ValidateTaintsInNodeAnnotations(annotations, fldPath)...)
 | |
| 	}
 | |
| 
 | |
| 	if annotations[core.PreferAvoidPodsAnnotationKey] != "" {
 | |
| 		allErrs = append(allErrs, ValidateAvoidPodsInNodeAnnotations(annotations, fldPath)...)
 | |
| 	}
 | |
| 	return allErrs
 | |
| }
 | |
| 
 | |
| // ValidateNode tests if required fields in the node are set.
 | |
| func ValidateNode(node *core.Node) field.ErrorList {
 | |
| 	fldPath := field.NewPath("metadata")
 | |
| 	allErrs := ValidateObjectMeta(&node.ObjectMeta, false, ValidateNodeName, fldPath)
 | |
| 	allErrs = append(allErrs, ValidateNodeSpecificAnnotations(node.ObjectMeta.Annotations, fldPath.Child("annotations"))...)
 | |
| 	if len(node.Spec.Taints) > 0 {
 | |
| 		allErrs = append(allErrs, validateNodeTaints(node.Spec.Taints, fldPath.Child("taints"))...)
 | |
| 	}
 | |
| 
 | |
| 	// Only validate spec.
 | |
| 	// All status fields are optional and can be updated later.
 | |
| 	// That said, if specified, we need to ensure they are valid.
 | |
| 	allErrs = append(allErrs, ValidateNodeResources(node)...)
 | |
| 
 | |
| 	// validate PodCIDRS only if we need to
 | |
| 	if len(node.Spec.PodCIDRs) > 0 {
 | |
| 		podCIDRsField := field.NewPath("spec", "podCIDRs")
 | |
| 
 | |
| 		// all PodCIDRs should be valid ones
 | |
| 		for idx, value := range node.Spec.PodCIDRs {
 | |
| 			if _, err := ValidateCIDR(value); err != nil {
 | |
| 				allErrs = append(allErrs, field.Invalid(podCIDRsField.Index(idx), node.Spec.PodCIDRs, "must be valid CIDR"))
 | |
| 			}
 | |
| 		}
 | |
| 
 | |
| 		// if more than PodCIDR then
 | |
| 		// - validate for dual stack
 | |
| 		// - validate for duplication
 | |
| 		if len(node.Spec.PodCIDRs) > 1 {
 | |
| 			dualStack, err := netutils.IsDualStackCIDRStrings(node.Spec.PodCIDRs)
 | |
| 			if err != nil {
 | |
| 				allErrs = append(allErrs, field.InternalError(podCIDRsField, fmt.Errorf("invalid PodCIDRs. failed to check with dual stack with error:%v", err)))
 | |
| 			}
 | |
| 			if !dualStack || len(node.Spec.PodCIDRs) > 2 {
 | |
| 				allErrs = append(allErrs, field.Invalid(podCIDRsField, node.Spec.PodCIDRs, "may specify no more than one CIDR for each IP family"))
 | |
| 			}
 | |
| 
 | |
| 			// PodCIDRs must not contain duplicates
 | |
| 			seen := sets.String{}
 | |
| 			for i, value := range node.Spec.PodCIDRs {
 | |
| 				if seen.Has(value) {
 | |
| 					allErrs = append(allErrs, field.Duplicate(podCIDRsField.Index(i), value))
 | |
| 				}
 | |
| 				seen.Insert(value)
 | |
| 			}
 | |
| 		}
 | |
| 	}
 | |
| 
 | |
| 	return allErrs
 | |
| }
 | |
| 
 | |
| // ValidateNodeResources is used to make sure a node has valid capacity and allocatable values.
 | |
| func ValidateNodeResources(node *core.Node) field.ErrorList {
 | |
| 	allErrs := field.ErrorList{}
 | |
| 
 | |
| 	// Validate resource quantities in capacity.
 | |
| 	for k, v := range node.Status.Capacity {
 | |
| 		resPath := field.NewPath("status", "capacity", string(k))
 | |
| 		allErrs = append(allErrs, ValidateResourceQuantityValue(string(k), v, resPath)...)
 | |
| 	}
 | |
| 
 | |
| 	// Validate resource quantities in allocatable.
 | |
| 	for k, v := range node.Status.Allocatable {
 | |
| 		resPath := field.NewPath("status", "allocatable", string(k))
 | |
| 		allErrs = append(allErrs, ValidateResourceQuantityValue(string(k), v, resPath)...)
 | |
| 	}
 | |
| 	return allErrs
 | |
| }
 | |
| 
 | |
| // ValidateNodeUpdate tests to make sure a node update can be applied.  Modifies oldNode.
 | |
| func ValidateNodeUpdate(node, oldNode *core.Node) field.ErrorList {
 | |
| 	fldPath := field.NewPath("metadata")
 | |
| 	allErrs := ValidateObjectMetaUpdate(&node.ObjectMeta, &oldNode.ObjectMeta, fldPath)
 | |
| 	allErrs = append(allErrs, ValidateNodeSpecificAnnotations(node.ObjectMeta.Annotations, fldPath.Child("annotations"))...)
 | |
| 
 | |
| 	// TODO: Enable the code once we have better core object.status update model. Currently,
 | |
| 	// anyone can update node status.
 | |
| 	// if !apiequality.Semantic.DeepEqual(node.Status, core.NodeStatus{}) {
 | |
| 	// 	allErrs = append(allErrs, field.Invalid("status", node.Status, "must be empty"))
 | |
| 	// }
 | |
| 
 | |
| 	allErrs = append(allErrs, ValidateNodeResources(node)...)
 | |
| 
 | |
| 	// Validate no duplicate addresses in node status.
 | |
| 	addresses := make(map[core.NodeAddress]bool)
 | |
| 	for i, address := range node.Status.Addresses {
 | |
| 		if _, ok := addresses[address]; ok {
 | |
| 			allErrs = append(allErrs, field.Duplicate(field.NewPath("status", "addresses").Index(i), address))
 | |
| 		}
 | |
| 		addresses[address] = true
 | |
| 	}
 | |
| 
 | |
| 	// Allow the controller manager to assign a CIDR to a node if it doesn't have one.
 | |
| 	if len(oldNode.Spec.PodCIDRs) > 0 {
 | |
| 		// compare the entire slice
 | |
| 		if len(oldNode.Spec.PodCIDRs) != len(node.Spec.PodCIDRs) {
 | |
| 			allErrs = append(allErrs, field.Forbidden(field.NewPath("spec", "podCIDRs"), "node updates may not change podCIDR except from \"\" to valid"))
 | |
| 		} else {
 | |
| 			for idx, value := range oldNode.Spec.PodCIDRs {
 | |
| 				if value != node.Spec.PodCIDRs[idx] {
 | |
| 					allErrs = append(allErrs, field.Forbidden(field.NewPath("spec", "podCIDRs"), "node updates may not change podCIDR except from \"\" to valid"))
 | |
| 				}
 | |
| 			}
 | |
| 		}
 | |
| 	}
 | |
| 
 | |
| 	// Allow controller manager updating provider ID when not set
 | |
| 	if len(oldNode.Spec.ProviderID) > 0 && oldNode.Spec.ProviderID != node.Spec.ProviderID {
 | |
| 		allErrs = append(allErrs, field.Forbidden(field.NewPath("spec", "providerID"), "node updates may not change providerID except from \"\" to valid"))
 | |
| 	}
 | |
| 
 | |
| 	if node.Spec.ConfigSource != nil {
 | |
| 		allErrs = append(allErrs, validateNodeConfigSourceSpec(node.Spec.ConfigSource, field.NewPath("spec", "configSource"))...)
 | |
| 	}
 | |
| 	if node.Status.Config != nil {
 | |
| 		allErrs = append(allErrs, validateNodeConfigStatus(node.Status.Config, field.NewPath("status", "config"))...)
 | |
| 	}
 | |
| 
 | |
| 	// update taints
 | |
| 	if len(node.Spec.Taints) > 0 {
 | |
| 		allErrs = append(allErrs, validateNodeTaints(node.Spec.Taints, fldPath.Child("taints"))...)
 | |
| 	}
 | |
| 
 | |
| 	if node.Spec.DoNotUseExternalID != oldNode.Spec.DoNotUseExternalID {
 | |
| 		allErrs = append(allErrs, field.Forbidden(field.NewPath("spec", "externalID"), "may not be updated"))
 | |
| 	}
 | |
| 
 | |
| 	// status and metadata are allowed change (barring restrictions above), so separately test spec field.
 | |
| 	// spec only has a few fields, so check the ones we don't allow changing
 | |
| 	//  1. PodCIDRs - immutable after first set - checked above
 | |
| 	//  2. ProviderID - immutable after first set - checked above
 | |
| 	//  3. Unschedulable - allowed to change
 | |
| 	//  4. Taints - allowed to change
 | |
| 	//  5. ConfigSource - allowed to change (and checked above)
 | |
| 	//  6. DoNotUseExternalID - immutable - checked above
 | |
| 
 | |
| 	return allErrs
 | |
| }
 | |
| 
 | |
| // validation specific to Node.Spec.ConfigSource
 | |
| func validateNodeConfigSourceSpec(source *core.NodeConfigSource, fldPath *field.Path) field.ErrorList {
 | |
| 	allErrs := field.ErrorList{}
 | |
| 	count := int(0)
 | |
| 	if source.ConfigMap != nil {
 | |
| 		count++
 | |
| 		allErrs = append(allErrs, validateConfigMapNodeConfigSourceSpec(source.ConfigMap, fldPath.Child("configMap"))...)
 | |
| 	}
 | |
| 	// add more subfields here in the future as they are added to NodeConfigSource
 | |
| 
 | |
| 	// exactly one reference subfield must be non-nil
 | |
| 	if count != 1 {
 | |
| 		allErrs = append(allErrs, field.Invalid(fldPath, source, "exactly one reference subfield must be non-nil"))
 | |
| 	}
 | |
| 	return allErrs
 | |
| }
 | |
| 
 | |
| // validation specific to Node.Spec.ConfigSource.ConfigMap
 | |
| func validateConfigMapNodeConfigSourceSpec(source *core.ConfigMapNodeConfigSource, fldPath *field.Path) field.ErrorList {
 | |
| 	allErrs := field.ErrorList{}
 | |
| 	// uid and resourceVersion must not be set in spec
 | |
| 	if string(source.UID) != "" {
 | |
| 		allErrs = append(allErrs, field.Forbidden(fldPath.Child("uid"), "uid must not be set in spec"))
 | |
| 	}
 | |
| 	if source.ResourceVersion != "" {
 | |
| 		allErrs = append(allErrs, field.Forbidden(fldPath.Child("resourceVersion"), "resourceVersion must not be set in spec"))
 | |
| 	}
 | |
| 	return append(allErrs, validateConfigMapNodeConfigSource(source, fldPath)...)
 | |
| }
 | |
| 
 | |
| // validation specififc to Node.Status.Config
 | |
| func validateNodeConfigStatus(status *core.NodeConfigStatus, fldPath *field.Path) field.ErrorList {
 | |
| 	allErrs := field.ErrorList{}
 | |
| 	if status.Assigned != nil {
 | |
| 		allErrs = append(allErrs, validateNodeConfigSourceStatus(status.Assigned, fldPath.Child("assigned"))...)
 | |
| 	}
 | |
| 	if status.Active != nil {
 | |
| 		allErrs = append(allErrs, validateNodeConfigSourceStatus(status.Active, fldPath.Child("active"))...)
 | |
| 	}
 | |
| 	if status.LastKnownGood != nil {
 | |
| 		allErrs = append(allErrs, validateNodeConfigSourceStatus(status.LastKnownGood, fldPath.Child("lastKnownGood"))...)
 | |
| 	}
 | |
| 	return allErrs
 | |
| }
 | |
| 
 | |
| // validation specific to Node.Status.Config.(Active|Assigned|LastKnownGood)
 | |
| func validateNodeConfigSourceStatus(source *core.NodeConfigSource, fldPath *field.Path) field.ErrorList {
 | |
| 	allErrs := field.ErrorList{}
 | |
| 	count := int(0)
 | |
| 	if source.ConfigMap != nil {
 | |
| 		count++
 | |
| 		allErrs = append(allErrs, validateConfigMapNodeConfigSourceStatus(source.ConfigMap, fldPath.Child("configMap"))...)
 | |
| 	}
 | |
| 	// add more subfields here in the future as they are added to NodeConfigSource
 | |
| 
 | |
| 	// exactly one reference subfield must be non-nil
 | |
| 	if count != 1 {
 | |
| 		allErrs = append(allErrs, field.Invalid(fldPath, source, "exactly one reference subfield must be non-nil"))
 | |
| 	}
 | |
| 	return allErrs
 | |
| }
 | |
| 
 | |
| // validation specific to Node.Status.Config.(Active|Assigned|LastKnownGood).ConfigMap
 | |
| func validateConfigMapNodeConfigSourceStatus(source *core.ConfigMapNodeConfigSource, fldPath *field.Path) field.ErrorList {
 | |
| 	allErrs := field.ErrorList{}
 | |
| 	// uid and resourceVersion must be set in status
 | |
| 	if string(source.UID) == "" {
 | |
| 		allErrs = append(allErrs, field.Required(fldPath.Child("uid"), "uid must be set in status"))
 | |
| 	}
 | |
| 	if source.ResourceVersion == "" {
 | |
| 		allErrs = append(allErrs, field.Required(fldPath.Child("resourceVersion"), "resourceVersion must be set in status"))
 | |
| 	}
 | |
| 	return append(allErrs, validateConfigMapNodeConfigSource(source, fldPath)...)
 | |
| }
 | |
| 
 | |
| // common validation
 | |
| func validateConfigMapNodeConfigSource(source *core.ConfigMapNodeConfigSource, fldPath *field.Path) field.ErrorList {
 | |
| 	allErrs := field.ErrorList{}
 | |
| 	// validate target configmap namespace
 | |
| 	if source.Namespace == "" {
 | |
| 		allErrs = append(allErrs, field.Required(fldPath.Child("namespace"), "namespace must be set"))
 | |
| 	} else {
 | |
| 		for _, msg := range ValidateNameFunc(ValidateNamespaceName)(source.Namespace, false) {
 | |
| 			allErrs = append(allErrs, field.Invalid(fldPath.Child("namespace"), source.Namespace, msg))
 | |
| 		}
 | |
| 	}
 | |
| 	// validate target configmap name
 | |
| 	if source.Name == "" {
 | |
| 		allErrs = append(allErrs, field.Required(fldPath.Child("name"), "name must be set"))
 | |
| 	} else {
 | |
| 		for _, msg := range ValidateNameFunc(ValidateConfigMapName)(source.Name, false) {
 | |
| 			allErrs = append(allErrs, field.Invalid(fldPath.Child("name"), source.Name, msg))
 | |
| 		}
 | |
| 	}
 | |
| 	// validate kubeletConfigKey against rules for configMap key names
 | |
| 	if source.KubeletConfigKey == "" {
 | |
| 		allErrs = append(allErrs, field.Required(fldPath.Child("kubeletConfigKey"), "kubeletConfigKey must be set"))
 | |
| 	} else {
 | |
| 		for _, msg := range validation.IsConfigMapKey(source.KubeletConfigKey) {
 | |
| 			allErrs = append(allErrs, field.Invalid(fldPath.Child("kubeletConfigKey"), source.KubeletConfigKey, msg))
 | |
| 		}
 | |
| 	}
 | |
| 	return allErrs
 | |
| }
 | |
| 
 | |
| // Validate compute resource typename.
 | |
| // Refer to docs/design/resources.md for more details.
 | |
| func validateResourceName(value string, fldPath *field.Path) field.ErrorList {
 | |
| 	allErrs := field.ErrorList{}
 | |
| 	for _, msg := range validation.IsQualifiedName(value) {
 | |
| 		allErrs = append(allErrs, field.Invalid(fldPath, value, msg))
 | |
| 	}
 | |
| 	if len(allErrs) != 0 {
 | |
| 		return allErrs
 | |
| 	}
 | |
| 
 | |
| 	if len(strings.Split(value, "/")) == 1 {
 | |
| 		if !helper.IsStandardResourceName(value) {
 | |
| 			return append(allErrs, field.Invalid(fldPath, value, "must be a standard resource type or fully qualified"))
 | |
| 		}
 | |
| 	}
 | |
| 
 | |
| 	return allErrs
 | |
| }
 | |
| 
 | |
| // Validate container resource name
 | |
| // Refer to docs/design/resources.md for more details.
 | |
| func validateContainerResourceName(value string, fldPath *field.Path) field.ErrorList {
 | |
| 	allErrs := validateResourceName(value, fldPath)
 | |
| 
 | |
| 	if len(strings.Split(value, "/")) == 1 {
 | |
| 		if !helper.IsStandardContainerResourceName(value) {
 | |
| 			return append(allErrs, field.Invalid(fldPath, value, "must be a standard resource for containers"))
 | |
| 		}
 | |
| 	} else if !helper.IsNativeResource(core.ResourceName(value)) {
 | |
| 		if !helper.IsExtendedResourceName(core.ResourceName(value)) {
 | |
| 			return append(allErrs, field.Invalid(fldPath, value, "doesn't follow extended resource name standard"))
 | |
| 		}
 | |
| 	}
 | |
| 	return allErrs
 | |
| }
 | |
| 
 | |
| // Validate resource names that can go in a resource quota
 | |
| // Refer to docs/design/resources.md for more details.
 | |
| func ValidateResourceQuotaResourceName(value string, fldPath *field.Path) field.ErrorList {
 | |
| 	allErrs := validateResourceName(value, fldPath)
 | |
| 
 | |
| 	if len(strings.Split(value, "/")) == 1 {
 | |
| 		if !helper.IsStandardQuotaResourceName(value) {
 | |
| 			return append(allErrs, field.Invalid(fldPath, value, isInvalidQuotaResource))
 | |
| 		}
 | |
| 	}
 | |
| 	return allErrs
 | |
| }
 | |
| 
 | |
| // Validate limit range types
 | |
| func validateLimitRangeTypeName(value string, fldPath *field.Path) field.ErrorList {
 | |
| 	allErrs := field.ErrorList{}
 | |
| 	for _, msg := range validation.IsQualifiedName(value) {
 | |
| 		allErrs = append(allErrs, field.Invalid(fldPath, value, msg))
 | |
| 	}
 | |
| 	if len(allErrs) != 0 {
 | |
| 		return allErrs
 | |
| 	}
 | |
| 
 | |
| 	if len(strings.Split(value, "/")) == 1 {
 | |
| 		if !helper.IsStandardLimitRangeType(value) {
 | |
| 			return append(allErrs, field.Invalid(fldPath, value, "must be a standard limit type or fully qualified"))
 | |
| 		}
 | |
| 	}
 | |
| 
 | |
| 	return allErrs
 | |
| }
 | |
| 
 | |
| // Validate limit range resource name
 | |
| // limit types (other than Pod/Container) could contain storage not just cpu or memory
 | |
| func validateLimitRangeResourceName(limitType core.LimitType, value string, fldPath *field.Path) field.ErrorList {
 | |
| 	switch limitType {
 | |
| 	case core.LimitTypePod, core.LimitTypeContainer:
 | |
| 		return validateContainerResourceName(value, fldPath)
 | |
| 	default:
 | |
| 		return validateResourceName(value, fldPath)
 | |
| 	}
 | |
| }
 | |
| 
 | |
| // ValidateLimitRange tests if required fields in the LimitRange are set.
 | |
| func ValidateLimitRange(limitRange *core.LimitRange) field.ErrorList {
 | |
| 	allErrs := ValidateObjectMeta(&limitRange.ObjectMeta, true, ValidateLimitRangeName, field.NewPath("metadata"))
 | |
| 
 | |
| 	// ensure resource names are properly qualified per docs/design/resources.md
 | |
| 	limitTypeSet := map[core.LimitType]bool{}
 | |
| 	fldPath := field.NewPath("spec", "limits")
 | |
| 	for i := range limitRange.Spec.Limits {
 | |
| 		idxPath := fldPath.Index(i)
 | |
| 		limit := &limitRange.Spec.Limits[i]
 | |
| 		allErrs = append(allErrs, validateLimitRangeTypeName(string(limit.Type), idxPath.Child("type"))...)
 | |
| 
 | |
| 		_, found := limitTypeSet[limit.Type]
 | |
| 		if found {
 | |
| 			allErrs = append(allErrs, field.Duplicate(idxPath.Child("type"), limit.Type))
 | |
| 		}
 | |
| 		limitTypeSet[limit.Type] = true
 | |
| 
 | |
| 		keys := sets.String{}
 | |
| 		min := map[string]resource.Quantity{}
 | |
| 		max := map[string]resource.Quantity{}
 | |
| 		defaults := map[string]resource.Quantity{}
 | |
| 		defaultRequests := map[string]resource.Quantity{}
 | |
| 		maxLimitRequestRatios := map[string]resource.Quantity{}
 | |
| 
 | |
| 		for k, q := range limit.Max {
 | |
| 			allErrs = append(allErrs, validateLimitRangeResourceName(limit.Type, string(k), idxPath.Child("max").Key(string(k)))...)
 | |
| 			keys.Insert(string(k))
 | |
| 			max[string(k)] = q
 | |
| 		}
 | |
| 		for k, q := range limit.Min {
 | |
| 			allErrs = append(allErrs, validateLimitRangeResourceName(limit.Type, string(k), idxPath.Child("min").Key(string(k)))...)
 | |
| 			keys.Insert(string(k))
 | |
| 			min[string(k)] = q
 | |
| 		}
 | |
| 
 | |
| 		if limit.Type == core.LimitTypePod {
 | |
| 			if len(limit.Default) > 0 {
 | |
| 				allErrs = append(allErrs, field.Forbidden(idxPath.Child("default"), "may not be specified when `type` is 'Pod'"))
 | |
| 			}
 | |
| 			if len(limit.DefaultRequest) > 0 {
 | |
| 				allErrs = append(allErrs, field.Forbidden(idxPath.Child("defaultRequest"), "may not be specified when `type` is 'Pod'"))
 | |
| 			}
 | |
| 		} else {
 | |
| 			for k, q := range limit.Default {
 | |
| 				allErrs = append(allErrs, validateLimitRangeResourceName(limit.Type, string(k), idxPath.Child("default").Key(string(k)))...)
 | |
| 				keys.Insert(string(k))
 | |
| 				defaults[string(k)] = q
 | |
| 			}
 | |
| 			for k, q := range limit.DefaultRequest {
 | |
| 				allErrs = append(allErrs, validateLimitRangeResourceName(limit.Type, string(k), idxPath.Child("defaultRequest").Key(string(k)))...)
 | |
| 				keys.Insert(string(k))
 | |
| 				defaultRequests[string(k)] = q
 | |
| 			}
 | |
| 		}
 | |
| 
 | |
| 		if limit.Type == core.LimitTypePersistentVolumeClaim {
 | |
| 			_, minQuantityFound := limit.Min[core.ResourceStorage]
 | |
| 			_, maxQuantityFound := limit.Max[core.ResourceStorage]
 | |
| 			if !minQuantityFound && !maxQuantityFound {
 | |
| 				allErrs = append(allErrs, field.Required(idxPath.Child("limits"), "either minimum or maximum storage value is required, but neither was provided"))
 | |
| 			}
 | |
| 		}
 | |
| 
 | |
| 		for k, q := range limit.MaxLimitRequestRatio {
 | |
| 			allErrs = append(allErrs, validateLimitRangeResourceName(limit.Type, string(k), idxPath.Child("maxLimitRequestRatio").Key(string(k)))...)
 | |
| 			keys.Insert(string(k))
 | |
| 			maxLimitRequestRatios[string(k)] = q
 | |
| 		}
 | |
| 
 | |
| 		for k := range keys {
 | |
| 			minQuantity, minQuantityFound := min[k]
 | |
| 			maxQuantity, maxQuantityFound := max[k]
 | |
| 			defaultQuantity, defaultQuantityFound := defaults[k]
 | |
| 			defaultRequestQuantity, defaultRequestQuantityFound := defaultRequests[k]
 | |
| 			maxRatio, maxRatioFound := maxLimitRequestRatios[k]
 | |
| 
 | |
| 			if minQuantityFound && maxQuantityFound && minQuantity.Cmp(maxQuantity) > 0 {
 | |
| 				allErrs = append(allErrs, field.Invalid(idxPath.Child("min").Key(string(k)), minQuantity, fmt.Sprintf("min value %s is greater than max value %s", minQuantity.String(), maxQuantity.String())))
 | |
| 			}
 | |
| 
 | |
| 			if defaultRequestQuantityFound && minQuantityFound && minQuantity.Cmp(defaultRequestQuantity) > 0 {
 | |
| 				allErrs = append(allErrs, field.Invalid(idxPath.Child("defaultRequest").Key(string(k)), defaultRequestQuantity, fmt.Sprintf("min value %s is greater than default request value %s", minQuantity.String(), defaultRequestQuantity.String())))
 | |
| 			}
 | |
| 
 | |
| 			if defaultRequestQuantityFound && maxQuantityFound && defaultRequestQuantity.Cmp(maxQuantity) > 0 {
 | |
| 				allErrs = append(allErrs, field.Invalid(idxPath.Child("defaultRequest").Key(string(k)), defaultRequestQuantity, fmt.Sprintf("default request value %s is greater than max value %s", defaultRequestQuantity.String(), maxQuantity.String())))
 | |
| 			}
 | |
| 
 | |
| 			if defaultRequestQuantityFound && defaultQuantityFound && defaultRequestQuantity.Cmp(defaultQuantity) > 0 {
 | |
| 				allErrs = append(allErrs, field.Invalid(idxPath.Child("defaultRequest").Key(string(k)), defaultRequestQuantity, fmt.Sprintf("default request value %s is greater than default limit value %s", defaultRequestQuantity.String(), defaultQuantity.String())))
 | |
| 			}
 | |
| 
 | |
| 			if defaultQuantityFound && minQuantityFound && minQuantity.Cmp(defaultQuantity) > 0 {
 | |
| 				allErrs = append(allErrs, field.Invalid(idxPath.Child("default").Key(string(k)), minQuantity, fmt.Sprintf("min value %s is greater than default value %s", minQuantity.String(), defaultQuantity.String())))
 | |
| 			}
 | |
| 
 | |
| 			if defaultQuantityFound && maxQuantityFound && defaultQuantity.Cmp(maxQuantity) > 0 {
 | |
| 				allErrs = append(allErrs, field.Invalid(idxPath.Child("default").Key(string(k)), maxQuantity, fmt.Sprintf("default value %s is greater than max value %s", defaultQuantity.String(), maxQuantity.String())))
 | |
| 			}
 | |
| 			if maxRatioFound && maxRatio.Cmp(*resource.NewQuantity(1, resource.DecimalSI)) < 0 {
 | |
| 				allErrs = append(allErrs, field.Invalid(idxPath.Child("maxLimitRequestRatio").Key(string(k)), maxRatio, fmt.Sprintf("ratio %s is less than 1", maxRatio.String())))
 | |
| 			}
 | |
| 			if maxRatioFound && minQuantityFound && maxQuantityFound {
 | |
| 				maxRatioValue := float64(maxRatio.Value())
 | |
| 				minQuantityValue := minQuantity.Value()
 | |
| 				maxQuantityValue := maxQuantity.Value()
 | |
| 				if maxRatio.Value() < resource.MaxMilliValue && minQuantityValue < resource.MaxMilliValue && maxQuantityValue < resource.MaxMilliValue {
 | |
| 					maxRatioValue = float64(maxRatio.MilliValue()) / 1000
 | |
| 					minQuantityValue = minQuantity.MilliValue()
 | |
| 					maxQuantityValue = maxQuantity.MilliValue()
 | |
| 				}
 | |
| 				maxRatioLimit := float64(maxQuantityValue) / float64(minQuantityValue)
 | |
| 				if maxRatioValue > maxRatioLimit {
 | |
| 					allErrs = append(allErrs, field.Invalid(idxPath.Child("maxLimitRequestRatio").Key(string(k)), maxRatio, fmt.Sprintf("ratio %s is greater than max/min = %f", maxRatio.String(), maxRatioLimit)))
 | |
| 				}
 | |
| 			}
 | |
| 
 | |
| 			// for GPU, hugepages and other resources that are not allowed to overcommit,
 | |
| 			// the default value and defaultRequest value must match if both are specified
 | |
| 			if !helper.IsOvercommitAllowed(core.ResourceName(k)) && defaultQuantityFound && defaultRequestQuantityFound && defaultQuantity.Cmp(defaultRequestQuantity) != 0 {
 | |
| 				allErrs = append(allErrs, field.Invalid(idxPath.Child("defaultRequest").Key(string(k)), defaultRequestQuantity, fmt.Sprintf("default value %s must equal to defaultRequest value %s in %s", defaultQuantity.String(), defaultRequestQuantity.String(), k)))
 | |
| 			}
 | |
| 		}
 | |
| 	}
 | |
| 
 | |
| 	return allErrs
 | |
| }
 | |
| 
 | |
| // ValidateServiceAccount tests if required fields in the ServiceAccount are set.
 | |
| func ValidateServiceAccount(serviceAccount *core.ServiceAccount) field.ErrorList {
 | |
| 	allErrs := ValidateObjectMeta(&serviceAccount.ObjectMeta, true, ValidateServiceAccountName, field.NewPath("metadata"))
 | |
| 	return allErrs
 | |
| }
 | |
| 
 | |
| // ValidateServiceAccountUpdate tests if required fields in the ServiceAccount are set.
 | |
| func ValidateServiceAccountUpdate(newServiceAccount, oldServiceAccount *core.ServiceAccount) field.ErrorList {
 | |
| 	allErrs := ValidateObjectMetaUpdate(&newServiceAccount.ObjectMeta, &oldServiceAccount.ObjectMeta, field.NewPath("metadata"))
 | |
| 	allErrs = append(allErrs, ValidateServiceAccount(newServiceAccount)...)
 | |
| 	return allErrs
 | |
| }
 | |
| 
 | |
| // ValidateSecret tests if required fields in the Secret are set.
 | |
| func ValidateSecret(secret *core.Secret) field.ErrorList {
 | |
| 	allErrs := ValidateObjectMeta(&secret.ObjectMeta, true, ValidateSecretName, field.NewPath("metadata"))
 | |
| 
 | |
| 	dataPath := field.NewPath("data")
 | |
| 	totalSize := 0
 | |
| 	for key, value := range secret.Data {
 | |
| 		for _, msg := range validation.IsConfigMapKey(key) {
 | |
| 			allErrs = append(allErrs, field.Invalid(dataPath.Key(key), key, msg))
 | |
| 		}
 | |
| 		totalSize += len(value)
 | |
| 	}
 | |
| 	if totalSize > core.MaxSecretSize {
 | |
| 		allErrs = append(allErrs, field.TooLong(dataPath, "", core.MaxSecretSize))
 | |
| 	}
 | |
| 
 | |
| 	switch secret.Type {
 | |
| 	case core.SecretTypeServiceAccountToken:
 | |
| 		// Only require Annotations[kubernetes.io/service-account.name]
 | |
| 		// Additional fields (like Annotations[kubernetes.io/service-account.uid] and Data[token]) might be contributed later by a controller loop
 | |
| 		if value := secret.Annotations[core.ServiceAccountNameKey]; len(value) == 0 {
 | |
| 			allErrs = append(allErrs, field.Required(field.NewPath("metadata", "annotations").Key(core.ServiceAccountNameKey), ""))
 | |
| 		}
 | |
| 	case core.SecretTypeOpaque, "":
 | |
| 	// no-op
 | |
| 	case core.SecretTypeDockercfg:
 | |
| 		dockercfgBytes, exists := secret.Data[core.DockerConfigKey]
 | |
| 		if !exists {
 | |
| 			allErrs = append(allErrs, field.Required(dataPath.Key(core.DockerConfigKey), ""))
 | |
| 			break
 | |
| 		}
 | |
| 
 | |
| 		// make sure that the content is well-formed json.
 | |
| 		if err := json.Unmarshal(dockercfgBytes, &map[string]interface{}{}); err != nil {
 | |
| 			allErrs = append(allErrs, field.Invalid(dataPath.Key(core.DockerConfigKey), "<secret contents redacted>", err.Error()))
 | |
| 		}
 | |
| 	case core.SecretTypeDockerConfigJSON:
 | |
| 		dockerConfigJSONBytes, exists := secret.Data[core.DockerConfigJSONKey]
 | |
| 		if !exists {
 | |
| 			allErrs = append(allErrs, field.Required(dataPath.Key(core.DockerConfigJSONKey), ""))
 | |
| 			break
 | |
| 		}
 | |
| 
 | |
| 		// make sure that the content is well-formed json.
 | |
| 		if err := json.Unmarshal(dockerConfigJSONBytes, &map[string]interface{}{}); err != nil {
 | |
| 			allErrs = append(allErrs, field.Invalid(dataPath.Key(core.DockerConfigJSONKey), "<secret contents redacted>", err.Error()))
 | |
| 		}
 | |
| 	case core.SecretTypeBasicAuth:
 | |
| 		_, usernameFieldExists := secret.Data[core.BasicAuthUsernameKey]
 | |
| 		_, passwordFieldExists := secret.Data[core.BasicAuthPasswordKey]
 | |
| 
 | |
| 		// username or password might be empty, but the field must be present
 | |
| 		if !usernameFieldExists && !passwordFieldExists {
 | |
| 			allErrs = append(allErrs, field.Required(field.NewPath("data[%s]").Key(core.BasicAuthUsernameKey), ""))
 | |
| 			allErrs = append(allErrs, field.Required(field.NewPath("data[%s]").Key(core.BasicAuthPasswordKey), ""))
 | |
| 			break
 | |
| 		}
 | |
| 	case core.SecretTypeSSHAuth:
 | |
| 		if len(secret.Data[core.SSHAuthPrivateKey]) == 0 {
 | |
| 			allErrs = append(allErrs, field.Required(field.NewPath("data[%s]").Key(core.SSHAuthPrivateKey), ""))
 | |
| 			break
 | |
| 		}
 | |
| 
 | |
| 	case core.SecretTypeTLS:
 | |
| 		if _, exists := secret.Data[core.TLSCertKey]; !exists {
 | |
| 			allErrs = append(allErrs, field.Required(dataPath.Key(core.TLSCertKey), ""))
 | |
| 		}
 | |
| 		if _, exists := secret.Data[core.TLSPrivateKeyKey]; !exists {
 | |
| 			allErrs = append(allErrs, field.Required(dataPath.Key(core.TLSPrivateKeyKey), ""))
 | |
| 		}
 | |
| 	// TODO: Verify that the key matches the cert.
 | |
| 	default:
 | |
| 		// no-op
 | |
| 	}
 | |
| 
 | |
| 	return allErrs
 | |
| }
 | |
| 
 | |
| // ValidateSecretUpdate tests if required fields in the Secret are set.
 | |
| func ValidateSecretUpdate(newSecret, oldSecret *core.Secret) field.ErrorList {
 | |
| 	allErrs := ValidateObjectMetaUpdate(&newSecret.ObjectMeta, &oldSecret.ObjectMeta, field.NewPath("metadata"))
 | |
| 
 | |
| 	allErrs = append(allErrs, ValidateImmutableField(newSecret.Type, oldSecret.Type, field.NewPath("type"))...)
 | |
| 	if oldSecret.Immutable != nil && *oldSecret.Immutable {
 | |
| 		if newSecret.Immutable == nil || !*newSecret.Immutable {
 | |
| 			allErrs = append(allErrs, field.Forbidden(field.NewPath("immutable"), "field is immutable when `immutable` is set"))
 | |
| 		}
 | |
| 		if !reflect.DeepEqual(newSecret.Data, oldSecret.Data) {
 | |
| 			allErrs = append(allErrs, field.Forbidden(field.NewPath("data"), "field is immutable when `immutable` is set"))
 | |
| 		}
 | |
| 		// We don't validate StringData, as it was already converted back to Data
 | |
| 		// before validation is happening.
 | |
| 	}
 | |
| 
 | |
| 	allErrs = append(allErrs, ValidateSecret(newSecret)...)
 | |
| 	return allErrs
 | |
| }
 | |
| 
 | |
| // ValidateConfigMapName can be used to check whether the given ConfigMap name is valid.
 | |
| // Prefix indicates this name will be used as part of generation, in which case
 | |
| // trailing dashes are allowed.
 | |
| var ValidateConfigMapName = apimachineryvalidation.NameIsDNSSubdomain
 | |
| 
 | |
| // ValidateConfigMap tests whether required fields in the ConfigMap are set.
 | |
| func ValidateConfigMap(cfg *core.ConfigMap) field.ErrorList {
 | |
| 	allErrs := field.ErrorList{}
 | |
| 	allErrs = append(allErrs, ValidateObjectMeta(&cfg.ObjectMeta, true, ValidateConfigMapName, field.NewPath("metadata"))...)
 | |
| 
 | |
| 	totalSize := 0
 | |
| 
 | |
| 	for key, value := range cfg.Data {
 | |
| 		for _, msg := range validation.IsConfigMapKey(key) {
 | |
| 			allErrs = append(allErrs, field.Invalid(field.NewPath("data").Key(key), key, msg))
 | |
| 		}
 | |
| 		// check if we have a duplicate key in the other bag
 | |
| 		if _, isValue := cfg.BinaryData[key]; isValue {
 | |
| 			msg := "duplicate of key present in binaryData"
 | |
| 			allErrs = append(allErrs, field.Invalid(field.NewPath("data").Key(key), key, msg))
 | |
| 		}
 | |
| 		totalSize += len(value)
 | |
| 	}
 | |
| 	for key, value := range cfg.BinaryData {
 | |
| 		for _, msg := range validation.IsConfigMapKey(key) {
 | |
| 			allErrs = append(allErrs, field.Invalid(field.NewPath("binaryData").Key(key), key, msg))
 | |
| 		}
 | |
| 		totalSize += len(value)
 | |
| 	}
 | |
| 	if totalSize > core.MaxSecretSize {
 | |
| 		// pass back "" to indicate that the error refers to the whole object.
 | |
| 		allErrs = append(allErrs, field.TooLong(field.NewPath(""), cfg, core.MaxSecretSize))
 | |
| 	}
 | |
| 
 | |
| 	return allErrs
 | |
| }
 | |
| 
 | |
| // ValidateConfigMapUpdate tests if required fields in the ConfigMap are set.
 | |
| func ValidateConfigMapUpdate(newCfg, oldCfg *core.ConfigMap) field.ErrorList {
 | |
| 	allErrs := field.ErrorList{}
 | |
| 	allErrs = append(allErrs, ValidateObjectMetaUpdate(&newCfg.ObjectMeta, &oldCfg.ObjectMeta, field.NewPath("metadata"))...)
 | |
| 
 | |
| 	if oldCfg.Immutable != nil && *oldCfg.Immutable {
 | |
| 		if newCfg.Immutable == nil || !*newCfg.Immutable {
 | |
| 			allErrs = append(allErrs, field.Forbidden(field.NewPath("immutable"), "field is immutable when `immutable` is set"))
 | |
| 		}
 | |
| 		if !reflect.DeepEqual(newCfg.Data, oldCfg.Data) {
 | |
| 			allErrs = append(allErrs, field.Forbidden(field.NewPath("data"), "field is immutable when `immutable` is set"))
 | |
| 		}
 | |
| 		if !reflect.DeepEqual(newCfg.BinaryData, oldCfg.BinaryData) {
 | |
| 			allErrs = append(allErrs, field.Forbidden(field.NewPath("binaryData"), "field is immutable when `immutable` is set"))
 | |
| 		}
 | |
| 	}
 | |
| 
 | |
| 	allErrs = append(allErrs, ValidateConfigMap(newCfg)...)
 | |
| 	return allErrs
 | |
| }
 | |
| 
 | |
| func validateBasicResource(quantity resource.Quantity, fldPath *field.Path) field.ErrorList {
 | |
| 	if quantity.Value() < 0 {
 | |
| 		return field.ErrorList{field.Invalid(fldPath, quantity.Value(), "must be a valid resource quantity")}
 | |
| 	}
 | |
| 	return field.ErrorList{}
 | |
| }
 | |
| 
 | |
| // Validates resource requirement spec.
 | |
| func ValidateResourceRequirements(requirements *core.ResourceRequirements, fldPath *field.Path, opts PodValidationOptions) field.ErrorList {
 | |
| 	allErrs := field.ErrorList{}
 | |
| 	limPath := fldPath.Child("limits")
 | |
| 	reqPath := fldPath.Child("requests")
 | |
| 	limContainsCPUOrMemory := false
 | |
| 	reqContainsCPUOrMemory := false
 | |
| 	limContainsHugePages := false
 | |
| 	reqContainsHugePages := false
 | |
| 	supportedQoSComputeResources := sets.NewString(string(core.ResourceCPU), string(core.ResourceMemory))
 | |
| 	for resourceName, quantity := range requirements.Limits {
 | |
| 
 | |
| 		fldPath := limPath.Key(string(resourceName))
 | |
| 		// Validate resource name.
 | |
| 		allErrs = append(allErrs, validateContainerResourceName(string(resourceName), fldPath)...)
 | |
| 
 | |
| 		// Validate resource quantity.
 | |
| 		allErrs = append(allErrs, ValidateResourceQuantityValue(string(resourceName), quantity, fldPath)...)
 | |
| 
 | |
| 		if helper.IsHugePageResourceName(resourceName) {
 | |
| 			limContainsHugePages = true
 | |
| 			if err := validateResourceQuantityHugePageValue(resourceName, quantity, opts); err != nil {
 | |
| 				allErrs = append(allErrs, field.Invalid(fldPath, quantity.String(), err.Error()))
 | |
| 			}
 | |
| 		}
 | |
| 
 | |
| 		if supportedQoSComputeResources.Has(string(resourceName)) {
 | |
| 			limContainsCPUOrMemory = true
 | |
| 		}
 | |
| 	}
 | |
| 	for resourceName, quantity := range requirements.Requests {
 | |
| 		fldPath := reqPath.Key(string(resourceName))
 | |
| 		// Validate resource name.
 | |
| 		allErrs = append(allErrs, validateContainerResourceName(string(resourceName), fldPath)...)
 | |
| 		// Validate resource quantity.
 | |
| 		allErrs = append(allErrs, ValidateResourceQuantityValue(string(resourceName), quantity, fldPath)...)
 | |
| 
 | |
| 		// Check that request <= limit.
 | |
| 		limitQuantity, exists := requirements.Limits[resourceName]
 | |
| 		if exists {
 | |
| 			// For non overcommitable resources, not only requests can't exceed limits, they also can't be lower, i.e. must be equal.
 | |
| 			if quantity.Cmp(limitQuantity) != 0 && !helper.IsOvercommitAllowed(resourceName) {
 | |
| 				allErrs = append(allErrs, field.Invalid(reqPath, quantity.String(), fmt.Sprintf("must be equal to %s limit", resourceName)))
 | |
| 			} else if quantity.Cmp(limitQuantity) > 0 {
 | |
| 				allErrs = append(allErrs, field.Invalid(reqPath, quantity.String(), fmt.Sprintf("must be less than or equal to %s limit", resourceName)))
 | |
| 			}
 | |
| 		} else if !helper.IsOvercommitAllowed(resourceName) {
 | |
| 			allErrs = append(allErrs, field.Required(limPath, "Limit must be set for non overcommitable resources"))
 | |
| 		}
 | |
| 		if helper.IsHugePageResourceName(resourceName) {
 | |
| 			reqContainsHugePages = true
 | |
| 			if err := validateResourceQuantityHugePageValue(resourceName, quantity, opts); err != nil {
 | |
| 				allErrs = append(allErrs, field.Invalid(fldPath, quantity.String(), err.Error()))
 | |
| 			}
 | |
| 		}
 | |
| 		if supportedQoSComputeResources.Has(string(resourceName)) {
 | |
| 			reqContainsCPUOrMemory = true
 | |
| 		}
 | |
| 
 | |
| 	}
 | |
| 	if !limContainsCPUOrMemory && !reqContainsCPUOrMemory && (reqContainsHugePages || limContainsHugePages) {
 | |
| 		allErrs = append(allErrs, field.Forbidden(fldPath, "HugePages require cpu or memory"))
 | |
| 	}
 | |
| 
 | |
| 	return allErrs
 | |
| }
 | |
| 
 | |
| func validateResourceQuantityHugePageValue(name core.ResourceName, quantity resource.Quantity, opts PodValidationOptions) error {
 | |
| 	if !helper.IsHugePageResourceName(name) {
 | |
| 		return nil
 | |
| 	}
 | |
| 
 | |
| 	if !opts.AllowIndivisibleHugePagesValues && !helper.IsHugePageResourceValueDivisible(name, quantity) {
 | |
| 		return fmt.Errorf("%s is not positive integer multiple of %s", quantity.String(), name)
 | |
| 	}
 | |
| 
 | |
| 	return nil
 | |
| }
 | |
| 
 | |
| // validateResourceQuotaScopes ensures that each enumerated hard resource constraint is valid for set of scopes
 | |
| func validateResourceQuotaScopes(resourceQuotaSpec *core.ResourceQuotaSpec, opts ResourceQuotaValidationOptions, fld *field.Path) field.ErrorList {
 | |
| 	allErrs := field.ErrorList{}
 | |
| 	if len(resourceQuotaSpec.Scopes) == 0 {
 | |
| 		return allErrs
 | |
| 	}
 | |
| 	hardLimits := sets.NewString()
 | |
| 	for k := range resourceQuotaSpec.Hard {
 | |
| 		hardLimits.Insert(string(k))
 | |
| 	}
 | |
| 	fldPath := fld.Child("scopes")
 | |
| 	scopeSet := sets.NewString()
 | |
| 	for _, scope := range resourceQuotaSpec.Scopes {
 | |
| 		if !helper.IsStandardResourceQuotaScope(string(scope), opts.AllowPodAffinityNamespaceSelector) {
 | |
| 			allErrs = append(allErrs, field.Invalid(fldPath, resourceQuotaSpec.Scopes, "unsupported scope"))
 | |
| 		}
 | |
| 		for _, k := range hardLimits.List() {
 | |
| 			if helper.IsStandardQuotaResourceName(k) && !helper.IsResourceQuotaScopeValidForResource(scope, k) {
 | |
| 				allErrs = append(allErrs, field.Invalid(fldPath, resourceQuotaSpec.Scopes, "unsupported scope applied to resource"))
 | |
| 			}
 | |
| 		}
 | |
| 		scopeSet.Insert(string(scope))
 | |
| 	}
 | |
| 	invalidScopePairs := []sets.String{
 | |
| 		sets.NewString(string(core.ResourceQuotaScopeBestEffort), string(core.ResourceQuotaScopeNotBestEffort)),
 | |
| 		sets.NewString(string(core.ResourceQuotaScopeTerminating), string(core.ResourceQuotaScopeNotTerminating)),
 | |
| 	}
 | |
| 	for _, invalidScopePair := range invalidScopePairs {
 | |
| 		if scopeSet.HasAll(invalidScopePair.List()...) {
 | |
| 			allErrs = append(allErrs, field.Invalid(fldPath, resourceQuotaSpec.Scopes, "conflicting scopes"))
 | |
| 		}
 | |
| 	}
 | |
| 	return allErrs
 | |
| }
 | |
| 
 | |
| // validateScopedResourceSelectorRequirement tests that the match expressions has valid data
 | |
| func validateScopedResourceSelectorRequirement(resourceQuotaSpec *core.ResourceQuotaSpec, opts ResourceQuotaValidationOptions, fld *field.Path) field.ErrorList {
 | |
| 	allErrs := field.ErrorList{}
 | |
| 	hardLimits := sets.NewString()
 | |
| 	for k := range resourceQuotaSpec.Hard {
 | |
| 		hardLimits.Insert(string(k))
 | |
| 	}
 | |
| 	fldPath := fld.Child("matchExpressions")
 | |
| 	scopeSet := sets.NewString()
 | |
| 	for _, req := range resourceQuotaSpec.ScopeSelector.MatchExpressions {
 | |
| 		if !helper.IsStandardResourceQuotaScope(string(req.ScopeName), opts.AllowPodAffinityNamespaceSelector) {
 | |
| 			allErrs = append(allErrs, field.Invalid(fldPath.Child("scopeName"), req.ScopeName, "unsupported scope"))
 | |
| 		}
 | |
| 		for _, k := range hardLimits.List() {
 | |
| 			if helper.IsStandardQuotaResourceName(k) && !helper.IsResourceQuotaScopeValidForResource(req.ScopeName, k) {
 | |
| 				allErrs = append(allErrs, field.Invalid(fldPath, resourceQuotaSpec.ScopeSelector, "unsupported scope applied to resource"))
 | |
| 			}
 | |
| 		}
 | |
| 		switch req.ScopeName {
 | |
| 		case core.ResourceQuotaScopeBestEffort, core.ResourceQuotaScopeNotBestEffort, core.ResourceQuotaScopeTerminating, core.ResourceQuotaScopeNotTerminating, core.ResourceQuotaScopeCrossNamespacePodAffinity:
 | |
| 			if req.Operator != core.ScopeSelectorOpExists {
 | |
| 				allErrs = append(allErrs, field.Invalid(fldPath.Child("operator"), req.Operator,
 | |
| 					"must be 'Exist' when scope is any of ResourceQuotaScopeTerminating, ResourceQuotaScopeNotTerminating, ResourceQuotaScopeBestEffort, ResourceQuotaScopeNotBestEffort or ResourceQuotaScopeCrossNamespacePodAffinity"))
 | |
| 			}
 | |
| 		}
 | |
| 
 | |
| 		switch req.Operator {
 | |
| 		case core.ScopeSelectorOpIn, core.ScopeSelectorOpNotIn:
 | |
| 			if len(req.Values) == 0 {
 | |
| 				allErrs = append(allErrs, field.Required(fldPath.Child("values"),
 | |
| 					"must be at least one value when `operator` is 'In' or 'NotIn' for scope selector"))
 | |
| 			}
 | |
| 		case core.ScopeSelectorOpExists, core.ScopeSelectorOpDoesNotExist:
 | |
| 			if len(req.Values) != 0 {
 | |
| 				allErrs = append(allErrs, field.Invalid(fldPath.Child("values"), req.Values,
 | |
| 					"must be no value when `operator` is 'Exist' or 'DoesNotExist' for scope selector"))
 | |
| 			}
 | |
| 		default:
 | |
| 			allErrs = append(allErrs, field.Invalid(fldPath.Child("operator"), req.Operator, "not a valid selector operator"))
 | |
| 		}
 | |
| 		scopeSet.Insert(string(req.ScopeName))
 | |
| 	}
 | |
| 	invalidScopePairs := []sets.String{
 | |
| 		sets.NewString(string(core.ResourceQuotaScopeBestEffort), string(core.ResourceQuotaScopeNotBestEffort)),
 | |
| 		sets.NewString(string(core.ResourceQuotaScopeTerminating), string(core.ResourceQuotaScopeNotTerminating)),
 | |
| 	}
 | |
| 	for _, invalidScopePair := range invalidScopePairs {
 | |
| 		if scopeSet.HasAll(invalidScopePair.List()...) {
 | |
| 			allErrs = append(allErrs, field.Invalid(fldPath, resourceQuotaSpec.Scopes, "conflicting scopes"))
 | |
| 		}
 | |
| 	}
 | |
| 
 | |
| 	return allErrs
 | |
| }
 | |
| 
 | |
| // validateScopeSelector tests that the specified scope selector has valid data
 | |
| func validateScopeSelector(resourceQuotaSpec *core.ResourceQuotaSpec, opts ResourceQuotaValidationOptions, fld *field.Path) field.ErrorList {
 | |
| 	allErrs := field.ErrorList{}
 | |
| 	if resourceQuotaSpec.ScopeSelector == nil {
 | |
| 		return allErrs
 | |
| 	}
 | |
| 	allErrs = append(allErrs, validateScopedResourceSelectorRequirement(resourceQuotaSpec, opts, fld.Child("scopeSelector"))...)
 | |
| 	return allErrs
 | |
| }
 | |
| 
 | |
| // ResourceQuotaValidationOptions contains the different settings for ResourceQuota validation
 | |
| type ResourceQuotaValidationOptions struct {
 | |
| 	// Allow pod-affinity namespace selector validation.
 | |
| 	AllowPodAffinityNamespaceSelector bool
 | |
| }
 | |
| 
 | |
| // ValidateResourceQuota tests if required fields in the ResourceQuota are set.
 | |
| func ValidateResourceQuota(resourceQuota *core.ResourceQuota, opts ResourceQuotaValidationOptions) field.ErrorList {
 | |
| 	allErrs := ValidateObjectMeta(&resourceQuota.ObjectMeta, true, ValidateResourceQuotaName, field.NewPath("metadata"))
 | |
| 
 | |
| 	allErrs = append(allErrs, ValidateResourceQuotaSpec(&resourceQuota.Spec, opts, field.NewPath("spec"))...)
 | |
| 	allErrs = append(allErrs, ValidateResourceQuotaStatus(&resourceQuota.Status, field.NewPath("status"))...)
 | |
| 
 | |
| 	return allErrs
 | |
| }
 | |
| 
 | |
| func ValidateResourceQuotaStatus(status *core.ResourceQuotaStatus, fld *field.Path) field.ErrorList {
 | |
| 	allErrs := field.ErrorList{}
 | |
| 
 | |
| 	fldPath := fld.Child("hard")
 | |
| 	for k, v := range status.Hard {
 | |
| 		resPath := fldPath.Key(string(k))
 | |
| 		allErrs = append(allErrs, ValidateResourceQuotaResourceName(string(k), resPath)...)
 | |
| 		allErrs = append(allErrs, ValidateResourceQuantityValue(string(k), v, resPath)...)
 | |
| 	}
 | |
| 	fldPath = fld.Child("used")
 | |
| 	for k, v := range status.Used {
 | |
| 		resPath := fldPath.Key(string(k))
 | |
| 		allErrs = append(allErrs, ValidateResourceQuotaResourceName(string(k), resPath)...)
 | |
| 		allErrs = append(allErrs, ValidateResourceQuantityValue(string(k), v, resPath)...)
 | |
| 	}
 | |
| 
 | |
| 	return allErrs
 | |
| }
 | |
| 
 | |
| func ValidateResourceQuotaSpec(resourceQuotaSpec *core.ResourceQuotaSpec, opts ResourceQuotaValidationOptions, fld *field.Path) field.ErrorList {
 | |
| 	allErrs := field.ErrorList{}
 | |
| 
 | |
| 	fldPath := fld.Child("hard")
 | |
| 	for k, v := range resourceQuotaSpec.Hard {
 | |
| 		resPath := fldPath.Key(string(k))
 | |
| 		allErrs = append(allErrs, ValidateResourceQuotaResourceName(string(k), resPath)...)
 | |
| 		allErrs = append(allErrs, ValidateResourceQuantityValue(string(k), v, resPath)...)
 | |
| 	}
 | |
| 
 | |
| 	allErrs = append(allErrs, validateResourceQuotaScopes(resourceQuotaSpec, opts, fld)...)
 | |
| 	allErrs = append(allErrs, validateScopeSelector(resourceQuotaSpec, opts, fld)...)
 | |
| 
 | |
| 	return allErrs
 | |
| }
 | |
| 
 | |
| // ValidateResourceQuantityValue enforces that specified quantity is valid for specified resource
 | |
| func ValidateResourceQuantityValue(resource string, value resource.Quantity, fldPath *field.Path) field.ErrorList {
 | |
| 	allErrs := field.ErrorList{}
 | |
| 	allErrs = append(allErrs, ValidateNonnegativeQuantity(value, fldPath)...)
 | |
| 	if helper.IsIntegerResourceName(resource) {
 | |
| 		if value.MilliValue()%int64(1000) != int64(0) {
 | |
| 			allErrs = append(allErrs, field.Invalid(fldPath, value, isNotIntegerErrorMsg))
 | |
| 		}
 | |
| 	}
 | |
| 	return allErrs
 | |
| }
 | |
| 
 | |
| // ValidateResourceQuotaUpdate tests to see if the update is legal for an end user to make.
 | |
| func ValidateResourceQuotaUpdate(newResourceQuota, oldResourceQuota *core.ResourceQuota, opts ResourceQuotaValidationOptions) field.ErrorList {
 | |
| 	allErrs := ValidateObjectMetaUpdate(&newResourceQuota.ObjectMeta, &oldResourceQuota.ObjectMeta, field.NewPath("metadata"))
 | |
| 	allErrs = append(allErrs, ValidateResourceQuotaSpec(&newResourceQuota.Spec, opts, field.NewPath("spec"))...)
 | |
| 
 | |
| 	// ensure scopes cannot change, and that resources are still valid for scope
 | |
| 	fldPath := field.NewPath("spec", "scopes")
 | |
| 	oldScopes := sets.NewString()
 | |
| 	newScopes := sets.NewString()
 | |
| 	for _, scope := range newResourceQuota.Spec.Scopes {
 | |
| 		newScopes.Insert(string(scope))
 | |
| 	}
 | |
| 	for _, scope := range oldResourceQuota.Spec.Scopes {
 | |
| 		oldScopes.Insert(string(scope))
 | |
| 	}
 | |
| 	if !oldScopes.Equal(newScopes) {
 | |
| 		allErrs = append(allErrs, field.Invalid(fldPath, newResourceQuota.Spec.Scopes, fieldImmutableErrorMsg))
 | |
| 	}
 | |
| 
 | |
| 	return allErrs
 | |
| }
 | |
| 
 | |
| // ValidateResourceQuotaStatusUpdate tests to see if the status update is legal for an end user to make.
 | |
| func ValidateResourceQuotaStatusUpdate(newResourceQuota, oldResourceQuota *core.ResourceQuota) field.ErrorList {
 | |
| 	allErrs := ValidateObjectMetaUpdate(&newResourceQuota.ObjectMeta, &oldResourceQuota.ObjectMeta, field.NewPath("metadata"))
 | |
| 	if len(newResourceQuota.ResourceVersion) == 0 {
 | |
| 		allErrs = append(allErrs, field.Required(field.NewPath("resourceVersion"), ""))
 | |
| 	}
 | |
| 	fldPath := field.NewPath("status", "hard")
 | |
| 	for k, v := range newResourceQuota.Status.Hard {
 | |
| 		resPath := fldPath.Key(string(k))
 | |
| 		allErrs = append(allErrs, ValidateResourceQuotaResourceName(string(k), resPath)...)
 | |
| 		allErrs = append(allErrs, ValidateResourceQuantityValue(string(k), v, resPath)...)
 | |
| 	}
 | |
| 	fldPath = field.NewPath("status", "used")
 | |
| 	for k, v := range newResourceQuota.Status.Used {
 | |
| 		resPath := fldPath.Key(string(k))
 | |
| 		allErrs = append(allErrs, ValidateResourceQuotaResourceName(string(k), resPath)...)
 | |
| 		allErrs = append(allErrs, ValidateResourceQuantityValue(string(k), v, resPath)...)
 | |
| 	}
 | |
| 	return allErrs
 | |
| }
 | |
| 
 | |
| // ValidateNamespace tests if required fields are set.
 | |
| func ValidateNamespace(namespace *core.Namespace) field.ErrorList {
 | |
| 	allErrs := ValidateObjectMeta(&namespace.ObjectMeta, false, ValidateNamespaceName, field.NewPath("metadata"))
 | |
| 	for i := range namespace.Spec.Finalizers {
 | |
| 		allErrs = append(allErrs, validateFinalizerName(string(namespace.Spec.Finalizers[i]), field.NewPath("spec", "finalizers"))...)
 | |
| 	}
 | |
| 	return allErrs
 | |
| }
 | |
| 
 | |
| // Validate finalizer names
 | |
| func validateFinalizerName(stringValue string, fldPath *field.Path) field.ErrorList {
 | |
| 	allErrs := apimachineryvalidation.ValidateFinalizerName(stringValue, fldPath)
 | |
| 	allErrs = append(allErrs, validateKubeFinalizerName(stringValue, fldPath)...)
 | |
| 	return allErrs
 | |
| }
 | |
| 
 | |
| // validateKubeFinalizerName checks for "standard" names of legacy finalizer
 | |
| func validateKubeFinalizerName(stringValue string, fldPath *field.Path) field.ErrorList {
 | |
| 	allErrs := field.ErrorList{}
 | |
| 	if len(strings.Split(stringValue, "/")) == 1 {
 | |
| 		if !helper.IsStandardFinalizerName(stringValue) {
 | |
| 			return append(allErrs, field.Invalid(fldPath, stringValue, "name is neither a standard finalizer name nor is it fully qualified"))
 | |
| 		}
 | |
| 	}
 | |
| 
 | |
| 	return allErrs
 | |
| }
 | |
| 
 | |
| // ValidateNamespaceUpdate tests to make sure a namespace update can be applied.
 | |
| func ValidateNamespaceUpdate(newNamespace *core.Namespace, oldNamespace *core.Namespace) field.ErrorList {
 | |
| 	allErrs := ValidateObjectMetaUpdate(&newNamespace.ObjectMeta, &oldNamespace.ObjectMeta, field.NewPath("metadata"))
 | |
| 	return allErrs
 | |
| }
 | |
| 
 | |
| // ValidateNamespaceStatusUpdate tests to see if the update is legal for an end user to make.
 | |
| func ValidateNamespaceStatusUpdate(newNamespace, oldNamespace *core.Namespace) field.ErrorList {
 | |
| 	allErrs := ValidateObjectMetaUpdate(&newNamespace.ObjectMeta, &oldNamespace.ObjectMeta, field.NewPath("metadata"))
 | |
| 	if newNamespace.DeletionTimestamp.IsZero() {
 | |
| 		if newNamespace.Status.Phase != core.NamespaceActive {
 | |
| 			allErrs = append(allErrs, field.Invalid(field.NewPath("status", "Phase"), newNamespace.Status.Phase, "may only be 'Active' if `deletionTimestamp` is empty"))
 | |
| 		}
 | |
| 	} else {
 | |
| 		if newNamespace.Status.Phase != core.NamespaceTerminating {
 | |
| 			allErrs = append(allErrs, field.Invalid(field.NewPath("status", "Phase"), newNamespace.Status.Phase, "may only be 'Terminating' if `deletionTimestamp` is not empty"))
 | |
| 		}
 | |
| 	}
 | |
| 	return allErrs
 | |
| }
 | |
| 
 | |
| // ValidateNamespaceFinalizeUpdate tests to see if the update is legal for an end user to make.
 | |
| func ValidateNamespaceFinalizeUpdate(newNamespace, oldNamespace *core.Namespace) field.ErrorList {
 | |
| 	allErrs := ValidateObjectMetaUpdate(&newNamespace.ObjectMeta, &oldNamespace.ObjectMeta, field.NewPath("metadata"))
 | |
| 
 | |
| 	fldPath := field.NewPath("spec", "finalizers")
 | |
| 	for i := range newNamespace.Spec.Finalizers {
 | |
| 		idxPath := fldPath.Index(i)
 | |
| 		allErrs = append(allErrs, validateFinalizerName(string(newNamespace.Spec.Finalizers[i]), idxPath)...)
 | |
| 	}
 | |
| 	return allErrs
 | |
| }
 | |
| 
 | |
| // ValidateEndpoints validates Endpoints on create and update.
 | |
| func ValidateEndpoints(endpoints *core.Endpoints) field.ErrorList {
 | |
| 	allErrs := ValidateObjectMeta(&endpoints.ObjectMeta, true, ValidateEndpointsName, field.NewPath("metadata"))
 | |
| 	allErrs = append(allErrs, ValidateEndpointsSpecificAnnotations(endpoints.Annotations, field.NewPath("annotations"))...)
 | |
| 	allErrs = append(allErrs, validateEndpointSubsets(endpoints.Subsets, field.NewPath("subsets"))...)
 | |
| 	return allErrs
 | |
| }
 | |
| 
 | |
| // ValidateEndpointsCreate validates Endpoints on create.
 | |
| func ValidateEndpointsCreate(endpoints *core.Endpoints) field.ErrorList {
 | |
| 	return ValidateEndpoints(endpoints)
 | |
| }
 | |
| 
 | |
| // ValidateEndpointsUpdate validates Endpoints on update. NodeName changes are
 | |
| // allowed during update to accommodate the case where nodeIP or PodCIDR is
 | |
| // reused. An existing endpoint ip will have a different nodeName if this
 | |
| // happens.
 | |
| func ValidateEndpointsUpdate(newEndpoints, oldEndpoints *core.Endpoints) field.ErrorList {
 | |
| 	allErrs := ValidateObjectMetaUpdate(&newEndpoints.ObjectMeta, &oldEndpoints.ObjectMeta, field.NewPath("metadata"))
 | |
| 	allErrs = append(allErrs, ValidateEndpoints(newEndpoints)...)
 | |
| 	return allErrs
 | |
| }
 | |
| 
 | |
| func validateEndpointSubsets(subsets []core.EndpointSubset, fldPath *field.Path) field.ErrorList {
 | |
| 	allErrs := field.ErrorList{}
 | |
| 	for i := range subsets {
 | |
| 		ss := &subsets[i]
 | |
| 		idxPath := fldPath.Index(i)
 | |
| 
 | |
| 		// EndpointSubsets must include endpoint address. For headless service, we allow its endpoints not to have ports.
 | |
| 		if len(ss.Addresses) == 0 && len(ss.NotReadyAddresses) == 0 {
 | |
| 			//TODO: consider adding a RequiredOneOf() error for this and similar cases
 | |
| 			allErrs = append(allErrs, field.Required(idxPath, "must specify `addresses` or `notReadyAddresses`"))
 | |
| 		}
 | |
| 		for addr := range ss.Addresses {
 | |
| 			allErrs = append(allErrs, validateEndpointAddress(&ss.Addresses[addr], idxPath.Child("addresses").Index(addr))...)
 | |
| 		}
 | |
| 		for addr := range ss.NotReadyAddresses {
 | |
| 			allErrs = append(allErrs, validateEndpointAddress(&ss.NotReadyAddresses[addr], idxPath.Child("notReadyAddresses").Index(addr))...)
 | |
| 		}
 | |
| 		for port := range ss.Ports {
 | |
| 			allErrs = append(allErrs, validateEndpointPort(&ss.Ports[port], len(ss.Ports) > 1, idxPath.Child("ports").Index(port))...)
 | |
| 		}
 | |
| 	}
 | |
| 
 | |
| 	return allErrs
 | |
| }
 | |
| 
 | |
| func validateEndpointAddress(address *core.EndpointAddress, fldPath *field.Path) field.ErrorList {
 | |
| 	allErrs := field.ErrorList{}
 | |
| 	for _, msg := range validation.IsValidIP(address.IP) {
 | |
| 		allErrs = append(allErrs, field.Invalid(fldPath.Child("ip"), address.IP, msg))
 | |
| 	}
 | |
| 	if len(address.Hostname) > 0 {
 | |
| 		allErrs = append(allErrs, ValidateDNS1123Label(address.Hostname, fldPath.Child("hostname"))...)
 | |
| 	}
 | |
| 	// During endpoint update, verify that NodeName is a DNS subdomain and transition rules allow the update
 | |
| 	if address.NodeName != nil {
 | |
| 		for _, msg := range ValidateNodeName(*address.NodeName, false) {
 | |
| 			allErrs = append(allErrs, field.Invalid(fldPath.Child("nodeName"), *address.NodeName, msg))
 | |
| 		}
 | |
| 	}
 | |
| 	allErrs = append(allErrs, ValidateNonSpecialIP(address.IP, fldPath.Child("ip"))...)
 | |
| 	return allErrs
 | |
| }
 | |
| 
 | |
| // ValidateNonSpecialIP is used to validate Endpoints, EndpointSlices, and
 | |
| // external IPs. Specifically, this disallows unspecified and loopback addresses
 | |
| // are nonsensical and link-local addresses tend to be used for node-centric
 | |
| // purposes (e.g. metadata service).
 | |
| //
 | |
| // IPv6 references
 | |
| // - https://www.iana.org/assignments/iana-ipv6-special-registry/iana-ipv6-special-registry.xhtml
 | |
| // - https://www.iana.org/assignments/ipv6-multicast-addresses/ipv6-multicast-addresses.xhtml
 | |
| func ValidateNonSpecialIP(ipAddress string, fldPath *field.Path) field.ErrorList {
 | |
| 	allErrs := field.ErrorList{}
 | |
| 	ip := net.ParseIP(ipAddress)
 | |
| 	if ip == nil {
 | |
| 		allErrs = append(allErrs, field.Invalid(fldPath, ipAddress, "must be a valid IP address"))
 | |
| 		return allErrs
 | |
| 	}
 | |
| 	if ip.IsUnspecified() {
 | |
| 		allErrs = append(allErrs, field.Invalid(fldPath, ipAddress, fmt.Sprintf("may not be unspecified (%v)", ipAddress)))
 | |
| 	}
 | |
| 	if ip.IsLoopback() {
 | |
| 		allErrs = append(allErrs, field.Invalid(fldPath, ipAddress, "may not be in the loopback range (127.0.0.0/8, ::1/128)"))
 | |
| 	}
 | |
| 	if ip.IsLinkLocalUnicast() {
 | |
| 		allErrs = append(allErrs, field.Invalid(fldPath, ipAddress, "may not be in the link-local range (169.254.0.0/16, fe80::/10)"))
 | |
| 	}
 | |
| 	if ip.IsLinkLocalMulticast() {
 | |
| 		allErrs = append(allErrs, field.Invalid(fldPath, ipAddress, "may not be in the link-local multicast range (224.0.0.0/24, ff02::/10)"))
 | |
| 	}
 | |
| 	return allErrs
 | |
| }
 | |
| 
 | |
| func validateEndpointPort(port *core.EndpointPort, requireName bool, fldPath *field.Path) field.ErrorList {
 | |
| 	allErrs := field.ErrorList{}
 | |
| 	if requireName && len(port.Name) == 0 {
 | |
| 		allErrs = append(allErrs, field.Required(fldPath.Child("name"), ""))
 | |
| 	} else if len(port.Name) != 0 {
 | |
| 		allErrs = append(allErrs, ValidateDNS1123Label(port.Name, fldPath.Child("name"))...)
 | |
| 	}
 | |
| 	for _, msg := range validation.IsValidPortNum(int(port.Port)) {
 | |
| 		allErrs = append(allErrs, field.Invalid(fldPath.Child("port"), port.Port, msg))
 | |
| 	}
 | |
| 	if len(port.Protocol) == 0 {
 | |
| 		allErrs = append(allErrs, field.Required(fldPath.Child("protocol"), ""))
 | |
| 	} else if !supportedPortProtocols.Has(string(port.Protocol)) {
 | |
| 		allErrs = append(allErrs, field.NotSupported(fldPath.Child("protocol"), port.Protocol, supportedPortProtocols.List()))
 | |
| 	}
 | |
| 	if port.AppProtocol != nil {
 | |
| 		for _, msg := range validation.IsQualifiedName(*port.AppProtocol) {
 | |
| 			allErrs = append(allErrs, field.Invalid(fldPath.Child("appProtocol"), port.AppProtocol, msg))
 | |
| 		}
 | |
| 	}
 | |
| 	return allErrs
 | |
| }
 | |
| 
 | |
| // ValidateSecurityContext ensures the security context contains valid settings
 | |
| func ValidateSecurityContext(sc *core.SecurityContext, fldPath *field.Path) field.ErrorList {
 | |
| 	allErrs := field.ErrorList{}
 | |
| 	//this should only be true for testing since SecurityContext is defaulted by the core
 | |
| 	if sc == nil {
 | |
| 		return allErrs
 | |
| 	}
 | |
| 
 | |
| 	if sc.Privileged != nil {
 | |
| 		if *sc.Privileged && !capabilities.Get().AllowPrivileged {
 | |
| 			allErrs = append(allErrs, field.Forbidden(fldPath.Child("privileged"), "disallowed by cluster policy"))
 | |
| 		}
 | |
| 	}
 | |
| 
 | |
| 	if sc.RunAsUser != nil {
 | |
| 		for _, msg := range validation.IsValidUserID(*sc.RunAsUser) {
 | |
| 			allErrs = append(allErrs, field.Invalid(fldPath.Child("runAsUser"), *sc.RunAsUser, msg))
 | |
| 		}
 | |
| 	}
 | |
| 
 | |
| 	if sc.RunAsGroup != nil {
 | |
| 		for _, msg := range validation.IsValidGroupID(*sc.RunAsGroup) {
 | |
| 			allErrs = append(allErrs, field.Invalid(fldPath.Child("runAsGroup"), *sc.RunAsGroup, msg))
 | |
| 		}
 | |
| 	}
 | |
| 
 | |
| 	if sc.ProcMount != nil {
 | |
| 		if err := ValidateProcMountType(fldPath.Child("procMount"), *sc.ProcMount); err != nil {
 | |
| 			allErrs = append(allErrs, err)
 | |
| 		}
 | |
| 
 | |
| 	}
 | |
| 	allErrs = append(allErrs, validateSeccompProfileField(sc.SeccompProfile, fldPath.Child("seccompProfile"))...)
 | |
| 	if sc.AllowPrivilegeEscalation != nil && !*sc.AllowPrivilegeEscalation {
 | |
| 		if sc.Privileged != nil && *sc.Privileged {
 | |
| 			allErrs = append(allErrs, field.Invalid(fldPath, sc, "cannot set `allowPrivilegeEscalation` to false and `privileged` to true"))
 | |
| 		}
 | |
| 
 | |
| 		if sc.Capabilities != nil {
 | |
| 			for _, cap := range sc.Capabilities.Add {
 | |
| 				if string(cap) == "CAP_SYS_ADMIN" {
 | |
| 					allErrs = append(allErrs, field.Invalid(fldPath, sc, "cannot set `allowPrivilegeEscalation` to false and `capabilities.Add` CAP_SYS_ADMIN"))
 | |
| 				}
 | |
| 			}
 | |
| 		}
 | |
| 	}
 | |
| 
 | |
| 	allErrs = append(allErrs, validateWindowsSecurityContextOptions(sc.WindowsOptions, fldPath.Child("windowsOptions"))...)
 | |
| 
 | |
| 	return allErrs
 | |
| }
 | |
| 
 | |
| // maxGMSACredentialSpecLength is the max length, in bytes, for the actual contents
 | |
| // of a GMSA cred spec. In general, those shouldn't be more than a few hundred bytes,
 | |
| // so we want to give plenty of room here while still providing an upper bound.
 | |
| // The runAsUserName field will be used to execute the given container's entrypoint, and
 | |
| // it can be formatted as "DOMAIN/USER", where the DOMAIN is optional, maxRunAsUserNameDomainLength
 | |
| // is the max character length for the user's DOMAIN, and maxRunAsUserNameUserLength
 | |
| // is the max character length for the USER itself. Both the DOMAIN and USER have their
 | |
| // own restrictions, and more information about them can be found here:
 | |
| // https://support.microsoft.com/en-us/help/909264/naming-conventions-in-active-directory-for-computers-domains-sites-and
 | |
| // https://docs.microsoft.com/en-us/previous-versions/windows/it-pro/windows-2000-server/bb726984(v=technet.10)
 | |
| const (
 | |
| 	maxGMSACredentialSpecLengthInKiB = 64
 | |
| 	maxGMSACredentialSpecLength      = maxGMSACredentialSpecLengthInKiB * 1024
 | |
| 	maxRunAsUserNameDomainLength     = 256
 | |
| 	maxRunAsUserNameUserLength       = 104
 | |
| )
 | |
| 
 | |
| var (
 | |
| 	// control characters are not permitted in the runAsUserName field.
 | |
| 	ctrlRegex = regexp.MustCompile(`[[:cntrl:]]+`)
 | |
| 
 | |
| 	// a valid NetBios Domain name cannot start with a dot, has at least 1 character,
 | |
| 	// at most 15 characters, and it cannot the characters: \ / : * ? " < > |
 | |
| 	validNetBiosRegex = regexp.MustCompile(`^[^\\/:\*\?"<>|\.][^\\/:\*\?"<>|]{0,14}$`)
 | |
| 
 | |
| 	// a valid DNS name contains only alphanumeric characters, dots, and dashes.
 | |
| 	dnsLabelFormat                 = `[a-zA-Z0-9](?:[a-zA-Z0-9-]{0,61}[a-zA-Z0-9])?`
 | |
| 	dnsSubdomainFormat             = fmt.Sprintf(`^%s(?:\.%s)*$`, dnsLabelFormat, dnsLabelFormat)
 | |
| 	validWindowsUserDomainDNSRegex = regexp.MustCompile(dnsSubdomainFormat)
 | |
| 
 | |
| 	// a username is invalid if it contains the characters: " / \ [ ] : ; | = , + * ? < > @
 | |
| 	// or it contains only dots or spaces.
 | |
| 	invalidUserNameCharsRegex      = regexp.MustCompile(`["/\\:;|=,\+\*\?<>@\[\]]`)
 | |
| 	invalidUserNameDotsSpacesRegex = regexp.MustCompile(`^[\. ]+$`)
 | |
| )
 | |
| 
 | |
| func validateWindowsSecurityContextOptions(windowsOptions *core.WindowsSecurityContextOptions, fieldPath *field.Path) field.ErrorList {
 | |
| 	allErrs := field.ErrorList{}
 | |
| 
 | |
| 	if windowsOptions == nil {
 | |
| 		return allErrs
 | |
| 	}
 | |
| 
 | |
| 	if windowsOptions.GMSACredentialSpecName != nil {
 | |
| 		// gmsaCredentialSpecName must be the name of a custom resource
 | |
| 		for _, msg := range validation.IsDNS1123Subdomain(*windowsOptions.GMSACredentialSpecName) {
 | |
| 			allErrs = append(allErrs, field.Invalid(fieldPath.Child("gmsaCredentialSpecName"), windowsOptions.GMSACredentialSpecName, msg))
 | |
| 		}
 | |
| 	}
 | |
| 
 | |
| 	if windowsOptions.GMSACredentialSpec != nil {
 | |
| 		if l := len(*windowsOptions.GMSACredentialSpec); l == 0 {
 | |
| 			allErrs = append(allErrs, field.Invalid(fieldPath.Child("gmsaCredentialSpec"), windowsOptions.GMSACredentialSpec, "gmsaCredentialSpec cannot be an empty string"))
 | |
| 		} else if l > maxGMSACredentialSpecLength {
 | |
| 			errMsg := fmt.Sprintf("gmsaCredentialSpec size must be under %d KiB", maxGMSACredentialSpecLengthInKiB)
 | |
| 			allErrs = append(allErrs, field.Invalid(fieldPath.Child("gmsaCredentialSpec"), windowsOptions.GMSACredentialSpec, errMsg))
 | |
| 		}
 | |
| 	}
 | |
| 
 | |
| 	if windowsOptions.RunAsUserName != nil {
 | |
| 		if l := len(*windowsOptions.RunAsUserName); l == 0 {
 | |
| 			allErrs = append(allErrs, field.Invalid(fieldPath.Child("runAsUserName"), windowsOptions.RunAsUserName, "runAsUserName cannot be an empty string"))
 | |
| 		} else if ctrlRegex.MatchString(*windowsOptions.RunAsUserName) {
 | |
| 			errMsg := "runAsUserName cannot contain control characters"
 | |
| 			allErrs = append(allErrs, field.Invalid(fieldPath.Child("runAsUserName"), windowsOptions.RunAsUserName, errMsg))
 | |
| 		} else if parts := strings.Split(*windowsOptions.RunAsUserName, "\\"); len(parts) > 2 {
 | |
| 			errMsg := "runAsUserName cannot contain more than one backslash"
 | |
| 			allErrs = append(allErrs, field.Invalid(fieldPath.Child("runAsUserName"), windowsOptions.RunAsUserName, errMsg))
 | |
| 		} else {
 | |
| 			var (
 | |
| 				hasDomain = false
 | |
| 				domain    = ""
 | |
| 				user      string
 | |
| 			)
 | |
| 			if len(parts) == 1 {
 | |
| 				user = parts[0]
 | |
| 			} else {
 | |
| 				hasDomain = true
 | |
| 				domain = parts[0]
 | |
| 				user = parts[1]
 | |
| 			}
 | |
| 
 | |
| 			if len(domain) >= maxRunAsUserNameDomainLength {
 | |
| 				errMsg := fmt.Sprintf("runAsUserName's Domain length must be under %d characters", maxRunAsUserNameDomainLength)
 | |
| 				allErrs = append(allErrs, field.Invalid(fieldPath.Child("runAsUserName"), windowsOptions.RunAsUserName, errMsg))
 | |
| 			}
 | |
| 
 | |
| 			if hasDomain && !(validNetBiosRegex.MatchString(domain) || validWindowsUserDomainDNSRegex.MatchString(domain)) {
 | |
| 				errMsg := "runAsUserName's Domain doesn't match the NetBios nor the DNS format"
 | |
| 				allErrs = append(allErrs, field.Invalid(fieldPath.Child("runAsUserName"), windowsOptions.RunAsUserName, errMsg))
 | |
| 			}
 | |
| 
 | |
| 			if l := len(user); l == 0 {
 | |
| 				errMsg := "runAsUserName's User cannot be empty"
 | |
| 				allErrs = append(allErrs, field.Invalid(fieldPath.Child("runAsUserName"), windowsOptions.RunAsUserName, errMsg))
 | |
| 			} else if l > maxRunAsUserNameUserLength {
 | |
| 				errMsg := fmt.Sprintf("runAsUserName's User length must not be longer than %d characters", maxRunAsUserNameUserLength)
 | |
| 				allErrs = append(allErrs, field.Invalid(fieldPath.Child("runAsUserName"), windowsOptions.RunAsUserName, errMsg))
 | |
| 			}
 | |
| 
 | |
| 			if invalidUserNameDotsSpacesRegex.MatchString(user) {
 | |
| 				errMsg := `runAsUserName's User cannot contain only periods or spaces`
 | |
| 				allErrs = append(allErrs, field.Invalid(fieldPath.Child("runAsUserName"), windowsOptions.RunAsUserName, errMsg))
 | |
| 			}
 | |
| 
 | |
| 			if invalidUserNameCharsRegex.MatchString(user) {
 | |
| 				errMsg := `runAsUserName's User cannot contain the following characters: "/\:;|=,+*?<>@[]`
 | |
| 				allErrs = append(allErrs, field.Invalid(fieldPath.Child("runAsUserName"), windowsOptions.RunAsUserName, errMsg))
 | |
| 			}
 | |
| 		}
 | |
| 	}
 | |
| 
 | |
| 	return allErrs
 | |
| }
 | |
| 
 | |
| func validateWindowsHostProcessPod(podSpec *core.PodSpec, fieldPath *field.Path, opts PodValidationOptions) field.ErrorList {
 | |
| 	allErrs := field.ErrorList{}
 | |
| 
 | |
| 	// Keep track of container and hostProcess container count for validate
 | |
| 	containerCount := 0
 | |
| 	hostProcessContainerCount := 0
 | |
| 
 | |
| 	var podHostProcess *bool
 | |
| 	if podSpec.SecurityContext != nil && podSpec.SecurityContext.WindowsOptions != nil {
 | |
| 		podHostProcess = podSpec.SecurityContext.WindowsOptions.HostProcess
 | |
| 	}
 | |
| 
 | |
| 	if !opts.AllowWindowsHostProcessField && podHostProcess != nil {
 | |
| 		// Do not allow pods to persist data that sets hostProcess (true or false)
 | |
| 		errMsg := "not allowed when feature gate 'WindowsHostProcessContainers' is not enabled"
 | |
| 		allErrs = append(allErrs, field.Forbidden(fieldPath.Child("securityContext", "windowsOptions", "hostProcess"), errMsg))
 | |
| 		return allErrs
 | |
| 	}
 | |
| 
 | |
| 	hostNetwork := false
 | |
| 	if podSpec.SecurityContext != nil {
 | |
| 		hostNetwork = podSpec.SecurityContext.HostNetwork
 | |
| 	}
 | |
| 
 | |
| 	podshelper.VisitContainersWithPath(podSpec, fieldPath, func(c *core.Container, cFieldPath *field.Path) bool {
 | |
| 		containerCount++
 | |
| 
 | |
| 		var containerHostProcess *bool = nil
 | |
| 		if c.SecurityContext != nil && c.SecurityContext.WindowsOptions != nil {
 | |
| 			containerHostProcess = c.SecurityContext.WindowsOptions.HostProcess
 | |
| 		}
 | |
| 
 | |
| 		if !opts.AllowWindowsHostProcessField && containerHostProcess != nil {
 | |
| 			// Do not allow pods to persist data that sets hostProcess (true or false)
 | |
| 			errMsg := "not allowed when feature gate 'WindowsHostProcessContainers' is not enabled"
 | |
| 			allErrs = append(allErrs, field.Forbidden(cFieldPath.Child("securityContext", "windowsOptions", "hostProcess"), errMsg))
 | |
| 		}
 | |
| 
 | |
| 		if podHostProcess != nil && containerHostProcess != nil && *podHostProcess != *containerHostProcess {
 | |
| 			errMsg := fmt.Sprintf("pod hostProcess value must be identical if both are specified, was %v", *podHostProcess)
 | |
| 			allErrs = append(allErrs, field.Invalid(cFieldPath.Child("securityContext", "windowsOptions", "hostProcess"), *containerHostProcess, errMsg))
 | |
| 		}
 | |
| 
 | |
| 		switch {
 | |
| 		case containerHostProcess != nil && *containerHostProcess:
 | |
| 			// Container explitly sets hostProcess=true
 | |
| 			hostProcessContainerCount++
 | |
| 		case containerHostProcess == nil && podHostProcess != nil && *podHostProcess:
 | |
| 			// Container inherits hostProcess=true from pod settings
 | |
| 			hostProcessContainerCount++
 | |
| 		}
 | |
| 
 | |
| 		return true
 | |
| 	})
 | |
| 
 | |
| 	if hostProcessContainerCount > 0 {
 | |
| 		// Fail Pod validation if feature is not enabled (unless podspec already exists and contains HostProcess fields) instead of dropping fields based on PRR reivew.
 | |
| 		if !opts.AllowWindowsHostProcessField {
 | |
| 			errMsg := "pod must not contain Windows hostProcess containers when feature gate 'WindowsHostProcessContainers' is not enabled"
 | |
| 			allErrs = append(allErrs, field.Forbidden(fieldPath, errMsg))
 | |
| 			return allErrs
 | |
| 		}
 | |
| 
 | |
| 		// At present, if a Windows Pods contains any HostProcess containers than all containers must be
 | |
| 		// HostProcess containers (explicitly set or inherited).
 | |
| 		if hostProcessContainerCount != containerCount {
 | |
| 			errMsg := "If pod contains any hostProcess containers then all containers must be HostProcess containers"
 | |
| 			allErrs = append(allErrs, field.Invalid(fieldPath, "", errMsg))
 | |
| 		}
 | |
| 
 | |
| 		// At present Windows Pods which contain HostProcess containers must also set HostNetwork.
 | |
| 		if hostNetwork != true {
 | |
| 			errMsg := "hostNetwork must be true if pod contains any hostProcess containers"
 | |
| 			allErrs = append(allErrs, field.Invalid(fieldPath.Child("hostNetwork"), hostNetwork, errMsg))
 | |
| 		}
 | |
| 
 | |
| 		if !capabilities.Get().AllowPrivileged {
 | |
| 			errMsg := "hostProcess containers are disallowed by cluster policy"
 | |
| 			allErrs = append(allErrs, field.Forbidden(fieldPath, errMsg))
 | |
| 		}
 | |
| 	}
 | |
| 
 | |
| 	return allErrs
 | |
| }
 | |
| 
 | |
| func ValidatePodLogOptions(opts *core.PodLogOptions) field.ErrorList {
 | |
| 	allErrs := field.ErrorList{}
 | |
| 	if opts.TailLines != nil && *opts.TailLines < 0 {
 | |
| 		allErrs = append(allErrs, field.Invalid(field.NewPath("tailLines"), *opts.TailLines, isNegativeErrorMsg))
 | |
| 	}
 | |
| 	if opts.LimitBytes != nil && *opts.LimitBytes < 1 {
 | |
| 		allErrs = append(allErrs, field.Invalid(field.NewPath("limitBytes"), *opts.LimitBytes, "must be greater than 0"))
 | |
| 	}
 | |
| 	switch {
 | |
| 	case opts.SinceSeconds != nil && opts.SinceTime != nil:
 | |
| 		allErrs = append(allErrs, field.Forbidden(field.NewPath(""), "at most one of `sinceTime` or `sinceSeconds` may be specified"))
 | |
| 	case opts.SinceSeconds != nil:
 | |
| 		if *opts.SinceSeconds < 1 {
 | |
| 			allErrs = append(allErrs, field.Invalid(field.NewPath("sinceSeconds"), *opts.SinceSeconds, "must be greater than 0"))
 | |
| 		}
 | |
| 	}
 | |
| 	return allErrs
 | |
| }
 | |
| 
 | |
| // ValidateLoadBalancerStatus validates required fields on a LoadBalancerStatus
 | |
| func ValidateLoadBalancerStatus(status *core.LoadBalancerStatus, fldPath *field.Path) field.ErrorList {
 | |
| 	allErrs := field.ErrorList{}
 | |
| 	for i, ingress := range status.Ingress {
 | |
| 		idxPath := fldPath.Child("ingress").Index(i)
 | |
| 		if len(ingress.IP) > 0 {
 | |
| 			if isIP := (net.ParseIP(ingress.IP) != nil); !isIP {
 | |
| 				allErrs = append(allErrs, field.Invalid(idxPath.Child("ip"), ingress.IP, "must be a valid IP address"))
 | |
| 			}
 | |
| 		}
 | |
| 		if len(ingress.Hostname) > 0 {
 | |
| 			for _, msg := range validation.IsDNS1123Subdomain(ingress.Hostname) {
 | |
| 				allErrs = append(allErrs, field.Invalid(idxPath.Child("hostname"), ingress.Hostname, msg))
 | |
| 			}
 | |
| 			if isIP := (net.ParseIP(ingress.Hostname) != nil); isIP {
 | |
| 				allErrs = append(allErrs, field.Invalid(idxPath.Child("hostname"), ingress.Hostname, "must be a DNS name, not an IP address"))
 | |
| 			}
 | |
| 		}
 | |
| 	}
 | |
| 	return allErrs
 | |
| }
 | |
| 
 | |
| // validateVolumeNodeAffinity tests that the PersistentVolume.NodeAffinity has valid data
 | |
| // returns:
 | |
| // - true if volumeNodeAffinity is set
 | |
| // - errorList if there are validation errors
 | |
| func validateVolumeNodeAffinity(nodeAffinity *core.VolumeNodeAffinity, fldPath *field.Path) (bool, field.ErrorList) {
 | |
| 	allErrs := field.ErrorList{}
 | |
| 
 | |
| 	if nodeAffinity == nil {
 | |
| 		return false, allErrs
 | |
| 	}
 | |
| 
 | |
| 	if nodeAffinity.Required != nil {
 | |
| 		allErrs = append(allErrs, ValidateNodeSelector(nodeAffinity.Required, fldPath.Child("required"))...)
 | |
| 	} else {
 | |
| 		allErrs = append(allErrs, field.Required(fldPath.Child("required"), "must specify required node constraints"))
 | |
| 	}
 | |
| 
 | |
| 	return true, allErrs
 | |
| }
 | |
| 
 | |
| // ValidateCIDR validates whether a CIDR matches the conventions expected by net.ParseCIDR
 | |
| func ValidateCIDR(cidr string) (*net.IPNet, error) {
 | |
| 	_, net, err := net.ParseCIDR(cidr)
 | |
| 	if err != nil {
 | |
| 		return nil, err
 | |
| 	}
 | |
| 	return net, nil
 | |
| }
 | |
| 
 | |
| func IsDecremented(update, old *int32) bool {
 | |
| 	if update == nil && old != nil {
 | |
| 		return true
 | |
| 	}
 | |
| 	if update == nil || old == nil {
 | |
| 		return false
 | |
| 	}
 | |
| 	return *update < *old
 | |
| }
 | |
| 
 | |
| // ValidateProcMountType tests that the argument is a valid ProcMountType.
 | |
| func ValidateProcMountType(fldPath *field.Path, procMountType core.ProcMountType) *field.Error {
 | |
| 	switch procMountType {
 | |
| 	case core.DefaultProcMount, core.UnmaskedProcMount:
 | |
| 		return nil
 | |
| 	default:
 | |
| 		return field.NotSupported(fldPath, procMountType, []string{string(core.DefaultProcMount), string(core.UnmaskedProcMount)})
 | |
| 	}
 | |
| }
 | |
| 
 | |
| var (
 | |
| 	supportedScheduleActions = sets.NewString(string(core.DoNotSchedule), string(core.ScheduleAnyway))
 | |
| )
 | |
| 
 | |
| // validateTopologySpreadConstraints validates given TopologySpreadConstraints.
 | |
| func validateTopologySpreadConstraints(constraints []core.TopologySpreadConstraint, fldPath *field.Path) field.ErrorList {
 | |
| 	allErrs := field.ErrorList{}
 | |
| 
 | |
| 	for i, constraint := range constraints {
 | |
| 		subFldPath := fldPath.Index(i)
 | |
| 		if err := ValidateMaxSkew(subFldPath.Child("maxSkew"), constraint.MaxSkew); err != nil {
 | |
| 			allErrs = append(allErrs, err)
 | |
| 		}
 | |
| 		if err := ValidateTopologyKey(subFldPath.Child("topologyKey"), constraint.TopologyKey); err != nil {
 | |
| 			allErrs = append(allErrs, err)
 | |
| 		}
 | |
| 		if err := ValidateWhenUnsatisfiable(subFldPath.Child("whenUnsatisfiable"), constraint.WhenUnsatisfiable); err != nil {
 | |
| 			allErrs = append(allErrs, err)
 | |
| 		}
 | |
| 		// tuple {topologyKey, whenUnsatisfiable} denotes one kind of spread constraint
 | |
| 		if err := ValidateSpreadConstraintNotRepeat(subFldPath.Child("{topologyKey, whenUnsatisfiable}"), constraint, constraints[i+1:]); err != nil {
 | |
| 			allErrs = append(allErrs, err)
 | |
| 		}
 | |
| 	}
 | |
| 
 | |
| 	return allErrs
 | |
| }
 | |
| 
 | |
| // ValidateMaxSkew tests that the argument is a valid MaxSkew.
 | |
| func ValidateMaxSkew(fldPath *field.Path, maxSkew int32) *field.Error {
 | |
| 	if maxSkew <= 0 {
 | |
| 		return field.Invalid(fldPath, maxSkew, isNotPositiveErrorMsg)
 | |
| 	}
 | |
| 	return nil
 | |
| }
 | |
| 
 | |
| // ValidateTopologyKey tests that the argument is a valid TopologyKey.
 | |
| func ValidateTopologyKey(fldPath *field.Path, topologyKey string) *field.Error {
 | |
| 	if len(topologyKey) == 0 {
 | |
| 		return field.Required(fldPath, "can not be empty")
 | |
| 	}
 | |
| 	return nil
 | |
| }
 | |
| 
 | |
| // ValidateWhenUnsatisfiable tests that the argument is a valid UnsatisfiableConstraintAction.
 | |
| func ValidateWhenUnsatisfiable(fldPath *field.Path, action core.UnsatisfiableConstraintAction) *field.Error {
 | |
| 	if !supportedScheduleActions.Has(string(action)) {
 | |
| 		return field.NotSupported(fldPath, action, supportedScheduleActions.List())
 | |
| 	}
 | |
| 	return nil
 | |
| }
 | |
| 
 | |
| // ValidateSpreadConstraintNotRepeat tests that if `constraint` duplicates with `existingConstraintPairs`
 | |
| // on TopologyKey and WhenUnsatisfiable fields.
 | |
| func ValidateSpreadConstraintNotRepeat(fldPath *field.Path, constraint core.TopologySpreadConstraint, restingConstraints []core.TopologySpreadConstraint) *field.Error {
 | |
| 	for _, restingConstraint := range restingConstraints {
 | |
| 		if constraint.TopologyKey == restingConstraint.TopologyKey &&
 | |
| 			constraint.WhenUnsatisfiable == restingConstraint.WhenUnsatisfiable {
 | |
| 			return field.Duplicate(fldPath, fmt.Sprintf("{%v, %v}", constraint.TopologyKey, constraint.WhenUnsatisfiable))
 | |
| 		}
 | |
| 	}
 | |
| 	return nil
 | |
| }
 | |
| 
 | |
| // validateServiceClusterIPsRelatedFields validates .spec.ClusterIPs,, .spec.IPFamilies, .spec.ipFamilyPolicy
 | |
| func validateServiceClusterIPsRelatedFields(service *core.Service) field.ErrorList {
 | |
| 	// ClusterIP, ClusterIPs, IPFamilyPolicy and IPFamilies are validated prior (all must be unset) for ExternalName service
 | |
| 	if service.Spec.Type == core.ServiceTypeExternalName {
 | |
| 		return field.ErrorList{}
 | |
| 	}
 | |
| 
 | |
| 	allErrs := field.ErrorList{}
 | |
| 	hasInvalidIPs := false
 | |
| 
 | |
| 	specPath := field.NewPath("spec")
 | |
| 	clusterIPsField := specPath.Child("clusterIPs")
 | |
| 	ipFamiliesField := specPath.Child("ipFamilies")
 | |
| 	ipFamilyPolicyField := specPath.Child("ipFamilyPolicy")
 | |
| 
 | |
| 	// Make sure ClusterIP and ClusterIPs are synced.  For most cases users can
 | |
| 	// just manage one or the other and we'll handle the rest (see PrepareFor*
 | |
| 	// in strategy).
 | |
| 	if len(service.Spec.ClusterIP) != 0 {
 | |
| 		// If ClusterIP is set, ClusterIPs[0] must match.
 | |
| 		if len(service.Spec.ClusterIPs) == 0 {
 | |
| 			allErrs = append(allErrs, field.Required(clusterIPsField, ""))
 | |
| 		} else if service.Spec.ClusterIPs[0] != service.Spec.ClusterIP {
 | |
| 			allErrs = append(allErrs, field.Invalid(clusterIPsField, service.Spec.ClusterIPs, "element [0] must match clusterIP"))
 | |
| 		}
 | |
| 	} else { // ClusterIP == ""
 | |
| 		// If ClusterIP is not set, ClusterIPs must also be unset.
 | |
| 		if len(service.Spec.ClusterIPs) != 0 {
 | |
| 			allErrs = append(allErrs, field.Invalid(clusterIPsField, service.Spec.ClusterIPs, "must be empty when clusterIP is empty"))
 | |
| 		}
 | |
| 	}
 | |
| 
 | |
| 	// ipfamilies stand alone validation
 | |
| 	// must be either IPv4 or IPv6
 | |
| 	seen := sets.String{}
 | |
| 	for i, ipFamily := range service.Spec.IPFamilies {
 | |
| 		if !supportedServiceIPFamily.Has(string(ipFamily)) {
 | |
| 			allErrs = append(allErrs, field.NotSupported(ipFamiliesField.Index(i), ipFamily, supportedServiceIPFamily.List()))
 | |
| 		}
 | |
| 		// no duplicate check also ensures that ipfamilies is dualstacked, in any order
 | |
| 		if seen.Has(string(ipFamily)) {
 | |
| 			allErrs = append(allErrs, field.Duplicate(ipFamiliesField.Index(i), ipFamily))
 | |
| 		}
 | |
| 		seen.Insert(string(ipFamily))
 | |
| 	}
 | |
| 
 | |
| 	// IPFamilyPolicy stand alone validation
 | |
| 	//note: nil is ok, defaulted in alloc check registry/core/service/*
 | |
| 	if service.Spec.IPFamilyPolicy != nil {
 | |
| 		// must have a supported value
 | |
| 		if !supportedServiceIPFamilyPolicy.Has(string(*(service.Spec.IPFamilyPolicy))) {
 | |
| 			allErrs = append(allErrs, field.NotSupported(ipFamilyPolicyField, service.Spec.IPFamilyPolicy, supportedServiceIPFamilyPolicy.List()))
 | |
| 		}
 | |
| 	}
 | |
| 
 | |
| 	// clusterIPs stand alone validation
 | |
| 	// valid ips with None and empty string handling
 | |
| 	// duplication check is done as part of DualStackvalidation below
 | |
| 	for i, clusterIP := range service.Spec.ClusterIPs {
 | |
| 		// valid at first location only. if and only if len(clusterIPs) == 1
 | |
| 		if i == 0 && clusterIP == core.ClusterIPNone {
 | |
| 			if len(service.Spec.ClusterIPs) > 1 {
 | |
| 				hasInvalidIPs = true
 | |
| 				allErrs = append(allErrs, field.Invalid(clusterIPsField, service.Spec.ClusterIPs, "'None' must be the first and only value"))
 | |
| 			}
 | |
| 			continue
 | |
| 		}
 | |
| 
 | |
| 		// is it valid ip?
 | |
| 		errorMessages := validation.IsValidIP(clusterIP)
 | |
| 		hasInvalidIPs = (len(errorMessages) != 0) || hasInvalidIPs
 | |
| 		for _, msg := range errorMessages {
 | |
| 			allErrs = append(allErrs, field.Invalid(clusterIPsField.Index(i), clusterIP, msg))
 | |
| 		}
 | |
| 	}
 | |
| 
 | |
| 	// max two
 | |
| 	if len(service.Spec.ClusterIPs) > 2 {
 | |
| 		allErrs = append(allErrs, field.Invalid(clusterIPsField, service.Spec.ClusterIPs, "may only hold up to 2 values"))
 | |
| 	}
 | |
| 
 | |
| 	// at this stage if there is an invalid ip or misplaced none/empty string
 | |
| 	// it will skew the error messages (bad index || dualstackness of already bad ips). so we
 | |
| 	// stop here if there are errors in clusterIPs validation
 | |
| 	if hasInvalidIPs {
 | |
| 		return allErrs
 | |
| 	}
 | |
| 
 | |
| 	// must be dual stacked ips if they are more than one ip
 | |
| 	if len(service.Spec.ClusterIPs) > 1 /* meaning: it does not have a None or empty string */ {
 | |
| 		dualStack, err := netutils.IsDualStackIPStrings(service.Spec.ClusterIPs)
 | |
| 		if err != nil { // though we check for that earlier. safe > sorry
 | |
| 			allErrs = append(allErrs, field.InternalError(clusterIPsField, fmt.Errorf("failed to check for dual stack with error:%v", err)))
 | |
| 		}
 | |
| 
 | |
| 		// We only support one from each IP family (i.e. max two IPs in this list).
 | |
| 		if !dualStack {
 | |
| 			allErrs = append(allErrs, field.Invalid(clusterIPsField, service.Spec.ClusterIPs, "may specify no more than one IP for each IP family"))
 | |
| 		}
 | |
| 	}
 | |
| 
 | |
| 	// match clusterIPs to their families, if they were provided
 | |
| 	if !isHeadlessService(service) && len(service.Spec.ClusterIPs) > 0 && len(service.Spec.IPFamilies) > 0 {
 | |
| 		for i, ip := range service.Spec.ClusterIPs {
 | |
| 			if i > (len(service.Spec.IPFamilies) - 1) {
 | |
| 				break // no more families to check
 | |
| 			}
 | |
| 
 | |
| 			// 4=>6
 | |
| 			if service.Spec.IPFamilies[i] == core.IPv4Protocol && netutils.IsIPv6String(ip) {
 | |
| 				allErrs = append(allErrs, field.Invalid(clusterIPsField.Index(i), ip, fmt.Sprintf("expected an IPv4 value as indicated by `ipFamilies[%v]`", i)))
 | |
| 			}
 | |
| 			// 6=>4
 | |
| 			if service.Spec.IPFamilies[i] == core.IPv6Protocol && !netutils.IsIPv6String(ip) {
 | |
| 				allErrs = append(allErrs, field.Invalid(clusterIPsField.Index(i), ip, fmt.Sprintf("expected an IPv6 value as indicated by `ipFamilies[%v]`", i)))
 | |
| 			}
 | |
| 		}
 | |
| 	}
 | |
| 
 | |
| 	return allErrs
 | |
| }
 | |
| 
 | |
| // specific validation for clusterIPs in cases of user upgrading or downgrading to/from dualstack
 | |
| func validateUpgradeDowngradeClusterIPs(oldService, service *core.Service) field.ErrorList {
 | |
| 	allErrs := make(field.ErrorList, 0)
 | |
| 
 | |
| 	// bail out early for ExternalName
 | |
| 	if service.Spec.Type == core.ServiceTypeExternalName || oldService.Spec.Type == core.ServiceTypeExternalName {
 | |
| 		return allErrs
 | |
| 	}
 | |
| 	newIsHeadless := isHeadlessService(service)
 | |
| 	oldIsHeadless := isHeadlessService(oldService)
 | |
| 
 | |
| 	if oldIsHeadless && newIsHeadless {
 | |
| 		return allErrs
 | |
| 	}
 | |
| 
 | |
| 	switch {
 | |
| 	// no change in ClusterIP lengths
 | |
| 	// compare each
 | |
| 	case len(oldService.Spec.ClusterIPs) == len(service.Spec.ClusterIPs):
 | |
| 		for i, ip := range oldService.Spec.ClusterIPs {
 | |
| 			if ip != service.Spec.ClusterIPs[i] {
 | |
| 				allErrs = append(allErrs, field.Invalid(field.NewPath("spec", "clusterIPs").Index(i), service.Spec.ClusterIPs, "may not change once set"))
 | |
| 			}
 | |
| 		}
 | |
| 
 | |
| 	// something has been released (downgraded)
 | |
| 	case len(oldService.Spec.ClusterIPs) > len(service.Spec.ClusterIPs):
 | |
| 		// primary ClusterIP has been released
 | |
| 		if len(service.Spec.ClusterIPs) == 0 {
 | |
| 			allErrs = append(allErrs, field.Invalid(field.NewPath("spec", "clusterIPs").Index(0), service.Spec.ClusterIPs, "primary clusterIP can not be unset"))
 | |
| 		}
 | |
| 
 | |
| 		// test if primary clusterIP has changed
 | |
| 		if len(oldService.Spec.ClusterIPs) > 0 &&
 | |
| 			len(service.Spec.ClusterIPs) > 0 &&
 | |
| 			service.Spec.ClusterIPs[0] != oldService.Spec.ClusterIPs[0] {
 | |
| 			allErrs = append(allErrs, field.Invalid(field.NewPath("spec", "clusterIPs").Index(0), service.Spec.ClusterIPs, "may not change once set"))
 | |
| 		}
 | |
| 
 | |
| 		// test if secondary ClusterIP has been released. has this service been downgraded correctly?
 | |
| 		// user *must* set IPFamilyPolicy == SingleStack
 | |
| 		if len(service.Spec.ClusterIPs) == 1 {
 | |
| 			if service.Spec.IPFamilyPolicy == nil || *(service.Spec.IPFamilyPolicy) != core.IPFamilyPolicySingleStack {
 | |
| 				allErrs = append(allErrs, field.Invalid(field.NewPath("spec", "clusterIPs").Index(0), service.Spec.ClusterIPs, "`ipFamilyPolicy` must be set to 'SingleStack' when releasing the secondary clusterIP"))
 | |
| 			}
 | |
| 		}
 | |
| 	case len(oldService.Spec.ClusterIPs) < len(service.Spec.ClusterIPs):
 | |
| 		// something has been added (upgraded)
 | |
| 		// test if primary clusterIP has changed
 | |
| 		if len(oldService.Spec.ClusterIPs) > 0 &&
 | |
| 			service.Spec.ClusterIPs[0] != oldService.Spec.ClusterIPs[0] {
 | |
| 			allErrs = append(allErrs, field.Invalid(field.NewPath("spec", "clusterIPs").Index(0), service.Spec.ClusterIPs, "may not change once set"))
 | |
| 		}
 | |
| 		// we don't check for Policy == RequireDualStack here since, Validation/Creation func takes care of it
 | |
| 	}
 | |
| 	return allErrs
 | |
| }
 | |
| 
 | |
| // specific validation for ipFamilies in cases of user upgrading or downgrading to/from dualstack
 | |
| func validateUpgradeDowngradeIPFamilies(oldService, service *core.Service) field.ErrorList {
 | |
| 	allErrs := make(field.ErrorList, 0)
 | |
| 	// bail out early for ExternalName
 | |
| 	if service.Spec.Type == core.ServiceTypeExternalName || oldService.Spec.Type == core.ServiceTypeExternalName {
 | |
| 		return allErrs
 | |
| 	}
 | |
| 
 | |
| 	oldIsHeadless := isHeadlessService(oldService)
 | |
| 	newIsHeadless := isHeadlessService(service)
 | |
| 
 | |
| 	// if changed to/from headless, then bail out
 | |
| 	if newIsHeadless != oldIsHeadless {
 | |
| 		return allErrs
 | |
| 	}
 | |
| 	// headless can change families
 | |
| 	if newIsHeadless {
 | |
| 		return allErrs
 | |
| 	}
 | |
| 
 | |
| 	switch {
 | |
| 	case len(oldService.Spec.IPFamilies) == len(service.Spec.IPFamilies):
 | |
| 		// no change in ClusterIP lengths
 | |
| 		// compare each
 | |
| 
 | |
| 		for i, ip := range oldService.Spec.IPFamilies {
 | |
| 			if ip != service.Spec.IPFamilies[i] {
 | |
| 				allErrs = append(allErrs, field.Invalid(field.NewPath("spec", "ipFamilies").Index(0), service.Spec.IPFamilies, "may not change once set"))
 | |
| 			}
 | |
| 		}
 | |
| 
 | |
| 	case len(oldService.Spec.IPFamilies) > len(service.Spec.IPFamilies):
 | |
| 		// something has been released (downgraded)
 | |
| 
 | |
| 		// test if primary ipfamily has been released
 | |
| 		if len(service.Spec.ClusterIPs) == 0 {
 | |
| 			allErrs = append(allErrs, field.Invalid(field.NewPath("spec", "ipFamilies").Index(0), service.Spec.IPFamilies, "primary ipFamily can not be unset"))
 | |
| 		}
 | |
| 
 | |
| 		// test if primary ipFamily has changed
 | |
| 		if len(service.Spec.IPFamilies) > 0 &&
 | |
| 			service.Spec.IPFamilies[0] != oldService.Spec.IPFamilies[0] {
 | |
| 			allErrs = append(allErrs, field.Invalid(field.NewPath("spec", "ipFamilies").Index(0), service.Spec.ClusterIPs, "may not change once set"))
 | |
| 		}
 | |
| 
 | |
| 		// test if secondary IPFamily has been released. has this service been downgraded correctly?
 | |
| 		// user *must* set IPFamilyPolicy == SingleStack
 | |
| 		if len(service.Spec.IPFamilies) == 1 {
 | |
| 			if service.Spec.IPFamilyPolicy == nil || *(service.Spec.IPFamilyPolicy) != core.IPFamilyPolicySingleStack {
 | |
| 				allErrs = append(allErrs, field.Invalid(field.NewPath("spec", "clusterIPs").Index(0), service.Spec.ClusterIPs, "`ipFamilyPolicy` must be set to 'SingleStack' when releasing the secondary ipFamily"))
 | |
| 			}
 | |
| 		}
 | |
| 	case len(oldService.Spec.IPFamilies) < len(service.Spec.IPFamilies):
 | |
| 		// something has been added (upgraded)
 | |
| 
 | |
| 		// test if primary ipFamily has changed
 | |
| 		if len(oldService.Spec.IPFamilies) > 0 &&
 | |
| 			len(service.Spec.IPFamilies) > 0 &&
 | |
| 			service.Spec.IPFamilies[0] != oldService.Spec.IPFamilies[0] {
 | |
| 			allErrs = append(allErrs, field.Invalid(field.NewPath("spec", "ipFamilies").Index(0), service.Spec.ClusterIPs, "may not change once set"))
 | |
| 		}
 | |
| 		// we don't check for Policy == RequireDualStack here since, Validation/Creation func takes care of it
 | |
| 	}
 | |
| 	return allErrs
 | |
| }
 | |
| 
 | |
| func isHeadlessService(service *core.Service) bool {
 | |
| 	return service != nil &&
 | |
| 		len(service.Spec.ClusterIPs) == 1 &&
 | |
| 		service.Spec.ClusterIPs[0] == core.ClusterIPNone
 | |
| }
 | |
| 
 | |
| // validateLoadBalancerClassField validation for loadBalancerClass
 | |
| func validateLoadBalancerClassField(oldService, service *core.Service) field.ErrorList {
 | |
| 	allErrs := make(field.ErrorList, 0)
 | |
| 	if oldService != nil {
 | |
| 		// validate update op
 | |
| 		if isTypeLoadBalancer(oldService) && isTypeLoadBalancer(service) {
 | |
| 			// old and new are both LoadBalancer
 | |
| 			if !sameLoadBalancerClass(oldService, service) {
 | |
| 				// can't change loadBalancerClass
 | |
| 				allErrs = append(allErrs, field.Invalid(field.NewPath("spec", "loadBalancerClass"), service.Spec.LoadBalancerClass, "may not change once set"))
 | |
| 			}
 | |
| 		}
 | |
| 	}
 | |
| 
 | |
| 	if isTypeLoadBalancer(service) {
 | |
| 		// check LoadBalancerClass format
 | |
| 		if service.Spec.LoadBalancerClass != nil {
 | |
| 			allErrs = append(allErrs, ValidateQualifiedName(*service.Spec.LoadBalancerClass, field.NewPath("spec", "loadBalancerClass"))...)
 | |
| 		}
 | |
| 	} else {
 | |
| 		// check if LoadBalancerClass set for non LoadBalancer type of service
 | |
| 		if service.Spec.LoadBalancerClass != nil {
 | |
| 			allErrs = append(allErrs, field.Forbidden(field.NewPath("spec", "loadBalancerClass"), "may only be used when `type` is 'LoadBalancer'"))
 | |
| 		}
 | |
| 	}
 | |
| 	return allErrs
 | |
| }
 | |
| 
 | |
| // isTypeLoadBalancer tests service type is loadBalancer or not
 | |
| func isTypeLoadBalancer(service *core.Service) bool {
 | |
| 	return service.Spec.Type == core.ServiceTypeLoadBalancer
 | |
| }
 | |
| 
 | |
| // sameLoadBalancerClass check two services have the same loadBalancerClass or not
 | |
| func sameLoadBalancerClass(oldService, service *core.Service) bool {
 | |
| 	if oldService.Spec.LoadBalancerClass == nil && service.Spec.LoadBalancerClass == nil {
 | |
| 		return true
 | |
| 	}
 | |
| 	if oldService.Spec.LoadBalancerClass == nil || service.Spec.LoadBalancerClass == nil {
 | |
| 		return false
 | |
| 	}
 | |
| 	return *oldService.Spec.LoadBalancerClass == *service.Spec.LoadBalancerClass
 | |
| }
 | 
