mirror of
				https://github.com/optim-enterprises-bv/kubernetes.git
				synced 2025-11-03 03:38:15 +00:00 
			
		
		
		
	
		
			
				
	
	
		
			309 lines
		
	
	
		
			9.2 KiB
		
	
	
	
		
			Go
		
	
	
	
	
	
			
		
		
	
	
			309 lines
		
	
	
		
			9.2 KiB
		
	
	
	
		
			Go
		
	
	
	
	
	
// +build integration,!no-etcd
 | 
						|
 | 
						|
/*
 | 
						|
Copyright 2015 The Kubernetes Authors.
 | 
						|
 | 
						|
Licensed under the Apache License, Version 2.0 (the "License");
 | 
						|
you may not use this file except in compliance with the License.
 | 
						|
You may obtain a copy of the License at
 | 
						|
 | 
						|
    http://www.apache.org/licenses/LICENSE-2.0
 | 
						|
 | 
						|
Unless required by applicable law or agreed to in writing, software
 | 
						|
distributed under the License is distributed on an "AS IS" BASIS,
 | 
						|
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
 | 
						|
See the License for the specific language governing permissions and
 | 
						|
limitations under the License.
 | 
						|
*/
 | 
						|
 | 
						|
package scheduler
 | 
						|
 | 
						|
// This file tests scheduler extender.
 | 
						|
 | 
						|
import (
 | 
						|
	"encoding/json"
 | 
						|
	"fmt"
 | 
						|
	"net/http"
 | 
						|
	"net/http/httptest"
 | 
						|
	"strings"
 | 
						|
	"testing"
 | 
						|
	"time"
 | 
						|
 | 
						|
	"k8s.io/apimachinery/pkg/api/resource"
 | 
						|
	metav1 "k8s.io/apimachinery/pkg/apis/meta/v1"
 | 
						|
	"k8s.io/apimachinery/pkg/util/wait"
 | 
						|
	v1core "k8s.io/client-go/kubernetes/typed/core/v1"
 | 
						|
	clientv1 "k8s.io/client-go/pkg/api/v1"
 | 
						|
	restclient "k8s.io/client-go/rest"
 | 
						|
	"k8s.io/client-go/tools/record"
 | 
						|
	"k8s.io/kubernetes/pkg/api"
 | 
						|
	"k8s.io/kubernetes/pkg/api/v1"
 | 
						|
	"k8s.io/kubernetes/pkg/client/clientset_generated/clientset"
 | 
						|
	"k8s.io/kubernetes/plugin/pkg/scheduler"
 | 
						|
	_ "k8s.io/kubernetes/plugin/pkg/scheduler/algorithmprovider"
 | 
						|
	schedulerapi "k8s.io/kubernetes/plugin/pkg/scheduler/api"
 | 
						|
	"k8s.io/kubernetes/plugin/pkg/scheduler/factory"
 | 
						|
	e2e "k8s.io/kubernetes/test/e2e/framework"
 | 
						|
	"k8s.io/kubernetes/test/integration/framework"
 | 
						|
)
 | 
						|
 | 
						|
const (
 | 
						|
	filter     = "filter"
 | 
						|
	prioritize = "prioritize"
 | 
						|
)
 | 
						|
 | 
						|
type fitPredicate func(pod *v1.Pod, node *v1.Node) (bool, error)
 | 
						|
type priorityFunc func(pod *v1.Pod, nodes *v1.NodeList) (*schedulerapi.HostPriorityList, error)
 | 
						|
 | 
						|
type priorityConfig struct {
 | 
						|
	function priorityFunc
 | 
						|
	weight   int
 | 
						|
}
 | 
						|
 | 
						|
type Extender struct {
 | 
						|
	name         string
 | 
						|
	predicates   []fitPredicate
 | 
						|
	prioritizers []priorityConfig
 | 
						|
}
 | 
						|
 | 
						|
func (e *Extender) serveHTTP(t *testing.T, w http.ResponseWriter, req *http.Request) {
 | 
						|
	var args schedulerapi.ExtenderArgs
 | 
						|
 | 
						|
	decoder := json.NewDecoder(req.Body)
 | 
						|
	defer req.Body.Close()
 | 
						|
 | 
						|
	if err := decoder.Decode(&args); err != nil {
 | 
						|
		http.Error(w, "Decode error", http.StatusBadRequest)
 | 
						|
		return
 | 
						|
	}
 | 
						|
 | 
						|
	encoder := json.NewEncoder(w)
 | 
						|
 | 
						|
	if strings.Contains(req.URL.Path, filter) {
 | 
						|
		resp := &schedulerapi.ExtenderFilterResult{}
 | 
						|
		nodes, failedNodes, err := e.Filter(&args.Pod, &args.Nodes)
 | 
						|
		if err != nil {
 | 
						|
			resp.Error = err.Error()
 | 
						|
		} else {
 | 
						|
			resp.Nodes = *nodes
 | 
						|
			resp.FailedNodes = failedNodes
 | 
						|
		}
 | 
						|
 | 
						|
		if err := encoder.Encode(resp); err != nil {
 | 
						|
			t.Fatalf("Failed to encode %+v", resp)
 | 
						|
		}
 | 
						|
	} else if strings.Contains(req.URL.Path, prioritize) {
 | 
						|
		// Prioritize errors are ignored. Default k8s priorities or another extender's
 | 
						|
		// priorities may be applied.
 | 
						|
		priorities, _ := e.Prioritize(&args.Pod, &args.Nodes)
 | 
						|
 | 
						|
		if err := encoder.Encode(priorities); err != nil {
 | 
						|
			t.Fatalf("Failed to encode %+v", priorities)
 | 
						|
		}
 | 
						|
	} else {
 | 
						|
		http.Error(w, "Unknown method", http.StatusNotFound)
 | 
						|
	}
 | 
						|
}
 | 
						|
 | 
						|
func (e *Extender) Filter(pod *v1.Pod, nodes *v1.NodeList) (*v1.NodeList, schedulerapi.FailedNodesMap, error) {
 | 
						|
	filtered := []v1.Node{}
 | 
						|
	failedNodesMap := schedulerapi.FailedNodesMap{}
 | 
						|
	for _, node := range nodes.Items {
 | 
						|
		fits := true
 | 
						|
		for _, predicate := range e.predicates {
 | 
						|
			fit, err := predicate(pod, &node)
 | 
						|
			if err != nil {
 | 
						|
				return &v1.NodeList{}, schedulerapi.FailedNodesMap{}, err
 | 
						|
			}
 | 
						|
			if !fit {
 | 
						|
				fits = false
 | 
						|
				break
 | 
						|
			}
 | 
						|
		}
 | 
						|
		if fits {
 | 
						|
			filtered = append(filtered, node)
 | 
						|
		} else {
 | 
						|
			failedNodesMap[node.Name] = fmt.Sprintf("extender failed: %s", e.name)
 | 
						|
		}
 | 
						|
	}
 | 
						|
	return &v1.NodeList{Items: filtered}, failedNodesMap, nil
 | 
						|
}
 | 
						|
 | 
						|
func (e *Extender) Prioritize(pod *v1.Pod, nodes *v1.NodeList) (*schedulerapi.HostPriorityList, error) {
 | 
						|
	result := schedulerapi.HostPriorityList{}
 | 
						|
	combinedScores := map[string]int{}
 | 
						|
	for _, prioritizer := range e.prioritizers {
 | 
						|
		weight := prioritizer.weight
 | 
						|
		if weight == 0 {
 | 
						|
			continue
 | 
						|
		}
 | 
						|
		priorityFunc := prioritizer.function
 | 
						|
		prioritizedList, err := priorityFunc(pod, nodes)
 | 
						|
		if err != nil {
 | 
						|
			return &schedulerapi.HostPriorityList{}, err
 | 
						|
		}
 | 
						|
		for _, hostEntry := range *prioritizedList {
 | 
						|
			combinedScores[hostEntry.Host] += hostEntry.Score * weight
 | 
						|
		}
 | 
						|
	}
 | 
						|
	for host, score := range combinedScores {
 | 
						|
		result = append(result, schedulerapi.HostPriority{Host: host, Score: score})
 | 
						|
	}
 | 
						|
	return &result, nil
 | 
						|
}
 | 
						|
 | 
						|
func machine_1_2_3_Predicate(pod *v1.Pod, node *v1.Node) (bool, error) {
 | 
						|
	if node.Name == "machine1" || node.Name == "machine2" || node.Name == "machine3" {
 | 
						|
		return true, nil
 | 
						|
	}
 | 
						|
	return false, nil
 | 
						|
}
 | 
						|
 | 
						|
func machine_2_3_5_Predicate(pod *v1.Pod, node *v1.Node) (bool, error) {
 | 
						|
	if node.Name == "machine2" || node.Name == "machine3" || node.Name == "machine5" {
 | 
						|
		return true, nil
 | 
						|
	}
 | 
						|
	return false, nil
 | 
						|
}
 | 
						|
 | 
						|
func machine_2_Prioritizer(pod *v1.Pod, nodes *v1.NodeList) (*schedulerapi.HostPriorityList, error) {
 | 
						|
	result := schedulerapi.HostPriorityList{}
 | 
						|
	for _, node := range nodes.Items {
 | 
						|
		score := 1
 | 
						|
		if node.Name == "machine2" {
 | 
						|
			score = 10
 | 
						|
		}
 | 
						|
		result = append(result, schedulerapi.HostPriority{node.Name, score})
 | 
						|
	}
 | 
						|
	return &result, nil
 | 
						|
}
 | 
						|
 | 
						|
func machine_3_Prioritizer(pod *v1.Pod, nodes *v1.NodeList) (*schedulerapi.HostPriorityList, error) {
 | 
						|
	result := schedulerapi.HostPriorityList{}
 | 
						|
	for _, node := range nodes.Items {
 | 
						|
		score := 1
 | 
						|
		if node.Name == "machine3" {
 | 
						|
			score = 10
 | 
						|
		}
 | 
						|
		result = append(result, schedulerapi.HostPriority{node.Name, score})
 | 
						|
	}
 | 
						|
	return &result, nil
 | 
						|
}
 | 
						|
 | 
						|
func TestSchedulerExtender(t *testing.T) {
 | 
						|
	_, s := framework.RunAMaster(nil)
 | 
						|
	defer s.Close()
 | 
						|
 | 
						|
	ns := framework.CreateTestingNamespace("scheduler-extender", s, t)
 | 
						|
	defer framework.DeleteTestingNamespace(ns, s, t)
 | 
						|
 | 
						|
	clientSet := clientset.NewForConfigOrDie(&restclient.Config{Host: s.URL, ContentConfig: restclient.ContentConfig{GroupVersion: &api.Registry.GroupOrDie(v1.GroupName).GroupVersion}})
 | 
						|
 | 
						|
	extender1 := &Extender{
 | 
						|
		name:         "extender1",
 | 
						|
		predicates:   []fitPredicate{machine_1_2_3_Predicate},
 | 
						|
		prioritizers: []priorityConfig{{machine_2_Prioritizer, 1}},
 | 
						|
	}
 | 
						|
	es1 := httptest.NewServer(http.HandlerFunc(func(w http.ResponseWriter, req *http.Request) {
 | 
						|
		extender1.serveHTTP(t, w, req)
 | 
						|
	}))
 | 
						|
	defer es1.Close()
 | 
						|
 | 
						|
	extender2 := &Extender{
 | 
						|
		name:         "extender2",
 | 
						|
		predicates:   []fitPredicate{machine_2_3_5_Predicate},
 | 
						|
		prioritizers: []priorityConfig{{machine_3_Prioritizer, 1}},
 | 
						|
	}
 | 
						|
	es2 := httptest.NewServer(http.HandlerFunc(func(w http.ResponseWriter, req *http.Request) {
 | 
						|
		extender2.serveHTTP(t, w, req)
 | 
						|
	}))
 | 
						|
	defer es2.Close()
 | 
						|
 | 
						|
	policy := schedulerapi.Policy{
 | 
						|
		ExtenderConfigs: []schedulerapi.ExtenderConfig{
 | 
						|
			{
 | 
						|
				URLPrefix:      es1.URL,
 | 
						|
				FilterVerb:     filter,
 | 
						|
				PrioritizeVerb: prioritize,
 | 
						|
				Weight:         3,
 | 
						|
				EnableHttps:    false,
 | 
						|
			},
 | 
						|
			{
 | 
						|
				URLPrefix:      es2.URL,
 | 
						|
				FilterVerb:     filter,
 | 
						|
				PrioritizeVerb: prioritize,
 | 
						|
				Weight:         4,
 | 
						|
				EnableHttps:    false,
 | 
						|
			},
 | 
						|
		},
 | 
						|
	}
 | 
						|
	policy.APIVersion = api.Registry.GroupOrDie(v1.GroupName).GroupVersion.String()
 | 
						|
 | 
						|
	schedulerConfigFactory := factory.NewConfigFactory(clientSet, v1.DefaultSchedulerName, v1.DefaultHardPodAffinitySymmetricWeight)
 | 
						|
	schedulerConfig, err := schedulerConfigFactory.CreateFromConfig(policy)
 | 
						|
	if err != nil {
 | 
						|
		t.Fatalf("Couldn't create scheduler config: %v", err)
 | 
						|
	}
 | 
						|
	eventBroadcaster := record.NewBroadcaster()
 | 
						|
	schedulerConfig.Recorder = eventBroadcaster.NewRecorder(api.Scheme, clientv1.EventSource{Component: v1.DefaultSchedulerName})
 | 
						|
	eventBroadcaster.StartRecordingToSink(&v1core.EventSinkImpl{Interface: v1core.New(clientSet.Core().RESTClient()).Events("")})
 | 
						|
	scheduler.New(schedulerConfig).Run()
 | 
						|
 | 
						|
	defer close(schedulerConfig.StopEverything)
 | 
						|
 | 
						|
	DoTestPodScheduling(ns, t, clientSet)
 | 
						|
}
 | 
						|
 | 
						|
func DoTestPodScheduling(ns *v1.Namespace, t *testing.T, cs clientset.Interface) {
 | 
						|
	// NOTE: This test cannot run in parallel, because it is creating and deleting
 | 
						|
	// non-namespaced objects (Nodes).
 | 
						|
	defer cs.Core().Nodes().DeleteCollection(nil, metav1.ListOptions{})
 | 
						|
 | 
						|
	goodCondition := v1.NodeCondition{
 | 
						|
		Type:              v1.NodeReady,
 | 
						|
		Status:            v1.ConditionTrue,
 | 
						|
		Reason:            fmt.Sprintf("schedulable condition"),
 | 
						|
		LastHeartbeatTime: metav1.Time{time.Now()},
 | 
						|
	}
 | 
						|
	node := &v1.Node{
 | 
						|
		Spec: v1.NodeSpec{Unschedulable: false},
 | 
						|
		Status: v1.NodeStatus{
 | 
						|
			Capacity: v1.ResourceList{
 | 
						|
				v1.ResourcePods: *resource.NewQuantity(32, resource.DecimalSI),
 | 
						|
			},
 | 
						|
			Conditions: []v1.NodeCondition{goodCondition},
 | 
						|
		},
 | 
						|
	}
 | 
						|
 | 
						|
	for ii := 0; ii < 5; ii++ {
 | 
						|
		node.Name = fmt.Sprintf("machine%d", ii+1)
 | 
						|
		if _, err := cs.Core().Nodes().Create(node); err != nil {
 | 
						|
			t.Fatalf("Failed to create nodes: %v", err)
 | 
						|
		}
 | 
						|
	}
 | 
						|
 | 
						|
	pod := &v1.Pod{
 | 
						|
		ObjectMeta: metav1.ObjectMeta{Name: "extender-test-pod"},
 | 
						|
		Spec: v1.PodSpec{
 | 
						|
			Containers: []v1.Container{{Name: "container", Image: e2e.GetPauseImageName(cs)}},
 | 
						|
		},
 | 
						|
	}
 | 
						|
 | 
						|
	myPod, err := cs.Core().Pods(ns.Name).Create(pod)
 | 
						|
	if err != nil {
 | 
						|
		t.Fatalf("Failed to create pod: %v", err)
 | 
						|
	}
 | 
						|
 | 
						|
	err = wait.Poll(time.Second, wait.ForeverTestTimeout, podScheduled(cs, myPod.Namespace, myPod.Name))
 | 
						|
	if err != nil {
 | 
						|
		t.Fatalf("Failed to schedule pod: %v", err)
 | 
						|
	}
 | 
						|
 | 
						|
	if myPod, err := cs.Core().Pods(ns.Name).Get(myPod.Name, metav1.GetOptions{}); err != nil {
 | 
						|
		t.Fatalf("Failed to get pod: %v", err)
 | 
						|
	} else if myPod.Spec.NodeName != "machine3" {
 | 
						|
		t.Fatalf("Failed to schedule using extender, expected machine3, got %v", myPod.Spec.NodeName)
 | 
						|
	}
 | 
						|
	t.Logf("Scheduled pod using extenders")
 | 
						|
}
 |