| // Copyright 2020 The Monogon Project Authors. | 
 | // | 
 | // SPDX-License-Identifier: Apache-2.0 | 
 | // | 
 | // Licensed under the Apache License, Version 2.0 (the "License"); | 
 | // you may not use this file except in compliance with the License. | 
 | // You may obtain a copy of the License at | 
 | // | 
 | //     http://www.apache.org/licenses/LICENSE-2.0 | 
 | // | 
 | // Unless required by applicable law or agreed to in writing, software | 
 | // distributed under the License is distributed on an "AS IS" BASIS, | 
 | // WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. | 
 | // See the License for the specific language governing permissions and | 
 | // limitations under the License. | 
 |  | 
 | package kubernetes | 
 |  | 
 | import ( | 
 | 	"context" | 
 | 	"fmt" | 
 | 	"net" | 
 | 	"time" | 
 |  | 
 | 	"google.golang.org/grpc/codes" | 
 | 	"google.golang.org/grpc/status" | 
 | 	"k8s.io/client-go/informers" | 
 | 	"k8s.io/client-go/kubernetes" | 
 | 	"k8s.io/client-go/tools/clientcmd" | 
 |  | 
 | 	"source.monogon.dev/metropolis/node/core/identity" | 
 | 	"source.monogon.dev/metropolis/node/core/localstorage" | 
 | 	"source.monogon.dev/metropolis/node/core/network" | 
 | 	"source.monogon.dev/metropolis/node/core/network/dns" | 
 | 	"source.monogon.dev/metropolis/node/kubernetes/authproxy" | 
 | 	"source.monogon.dev/metropolis/node/kubernetes/clusternet" | 
 | 	"source.monogon.dev/metropolis/node/kubernetes/nfproxy" | 
 | 	"source.monogon.dev/metropolis/node/kubernetes/pki" | 
 | 	"source.monogon.dev/metropolis/node/kubernetes/plugins/kvmdevice" | 
 | 	"source.monogon.dev/metropolis/node/kubernetes/reconciler" | 
 | 	"source.monogon.dev/metropolis/pkg/supervisor" | 
 | 	apb "source.monogon.dev/metropolis/proto/api" | 
 | ) | 
 |  | 
 | type Config struct { | 
 | 	ServiceIPRange net.IPNet | 
 | 	ClusterNet     net.IPNet | 
 | 	ClusterDomain  string | 
 |  | 
 | 	KPKI    *pki.PKI | 
 | 	Root    *localstorage.Root | 
 | 	Network *network.Service | 
 | 	Node    *identity.Node | 
 | } | 
 |  | 
 | type Service struct { | 
 | 	c Config | 
 | } | 
 |  | 
 | func New(c Config) *Service { | 
 | 	s := &Service{ | 
 | 		c: c, | 
 | 	} | 
 | 	return s | 
 | } | 
 |  | 
 | func (s *Service) Run(ctx context.Context) error { | 
 | 	controllerManagerConfig, err := getPKIControllerManagerConfig(ctx, s.c.KPKI) | 
 | 	if err != nil { | 
 | 		return fmt.Errorf("could not generate controller manager pki config: %w", err) | 
 | 	} | 
 | 	controllerManagerConfig.clusterNet = s.c.ClusterNet | 
 | 	schedulerConfig, err := getPKISchedulerConfig(ctx, s.c.KPKI) | 
 | 	if err != nil { | 
 | 		return fmt.Errorf("could not generate scheduler pki config: %w", err) | 
 | 	} | 
 |  | 
 | 	masterKubeconfig, err := s.c.KPKI.Kubeconfig(ctx, pki.Master) | 
 | 	if err != nil { | 
 | 		return fmt.Errorf("could not generate master kubeconfig: %w", err) | 
 | 	} | 
 |  | 
 | 	rawClientConfig, err := clientcmd.NewClientConfigFromBytes(masterKubeconfig) | 
 | 	if err != nil { | 
 | 		return fmt.Errorf("could not generate kubernetes client config: %w", err) | 
 | 	} | 
 |  | 
 | 	clientConfig, err := rawClientConfig.ClientConfig() | 
 | 	clientSet, err := kubernetes.NewForConfig(clientConfig) | 
 | 	if err != nil { | 
 | 		return fmt.Errorf("could not generate kubernetes client: %w", err) | 
 | 	} | 
 |  | 
 | 	informerFactory := informers.NewSharedInformerFactory(clientSet, 5*time.Minute) | 
 |  | 
 | 	// Sub-runnable which starts all parts of Kubernetes that depend on the | 
 | 	// machine's external IP address. If it changes, the runnable will exit. | 
 | 	// TODO(q3k): test this | 
 | 	startKubelet := make(chan struct{}) | 
 | 	supervisor.Run(ctx, "networked", func(ctx context.Context) error { | 
 | 		networkWatch := s.c.Network.Watch() | 
 | 		defer networkWatch.Close() | 
 |  | 
 | 		var status *network.Status | 
 |  | 
 | 		supervisor.Logger(ctx).Info("Waiting for node networking...") | 
 | 		for status == nil || status.ExternalAddress == nil { | 
 | 			status, err = networkWatch.Get(ctx) | 
 | 			if err != nil { | 
 | 				return fmt.Errorf("failed to get network status: %w", err) | 
 | 			} | 
 | 		} | 
 | 		address := status.ExternalAddress | 
 | 		supervisor.Logger(ctx).Info("Node has active networking, starting apiserver/kubelet") | 
 |  | 
 | 		apiserver := &apiserverService{ | 
 | 			KPKI:                        s.c.KPKI, | 
 | 			AdvertiseAddress:            address, | 
 | 			ServiceIPRange:              s.c.ServiceIPRange, | 
 | 			EphemeralConsensusDirectory: &s.c.Root.Ephemeral.Consensus, | 
 | 		} | 
 |  | 
 | 		kubelet := kubeletService{ | 
 | 			NodeName:           s.c.Node.ID(), | 
 | 			ClusterDNS:         []net.IP{address}, | 
 | 			ClusterDomain:      s.c.ClusterDomain, | 
 | 			KubeletDirectory:   &s.c.Root.Data.Kubernetes.Kubelet, | 
 | 			EphemeralDirectory: &s.c.Root.Ephemeral, | 
 | 			KPKI:               s.c.KPKI, | 
 | 		} | 
 |  | 
 | 		err := supervisor.RunGroup(ctx, map[string]supervisor.Runnable{ | 
 | 			"apiserver": apiserver.Run, | 
 | 			"kubelet": func(ctx context.Context) error { | 
 | 				<-startKubelet | 
 | 				return kubelet.Run(ctx) | 
 | 			}, | 
 | 		}) | 
 | 		if err != nil { | 
 | 			return fmt.Errorf("when starting apiserver/kubelet: %w", err) | 
 | 		} | 
 |  | 
 | 		supervisor.Signal(ctx, supervisor.SignalHealthy) | 
 |  | 
 | 		for status.ExternalAddress.Equal(address) { | 
 | 			status, err = networkWatch.Get(ctx) | 
 | 			if err != nil { | 
 | 				return fmt.Errorf("when watching for network changes: %w", err) | 
 | 			} | 
 | 		} | 
 | 		return fmt.Errorf("network configuration changed (%s -> %s)", address.String(), status.ExternalAddress.String()) | 
 | 	}) | 
 |  | 
 | 	// Before we start anything else, make sure reconciliation passes at least once. | 
 | 	// This makes the initial startup of a cluster much cleaner as we don't end up | 
 | 	// starting the scheduler/controller-manager/etc just to get them to immediately | 
 | 	// fail and back off with 'unauthorized'. | 
 | 	startLogging := time.Now().Add(2 * time.Second) | 
 | 	supervisor.Logger(ctx).Infof("Performing initial resource reconciliation...") | 
 | 	for { | 
 | 		err := reconciler.ReconcileAll(ctx, clientSet) | 
 | 		if err == nil { | 
 | 			supervisor.Logger(ctx).Infof("Initial resource reconciliation succeeded.") | 
 | 			close(startKubelet) | 
 | 			break | 
 | 		} | 
 | 		if time.Now().After(startLogging) { | 
 | 			supervisor.Logger(ctx).Errorf("Still couldn't do initial reconciliation: %v", err) | 
 | 		} | 
 | 		time.Sleep(100 * time.Millisecond) | 
 | 	} | 
 |  | 
 | 	csiPlugin := csiPluginServer{ | 
 | 		KubeletDirectory: &s.c.Root.Data.Kubernetes.Kubelet, | 
 | 		VolumesDirectory: &s.c.Root.Data.Volumes, | 
 | 	} | 
 |  | 
 | 	csiProvisioner := csiProvisionerServer{ | 
 | 		NodeName:         s.c.Node.ID(), | 
 | 		Kubernetes:       clientSet, | 
 | 		InformerFactory:  informerFactory, | 
 | 		VolumesDirectory: &s.c.Root.Data.Volumes, | 
 | 	} | 
 |  | 
 | 	clusternet := clusternet.Service{ | 
 | 		NodeName:        s.c.Node.ID(), | 
 | 		Kubernetes:      clientSet, | 
 | 		ClusterNet:      s.c.ClusterNet, | 
 | 		InformerFactory: informerFactory, | 
 | 		DataDirectory:   &s.c.Root.Data.Kubernetes.ClusterNetworking, | 
 | 	} | 
 |  | 
 | 	nfproxy := nfproxy.Service{ | 
 | 		ClusterCIDR: s.c.ClusterNet, | 
 | 		ClientSet:   clientSet, | 
 | 	} | 
 |  | 
 | 	kvmDevicePlugin := kvmdevice.Plugin{ | 
 | 		KubeletDirectory: &s.c.Root.Data.Kubernetes.Kubelet, | 
 | 	} | 
 |  | 
 | 	authProxy := authproxy.Service{ | 
 | 		KPKI: s.c.KPKI, | 
 | 		Node: s.c.Node, | 
 | 	} | 
 |  | 
 | 	for _, sub := range []struct { | 
 | 		name     string | 
 | 		runnable supervisor.Runnable | 
 | 	}{ | 
 | 		{"controller-manager", runControllerManager(*controllerManagerConfig)}, | 
 | 		{"scheduler", runScheduler(*schedulerConfig)}, | 
 | 		{"reconciler", reconciler.Maintain(clientSet)}, | 
 | 		{"csi-plugin", csiPlugin.Run}, | 
 | 		{"csi-provisioner", csiProvisioner.Run}, | 
 | 		{"clusternet", clusternet.Run}, | 
 | 		{"nfproxy", nfproxy.Run}, | 
 | 		{"kvmdeviceplugin", kvmDevicePlugin.Run}, | 
 | 		{"authproxy", authProxy.Run}, | 
 | 	} { | 
 | 		err := supervisor.Run(ctx, sub.name, sub.runnable) | 
 | 		if err != nil { | 
 | 			return fmt.Errorf("could not run sub-service %q: %w", sub.name, err) | 
 | 		} | 
 | 	} | 
 |  | 
 | 	supervisor.Logger(ctx).Info("Registering K8s CoreDNS") | 
 | 	clusterDNSDirective := dns.NewKubernetesDirective(s.c.ClusterDomain, masterKubeconfig) | 
 | 	s.c.Network.ConfigureDNS(clusterDNSDirective) | 
 |  | 
 | 	supervisor.Signal(ctx, supervisor.SignalHealthy) | 
 | 	<-ctx.Done() | 
 | 	s.c.Network.ConfigureDNS(dns.CancelDirective(clusterDNSDirective)) | 
 | 	return nil | 
 | } | 
 |  | 
 | // GetDebugKubeconfig issues a kubeconfig for an arbitrary given identity. | 
 | // Useful for debugging and testing. | 
 | func (s *Service) GetDebugKubeconfig(ctx context.Context, request *apb.GetDebugKubeconfigRequest) (*apb.GetDebugKubeconfigResponse, error) { | 
 | 	client, err := s.c.KPKI.VolatileClient(ctx, request.Id, request.Groups) | 
 | 	if err != nil { | 
 | 		return nil, status.Errorf(codes.Unavailable, "Failed to get volatile client certificate: %v", err) | 
 | 	} | 
 | 	kubeconfig, err := pki.Kubeconfig(ctx, s.c.KPKI.KV, client) | 
 | 	if err != nil { | 
 | 		return nil, status.Errorf(codes.Unavailable, "Failed to generate kubeconfig: %v", err) | 
 | 	} | 
 | 	return &apb.GetDebugKubeconfigResponse{DebugKubeconfig: string(kubeconfig)}, nil | 
 | } |