blob: 03be33cb26769deee699d5cbca97571b951920db [file] [log] [blame]
// Copyright 2020 The Monogon Project Authors.
//
// SPDX-License-Identifier: Apache-2.0
//
// Licensed under the Apache License, Version 2.0 (the "License");
// you may not use this file except in compliance with the License.
// You may obtain a copy of the License at
//
// http://www.apache.org/licenses/LICENSE-2.0
//
// Unless required by applicable law or agreed to in writing, software
// distributed under the License is distributed on an "AS IS" BASIS,
// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
// See the License for the specific language governing permissions and
// limitations under the License.
package kubernetes
import (
"context"
"fmt"
"net"
"time"
"google.golang.org/grpc/codes"
"google.golang.org/grpc/status"
"k8s.io/client-go/informers"
"k8s.io/client-go/kubernetes"
"k8s.io/client-go/tools/clientcmd"
"source.monogon.dev/metropolis/node/core/identity"
"source.monogon.dev/metropolis/node/core/localstorage"
"source.monogon.dev/metropolis/node/core/network"
"source.monogon.dev/metropolis/node/core/network/dns"
"source.monogon.dev/metropolis/node/kubernetes/authproxy"
"source.monogon.dev/metropolis/node/kubernetes/clusternet"
"source.monogon.dev/metropolis/node/kubernetes/nfproxy"
"source.monogon.dev/metropolis/node/kubernetes/pki"
"source.monogon.dev/metropolis/node/kubernetes/plugins/kvmdevice"
"source.monogon.dev/metropolis/node/kubernetes/reconciler"
"source.monogon.dev/metropolis/pkg/supervisor"
apb "source.monogon.dev/metropolis/proto/api"
)
type Config struct {
ServiceIPRange net.IPNet
ClusterNet net.IPNet
KPKI *pki.PKI
Root *localstorage.Root
Network *network.Service
Node *identity.Node
}
type Service struct {
c Config
}
func New(c Config) *Service {
s := &Service{
c: c,
}
return s
}
func (s *Service) Run(ctx context.Context) error {
controllerManagerConfig, err := getPKIControllerManagerConfig(ctx, s.c.KPKI)
if err != nil {
return fmt.Errorf("could not generate controller manager pki config: %w", err)
}
controllerManagerConfig.clusterNet = s.c.ClusterNet
schedulerConfig, err := getPKISchedulerConfig(ctx, s.c.KPKI)
if err != nil {
return fmt.Errorf("could not generate scheduler pki config: %w", err)
}
masterKubeconfig, err := s.c.KPKI.Kubeconfig(ctx, pki.Master)
if err != nil {
return fmt.Errorf("could not generate master kubeconfig: %w", err)
}
rawClientConfig, err := clientcmd.NewClientConfigFromBytes(masterKubeconfig)
if err != nil {
return fmt.Errorf("could not generate kubernetes client config: %w", err)
}
clientConfig, err := rawClientConfig.ClientConfig()
clientSet, err := kubernetes.NewForConfig(clientConfig)
if err != nil {
return fmt.Errorf("could not generate kubernetes client: %w", err)
}
informerFactory := informers.NewSharedInformerFactory(clientSet, 5*time.Minute)
// Sub-runnable which starts all parts of Kubernetes that depend on the
// machine's external IP address. If it changes, the runnable will exit.
// TODO(q3k): test this
supervisor.Run(ctx, "networked", func(ctx context.Context) error {
networkWatch := s.c.Network.Watch()
defer networkWatch.Close()
var status *network.Status
supervisor.Logger(ctx).Info("Waiting for node networking...")
for status == nil || status.ExternalAddress == nil {
status, err = networkWatch.Get(ctx)
if err != nil {
return fmt.Errorf("failed to get network status: %w", err)
}
}
address := status.ExternalAddress
supervisor.Logger(ctx).Info("Node has active networking, starting apiserver/kubelet")
apiserver := &apiserverService{
KPKI: s.c.KPKI,
AdvertiseAddress: address,
ServiceIPRange: s.c.ServiceIPRange,
EphemeralConsensusDirectory: &s.c.Root.Ephemeral.Consensus,
}
kubelet := kubeletService{
NodeName: s.c.Node.ID(),
ClusterDNS: []net.IP{address},
KubeletDirectory: &s.c.Root.Data.Kubernetes.Kubelet,
EphemeralDirectory: &s.c.Root.Ephemeral,
KPKI: s.c.KPKI,
}
err := supervisor.RunGroup(ctx, map[string]supervisor.Runnable{
"apiserver": apiserver.Run,
"kubelet": kubelet.Run,
})
if err != nil {
return fmt.Errorf("when starting apiserver/kubelet: %w", err)
}
supervisor.Signal(ctx, supervisor.SignalHealthy)
for status.ExternalAddress.Equal(address) {
status, err = networkWatch.Get(ctx)
if err != nil {
return fmt.Errorf("when watching for network changes: %w", err)
}
}
return fmt.Errorf("network configuration changed (%s -> %s)", address.String(), status.ExternalAddress.String())
})
csiPlugin := csiPluginServer{
KubeletDirectory: &s.c.Root.Data.Kubernetes.Kubelet,
VolumesDirectory: &s.c.Root.Data.Volumes,
}
csiProvisioner := csiProvisionerServer{
NodeName: s.c.Node.ID(),
Kubernetes: clientSet,
InformerFactory: informerFactory,
VolumesDirectory: &s.c.Root.Data.Volumes,
}
clusternet := clusternet.Service{
NodeName: s.c.Node.ID(),
Kubernetes: clientSet,
ClusterNet: s.c.ClusterNet,
InformerFactory: informerFactory,
DataDirectory: &s.c.Root.Data.Kubernetes.ClusterNetworking,
}
nfproxy := nfproxy.Service{
ClusterCIDR: s.c.ClusterNet,
ClientSet: clientSet,
}
kvmDevicePlugin := kvmdevice.Plugin{
KubeletDirectory: &s.c.Root.Data.Kubernetes.Kubelet,
}
authProxy := authproxy.Service{
KPKI: s.c.KPKI,
Node: s.c.Node,
}
for _, sub := range []struct {
name string
runnable supervisor.Runnable
}{
{"controller-manager", runControllerManager(*controllerManagerConfig)},
{"scheduler", runScheduler(*schedulerConfig)},
{"reconciler", reconciler.Run(clientSet)},
{"csi-plugin", csiPlugin.Run},
{"csi-provisioner", csiProvisioner.Run},
{"clusternet", clusternet.Run},
{"nfproxy", nfproxy.Run},
{"kvmdeviceplugin", kvmDevicePlugin.Run},
{"authproxy", authProxy.Run},
} {
err := supervisor.Run(ctx, sub.name, sub.runnable)
if err != nil {
return fmt.Errorf("could not run sub-service %q: %w", sub.name, err)
}
}
supervisor.Logger(ctx).Info("Registering K8s CoreDNS")
clusterDNSDirective := dns.NewKubernetesDirective("cluster.local", masterKubeconfig)
s.c.Network.ConfigureDNS(clusterDNSDirective)
supervisor.Signal(ctx, supervisor.SignalHealthy)
<-ctx.Done()
s.c.Network.ConfigureDNS(dns.CancelDirective(clusterDNSDirective))
return nil
}
// GetDebugKubeconfig issues a kubeconfig for an arbitrary given identity.
// Useful for debugging and testing.
func (s *Service) GetDebugKubeconfig(ctx context.Context, request *apb.GetDebugKubeconfigRequest) (*apb.GetDebugKubeconfigResponse, error) {
client, err := s.c.KPKI.VolatileClient(ctx, request.Id, request.Groups)
if err != nil {
return nil, status.Errorf(codes.Unavailable, "Failed to get volatile client certificate: %v", err)
}
kubeconfig, err := pki.Kubeconfig(ctx, s.c.KPKI.KV, client)
if err != nil {
return nil, status.Errorf(codes.Unavailable, "Failed to generate kubeconfig: %v", err)
}
return &apb.GetDebugKubeconfigResponse{DebugKubeconfig: string(kubeconfig)}, nil
}