/* Copyright © 2020 The k3d Author(s) Permission is hereby granted, free of charge, to any person obtaining a copy of this software and associated documentation files (the "Software"), to deal in the Software without restriction, including without limitation the rights to use, copy, modify, merge, publish, distribute, sublicense, and/or sell copies of the Software, and to permit persons to whom the Software is furnished to do so, subject to the following conditions: The above copyright notice and this permission notice shall be included in all copies or substantial portions of the Software. THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE SOFTWARE. */ package cluster import ( "bytes" "context" "errors" "fmt" "sort" "strconv" "strings" "time" "github.com/imdario/mergo" k3drt "github.com/rancher/k3d/v3/pkg/runtimes" "github.com/rancher/k3d/v3/pkg/types" k3d "github.com/rancher/k3d/v3/pkg/types" "github.com/rancher/k3d/v3/pkg/util" "github.com/rancher/k3d/v3/version" log "github.com/sirupsen/logrus" "golang.org/x/sync/errgroup" ) // ClusterCreate creates a new cluster consisting of // - some containerized k3s nodes // - a docker network func ClusterCreate(ctx context.Context, runtime k3drt.Runtime, cluster *k3d.Cluster) error { clusterCreateCtx := ctx clusterPrepCtx := ctx if cluster.CreateClusterOpts.Timeout > 0*time.Second { var cancelClusterCreateCtx context.CancelFunc var cancelClusterPrepCtx context.CancelFunc clusterCreateCtx, cancelClusterCreateCtx = context.WithTimeout(ctx, cluster.CreateClusterOpts.Timeout) clusterPrepCtx, cancelClusterPrepCtx = context.WithTimeout(ctx, cluster.CreateClusterOpts.Timeout) defer cancelClusterCreateCtx() defer cancelClusterPrepCtx() } /* * Network */ // error out if external cluster network should be used but no name was set if cluster.Network.Name == "" && cluster.Network.External { return fmt.Errorf("Failed to use external network because no name was specified") } // generate cluster network name, if not set if cluster.Network.Name == "" && !cluster.Network.External { cluster.Network.Name = fmt.Sprintf("%s-%s", k3d.DefaultObjectNamePrefix, cluster.Name) } // handle hostnetwork useHostNet := false if cluster.Network.Name == "host" { useHostNet = true if len(cluster.Nodes) > 1 { return fmt.Errorf("Only one server node supported when using host network") } } // create cluster network or use an existing one networkID, networkExists, err := runtime.CreateNetworkIfNotPresent(clusterCreateCtx, cluster.Network.Name) if err != nil { log.Errorln("Failed to create cluster network") return err } cluster.Network.Name = networkID extraLabels := map[string]string{ k3d.LabelNetwork: networkID, k3d.LabelNetworkExternal: strconv.FormatBool(cluster.Network.External), } if networkExists { extraLabels[k3d.LabelNetworkExternal] = "true" // if the network wasn't created, we say that it's managed externally (important for cluster deletion) } /* * Cluster Token */ if cluster.Token == "" { cluster.Token = GenerateClusterToken() } /* * Cluster-Wide volumes * - image volume (for importing images) */ if !cluster.CreateClusterOpts.DisableImageVolume { imageVolumeName := fmt.Sprintf("%s-%s-images", k3d.DefaultObjectNamePrefix, cluster.Name) if err := runtime.CreateVolume(clusterCreateCtx, imageVolumeName, map[string]string{k3d.LabelClusterName: cluster.Name}); err != nil { log.Errorf("Failed to create image volume '%s' for cluster '%s'", imageVolumeName, cluster.Name) return err } extraLabels[k3d.LabelImageVolume] = imageVolumeName // attach volume to nodes for _, node := range cluster.Nodes { node.Volumes = append(node.Volumes, fmt.Sprintf("%s:%s", imageVolumeName, k3d.DefaultImageVolumeMountPath)) } } /* * Nodes */ // agent defaults (per cluster) // connection url is always the name of the first server node (index 0) connectionURL := fmt.Sprintf("https://%s:%s", generateNodeName(cluster.Name, k3d.ServerRole, 0), k3d.DefaultAPIPort) nodeSetup := func(node *k3d.Node, suffix int) error { // cluster specific settings if node.Labels == nil { node.Labels = make(map[string]string) // TODO: maybe create an init function? } node.Labels[k3d.LabelClusterName] = cluster.Name node.Env = append(node.Env, fmt.Sprintf("K3S_TOKEN=%s", cluster.Token)) node.Labels[k3d.LabelClusterToken] = cluster.Token node.Labels[k3d.LabelClusterURL] = connectionURL // append extra labels for k, v := range extraLabels { node.Labels[k] = v } // node role specific settings if node.Role == k3d.ServerRole { node.ServerOpts.ExposeAPI = cluster.ExposeAPI // the cluster has an init server node, but its not this one, so connect it to the init node if cluster.InitNode != nil && !node.ServerOpts.IsInit { node.Env = append(node.Env, fmt.Sprintf("K3S_URL=%s", connectionURL)) } } else if node.Role == k3d.AgentRole { node.Env = append(node.Env, fmt.Sprintf("K3S_URL=%s", connectionURL)) } node.Name = generateNodeName(cluster.Name, node.Role, suffix) node.Network = cluster.Network.Name // create node log.Infof("Creating node '%s'", node.Name) if err := NodeCreate(clusterCreateCtx, runtime, node, k3d.NodeCreateOpts{}); err != nil { log.Errorln("Failed to create node") return err } log.Debugf("Created node '%s'", node.Name) return err } // used for node suffices serverCount := 0 agentCount := 0 suffix := 0 // create init node first if cluster.InitNode != nil { log.Infoln("Creating initializing server node") cluster.InitNode.Args = append(cluster.InitNode.Args, "--cluster-init") // in case the LoadBalancer was disabled, expose the API Port on the initializing server node if cluster.CreateClusterOpts.DisableLoadBalancer { cluster.InitNode.Ports = append(cluster.InitNode.Ports, fmt.Sprintf("%s:%s:%s/tcp", cluster.ExposeAPI.Host, cluster.ExposeAPI.Port, k3d.DefaultAPIPort)) } if err := nodeSetup(cluster.InitNode, serverCount); err != nil { return err } serverCount++ // wait for the initnode to come up before doing anything else for { select { case <-clusterCreateCtx.Done(): log.Errorln("Failed to bring up initializing server node in time") return fmt.Errorf(">>> %w", clusterCreateCtx.Err()) default: } log.Debugln("Waiting for initializing server node...") logreader, err := runtime.GetNodeLogs(clusterCreateCtx, cluster.InitNode, time.Time{}) if err != nil { if logreader != nil { logreader.Close() } log.Errorln(err) log.Errorln("Failed to get logs from the initializig server node.. waiting for 3 seconds instead") time.Sleep(3 * time.Second) break } defer logreader.Close() buf := new(bytes.Buffer) nRead, _ := buf.ReadFrom(logreader) logreader.Close() if nRead > 0 && strings.Contains(buf.String(), k3d.ReadyLogMessageByRole[k3d.ServerRole]) { log.Debugln("Initializing server node is up... continuing") break } time.Sleep(time.Second) } } // vars to support waiting for server nodes to be ready waitForServerWaitgroup, clusterCreateCtx := errgroup.WithContext(clusterCreateCtx) // create all other nodes, but skip the init node for _, node := range cluster.Nodes { if node.Role == k3d.ServerRole { // skip the init node here if node == cluster.InitNode { continue } else if serverCount == 0 && cluster.CreateClusterOpts.DisableLoadBalancer { // if this is the first server node and the server loadbalancer is disabled, expose the API Port on this server node node.Ports = append(node.Ports, fmt.Sprintf("%s:%s:%s/tcp", cluster.ExposeAPI.Host, cluster.ExposeAPI.Port, k3d.DefaultAPIPort)) } time.Sleep(1 * time.Second) // FIXME: arbitrary wait for one second to avoid race conditions of servers registering // name suffix suffix = serverCount serverCount++ } else if node.Role == k3d.AgentRole { // name suffix suffix = agentCount agentCount++ } if node.Role == k3d.ServerRole || node.Role == k3d.AgentRole { if err := nodeSetup(node, suffix); err != nil { return err } } // asynchronously wait for this server node to be ready (by checking the logs for a specific log mesage) if node.Role == k3d.ServerRole && cluster.CreateClusterOpts.WaitForServer { log.Debugf("Waiting for server node '%s' to get ready", node.Name) if err := NodeWaitForLogMessage(clusterCreateCtx, runtime, node, k3d.ReadyLogMessageByRole[k3d.ServerRole], time.Time{}); err != nil { return fmt.Errorf("Server node '%s' failed to get ready: %+v", node.Name, err) } } } /* * Auxiliary Containers */ // *** ServerLoadBalancer *** if !cluster.CreateClusterOpts.DisableLoadBalancer { if !useHostNet { // serverlb not supported in hostnetwork mode due to port collisions with server node // Generate a comma-separated list of server/server names to pass to the LB container servers := "" for _, node := range cluster.Nodes { if node.Role == k3d.ServerRole { if servers == "" { servers = node.Name } else { servers = fmt.Sprintf("%s,%s", servers, node.Name) } } } // generate comma-separated list of extra ports to forward ports := k3d.DefaultAPIPort for _, portString := range cluster.ServerLoadBalancer.Ports { split := strings.Split(portString, ":") port := split[len(split)-1] if strings.Contains(port, "-") { split := strings.Split(port, "-") start, err := strconv.Atoi(split[0]) if err != nil { log.Errorf("Failed to parse port mapping for loadbalancer '%s'", port) return err } end, err := strconv.Atoi(split[1]) if err != nil { log.Errorf("Failed to parse port mapping for loadbalancer '%s'", port) return err } for i := start; i <= end; i++ { ports += "," + strconv.Itoa(i) } } else { ports += "," + port } } // Create LB as a modified node with loadbalancerRole lbNode := &k3d.Node{ Name: fmt.Sprintf("%s-%s-serverlb", k3d.DefaultObjectNamePrefix, cluster.Name), Image: fmt.Sprintf("%s:%s", k3d.DefaultLBImageRepo, version.GetHelperImageVersion()), Ports: append(cluster.ServerLoadBalancer.Ports, fmt.Sprintf("%s:%s:%s/tcp", cluster.ExposeAPI.Host, cluster.ExposeAPI.Port, k3d.DefaultAPIPort)), Env: []string{ fmt.Sprintf("SERVERS=%s", servers), fmt.Sprintf("PORTS=%s", ports), fmt.Sprintf("WORKER_PROCESSES=%d", len(strings.Split(ports, ","))), }, Role: k3d.LoadBalancerRole, Labels: k3d.DefaultObjectLabels, // TODO: createLoadBalancer: add more expressive labels Network: cluster.Network.Name, } cluster.Nodes = append(cluster.Nodes, lbNode) // append lbNode to list of cluster nodes, so it will be considered during rollback log.Infof("Creating LoadBalancer '%s'", lbNode.Name) if err := NodeCreate(clusterCreateCtx, runtime, lbNode, k3d.NodeCreateOpts{}); err != nil { log.Errorln("Failed to create loadbalancer") return err } if cluster.CreateClusterOpts.WaitForServer { waitForServerWaitgroup.Go(func() error { // TODO: avoid `level=fatal msg="starting kubernetes: preparing server: post join: a configuration change is already in progress (5)"` // ... by scanning for this line in logs and restarting the container in case it appears log.Debugf("Starting to wait for loadbalancer node '%s'", lbNode.Name) return NodeWaitForLogMessage(clusterCreateCtx, runtime, lbNode, k3d.ReadyLogMessageByRole[k3d.LoadBalancerRole], time.Time{}) }) } } else { log.Infoln("Hostnetwork selected -> Skipping creation of server LoadBalancer") } } if err := waitForServerWaitgroup.Wait(); err != nil { log.Errorln("Failed to bring up all server nodes (and loadbalancer) in time. Check the logs:") log.Errorf(">>> %+v", err) return fmt.Errorf("Failed to bring up cluster") } /********************************** * Additional Cluster Preparation * **********************************/ /* * Networking Magic */ // add /etc/hosts and CoreDNS entry for host.k3d.internal, referring to the host system if !cluster.CreateClusterOpts.PrepDisableHostIPInjection { prepInjectHostIP(clusterPrepCtx, runtime, cluster) } return nil } // ClusterDelete deletes an existing cluster func ClusterDelete(ctx context.Context, runtime k3drt.Runtime, cluster *k3d.Cluster) error { log.Infof("Deleting cluster '%s'", cluster.Name) log.Debugf("Cluster Details: %+v", cluster) failed := 0 for _, node := range cluster.Nodes { if err := runtime.DeleteNode(ctx, node); err != nil { log.Warningf("Failed to delete node '%s': Try to delete it manually", node.Name) failed++ continue } } // Delete the cluster network, if it was created for/by this cluster (and if it's not in use anymore) if cluster.Network.Name != "" { if !cluster.Network.External { log.Infof("Deleting cluster network '%s'", cluster.Network.Name) if err := runtime.DeleteNetwork(ctx, cluster.Network.Name); err != nil { if strings.HasSuffix(err.Error(), "active endpoints") { log.Warningf("Failed to delete cluster network '%s' because it's still in use: is there another cluster using it?", cluster.Network.Name) } else { log.Warningf("Failed to delete cluster network '%s': '%+v'", cluster.Network.Name, err) } } } else if cluster.Network.External { log.Debugf("Skip deletion of cluster network '%s' because it's managed externally", cluster.Network.Name) } } // delete image volume if cluster.ImageVolume != "" { log.Infof("Deleting image volume '%s'", cluster.ImageVolume) if err := runtime.DeleteVolume(ctx, cluster.ImageVolume); err != nil { log.Warningf("Failed to delete image volume '%s' of cluster '%s': Try to delete it manually", cluster.ImageVolume, cluster.Name) } } // return error if we failed to delete a node if failed > 0 { return fmt.Errorf("Failed to delete %d nodes: Try to delete them manually", failed) } return nil } // ClusterList returns a list of all existing clusters func ClusterList(ctx context.Context, runtime k3drt.Runtime) ([]*k3d.Cluster, error) { nodes, err := runtime.GetNodesByLabel(ctx, k3d.DefaultObjectLabels) if err != nil { log.Errorln("Failed to get clusters") return nil, err } clusters := []*k3d.Cluster{} // for each node, check, if we can add it to a cluster or add the cluster if it doesn't exist yet for _, node := range nodes { clusterExists := false for _, cluster := range clusters { if node.Labels[k3d.LabelClusterName] == cluster.Name { // TODO: handle case, where this label doesn't exist cluster.Nodes = append(cluster.Nodes, node) clusterExists = true break } } // cluster is not in the list yet, so we add it with the current node as its first member if !clusterExists { clusters = append(clusters, &k3d.Cluster{ Name: node.Labels[k3d.LabelClusterName], Nodes: []*k3d.Node{node}, }) } } // enrich cluster structs with label values for _, cluster := range clusters { if err := populateClusterFieldsFromLabels(cluster); err != nil { log.Warnf("Failed to populate cluster fields from node label values for cluster '%s'", cluster.Name) log.Warnln(err) } } return clusters, nil } // populateClusterFieldsFromLabels inspects labels attached to nodes and translates them to struct fields func populateClusterFieldsFromLabels(cluster *k3d.Cluster) error { networkExternalSet := false for _, node := range cluster.Nodes { // get the name of the cluster network if cluster.Network.Name == "" { if networkName, ok := node.Labels[k3d.LabelNetwork]; ok { cluster.Network.Name = networkName } } // check if the network is external // since the struct value is a bool, initialized as false, we cannot check if it's unset if !cluster.Network.External && !networkExternalSet { if networkExternalString, ok := node.Labels[k3d.LabelNetworkExternal]; ok { if networkExternal, err := strconv.ParseBool(networkExternalString); err == nil { cluster.Network.External = networkExternal networkExternalSet = true } } } // get image volume // TODO: enable external image volumes the same way we do it with networks if cluster.ImageVolume == "" { if imageVolumeName, ok := node.Labels[k3d.LabelImageVolume]; ok { cluster.ImageVolume = imageVolumeName } } // get k3s cluster's token if cluster.Token == "" { if token, ok := node.Labels[k3d.LabelClusterToken]; ok { cluster.Token = token } } } return nil } var ClusterGetNoNodesFoundError = errors.New("No nodes found for given cluster") // ClusterGet returns an existing cluster with all fields and node lists populated func ClusterGet(ctx context.Context, runtime k3drt.Runtime, cluster *k3d.Cluster) (*k3d.Cluster, error) { // get nodes that belong to the selected cluster nodes, err := runtime.GetNodesByLabel(ctx, map[string]string{k3d.LabelClusterName: cluster.Name}) if err != nil { log.Errorf("Failed to get nodes for cluster '%s'", cluster.Name) } if len(nodes) == 0 { return nil, ClusterGetNoNodesFoundError } // append nodes for _, node := range nodes { // check if there's already a node in the struct overwroteExisting := false for _, existingNode := range cluster.Nodes { // overwrite existing node if existingNode.Name == node.Name { mergo.MergeWithOverwrite(existingNode, node) overwroteExisting = true } } // no existing node overwritten: append new node if !overwroteExisting { cluster.Nodes = append(cluster.Nodes, node) } } if err := populateClusterFieldsFromLabels(cluster); err != nil { log.Warnf("Failed to populate cluster fields from node labels") log.Warnln(err) } return cluster, nil } // GenerateClusterToken generates a random 20 character string func GenerateClusterToken() string { return util.GenerateRandomString(20) } func generateNodeName(cluster string, role k3d.Role, suffix int) string { return fmt.Sprintf("%s-%s-%s-%d", k3d.DefaultObjectNamePrefix, cluster, role, suffix) } // ClusterStart starts a whole cluster (i.e. all nodes of the cluster) func ClusterStart(ctx context.Context, runtime k3drt.Runtime, cluster *k3d.Cluster, startClusterOpts types.ClusterStartOpts) error { log.Infof("Starting cluster '%s'", cluster.Name) start := time.Now() if startClusterOpts.Timeout > 0*time.Second { var cancel context.CancelFunc ctx, cancel = context.WithTimeout(ctx, startClusterOpts.Timeout) defer cancel() } // vars to support waiting for server nodes to be ready waitForServerWaitgroup, ctx := errgroup.WithContext(ctx) failed := 0 var serverlb *k3d.Node for _, node := range cluster.Nodes { // skip the LB, because we want to start it last if node.Role == k3d.LoadBalancerRole { serverlb = node continue } // check if node is running already to avoid waiting forever when checking for the node log message if !node.State.Running { // start node if err := runtime.StartNode(ctx, node); err != nil { log.Warningf("Failed to start node '%s': Try to start it manually", node.Name) failed++ continue } // asynchronously wait for this server node to be ready (by checking the logs for a specific log mesage) if node.Role == k3d.ServerRole && startClusterOpts.WaitForServer { serverNode := node waitForServerWaitgroup.Go(func() error { // TODO: avoid `level=fatal msg="starting kubernetes: preparing server: post join: a configuration change is already in progress (5)"` // ... by scanning for this line in logs and restarting the container in case it appears log.Debugf("Starting to wait for server node '%s'", serverNode.Name) return NodeWaitForLogMessage(ctx, runtime, serverNode, k3d.ReadyLogMessageByRole[k3d.ServerRole], start) }) } } else { log.Infof("Node '%s' already running", node.Name) } } // start serverlb if serverlb != nil { if !serverlb.State.Running { log.Debugln("Starting serverlb...") if err := runtime.StartNode(ctx, serverlb); err != nil { // FIXME: we could run into a nullpointer exception here log.Warningf("Failed to start serverlb '%s': Try to start it manually", serverlb.Name) failed++ } waitForServerWaitgroup.Go(func() error { // TODO: avoid `level=fatal msg="starting kubernetes: preparing server: post join: a configuration change is already in progress (5)"` // ... by scanning for this line in logs and restarting the container in case it appears log.Debugf("Starting to wait for loadbalancer node '%s'", serverlb.Name) return NodeWaitForLogMessage(ctx, runtime, serverlb, k3d.ReadyLogMessageByRole[k3d.LoadBalancerRole], start) }) } else { log.Infof("Serverlb '%s' already running", serverlb.Name) } } if err := waitForServerWaitgroup.Wait(); err != nil { log.Errorln("Failed to bring up all nodes in time. Check the logs:") log.Errorln(">>> ", err) return fmt.Errorf("Failed to bring up cluster") } if failed > 0 { return fmt.Errorf("Failed to start %d nodes: Try to start them manually", failed) } return nil } // ClusterStop stops a whole cluster (i.e. all nodes of the cluster) func ClusterStop(ctx context.Context, runtime k3drt.Runtime, cluster *k3d.Cluster) error { log.Infof("Stopping cluster '%s'", cluster.Name) failed := 0 for _, node := range cluster.Nodes { if err := runtime.StopNode(ctx, node); err != nil { log.Warningf("Failed to stop node '%s': Try to stop it manually", node.Name) failed++ continue } } if failed > 0 { return fmt.Errorf("Failed to stop %d nodes: Try to stop them manually", failed) } return nil } // SortClusters : in place sort cluster list by cluster name alphabetical order func SortClusters(clusters []*k3d.Cluster) []*k3d.Cluster { sort.Slice(clusters, func(i, j int) bool { return clusters[i].Name < clusters[j].Name }) return clusters } // prepInjectHostIP adds /etc/hosts and CoreDNS entry for host.k3d.internal, referring to the host system func prepInjectHostIP(ctx context.Context, runtime k3drt.Runtime, cluster *k3d.Cluster) { log.Infoln("(Optional) Trying to get IP of the docker host and inject it into the cluster as 'host.k3d.internal' for easy access") hostIP, err := GetHostIP(ctx, runtime, cluster) if err != nil { log.Warnf("Failed to get HostIP: %+v", err) } if hostIP != nil { hostRecordSuccessMessage := "" etcHostsFailureCount := 0 hostsEntry := fmt.Sprintf("%s %s", hostIP, k3d.DefaultK3dInternalHostRecord) log.Debugf("Adding extra host entry '%s'...", hostsEntry) for _, node := range cluster.Nodes { if err := runtime.ExecInNode(ctx, node, []string{"sh", "-c", fmt.Sprintf("echo '%s' >> /etc/hosts", hostsEntry)}); err != nil { log.Warnf("Failed to add extra entry '%s' to /etc/hosts in node '%s'", hostsEntry, node.Name) etcHostsFailureCount++ } } if etcHostsFailureCount < len(cluster.Nodes) { hostRecordSuccessMessage += fmt.Sprintf("Successfully added host record to /etc/hosts in %d/%d nodes", (len(cluster.Nodes) - etcHostsFailureCount), len(cluster.Nodes)) } patchCmd := `test=$(kubectl get cm coredns -n kube-system --template='{{.data.NodeHosts}}' | sed -n -E -e '/[0-9\.]{4,12}\s+host\.k3d\.internal$/!p' -e '$a` + hostsEntry + `' | tr '\n' '^' | busybox xargs -0 printf '{"data": {"NodeHosts":"%s"}}'| sed -E 's%\^%\\n%g') && kubectl patch cm coredns -n kube-system -p="$test"` if err = runtime.ExecInNode(ctx, cluster.Nodes[0], []string{"sh", "-c", patchCmd}); err != nil { log.Warnf("Failed to patch CoreDNS ConfigMap to include entry '%s': %+v", hostsEntry, err) } else { hostRecordSuccessMessage += " and to the CoreDNS ConfigMap" } if hostRecordSuccessMessage != "" { log.Infoln(hostRecordSuccessMessage) } } }