зеркало из https://github.com/docker/compose-cli.git
325 строки
8.5 KiB
Go
325 строки
8.5 KiB
Go
//go:build kube
|
|
// +build kube
|
|
|
|
/*
|
|
Copyright 2020 Docker Compose CLI authors
|
|
|
|
Licensed under the Apache License, Version 2.0 (the "License");
|
|
you may not use this file except in compliance with the License.
|
|
You may obtain a copy of the License at
|
|
|
|
http://www.apache.org/licenses/LICENSE-2.0
|
|
|
|
Unless required by applicable law or agreed to in writing, software
|
|
distributed under the License is distributed on an "AS IS" BASIS,
|
|
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
See the License for the specific language governing permissions and
|
|
limitations under the License.
|
|
*/
|
|
|
|
package client
|
|
|
|
import (
|
|
"context"
|
|
"fmt"
|
|
"io"
|
|
"net/http"
|
|
"os"
|
|
"strings"
|
|
"time"
|
|
|
|
"github.com/docker/compose/v2/pkg/api"
|
|
"github.com/docker/compose/v2/pkg/utils"
|
|
"golang.org/x/sync/errgroup"
|
|
corev1 "k8s.io/api/core/v1"
|
|
metav1 "k8s.io/apimachinery/pkg/apis/meta/v1"
|
|
"k8s.io/apimachinery/pkg/runtime"
|
|
"k8s.io/cli-runtime/pkg/genericclioptions"
|
|
"k8s.io/client-go/kubernetes"
|
|
"k8s.io/client-go/rest"
|
|
"k8s.io/client-go/tools/portforward"
|
|
"k8s.io/client-go/tools/remotecommand"
|
|
"k8s.io/client-go/transport/spdy"
|
|
)
|
|
|
|
// KubeClient API to access kube objects
|
|
type KubeClient struct {
|
|
client *kubernetes.Clientset
|
|
namespace string
|
|
config *rest.Config
|
|
ioStreams genericclioptions.IOStreams
|
|
}
|
|
|
|
// NewKubeClient new kubernetes client
|
|
func NewKubeClient(config genericclioptions.RESTClientGetter) (*KubeClient, error) {
|
|
restConfig, err := config.ToRESTConfig()
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
|
|
clientset, err := kubernetes.NewForConfig(restConfig)
|
|
if err != nil {
|
|
return nil, fmt.Errorf("failed creating clientset. Error: %+v", err)
|
|
}
|
|
|
|
namespace, _, err := config.ToRawKubeConfigLoader().Namespace()
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
|
|
return &KubeClient{
|
|
client: clientset,
|
|
namespace: namespace,
|
|
config: restConfig,
|
|
ioStreams: genericclioptions.IOStreams{In: os.Stdin, Out: os.Stdout, ErrOut: os.Stderr},
|
|
}, nil
|
|
}
|
|
|
|
// GetPod retrieves a service pod
|
|
func (kc KubeClient) GetPod(ctx context.Context, projectName, serviceName string) (*corev1.Pod, error) {
|
|
pods, err := kc.client.CoreV1().Pods(kc.namespace).List(ctx, metav1.ListOptions{
|
|
LabelSelector: fmt.Sprintf("%s=%s", api.ProjectLabel, projectName),
|
|
})
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
if pods == nil {
|
|
return nil, nil
|
|
}
|
|
|
|
for _, p := range pods.Items {
|
|
if p.Name == serviceName {
|
|
return &p, nil
|
|
}
|
|
}
|
|
|
|
for _, p := range pods.Items {
|
|
service := p.Labels[api.ServiceLabel]
|
|
if service == serviceName {
|
|
return &p, nil
|
|
}
|
|
}
|
|
|
|
return nil, nil
|
|
}
|
|
|
|
// Exec executes a command in a container
|
|
func (kc KubeClient) Exec(ctx context.Context, projectName string, opts api.RunOptions) error {
|
|
pod, err := kc.GetPod(ctx, projectName, opts.Service)
|
|
if err != nil || pod == nil {
|
|
return err
|
|
}
|
|
if len(pod.Spec.Containers) == 0 {
|
|
return fmt.Errorf("no containers running in pod %s", pod.Name)
|
|
}
|
|
// get first container in the pod
|
|
container := &pod.Spec.Containers[0]
|
|
containerName := container.Name
|
|
|
|
req := kc.client.CoreV1().RESTClient().Post().
|
|
Resource("pods").
|
|
Name(pod.Name).
|
|
Namespace(kc.namespace).
|
|
SubResource("exec")
|
|
|
|
option := &corev1.PodExecOptions{
|
|
Container: containerName,
|
|
Command: opts.Command,
|
|
Stdin: true,
|
|
Stdout: true,
|
|
Stderr: true,
|
|
TTY: opts.Tty,
|
|
}
|
|
|
|
if opts.Stdin == nil {
|
|
option.Stdin = false
|
|
}
|
|
|
|
scheme := runtime.NewScheme()
|
|
if err := corev1.AddToScheme(scheme); err != nil {
|
|
return fmt.Errorf("error adding to scheme: %v", err)
|
|
}
|
|
parameterCodec := runtime.NewParameterCodec(scheme)
|
|
req.VersionedParams(option, parameterCodec)
|
|
|
|
exec, err := remotecommand.NewSPDYExecutor(kc.config, "POST", req.URL())
|
|
if err != nil {
|
|
return err
|
|
}
|
|
return exec.Stream(remotecommand.StreamOptions{
|
|
Stdin: opts.Stdin,
|
|
Stdout: opts.Stdout,
|
|
Stderr: opts.Stdout,
|
|
Tty: opts.Tty,
|
|
})
|
|
}
|
|
|
|
// GetContainers get containers for a given compose project
|
|
func (kc KubeClient) GetContainers(ctx context.Context, projectName string, all bool) ([]api.ContainerSummary, error) {
|
|
fieldSelector := ""
|
|
if !all {
|
|
fieldSelector = "status.phase=Running"
|
|
}
|
|
|
|
pods, err := kc.client.CoreV1().Pods(kc.namespace).List(ctx, metav1.ListOptions{
|
|
LabelSelector: fmt.Sprintf("%s=%s", api.ProjectLabel, projectName),
|
|
FieldSelector: fieldSelector,
|
|
})
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
services := map[string][]api.PortPublisher{}
|
|
result := []api.ContainerSummary{}
|
|
for _, pod := range pods.Items {
|
|
summary := podToContainerSummary(pod)
|
|
serviceName := pod.GetObjectMeta().GetLabels()[api.ServiceLabel]
|
|
ports, ok := services[serviceName]
|
|
if !ok {
|
|
s, err := kc.client.CoreV1().Services(kc.namespace).Get(ctx, serviceName, metav1.GetOptions{})
|
|
if err != nil {
|
|
if !strings.Contains(err.Error(), "not found") {
|
|
return nil, err
|
|
}
|
|
result = append(result, summary)
|
|
continue
|
|
}
|
|
ports = []api.PortPublisher{}
|
|
if s != nil {
|
|
if s.Spec.Type == corev1.ServiceTypeLoadBalancer {
|
|
if len(s.Status.LoadBalancer.Ingress) > 0 {
|
|
port := api.PortPublisher{URL: s.Status.LoadBalancer.Ingress[0].IP}
|
|
if len(s.Spec.Ports) > 0 {
|
|
port.URL = fmt.Sprintf("%s:%d", port.URL, s.Spec.Ports[0].Port)
|
|
port.TargetPort = s.Spec.Ports[0].TargetPort.IntValue()
|
|
port.Protocol = string(s.Spec.Ports[0].Protocol)
|
|
}
|
|
ports = append(ports, port)
|
|
}
|
|
}
|
|
}
|
|
services[serviceName] = ports
|
|
}
|
|
summary.Publishers = ports
|
|
result = append(result, summary)
|
|
}
|
|
|
|
return result, nil
|
|
}
|
|
|
|
// GetLogs retrieves pod logs
|
|
func (kc *KubeClient) GetLogs(ctx context.Context, projectName string, consumer api.LogConsumer, follow bool) error {
|
|
pods, err := kc.client.CoreV1().Pods(kc.namespace).List(ctx, metav1.ListOptions{
|
|
LabelSelector: fmt.Sprintf("%s=%s", api.ProjectLabel, projectName),
|
|
})
|
|
if err != nil {
|
|
return err
|
|
}
|
|
eg, ctx := errgroup.WithContext(ctx)
|
|
for _, pod := range pods.Items {
|
|
podName := pod.Name
|
|
request := kc.client.CoreV1().Pods(kc.namespace).GetLogs(podName, &corev1.PodLogOptions{Follow: follow})
|
|
service := pod.Labels[api.ServiceLabel]
|
|
w := utils.GetWriter(func(line string) {
|
|
consumer.Log(podName, service, line)
|
|
})
|
|
|
|
eg.Go(func() error {
|
|
r, err := request.Stream(ctx)
|
|
if err != nil {
|
|
return err
|
|
}
|
|
|
|
defer r.Close() // nolint errcheck
|
|
_, err = io.Copy(w, r)
|
|
return err
|
|
})
|
|
}
|
|
return eg.Wait()
|
|
}
|
|
|
|
// WaitForPodState blocks until pods reach desired state
|
|
func (kc KubeClient) WaitForPodState(ctx context.Context, opts WaitForStatusOptions) error {
|
|
var timeout = time.Minute
|
|
if opts.Timeout != nil {
|
|
timeout = *opts.Timeout
|
|
}
|
|
|
|
errch := make(chan error, 1)
|
|
done := make(chan bool)
|
|
go func() {
|
|
for {
|
|
time.Sleep(500 * time.Millisecond)
|
|
|
|
pods, err := kc.client.CoreV1().Pods(kc.namespace).List(ctx, metav1.ListOptions{
|
|
LabelSelector: fmt.Sprintf("%s=%s", api.ProjectLabel, opts.ProjectName),
|
|
})
|
|
if err != nil {
|
|
errch <- err
|
|
}
|
|
stateReached, servicePods, err := checkPodsState(opts.Services, pods.Items, opts.Status)
|
|
if err != nil {
|
|
errch <- err
|
|
}
|
|
if opts.Log != nil {
|
|
for p, m := range servicePods {
|
|
opts.Log(p, stateReached, m)
|
|
}
|
|
}
|
|
|
|
if stateReached {
|
|
done <- true
|
|
}
|
|
}
|
|
}()
|
|
|
|
select {
|
|
case <-time.After(timeout):
|
|
return fmt.Errorf("timeout: pods did not reach expected state")
|
|
case err := <-errch:
|
|
if err != nil {
|
|
return err
|
|
}
|
|
case <-done:
|
|
return nil
|
|
}
|
|
return nil
|
|
}
|
|
|
|
// MapPortsToLocalhost runs a port-forwarder daemon process
|
|
func (kc KubeClient) MapPortsToLocalhost(ctx context.Context, opts PortMappingOptions) error {
|
|
stopChannel := make(chan struct{}, 1)
|
|
readyChannel := make(chan struct{})
|
|
|
|
eg, ctx := errgroup.WithContext(ctx)
|
|
for serviceName, servicePorts := range opts.Services {
|
|
serviceName, servicePorts := serviceName, servicePorts
|
|
pod, err := kc.GetPod(ctx, opts.ProjectName, serviceName)
|
|
if err != nil {
|
|
return err
|
|
}
|
|
eg.Go(func() error {
|
|
ports := []string{}
|
|
for _, p := range servicePorts {
|
|
ports = append(ports, fmt.Sprintf("%d:%d", p.PublishedPort, p.TargetPort))
|
|
}
|
|
|
|
req := kc.client.CoreV1().RESTClient().Post().
|
|
Resource("pods").
|
|
Name(pod.Name).
|
|
Namespace(kc.namespace).
|
|
SubResource("portforward")
|
|
transport, upgrader, err := spdy.RoundTripperFor(kc.config)
|
|
if err != nil {
|
|
return err
|
|
}
|
|
dialer := spdy.NewDialer(upgrader, &http.Client{Transport: transport}, "POST", req.URL())
|
|
fw, err := portforward.New(dialer, ports, stopChannel, readyChannel, os.Stdout, os.Stderr)
|
|
if err != nil {
|
|
return err
|
|
}
|
|
return fw.ForwardPorts()
|
|
})
|
|
}
|
|
return eg.Wait()
|
|
}
|