compose/local/compose.go

646 lines
17 KiB
Go
Raw Normal View History

// +build local
/*
Copyright 2020 Docker Compose CLI authors
Licensed under the Apache License, Version 2.0 (the "License");
you may not use this file except in compliance with the License.
You may obtain a copy of the License at
http://www.apache.org/licenses/LICENSE-2.0
Unless required by applicable law or agreed to in writing, software
distributed under the License is distributed on an "AS IS" BASIS,
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
See the License for the specific language governing permissions and
limitations under the License.
*/
package local
import (
"context"
"encoding/json"
"fmt"
"io"
"path/filepath"
"sort"
"strconv"
"strings"
"sync"
"github.com/compose-spec/compose-go/types"
moby "github.com/docker/docker/api/types"
"github.com/docker/docker/api/types/container"
"github.com/docker/docker/api/types/filters"
"github.com/docker/docker/api/types/mount"
"github.com/docker/docker/api/types/network"
"github.com/docker/docker/api/types/strslice"
"github.com/docker/docker/errdefs"
"github.com/docker/docker/pkg/jsonmessage"
"github.com/docker/go-connections/nat"
"github.com/pkg/errors"
"github.com/sanathkr/go-yaml"
"golang.org/x/sync/errgroup"
"github.com/docker/compose-cli/api/compose"
"github.com/docker/compose-cli/api/containers"
"github.com/docker/compose-cli/formatter"
"github.com/docker/compose-cli/progress"
)
func (s *local) Up(ctx context.Context, project *types.Project, detach bool) error {
for k, network := range project.Networks {
if !network.External.External && network.Name != "" {
network.Name = fmt.Sprintf("%s_%s", project.Name, k)
project.Networks[k] = network
}
err := s.ensureNetwork(ctx, network)
if err != nil {
return err
}
}
for k, volume := range project.Volumes {
if !volume.External.External && volume.Name != "" {
volume.Name = fmt.Sprintf("%s_%s", project.Name, k)
project.Volumes[k] = volume
}
err := s.ensureVolume(ctx, volume)
if err != nil {
return err
}
}
for _, service := range project.Services {
err := s.applyPullPolicy(ctx, service)
if err != nil {
return err
}
}
err := inDependencyOrder(ctx, project, func(c context.Context, service types.ServiceConfig) error {
return s.ensureService(c, project, service)
})
return err
}
func getContainerName(c moby.Container) string {
// Names return container canonical name /foo + link aliases /linked_by/foo
for _, name := range c.Names {
if strings.LastIndex(name, "/") == 0 {
return name[1:]
}
}
return c.Names[0][1:]
}
func (s *local) applyPullPolicy(ctx context.Context, service types.ServiceConfig) error {
w := progress.ContextWriter(ctx)
// TODO build vs pull should be controlled by pull policy
// if service.Build {}
if service.Image != "" {
_, _, err := s.containerService.apiClient.ImageInspectWithRaw(ctx, service.Image)
if err != nil {
if errdefs.IsNotFound(err) {
stream, err := s.containerService.apiClient.ImagePull(ctx, service.Image, moby.ImagePullOptions{})
if err != nil {
return err
}
dec := json.NewDecoder(stream)
for {
var jm jsonmessage.JSONMessage
if err := dec.Decode(&jm); err != nil {
if err == io.EOF {
break
}
return err
}
toProgressEvent(jm, w)
}
}
}
}
return nil
}
func toProgressEvent(jm jsonmessage.JSONMessage, w progress.Writer) {
if jm.Progress != nil {
if jm.Progress.Total != 0 {
percentage := int(float64(jm.Progress.Current)/float64(jm.Progress.Total)*100) / 2
numSpaces := 50 - percentage
w.Event(progress.Event{
ID: jm.ID,
Text: jm.Status,
Status: 0,
StatusText: fmt.Sprintf("[%s>%s] ", strings.Repeat("=", percentage), strings.Repeat(" ", numSpaces)),
Done: jm.Status == "Pull complete",
})
} else {
if jm.Error != nil {
w.Event(progress.Event{
ID: jm.ID,
Text: jm.Status,
Status: progress.Error,
StatusText: jm.Error.Message,
Done: true,
})
} else if jm.Status == "Pull complete" || jm.Status == "Already exists" {
w.Event(progress.Event{
ID: jm.ID,
Text: jm.Status,
Status: progress.Done,
Done: true,
})
} else {
w.Event(progress.Event{
ID: jm.ID,
Text: jm.Status,
Status: progress.Working,
Done: false,
})
}
}
}
}
func (s *local) Down(ctx context.Context, projectName string) error {
list, err := s.containerService.apiClient.ContainerList(ctx, moby.ContainerListOptions{
Filters: filters.NewArgs(
projectFilter(projectName),
),
})
if err != nil {
return err
}
eg, ctx := errgroup.WithContext(ctx)
w := progress.ContextWriter(ctx)
for _, c := range list {
container := c
eg.Go(func() error {
w.Event(progress.Event{
ID: getContainerName(container),
Text: "Stopping",
Status: progress.Working,
Done: false,
})
err := s.containerService.Stop(ctx, container.ID, nil)
if err != nil {
return err
}
w.Event(progress.Event{
ID: getContainerName(container),
Text: "Removing",
Status: progress.Working,
Done: false,
})
err = s.containerService.Delete(ctx, container.ID, containers.DeleteRequest{})
if err != nil {
return err
}
w.Event(progress.Event{
ID: getContainerName(container),
Text: "Removed",
Status: progress.Done,
Done: true,
})
return nil
})
}
return eg.Wait()
}
func (s *local) Logs(ctx context.Context, projectName string, w io.Writer) error {
list, err := s.containerService.apiClient.ContainerList(ctx, moby.ContainerListOptions{
Filters: filters.NewArgs(
projectFilter(projectName),
),
})
if err != nil {
return err
}
var wg sync.WaitGroup
consumer := formatter.NewLogConsumer(w)
for _, c := range list {
service := c.Labels[serviceLabel]
containerID := c.ID
go func() {
_ = s.containerService.Logs(ctx, containerID, containers.LogsRequest{
Follow: true,
Writer: consumer.GetWriter(service, containerID),
})
wg.Done()
}()
wg.Add(1)
}
wg.Wait()
return nil
}
func (s *local) Ps(ctx context.Context, projectName string) ([]compose.ServiceStatus, error) {
list, err := s.containerService.apiClient.ContainerList(ctx, moby.ContainerListOptions{
Filters: filters.NewArgs(
projectFilter(projectName),
),
})
if err != nil {
return nil, err
}
var status []compose.ServiceStatus
for _, c := range list {
// TODO group by service
status = append(status, compose.ServiceStatus{
ID: c.ID,
Name: c.Labels[serviceLabel],
Replicas: 0,
Desired: 0,
Ports: nil,
Publishers: nil,
})
}
return status, nil
}
func (s *local) List(ctx context.Context, projectName string) ([]compose.Stack, error) {
list, err := s.containerService.apiClient.ContainerList(ctx, moby.ContainerListOptions{
Filters: filters.NewArgs(hasProjectLabelFilter()),
})
if err != nil {
return nil, err
}
return containersToStacks(list)
}
func containersToStacks(containers []moby.Container) ([]compose.Stack, error) {
statusesByProject := map[string][]string{}
keys := []string{}
for _, c := range containers {
project, ok := c.Labels[projectLabel]
if !ok {
return nil, fmt.Errorf("No label %q set on container %q of compose project", serviceLabel, c.ID)
}
projectStatuses, ok := statusesByProject[project]
if !ok {
projectStatuses = []string{}
keys = append(keys, project)
}
projectStatuses = append(projectStatuses, c.State)
statusesByProject[project] = projectStatuses
}
sort.Strings(keys)
var projects []compose.Stack
for _, project := range keys {
statuses := statusesByProject[project]
projects = append(projects, compose.Stack{
ID: project,
Name: project,
Status: combinedStatus(statuses),
})
}
return projects, nil
}
func combinedStatus(statuses []string) string {
nbByStatus := map[string]int{}
keys := []string{}
for _, status := range statuses {
nb, ok := nbByStatus[status]
if !ok {
nb = 0
keys = append(keys, status)
}
nbByStatus[status] = nb + 1
}
sort.Strings(keys)
result := ""
for _, status := range keys {
nb := nbByStatus[status]
if result != "" {
result = result + ", "
}
result = result + fmt.Sprintf("%s(%d)", status, nb)
}
return result
}
func (s *local) Convert(ctx context.Context, project *types.Project, format string) ([]byte, error) {
switch format {
case "json":
return json.MarshalIndent(project, "", " ")
case "yaml":
return yaml.Marshal(project)
default:
return nil, fmt.Errorf("unsupported format %q", format)
}
}
func getContainerCreateOptions(p *types.Project, s types.ServiceConfig, number int, inherit *moby.Container) (*container.Config, *container.HostConfig, *network.NetworkingConfig, error) {
hash, err := jsonHash(s)
if err != nil {
return nil, nil, nil, err
}
labels := map[string]string{
projectLabel: p.Name,
serviceLabel: s.Name,
configHashLabel: hash,
containerNumberLabel: strconv.Itoa(number),
}
var (
runCmd strslice.StrSlice
entrypoint strslice.StrSlice
)
if len(s.Command) > 0 {
runCmd = strslice.StrSlice(s.Command)
}
if len(s.Entrypoint) > 0 {
entrypoint = strslice.StrSlice(s.Entrypoint)
}
image := s.Image
if s.Image == "" {
image = fmt.Sprintf("%s_%s", p.Name, s.Name)
}
var (
tty = s.Tty
stdinOpen = s.StdinOpen
attachStdin = false
)
containerConfig := container.Config{
Hostname: s.Hostname,
Domainname: s.DomainName,
User: s.User,
ExposedPorts: buildContainerPorts(s),
Tty: tty,
OpenStdin: stdinOpen,
StdinOnce: true,
AttachStdin: attachStdin,
AttachStderr: true,
AttachStdout: true,
Cmd: runCmd,
Image: image,
WorkingDir: s.WorkingDir,
Entrypoint: entrypoint,
NetworkDisabled: s.NetworkMode == "disabled",
MacAddress: s.MacAddress,
Labels: labels,
StopSignal: s.StopSignal,
Env: toMobyEnv(s.Environment),
Healthcheck: toMobyHealthCheck(s.HealthCheck),
// Volumes: // FIXME unclear to me the overlap with HostConfig.Mounts
StopTimeout: toSeconds(s.StopGracePeriod),
}
mountOptions := buildContainerMountOptions(p, s, inherit)
bindings := buildContainerBindingOptions(s)
networkMode := getNetworkMode(p, s)
hostConfig := container.HostConfig{
Mounts: mountOptions,
CapAdd: strslice.StrSlice(s.CapAdd),
CapDrop: strslice.StrSlice(s.CapDrop),
NetworkMode: networkMode,
Init: s.Init,
ReadonlyRootfs: s.ReadOnly,
// ShmSize: , TODO
Sysctls: s.Sysctls,
PortBindings: bindings,
}
networkConfig := buildDefaultNetworkConfig(s, networkMode)
return &containerConfig, &hostConfig, networkConfig, nil
}
func buildContainerPorts(s types.ServiceConfig) nat.PortSet {
ports := nat.PortSet{}
for _, p := range s.Ports {
p := nat.Port(fmt.Sprintf("%d/%s", p.Target, p.Protocol))
ports[p] = struct{}{}
}
return ports
}
func buildContainerBindingOptions(s types.ServiceConfig) nat.PortMap {
bindings := nat.PortMap{}
for _, port := range s.Ports {
p := nat.Port(fmt.Sprintf("%d/%s", port.Target, port.Protocol))
bind := []nat.PortBinding{}
binding := nat.PortBinding{}
if port.Published > 0 {
binding.HostPort = fmt.Sprint(port.Published)
}
bind = append(bind, binding)
bindings[p] = bind
}
return bindings
}
func buildContainerMountOptions(p *types.Project, s types.ServiceConfig, inherit *moby.Container) []mount.Mount {
mounts := []mount.Mount{}
var inherited []string
if inherit != nil {
for _, m := range inherit.Mounts {
if m.Type == "tmpfs" {
continue
}
src := m.Source
if m.Type == "volume" {
src = m.Name
}
mounts = append(mounts, mount.Mount{
Type: m.Type,
Source: src,
Target: m.Destination,
ReadOnly: !m.RW,
})
inherited = append(inherited, m.Destination)
}
}
for _, v := range s.Volumes {
if contains(inherited, v.Target) {
continue
}
source := v.Source
if v.Type == "bind" && !filepath.IsAbs(source) {
// FIXME handle ~/
source = filepath.Join(p.WorkingDir, source)
}
mounts = append(mounts, mount.Mount{
Type: mount.Type(v.Type),
Source: source,
Target: v.Target,
ReadOnly: v.ReadOnly,
Consistency: mount.Consistency(v.Consistency),
BindOptions: buildBindOption(v.Bind),
VolumeOptions: buildVolumeOptions(v.Volume),
TmpfsOptions: buildTmpfsOptions(v.Tmpfs),
})
}
return mounts
}
func buildBindOption(bind *types.ServiceVolumeBind) *mount.BindOptions {
if bind == nil {
return nil
}
return &mount.BindOptions{
Propagation: mount.Propagation(bind.Propagation),
// NonRecursive: false, FIXME missing from model ?
}
}
func buildVolumeOptions(vol *types.ServiceVolumeVolume) *mount.VolumeOptions {
if vol == nil {
return nil
}
return &mount.VolumeOptions{
NoCopy: vol.NoCopy,
// Labels: , // FIXME missing from model ?
// DriverConfig: , // FIXME missing from model ?
}
}
func buildTmpfsOptions(tmpfs *types.ServiceVolumeTmpfs) *mount.TmpfsOptions {
if tmpfs == nil {
return nil
}
return &mount.TmpfsOptions{
SizeBytes: tmpfs.Size,
// Mode: , // FIXME missing from model ?
}
}
func buildDefaultNetworkConfig(s types.ServiceConfig, networkMode container.NetworkMode) *network.NetworkingConfig {
config := map[string]*network.EndpointSettings{}
net := string(networkMode)
config[net] = &network.EndpointSettings{
Aliases: getAliases(s, s.Networks[net]),
}
return &network.NetworkingConfig{
EndpointsConfig: config,
}
}
func getAliases(s types.ServiceConfig, c *types.ServiceNetworkConfig) []string {
aliases := []string{s.Name}
if c != nil {
aliases = append(aliases, c.Aliases...)
}
return aliases
}
func getNetworkMode(p *types.Project, service types.ServiceConfig) container.NetworkMode {
mode := service.NetworkMode
if mode == "" {
if len(p.Networks) > 0 {
for name := range getNetworksForService(service) {
return container.NetworkMode(p.Networks[name].Name)
}
}
return container.NetworkMode("none")
}
/// FIXME incomplete implementation
if strings.HasPrefix(mode, "service:") {
panic("Not yet implemented")
}
if strings.HasPrefix(mode, "container:") {
panic("Not yet implemented")
}
return container.NetworkMode(mode)
}
func getNetworksForService(s types.ServiceConfig) map[string]*types.ServiceNetworkConfig {
if len(s.Networks) > 0 {
return s.Networks
}
return map[string]*types.ServiceNetworkConfig{"default": nil}
}
func (s *local) ensureNetwork(ctx context.Context, n types.NetworkConfig) error {
_, err := s.containerService.apiClient.NetworkInspect(ctx, n.Name, moby.NetworkInspectOptions{})
if err != nil {
if errdefs.IsNotFound(err) {
createOpts := moby.NetworkCreate{
// TODO NameSpace Labels
Labels: n.Labels,
Driver: n.Driver,
Options: n.DriverOpts,
Internal: n.Internal,
Attachable: n.Attachable,
}
if n.Ipam.Driver != "" || len(n.Ipam.Config) > 0 {
createOpts.IPAM = &network.IPAM{}
}
if n.Ipam.Driver != "" {
createOpts.IPAM.Driver = n.Ipam.Driver
}
for _, ipamConfig := range n.Ipam.Config {
config := network.IPAMConfig{
Subnet: ipamConfig.Subnet,
}
createOpts.IPAM.Config = append(createOpts.IPAM.Config, config)
}
w := progress.ContextWriter(ctx)
w.Event(progress.Event{
ID: fmt.Sprintf("Network %q", n.Name),
Status: progress.Working,
StatusText: "Create",
Done: false,
})
if _, err := s.containerService.apiClient.NetworkCreate(context.Background(), n.Name, createOpts); err != nil {
return errors.Wrapf(err, "failed to create network %s", n.Name)
}
w.Event(progress.Event{
ID: fmt.Sprintf("Network %q", n.Name),
Status: progress.Done,
StatusText: "Created",
Done: true,
})
return nil
}
return err
}
return nil
}
func (s *local) ensureVolume(ctx context.Context, volume types.VolumeConfig) error {
// TODO could identify volume by label vs name
_, err := s.volumeService.Inspect(ctx, volume.Name)
if err != nil {
if errdefs.IsNotFound(err) {
w := progress.ContextWriter(ctx)
w.Event(progress.Event{
ID: fmt.Sprintf("Volume %q", volume.Name),
Status: progress.Working,
StatusText: "Create",
Done: false,
})
// TODO we miss support for driver_opts and labels
_, err := s.volumeService.Create(ctx, volume.Name, nil)
w.Event(progress.Event{
ID: fmt.Sprintf("Volume %q", volume.Name),
Status: progress.Done,
StatusText: "Created",
Done: true,
})
if err != nil {
return err
}
}
return err
}
return nil
}