12345678910111213141516171819202122232425262728293031323334353637383940414243444546474849505152535455565758596061626364656667686970717273747576777879808182838485868788899091929394959697989910010110210310410510610710810911011111211311411511611711811912012112212312412512612712812913013113213313413513613713813914014114214314414514614714814915015115215315415515615715815916016116216316416516616716816917017117217317417517617717817918018118218318418518618718818919019119219319419519619719819920020120220320420520620720820921021121221321421521621721821922022122222322422522622722822923023123223323423523623723823924024124224324424524624724824925025125225325425525625725825926026126226326426526626726826927027127227327427527627727827928028128228328428528628728828929029129229329429529629729829930030130230330430530630730830931031131231331431531631731831932032132232332432532632732832933033133233333433533633733833934034134234334434534634734834935035135235335435535635735835936036136236336436536636736836937037137237337437537637737837938038138238338438538638738838939039139239339439539639739839940040140240340440540640740840941041141241341441541641741841942042142242342442542642742842943043143243343443543643743843944044144244344444544644744844945045145245345445545645745845946046146246346446546646746846947047147247347447547647747847948048148248348448548648748848949049149249349449549649749849950050150250350450550650750850951051151251351451551651751851952052152252352452552652752852953053153253353453553653753853954054154254354454554654754854955055155255355455555655755855956056156256356456556656756856957057157257357457557657757857958058158258358458558658758858959059159259359459559659759859960060160260360460560660760860961061161261361461561661761861962062162262362462562662762862963063163263363463563663763863964064164264364464564664764864965065165265365465565665765865966066166266366466566666766866967067167267367467567667767867968068168268368468568668768868969069169269369469569669769869970070170270370470570670770870971071171271371471571671771871972072172272372472572672772872973073173273373473573673773873974074174274374474574674774874975075175275375475575675775875976076176276376476576676776876977077177277377477577677777877978078178278378478578678778878979079179279379479579679779879980080180280380480580680780880981081181281381481581681781881982082182282382482582682782882983083183283383483583683783883984084184284384484584684784884985085185285385485585685785885986086186286386486586686786886987087187287387487587687787887988088188288388488588688788888989089189289389489589689789889990090190290390490590690790890991091191291391491591691791891992092192292392492592692792892993093193293393493593693793893994094194294394494594694794894995095195295395495595695795895996096196296396496596696796896997097197297397497597697797897998098198298398498598698798898999099199299399499599699799899910001001100210031004100510061007100810091010101110121013101410151016101710181019102010211022102310241025102610271028102910301031103210331034103510361037103810391040104110421043104410451046104710481049105010511052105310541055105610571058105910601061106210631064106510661067106810691070107110721073107410751076107710781079108010811082108310841085108610871088108910901091109210931094109510961097109810991100110111021103110411051106110711081109111011111112111311141115111611171118111911201121112211231124112511261127112811291130113111321133113411351136113711381139114011411142114311441145114611471148114911501151115211531154115511561157115811591160116111621163116411651166116711681169117011711172117311741175117611771178117911801181118211831184118511861187118811891190119111921193119411951196119711981199120012011202120312041205120612071208120912101211121212131214121512161217121812191220122112221223122412251226122712281229123012311232123312341235123612371238123912401241124212431244124512461247124812491250125112521253125412551256125712581259126012611262126312641265126612671268126912701271127212731274127512761277127812791280128112821283128412851286128712881289129012911292129312941295129612971298129913001301130213031304130513061307130813091310131113121313131413151316131713181319132013211322132313241325132613271328132913301331133213331334133513361337133813391340134113421343134413451346134713481349135013511352135313541355135613571358135913601361136213631364136513661367136813691370137113721373137413751376137713781379138013811382138313841385138613871388138913901391139213931394139513961397139813991400140114021403140414051406140714081409141014111412141314141415141614171418141914201421142214231424142514261427142814291430143114321433143414351436143714381439144014411442144314441445144614471448144914501451145214531454145514561457145814591460146114621463146414651466146714681469147014711472147314741475147614771478147914801481148214831484148514861487148814891490149114921493149414951496149714981499150015011502150315041505150615071508150915101511151215131514151515161517151815191520152115221523152415251526152715281529153015311532153315341535153615371538153915401541154215431544154515461547154815491550155115521553155415551556155715581559156015611562156315641565156615671568156915701571157215731574157515761577157815791580158115821583158415851586158715881589159015911592159315941595159615971598159916001601160216031604160516061607160816091610161116121613161416151616161716181619162016211622162316241625162616271628162916301631163216331634163516361637163816391640164116421643164416451646164716481649165016511652165316541655165616571658165916601661166216631664166516661667166816691670167116721673167416751676167716781679168016811682168316841685168616871688 |
- from __future__ import absolute_import
- from __future__ import unicode_literals
- import itertools
- import logging
- import os
- import re
- import sys
- from collections import namedtuple
- from collections import OrderedDict
- from operator import attrgetter
- import enum
- import six
- from docker.errors import APIError
- from docker.errors import ImageNotFound
- from docker.errors import NotFound
- from docker.types import LogConfig
- from docker.types import Mount
- from docker.utils import version_gte
- from docker.utils import version_lt
- from docker.utils.ports import build_port_bindings
- from docker.utils.ports import split_port
- from docker.utils.utils import convert_tmpfs_mounts
- from . import __version__
- from . import const
- from . import progress_stream
- from .config import DOCKER_CONFIG_KEYS
- from .config import is_url
- from .config import merge_environment
- from .config import merge_labels
- from .config.errors import DependencyError
- from .config.types import MountSpec
- from .config.types import ServicePort
- from .config.types import VolumeSpec
- from .const import DEFAULT_TIMEOUT
- from .const import IS_WINDOWS_PLATFORM
- from .const import LABEL_CONFIG_HASH
- from .const import LABEL_CONTAINER_NUMBER
- from .const import LABEL_ONE_OFF
- from .const import LABEL_PROJECT
- from .const import LABEL_SERVICE
- from .const import LABEL_SLUG
- from .const import LABEL_VERSION
- from .const import NANOCPUS_SCALE
- from .const import WINDOWS_LONGPATH_PREFIX
- from .container import Container
- from .errors import HealthCheckFailed
- from .errors import NoHealthCheckConfigured
- from .errors import OperationFailedError
- from .parallel import parallel_execute
- from .progress_stream import stream_output
- from .progress_stream import StreamOutputError
- from .utils import generate_random_id
- from .utils import json_hash
- from .utils import parse_bytes
- from .utils import parse_seconds_float
- from .utils import truncate_id
- from .utils import unique_everseen
- log = logging.getLogger(__name__)
- HOST_CONFIG_KEYS = [
- 'cap_add',
- 'cap_drop',
- 'cgroup_parent',
- 'cpu_count',
- 'cpu_percent',
- 'cpu_period',
- 'cpu_quota',
- 'cpu_rt_period',
- 'cpu_rt_runtime',
- 'cpu_shares',
- 'cpus',
- 'cpuset',
- 'device_cgroup_rules',
- 'devices',
- 'dns',
- 'dns_search',
- 'dns_opt',
- 'env_file',
- 'extra_hosts',
- 'group_add',
- 'init',
- 'ipc',
- 'isolation',
- 'read_only',
- 'log_driver',
- 'log_opt',
- 'mem_limit',
- 'mem_reservation',
- 'memswap_limit',
- 'mem_swappiness',
- 'oom_kill_disable',
- 'oom_score_adj',
- 'pid',
- 'pids_limit',
- 'privileged',
- 'restart',
- 'runtime',
- 'security_opt',
- 'shm_size',
- 'storage_opt',
- 'sysctls',
- 'userns_mode',
- 'volumes_from',
- 'volume_driver',
- ]
- CONDITION_STARTED = 'service_started'
- CONDITION_HEALTHY = 'service_healthy'
- class BuildError(Exception):
- def __init__(self, service, reason):
- self.service = service
- self.reason = reason
- class NeedsBuildError(Exception):
- def __init__(self, service):
- self.service = service
- class NoSuchImageError(Exception):
- pass
- ServiceName = namedtuple('ServiceName', 'project service number')
- ConvergencePlan = namedtuple('ConvergencePlan', 'action containers')
- @enum.unique
- class ConvergenceStrategy(enum.Enum):
- """Enumeration for all possible convergence strategies. Values refer to
- when containers should be recreated.
- """
- changed = 1
- always = 2
- never = 3
- @property
- def allows_recreate(self):
- return self is not type(self).never
- @enum.unique
- class ImageType(enum.Enum):
- """Enumeration for the types of images known to compose."""
- none = 0
- local = 1
- all = 2
- @enum.unique
- class BuildAction(enum.Enum):
- """Enumeration for the possible build actions."""
- none = 0
- force = 1
- skip = 2
- class Service(object):
- def __init__(
- self,
- name,
- client=None,
- project='default',
- use_networking=False,
- links=None,
- volumes_from=None,
- network_mode=None,
- networks=None,
- secrets=None,
- scale=None,
- pid_mode=None,
- default_platform=None,
- **options
- ):
- self.name = name
- self.client = client
- self.project = project
- self.use_networking = use_networking
- self.links = links or []
- self.volumes_from = volumes_from or []
- self.network_mode = network_mode or NetworkMode(None)
- self.pid_mode = pid_mode or PidMode(None)
- self.networks = networks or {}
- self.secrets = secrets or []
- self.scale_num = scale or 1
- self.default_platform = default_platform
- self.options = options
- def __repr__(self):
- return '<Service: {}>'.format(self.name)
- def containers(self, stopped=False, one_off=False, filters=None, labels=None):
- if filters is None:
- filters = {}
- filters.update({'label': self.labels(one_off=one_off) + (labels or [])})
- result = list(filter(None, [
- Container.from_ps(self.client, container)
- for container in self.client.containers(
- all=stopped,
- filters=filters)])
- )
- if result:
- return result
- filters.update({'label': self.labels(one_off=one_off, legacy=True) + (labels or [])})
- return list(
- filter(
- lambda c: c.has_legacy_proj_name(self.project), filter(None, [
- Container.from_ps(self.client, container)
- for container in self.client.containers(
- all=stopped,
- filters=filters)])
- )
- )
- def get_container(self, number=1):
- """Return a :class:`compose.container.Container` for this service. The
- container must be active, and match `number`.
- """
- for container in self.containers(labels=['{0}={1}'.format(LABEL_CONTAINER_NUMBER, number)]):
- return container
- raise ValueError("No container found for %s_%s" % (self.name, number))
- def start(self, **options):
- containers = self.containers(stopped=True)
- for c in containers:
- self.start_container_if_stopped(c, **options)
- return containers
- def show_scale_warnings(self, desired_num):
- if self.custom_container_name and desired_num > 1:
- log.warn('The "%s" service is using the custom container name "%s". '
- 'Docker requires each container to have a unique name. '
- 'Remove the custom name to scale the service.'
- % (self.name, self.custom_container_name))
- if self.specifies_host_port() and desired_num > 1:
- log.warn('The "%s" service specifies a port on the host. If multiple containers '
- 'for this service are created on a single host, the port will clash.'
- % self.name)
- def scale(self, desired_num, timeout=None):
- """
- Adjusts the number of containers to the specified number and ensures
- they are running.
- - creates containers until there are at least `desired_num`
- - stops containers until there are at most `desired_num` running
- - starts containers until there are at least `desired_num` running
- - removes all stopped containers
- """
- self.show_scale_warnings(desired_num)
- running_containers = self.containers(stopped=False)
- num_running = len(running_containers)
- for c in running_containers:
- if not c.has_legacy_proj_name(self.project):
- continue
- log.info('Recreating container with legacy name %s' % c.name)
- self.recreate_container(c, timeout, start_new_container=False)
- if desired_num == num_running:
- # do nothing as we already have the desired number
- log.info('Desired container number already achieved')
- return
- if desired_num > num_running:
- all_containers = self.containers(stopped=True)
- if num_running != len(all_containers):
- # we have some stopped containers, check for divergences
- stopped_containers = [
- c for c in all_containers if not c.is_running
- ]
- # Remove containers that have diverged
- divergent_containers = [
- c for c in stopped_containers if self._containers_have_diverged([c])
- ]
- for c in divergent_containers:
- c.remove()
- all_containers = list(set(all_containers) - set(divergent_containers))
- sorted_containers = sorted(all_containers, key=attrgetter('number'))
- self._execute_convergence_start(
- sorted_containers, desired_num, timeout, True, True
- )
- if desired_num < num_running:
- num_to_stop = num_running - desired_num
- sorted_running_containers = sorted(
- running_containers,
- key=attrgetter('number'))
- self._downscale(sorted_running_containers[-num_to_stop:], timeout)
- def create_container(self,
- one_off=False,
- previous_container=None,
- number=None,
- quiet=False,
- **override_options):
- """
- Create a container for this service. If the image doesn't exist, attempt to pull
- it.
- """
- # This is only necessary for `scale` and `volumes_from`
- # auto-creating containers to satisfy the dependency.
- self.ensure_image_exists()
- container_options = self._get_container_create_options(
- override_options,
- number or self._next_container_number(one_off=one_off),
- one_off=one_off,
- previous_container=previous_container,
- )
- if 'name' in container_options and not quiet:
- log.info("Creating %s" % container_options['name'])
- try:
- return Container.create(self.client, **container_options)
- except APIError as ex:
- raise OperationFailedError("Cannot create container for service %s: %s" %
- (self.name, ex.explanation))
- def ensure_image_exists(self, do_build=BuildAction.none, silent=False):
- if self.can_be_built() and do_build == BuildAction.force:
- self.build()
- return
- try:
- self.image()
- return
- except NoSuchImageError:
- pass
- if not self.can_be_built():
- self.pull(silent=silent)
- return
- if do_build == BuildAction.skip:
- raise NeedsBuildError(self)
- self.build()
- log.warn(
- "Image for service {} was built because it did not already exist. To "
- "rebuild this image you must use `docker-compose build` or "
- "`docker-compose up --build`.".format(self.name))
- def image(self):
- try:
- return self.client.inspect_image(self.image_name)
- except ImageNotFound:
- raise NoSuchImageError("Image '{}' not found".format(self.image_name))
- @property
- def image_name(self):
- return self.options.get('image', '{project}_{s.name}'.format(
- s=self, project=self.project.lstrip('_-')
- ))
- @property
- def platform(self):
- platform = self.options.get('platform')
- if not platform and version_gte(self.client.api_version, '1.35'):
- platform = self.default_platform
- return platform
- def convergence_plan(self, strategy=ConvergenceStrategy.changed):
- containers = self.containers(stopped=True)
- if not containers:
- return ConvergencePlan('create', [])
- if strategy is ConvergenceStrategy.never:
- return ConvergencePlan('start', containers)
- if (
- strategy is ConvergenceStrategy.always or
- self._containers_have_diverged(containers)
- ):
- return ConvergencePlan('recreate', containers)
- stopped = [c for c in containers if not c.is_running]
- if stopped:
- return ConvergencePlan('start', stopped)
- return ConvergencePlan('noop', containers)
- def _containers_have_diverged(self, containers):
- config_hash = None
- try:
- config_hash = self.config_hash
- except NoSuchImageError as e:
- log.debug(
- 'Service %s has diverged: %s',
- self.name, six.text_type(e),
- )
- return True
- has_diverged = False
- for c in containers:
- if c.has_legacy_proj_name(self.project):
- log.debug('%s has diverged: Legacy project name' % c.name)
- has_diverged = True
- continue
- container_config_hash = c.labels.get(LABEL_CONFIG_HASH, None)
- if container_config_hash != config_hash:
- log.debug(
- '%s has diverged: %s != %s',
- c.name, container_config_hash, config_hash,
- )
- has_diverged = True
- return has_diverged
- def _execute_convergence_create(self, scale, detached, start):
- i = self._next_container_number()
- def create_and_start(service, n):
- container = service.create_container(number=n, quiet=True)
- if not detached:
- container.attach_log_stream()
- if start:
- self.start_container(container)
- return container
- containers, errors = parallel_execute(
- [
- ServiceName(self.project, self.name, index)
- for index in range(i, i + scale)
- ],
- lambda service_name: create_and_start(self, service_name.number),
- lambda service_name: self.get_container_name(service_name.service, service_name.number),
- "Creating"
- )
- for error in errors.values():
- raise OperationFailedError(error)
- return containers
- def _execute_convergence_recreate(self, containers, scale, timeout, detached, start,
- renew_anonymous_volumes):
- if scale is not None and len(containers) > scale:
- self._downscale(containers[scale:], timeout)
- containers = containers[:scale]
- def recreate(container):
- return self.recreate_container(
- container, timeout=timeout, attach_logs=not detached,
- start_new_container=start, renew_anonymous_volumes=renew_anonymous_volumes
- )
- containers, errors = parallel_execute(
- containers,
- recreate,
- lambda c: c.name,
- "Recreating",
- )
- for error in errors.values():
- raise OperationFailedError(error)
- if scale is not None and len(containers) < scale:
- containers.extend(self._execute_convergence_create(
- scale - len(containers), detached, start
- ))
- return containers
- def _execute_convergence_start(self, containers, scale, timeout, detached, start):
- if scale is not None and len(containers) > scale:
- self._downscale(containers[scale:], timeout)
- containers = containers[:scale]
- if start:
- _, errors = parallel_execute(
- containers,
- lambda c: self.start_container_if_stopped(c, attach_logs=not detached, quiet=True),
- lambda c: c.name,
- "Starting",
- )
- for error in errors.values():
- raise OperationFailedError(error)
- if scale is not None and len(containers) < scale:
- containers.extend(self._execute_convergence_create(
- scale - len(containers), detached, start
- ))
- return containers
- def _downscale(self, containers, timeout=None):
- def stop_and_remove(container):
- container.stop(timeout=self.stop_timeout(timeout))
- container.remove()
- parallel_execute(
- containers,
- stop_and_remove,
- lambda c: c.name,
- "Stopping and removing",
- )
- def execute_convergence_plan(self, plan, timeout=None, detached=False,
- start=True, scale_override=None,
- rescale=True, reset_container_image=False,
- renew_anonymous_volumes=False):
- (action, containers) = plan
- scale = scale_override if scale_override is not None else self.scale_num
- containers = sorted(containers, key=attrgetter('number'))
- self.show_scale_warnings(scale)
- if action == 'create':
- return self._execute_convergence_create(
- scale, detached, start
- )
- # The create action needs always needs an initial scale, but otherwise,
- # we set scale to none in no-rescale scenarios (`run` dependencies)
- if not rescale:
- scale = None
- if action == 'recreate':
- if reset_container_image:
- # Updating the image ID on the container object lets us recover old volumes if
- # the new image uses them as well
- img_id = self.image()['Id']
- for c in containers:
- c.reset_image(img_id)
- return self._execute_convergence_recreate(
- containers, scale, timeout, detached, start,
- renew_anonymous_volumes,
- )
- if action == 'start':
- return self._execute_convergence_start(
- containers, scale, timeout, detached, start
- )
- if action == 'noop':
- if scale != len(containers):
- return self._execute_convergence_start(
- containers, scale, timeout, detached, start
- )
- for c in containers:
- log.info("%s is up-to-date" % c.name)
- return containers
- raise Exception("Invalid action: {}".format(action))
- def recreate_container(self, container, timeout=None, attach_logs=False, start_new_container=True,
- renew_anonymous_volumes=False):
- """Recreate a container.
- The original container is renamed to a temporary name so that data
- volumes can be copied to the new container, before the original
- container is removed.
- """
- container.stop(timeout=self.stop_timeout(timeout))
- container.rename_to_tmp_name()
- new_container = self.create_container(
- previous_container=container if not renew_anonymous_volumes else None,
- number=container.number,
- quiet=True,
- )
- if attach_logs:
- new_container.attach_log_stream()
- if start_new_container:
- self.start_container(new_container)
- container.remove()
- return new_container
- def stop_timeout(self, timeout):
- if timeout is not None:
- return timeout
- timeout = parse_seconds_float(self.options.get('stop_grace_period'))
- if timeout is not None:
- return timeout
- return DEFAULT_TIMEOUT
- def start_container_if_stopped(self, container, attach_logs=False, quiet=False):
- if not container.is_running:
- if not quiet:
- log.info("Starting %s" % container.name)
- if attach_logs:
- container.attach_log_stream()
- return self.start_container(container)
- def start_container(self, container, use_network_aliases=True):
- self.connect_container_to_networks(container, use_network_aliases)
- try:
- container.start()
- except APIError as ex:
- raise OperationFailedError("Cannot start service %s: %s" % (self.name, ex.explanation))
- return container
- @property
- def prioritized_networks(self):
- return OrderedDict(
- sorted(
- self.networks.items(),
- key=lambda t: t[1].get('priority') or 0, reverse=True
- )
- )
- def connect_container_to_networks(self, container, use_network_aliases=True):
- connected_networks = container.get('NetworkSettings.Networks')
- for network, netdefs in self.prioritized_networks.items():
- if network in connected_networks:
- if short_id_alias_exists(container, network):
- continue
- self.client.disconnect_container_from_network(container.id, network)
- aliases = self._get_aliases(netdefs, container) if use_network_aliases else []
- self.client.connect_container_to_network(
- container.id, network,
- aliases=aliases,
- ipv4_address=netdefs.get('ipv4_address', None),
- ipv6_address=netdefs.get('ipv6_address', None),
- links=self._get_links(False),
- link_local_ips=netdefs.get('link_local_ips', None),
- )
- def remove_duplicate_containers(self, timeout=None):
- for c in self.duplicate_containers():
- log.info('Removing %s' % c.name)
- c.stop(timeout=self.stop_timeout(timeout))
- c.remove()
- def duplicate_containers(self):
- containers = sorted(
- self.containers(stopped=True),
- key=lambda c: c.get('Created'),
- )
- numbers = set()
- for c in containers:
- if c.number in numbers:
- yield c
- else:
- numbers.add(c.number)
- @property
- def config_hash(self):
- return json_hash(self.config_dict())
- def config_dict(self):
- def image_id():
- try:
- return self.image()['Id']
- except NoSuchImageError:
- return None
- return {
- 'options': self.options,
- 'image_id': image_id(),
- 'links': self.get_link_names(),
- 'net': self.network_mode.id,
- 'networks': self.networks,
- 'volumes_from': [
- (v.source.name, v.mode)
- for v in self.volumes_from if isinstance(v.source, Service)
- ],
- }
- def get_dependency_names(self):
- net_name = self.network_mode.service_name
- pid_namespace = self.pid_mode.service_name
- return (
- self.get_linked_service_names() +
- self.get_volumes_from_names() +
- ([net_name] if net_name else []) +
- ([pid_namespace] if pid_namespace else []) +
- list(self.options.get('depends_on', {}).keys())
- )
- def get_dependency_configs(self):
- net_name = self.network_mode.service_name
- pid_namespace = self.pid_mode.service_name
- configs = dict(
- [(name, None) for name in self.get_linked_service_names()]
- )
- configs.update(dict(
- [(name, None) for name in self.get_volumes_from_names()]
- ))
- configs.update({net_name: None} if net_name else {})
- configs.update({pid_namespace: None} if pid_namespace else {})
- configs.update(self.options.get('depends_on', {}))
- for svc, config in self.options.get('depends_on', {}).items():
- if config['condition'] == CONDITION_STARTED:
- configs[svc] = lambda s: True
- elif config['condition'] == CONDITION_HEALTHY:
- configs[svc] = lambda s: s.is_healthy()
- else:
- # The config schema already prevents this, but it might be
- # bypassed if Compose is called programmatically.
- raise ValueError(
- 'depends_on condition "{}" is invalid.'.format(
- config['condition']
- )
- )
- return configs
- def get_linked_service_names(self):
- return [service.name for (service, _) in self.links]
- def get_link_names(self):
- return [(service.name, alias) for service, alias in self.links]
- def get_volumes_from_names(self):
- return [s.source.name for s in self.volumes_from if isinstance(s.source, Service)]
- def _next_container_number(self, one_off=False):
- if one_off:
- return None
- containers = itertools.chain(
- self._fetch_containers(
- all=True,
- filters={'label': self.labels(one_off=False)}
- ), self._fetch_containers(
- all=True,
- filters={'label': self.labels(one_off=False, legacy=True)}
- )
- )
- numbers = [c.number for c in containers if c.number is not None]
- return 1 if not numbers else max(numbers) + 1
- def _fetch_containers(self, **fetch_options):
- # Account for containers that might have been removed since we fetched
- # the list.
- def soft_inspect(container):
- try:
- return Container.from_id(self.client, container['Id'])
- except NotFound:
- return None
- return filter(None, [
- soft_inspect(container)
- for container in self.client.containers(**fetch_options)
- ])
- def _get_aliases(self, network, container=None):
- return list(
- {self.name} |
- ({container.short_id} if container else set()) |
- set(network.get('aliases', ()))
- )
- def build_default_networking_config(self):
- if not self.networks:
- return {}
- network = self.networks[self.network_mode.id]
- endpoint = {
- 'Aliases': self._get_aliases(network),
- 'IPAMConfig': {},
- }
- if network.get('ipv4_address'):
- endpoint['IPAMConfig']['IPv4Address'] = network.get('ipv4_address')
- if network.get('ipv6_address'):
- endpoint['IPAMConfig']['IPv6Address'] = network.get('ipv6_address')
- return {"EndpointsConfig": {self.network_mode.id: endpoint}}
- def _get_links(self, link_to_self):
- links = {}
- for service, link_name in self.links:
- for container in service.containers():
- links[link_name or service.name] = container.name
- links[container.name] = container.name
- links[container.name_without_project] = container.name
- if link_to_self:
- for container in self.containers():
- links[self.name] = container.name
- links[container.name] = container.name
- links[container.name_without_project] = container.name
- for external_link in self.options.get('external_links') or []:
- if ':' not in external_link:
- link_name = external_link
- else:
- external_link, link_name = external_link.split(':')
- links[link_name] = external_link
- return [
- (alias, container_name)
- for (container_name, alias) in links.items()
- ]
- def _get_volumes_from(self):
- return [build_volume_from(spec) for spec in self.volumes_from]
- def _get_container_create_options(
- self,
- override_options,
- number,
- one_off=False,
- previous_container=None):
- add_config_hash = (not one_off and not override_options)
- slug = generate_random_id() if one_off else None
- container_options = dict(
- (k, self.options[k])
- for k in DOCKER_CONFIG_KEYS if k in self.options)
- override_volumes = override_options.pop('volumes', [])
- container_options.update(override_options)
- if not container_options.get('name'):
- container_options['name'] = self.get_container_name(self.name, number, slug)
- container_options.setdefault('detach', True)
- # If a qualified hostname was given, split it into an
- # unqualified hostname and a domainname unless domainname
- # was also given explicitly. This matches behavior
- # until Docker Engine 1.11.0 - Docker API 1.23.
- if (version_lt(self.client.api_version, '1.23') and
- 'hostname' in container_options and
- 'domainname' not in container_options and
- '.' in container_options['hostname']):
- parts = container_options['hostname'].partition('.')
- container_options['hostname'] = parts[0]
- container_options['domainname'] = parts[2]
- if (version_gte(self.client.api_version, '1.25') and
- 'stop_grace_period' in self.options):
- container_options['stop_timeout'] = self.stop_timeout(None)
- if 'ports' in container_options or 'expose' in self.options:
- container_options['ports'] = build_container_ports(
- formatted_ports(container_options.get('ports', [])),
- self.options)
- if 'volumes' in container_options or override_volumes:
- container_options['volumes'] = list(set(
- container_options.get('volumes', []) + override_volumes
- ))
- container_options['environment'] = merge_environment(
- self._parse_proxy_config(),
- merge_environment(
- self.options.get('environment'),
- override_options.get('environment')
- )
- )
- container_options['labels'] = merge_labels(
- self.options.get('labels'),
- override_options.get('labels'))
- container_options, override_options = self._build_container_volume_options(
- previous_container, container_options, override_options
- )
- container_options['image'] = self.image_name
- container_options['labels'] = build_container_labels(
- container_options.get('labels', {}),
- self.labels(one_off=one_off),
- number,
- self.config_hash if add_config_hash else None,
- slug
- )
- # Delete options which are only used in HostConfig
- for key in HOST_CONFIG_KEYS:
- container_options.pop(key, None)
- container_options['host_config'] = self._get_container_host_config(
- override_options,
- one_off=one_off)
- networking_config = self.build_default_networking_config()
- if networking_config:
- container_options['networking_config'] = networking_config
- container_options['environment'] = format_environment(
- container_options['environment'])
- return container_options
- def _build_container_volume_options(self, previous_container, container_options, override_options):
- container_volumes = []
- container_mounts = []
- if 'volumes' in container_options:
- container_volumes = [
- v for v in container_options.get('volumes') if isinstance(v, VolumeSpec)
- ]
- container_mounts = [v for v in container_options.get('volumes') if isinstance(v, MountSpec)]
- binds, affinity = merge_volume_bindings(
- container_volumes, self.options.get('tmpfs') or [], previous_container,
- container_mounts
- )
- container_options['environment'].update(affinity)
- container_options['volumes'] = dict((v.internal, {}) for v in container_volumes or {})
- if version_gte(self.client.api_version, '1.30'):
- override_options['mounts'] = [build_mount(v) for v in container_mounts] or None
- else:
- # Workaround for 3.2 format
- override_options['tmpfs'] = self.options.get('tmpfs') or []
- for m in container_mounts:
- if m.is_tmpfs:
- override_options['tmpfs'].append(m.target)
- else:
- binds.append(m.legacy_repr())
- container_options['volumes'][m.target] = {}
- secret_volumes = self.get_secret_volumes()
- if secret_volumes:
- if version_lt(self.client.api_version, '1.30'):
- binds.extend(v.legacy_repr() for v in secret_volumes)
- container_options['volumes'].update(
- (v.target, {}) for v in secret_volumes
- )
- else:
- override_options['mounts'] = override_options.get('mounts') or []
- override_options['mounts'].extend([build_mount(v) for v in secret_volumes])
- # Remove possible duplicates (see e.g. https://github.com/docker/compose/issues/5885).
- # unique_everseen preserves order. (see https://github.com/docker/compose/issues/6091).
- override_options['binds'] = list(unique_everseen(binds))
- return container_options, override_options
- def _get_container_host_config(self, override_options, one_off=False):
- options = dict(self.options, **override_options)
- logging_dict = options.get('logging', None)
- blkio_config = convert_blkio_config(options.get('blkio_config', None))
- log_config = get_log_config(logging_dict)
- init_path = None
- if isinstance(options.get('init'), six.string_types):
- init_path = options.get('init')
- options['init'] = True
- security_opt = [
- o.value for o in options.get('security_opt')
- ] if options.get('security_opt') else None
- nano_cpus = None
- if 'cpus' in options:
- nano_cpus = int(options.get('cpus') * NANOCPUS_SCALE)
- return self.client.create_host_config(
- links=self._get_links(link_to_self=one_off),
- port_bindings=build_port_bindings(
- formatted_ports(options.get('ports', []))
- ),
- binds=options.get('binds'),
- volumes_from=self._get_volumes_from(),
- privileged=options.get('privileged', False),
- network_mode=self.network_mode.mode,
- devices=options.get('devices'),
- dns=options.get('dns'),
- dns_opt=options.get('dns_opt'),
- dns_search=options.get('dns_search'),
- restart_policy=options.get('restart'),
- runtime=options.get('runtime'),
- cap_add=options.get('cap_add'),
- cap_drop=options.get('cap_drop'),
- mem_limit=options.get('mem_limit'),
- mem_reservation=options.get('mem_reservation'),
- memswap_limit=options.get('memswap_limit'),
- ulimits=build_ulimits(options.get('ulimits')),
- log_config=log_config,
- extra_hosts=options.get('extra_hosts'),
- read_only=options.get('read_only'),
- pid_mode=self.pid_mode.mode,
- security_opt=security_opt,
- ipc_mode=options.get('ipc'),
- cgroup_parent=options.get('cgroup_parent'),
- cpu_quota=options.get('cpu_quota'),
- shm_size=options.get('shm_size'),
- sysctls=options.get('sysctls'),
- pids_limit=options.get('pids_limit'),
- tmpfs=options.get('tmpfs'),
- oom_kill_disable=options.get('oom_kill_disable'),
- oom_score_adj=options.get('oom_score_adj'),
- mem_swappiness=options.get('mem_swappiness'),
- group_add=options.get('group_add'),
- userns_mode=options.get('userns_mode'),
- init=options.get('init', None),
- init_path=init_path,
- isolation=options.get('isolation'),
- cpu_count=options.get('cpu_count'),
- cpu_percent=options.get('cpu_percent'),
- nano_cpus=nano_cpus,
- volume_driver=options.get('volume_driver'),
- cpuset_cpus=options.get('cpuset'),
- cpu_shares=options.get('cpu_shares'),
- storage_opt=options.get('storage_opt'),
- blkio_weight=blkio_config.get('weight'),
- blkio_weight_device=blkio_config.get('weight_device'),
- device_read_bps=blkio_config.get('device_read_bps'),
- device_read_iops=blkio_config.get('device_read_iops'),
- device_write_bps=blkio_config.get('device_write_bps'),
- device_write_iops=blkio_config.get('device_write_iops'),
- mounts=options.get('mounts'),
- device_cgroup_rules=options.get('device_cgroup_rules'),
- cpu_period=options.get('cpu_period'),
- cpu_rt_period=options.get('cpu_rt_period'),
- cpu_rt_runtime=options.get('cpu_rt_runtime'),
- )
- def get_secret_volumes(self):
- def build_spec(secret):
- target = secret['secret'].target
- if target is None:
- target = '{}/{}'.format(const.SECRETS_PATH, secret['secret'].source)
- elif not os.path.isabs(target):
- target = '{}/{}'.format(const.SECRETS_PATH, target)
- return MountSpec('bind', secret['file'], target, read_only=True)
- return [build_spec(secret) for secret in self.secrets]
- def build(self, no_cache=False, pull=False, force_rm=False, memory=None, build_args_override=None,
- gzip=False):
- log.info('Building %s' % self.name)
- build_opts = self.options.get('build', {})
- build_args = build_opts.get('args', {}).copy()
- if build_args_override:
- build_args.update(build_args_override)
- for k, v in self._parse_proxy_config().items():
- build_args.setdefault(k, v)
- path = rewrite_build_path(build_opts.get('context'))
- if self.platform and version_lt(self.client.api_version, '1.35'):
- raise OperationFailedError(
- 'Impossible to perform platform-targeted builds for API version < 1.35'
- )
- build_output = self.client.build(
- path=path,
- tag=self.image_name,
- rm=True,
- forcerm=force_rm,
- pull=pull,
- nocache=no_cache,
- dockerfile=build_opts.get('dockerfile', None),
- cache_from=build_opts.get('cache_from', None),
- labels=build_opts.get('labels', None),
- buildargs=build_args,
- network_mode=build_opts.get('network', None),
- target=build_opts.get('target', None),
- shmsize=parse_bytes(build_opts.get('shm_size')) if build_opts.get('shm_size') else None,
- extra_hosts=build_opts.get('extra_hosts', None),
- container_limits={
- 'memory': parse_bytes(memory) if memory else None
- },
- gzip=gzip,
- isolation=build_opts.get('isolation', self.options.get('isolation', None)),
- platform=self.platform,
- )
- try:
- all_events = list(stream_output(build_output, sys.stdout))
- except StreamOutputError as e:
- raise BuildError(self, six.text_type(e))
- # Ensure the HTTP connection is not reused for another
- # streaming command, as the Docker daemon can sometimes
- # complain about it
- self.client.close()
- image_id = None
- for event in all_events:
- if 'stream' in event:
- match = re.search(r'Successfully built ([0-9a-f]+)', event.get('stream', ''))
- if match:
- image_id = match.group(1)
- if image_id is None:
- raise BuildError(self, event if all_events else 'Unknown')
- return image_id
- def can_be_built(self):
- return 'build' in self.options
- def labels(self, one_off=False, legacy=False):
- proj_name = self.project if not legacy else re.sub(r'[_-]', '', self.project)
- return [
- '{0}={1}'.format(LABEL_PROJECT, proj_name),
- '{0}={1}'.format(LABEL_SERVICE, self.name),
- '{0}={1}'.format(LABEL_ONE_OFF, "True" if one_off else "False"),
- ]
- @property
- def custom_container_name(self):
- return self.options.get('container_name')
- def get_container_name(self, service_name, number, slug=None):
- if self.custom_container_name and slug is None:
- return self.custom_container_name
- container_name = build_container_name(
- self.project, service_name, number, slug,
- )
- ext_links_origins = [l.split(':')[0] for l in self.options.get('external_links', [])]
- if container_name in ext_links_origins:
- raise DependencyError(
- 'Service {0} has a self-referential external link: {1}'.format(
- self.name, container_name
- )
- )
- return container_name
- def remove_image(self, image_type):
- if not image_type or image_type == ImageType.none:
- return False
- if image_type == ImageType.local and self.options.get('image'):
- return False
- log.info("Removing image %s", self.image_name)
- try:
- self.client.remove_image(self.image_name)
- return True
- except ImageNotFound:
- log.warning("Image %s not found.", self.image_name)
- return False
- except APIError as e:
- log.error("Failed to remove image for service %s: %s", self.name, e)
- return False
- def specifies_host_port(self):
- def has_host_port(binding):
- if isinstance(binding, dict):
- external_bindings = binding.get('published')
- else:
- _, external_bindings = split_port(binding)
- # there are no external bindings
- if external_bindings is None:
- return False
- # we only need to check the first binding from the range
- external_binding = external_bindings[0]
- # non-tuple binding means there is a host port specified
- if not isinstance(external_binding, tuple):
- return True
- # extract actual host port from tuple of (host_ip, host_port)
- _, host_port = external_binding
- if host_port is not None:
- return True
- return False
- return any(has_host_port(binding) for binding in self.options.get('ports', []))
- def _do_pull(self, repo, pull_kwargs, silent, ignore_pull_failures):
- try:
- output = self.client.pull(repo, **pull_kwargs)
- if silent:
- with open(os.devnull, 'w') as devnull:
- for event in stream_output(output, devnull):
- yield event
- else:
- for event in stream_output(output, sys.stdout):
- yield event
- except (StreamOutputError, NotFound) as e:
- if not ignore_pull_failures:
- raise
- else:
- log.error(six.text_type(e))
- def pull(self, ignore_pull_failures=False, silent=False, stream=False):
- if 'image' not in self.options:
- return
- repo, tag, separator = parse_repository_tag(self.options['image'])
- kwargs = {
- 'tag': tag or 'latest',
- 'stream': True,
- 'platform': self.platform,
- }
- if not silent:
- log.info('Pulling %s (%s%s%s)...' % (self.name, repo, separator, tag))
- if kwargs['platform'] and version_lt(self.client.api_version, '1.35'):
- raise OperationFailedError(
- 'Impossible to perform platform-targeted pulls for API version < 1.35'
- )
- event_stream = self._do_pull(repo, kwargs, silent, ignore_pull_failures)
- if stream:
- return event_stream
- return progress_stream.get_digest_from_pull(event_stream)
- def push(self, ignore_push_failures=False):
- if 'image' not in self.options or 'build' not in self.options:
- return
- repo, tag, separator = parse_repository_tag(self.options['image'])
- tag = tag or 'latest'
- log.info('Pushing %s (%s%s%s)...' % (self.name, repo, separator, tag))
- output = self.client.push(repo, tag=tag, stream=True)
- try:
- return progress_stream.get_digest_from_push(
- stream_output(output, sys.stdout))
- except StreamOutputError as e:
- if not ignore_push_failures:
- raise
- else:
- log.error(six.text_type(e))
- def is_healthy(self):
- """ Check that all containers for this service report healthy.
- Returns false if at least one healthcheck is pending.
- If an unhealthy container is detected, raise a HealthCheckFailed
- exception.
- """
- result = True
- for ctnr in self.containers():
- ctnr.inspect()
- status = ctnr.get('State.Health.Status')
- if status is None:
- raise NoHealthCheckConfigured(self.name)
- elif status == 'starting':
- result = False
- elif status == 'unhealthy':
- raise HealthCheckFailed(ctnr.short_id)
- return result
- def _parse_proxy_config(self):
- client = self.client
- if 'proxies' not in client._general_configs:
- return {}
- docker_host = getattr(client, '_original_base_url', client.base_url)
- proxy_config = client._general_configs['proxies'].get(
- docker_host, client._general_configs['proxies'].get('default')
- ) or {}
- permitted = {
- 'ftpProxy': 'FTP_PROXY',
- 'httpProxy': 'HTTP_PROXY',
- 'httpsProxy': 'HTTPS_PROXY',
- 'noProxy': 'NO_PROXY',
- }
- result = {}
- for k, v in proxy_config.items():
- if k not in permitted:
- continue
- result[permitted[k]] = result[permitted[k].lower()] = v
- return result
- def short_id_alias_exists(container, network):
- aliases = container.get(
- 'NetworkSettings.Networks.{net}.Aliases'.format(net=network)) or ()
- return container.short_id in aliases
- class PidMode(object):
- def __init__(self, mode):
- self._mode = mode
- @property
- def mode(self):
- return self._mode
- @property
- def service_name(self):
- return None
- class ServicePidMode(PidMode):
- def __init__(self, service):
- self.service = service
- @property
- def service_name(self):
- return self.service.name
- @property
- def mode(self):
- containers = self.service.containers()
- if containers:
- return 'container:' + containers[0].id
- log.warn(
- "Service %s is trying to use reuse the PID namespace "
- "of another service that is not running." % (self.service_name)
- )
- return None
- class ContainerPidMode(PidMode):
- def __init__(self, container):
- self.container = container
- self._mode = 'container:{}'.format(container.id)
- class NetworkMode(object):
- """A `standard` network mode (ex: host, bridge)"""
- service_name = None
- def __init__(self, network_mode):
- self.network_mode = network_mode
- @property
- def id(self):
- return self.network_mode
- mode = id
- class ContainerNetworkMode(object):
- """A network mode that uses a container's network stack."""
- service_name = None
- def __init__(self, container):
- self.container = container
- @property
- def id(self):
- return self.container.id
- @property
- def mode(self):
- return 'container:' + self.container.id
- class ServiceNetworkMode(object):
- """A network mode that uses a service's network stack."""
- def __init__(self, service):
- self.service = service
- @property
- def id(self):
- return self.service.name
- service_name = id
- @property
- def mode(self):
- containers = self.service.containers()
- if containers:
- return 'container:' + containers[0].id
- log.warn("Service %s is trying to use reuse the network stack "
- "of another service that is not running." % (self.id))
- return None
- # Names
- def build_container_name(project, service, number, slug=None):
- bits = [project.lstrip('-_'), service]
- if slug:
- bits.extend(['run', truncate_id(slug)])
- else:
- bits.append(str(number))
- return '_'.join(bits)
- # Images
- def parse_repository_tag(repo_path):
- """Splits image identification into base image path, tag/digest
- and it's separator.
- Example:
- >>> parse_repository_tag('user/repo@sha256:digest')
- ('user/repo', 'sha256:digest', '@')
- >>> parse_repository_tag('user/repo:v1')
- ('user/repo', 'v1', ':')
- """
- tag_separator = ":"
- digest_separator = "@"
- if digest_separator in repo_path:
- repo, tag = repo_path.rsplit(digest_separator, 1)
- return repo, tag, digest_separator
- repo, tag = repo_path, ""
- if tag_separator in repo_path:
- repo, tag = repo_path.rsplit(tag_separator, 1)
- if "/" in tag:
- repo, tag = repo_path, ""
- return repo, tag, tag_separator
- # Volumes
- def merge_volume_bindings(volumes, tmpfs, previous_container, mounts):
- """
- Return a list of volume bindings for a container. Container data volumes
- are replaced by those from the previous container.
- Anonymous mounts are updated in place.
- """
- affinity = {}
- volume_bindings = OrderedDict(
- build_volume_binding(volume)
- for volume in volumes
- if volume.external
- )
- if previous_container:
- old_volumes, old_mounts = get_container_data_volumes(
- previous_container, volumes, tmpfs, mounts
- )
- warn_on_masked_volume(volumes, old_volumes, previous_container.service)
- volume_bindings.update(
- build_volume_binding(volume) for volume in old_volumes
- )
- if old_volumes or old_mounts:
- affinity = {'affinity:container': '=' + previous_container.id}
- return list(volume_bindings.values()), affinity
- def get_container_data_volumes(container, volumes_option, tmpfs_option, mounts_option):
- """
- Find the container data volumes that are in `volumes_option`, and return
- a mapping of volume bindings for those volumes.
- Anonymous volume mounts are updated in place instead.
- """
- volumes = []
- volumes_option = volumes_option or []
- container_mounts = dict(
- (mount['Destination'], mount)
- for mount in container.get('Mounts') or {}
- )
- image_volumes = [
- VolumeSpec.parse(volume)
- for volume in
- container.image_config['ContainerConfig'].get('Volumes') or {}
- ]
- for volume in set(volumes_option + image_volumes):
- # No need to preserve host volumes
- if volume.external:
- continue
- # Attempting to rebind tmpfs volumes breaks: https://github.com/docker/compose/issues/4751
- if volume.internal in convert_tmpfs_mounts(tmpfs_option).keys():
- continue
- mount = container_mounts.get(volume.internal)
- # New volume, doesn't exist in the old container
- if not mount:
- continue
- # Volume was previously a host volume, now it's a container volume
- if not mount.get('Name'):
- continue
- # Volume (probably an image volume) is overridden by a mount in the service's config
- # and would cause a duplicate mountpoint error
- if volume.internal in [m.target for m in mounts_option]:
- continue
- # Copy existing volume from old container
- volume = volume._replace(external=mount['Name'])
- volumes.append(volume)
- updated_mounts = False
- for mount in mounts_option:
- if mount.type != 'volume':
- continue
- ctnr_mount = container_mounts.get(mount.target)
- if not ctnr_mount or not ctnr_mount.get('Name'):
- continue
- mount.source = ctnr_mount['Name']
- updated_mounts = True
- return volumes, updated_mounts
- def warn_on_masked_volume(volumes_option, container_volumes, service):
- container_volumes = dict(
- (volume.internal, volume.external)
- for volume in container_volumes)
- for volume in volumes_option:
- if (
- volume.external and
- volume.internal in container_volumes and
- container_volumes.get(volume.internal) != volume.external
- ):
- log.warn((
- "Service \"{service}\" is using volume \"{volume}\" from the "
- "previous container. Host mapping \"{host_path}\" has no effect. "
- "Remove the existing containers (with `docker-compose rm {service}`) "
- "to use the host volume mapping."
- ).format(
- service=service,
- volume=volume.internal,
- host_path=volume.external))
- def build_volume_binding(volume_spec):
- return volume_spec.internal, volume_spec.repr()
- def build_volume_from(volume_from_spec):
- """
- volume_from can be either a service or a container. We want to return the
- container.id and format it into a string complete with the mode.
- """
- if isinstance(volume_from_spec.source, Service):
- containers = volume_from_spec.source.containers(stopped=True)
- if not containers:
- return "{}:{}".format(
- volume_from_spec.source.create_container().id,
- volume_from_spec.mode)
- container = containers[0]
- return "{}:{}".format(container.id, volume_from_spec.mode)
- elif isinstance(volume_from_spec.source, Container):
- return "{}:{}".format(volume_from_spec.source.id, volume_from_spec.mode)
- def build_mount(mount_spec):
- kwargs = {}
- if mount_spec.options:
- for option, sdk_name in mount_spec.options_map[mount_spec.type].items():
- if option in mount_spec.options:
- kwargs[sdk_name] = mount_spec.options[option]
- return Mount(
- type=mount_spec.type, target=mount_spec.target, source=mount_spec.source,
- read_only=mount_spec.read_only, consistency=mount_spec.consistency, **kwargs
- )
- # Labels
- def build_container_labels(label_options, service_labels, number, config_hash, slug):
- labels = dict(label_options or {})
- labels.update(label.split('=', 1) for label in service_labels)
- if number is not None:
- labels[LABEL_CONTAINER_NUMBER] = str(number)
- if slug is not None:
- labels[LABEL_SLUG] = slug
- labels[LABEL_VERSION] = __version__
- if config_hash:
- log.debug("Added config hash: %s" % config_hash)
- labels[LABEL_CONFIG_HASH] = config_hash
- return labels
- # Ulimits
- def build_ulimits(ulimit_config):
- if not ulimit_config:
- return None
- ulimits = []
- for limit_name, soft_hard_values in six.iteritems(ulimit_config):
- if isinstance(soft_hard_values, six.integer_types):
- ulimits.append({'name': limit_name, 'soft': soft_hard_values, 'hard': soft_hard_values})
- elif isinstance(soft_hard_values, dict):
- ulimit_dict = {'name': limit_name}
- ulimit_dict.update(soft_hard_values)
- ulimits.append(ulimit_dict)
- return ulimits
- def get_log_config(logging_dict):
- log_driver = logging_dict.get('driver', "") if logging_dict else ""
- log_options = logging_dict.get('options', None) if logging_dict else None
- return LogConfig(
- type=log_driver,
- config=log_options
- )
- # TODO: remove once fix is available in docker-py
- def format_environment(environment):
- def format_env(key, value):
- if value is None:
- return key
- if isinstance(value, six.binary_type):
- value = value.decode('utf-8')
- return '{key}={value}'.format(key=key, value=value)
- return [format_env(*item) for item in environment.items()]
- # Ports
- def formatted_ports(ports):
- result = []
- for port in ports:
- if isinstance(port, ServicePort):
- result.append(port.legacy_repr())
- else:
- result.append(port)
- return result
- def build_container_ports(container_ports, options):
- ports = []
- all_ports = container_ports + options.get('expose', [])
- for port_range in all_ports:
- internal_range, _ = split_port(port_range)
- for port in internal_range:
- port = str(port)
- if '/' in port:
- port = tuple(port.split('/'))
- ports.append(port)
- return ports
- def convert_blkio_config(blkio_config):
- result = {}
- if blkio_config is None:
- return result
- result['weight'] = blkio_config.get('weight')
- for field in [
- "device_read_bps", "device_read_iops", "device_write_bps",
- "device_write_iops", "weight_device",
- ]:
- if field not in blkio_config:
- continue
- arr = []
- for item in blkio_config[field]:
- arr.append(dict([(k.capitalize(), v) for k, v in item.items()]))
- result[field] = arr
- return result
- def rewrite_build_path(path):
- # python2 os.stat() doesn't support unicode on some UNIX, so we
- # encode it to a bytestring to be safe
- if not six.PY3 and not IS_WINDOWS_PLATFORM:
- path = path.encode('utf8')
- if IS_WINDOWS_PLATFORM and not is_url(path) and not path.startswith(WINDOWS_LONGPATH_PREFIX):
- path = WINDOWS_LONGPATH_PREFIX + os.path.normpath(path)
- return path
|