mirror of
https://github.com/containers/podman-compose.git
synced 2025-04-04 05:51:09 +02:00
formatting with black so pre-commit does not fail
This commit is contained in:
parent
888be2383e
commit
a5c80a6cc7
@ -7,7 +7,8 @@
|
|||||||
# https://docs.docker.com/compose/django/
|
# https://docs.docker.com/compose/django/
|
||||||
# https://docs.docker.com/compose/wordpress/
|
# https://docs.docker.com/compose/wordpress/
|
||||||
# TODO: podman pod logs --color -n -f pod_testlogs
|
# TODO: podman pod logs --color -n -f pod_testlogs
|
||||||
from __future__ import annotations # If you see an error here, use Python 3.7 or greater
|
# If you see an error here, use Python 3.7 or greater
|
||||||
|
from __future__ import annotations
|
||||||
|
|
||||||
import argparse
|
import argparse
|
||||||
import asyncio.exceptions
|
import asyncio.exceptions
|
||||||
@ -160,8 +161,10 @@ def parse_short_mount(mount_str, basedir):
|
|||||||
# User-relative path
|
# User-relative path
|
||||||
# - ~/configs:/etc/configs/:ro
|
# - ~/configs:/etc/configs/:ro
|
||||||
mount_type = "bind"
|
mount_type = "bind"
|
||||||
if os.name != 'nt' or (os.name == 'nt' and ".sock" not in mount_src):
|
if os.name != "nt" or (os.name == "nt" and ".sock" not in mount_src):
|
||||||
mount_src = os.path.abspath(os.path.join(basedir, os.path.expanduser(mount_src)))
|
mount_src = os.path.abspath(
|
||||||
|
os.path.join(basedir, os.path.expanduser(mount_src))
|
||||||
|
)
|
||||||
else:
|
else:
|
||||||
# Named volume
|
# Named volume
|
||||||
# - datavolume:/var/lib/mysql
|
# - datavolume:/var/lib/mysql
|
||||||
@ -214,11 +217,13 @@ def fix_mount_dict(compose, mount_dict, srv_name):
|
|||||||
# handle anonymous or implied volume
|
# handle anonymous or implied volume
|
||||||
if not source:
|
if not source:
|
||||||
# missing source
|
# missing source
|
||||||
vol["name"] = "_".join([
|
vol["name"] = "_".join(
|
||||||
compose.project_name,
|
[
|
||||||
srv_name,
|
compose.project_name,
|
||||||
hashlib.sha256(mount_dict["target"].encode("utf-8")).hexdigest(),
|
srv_name,
|
||||||
])
|
hashlib.sha256(mount_dict["target"].encode("utf-8")).hexdigest(),
|
||||||
|
]
|
||||||
|
)
|
||||||
elif not name:
|
elif not name:
|
||||||
external = vol.get("external")
|
external = vol.get("external")
|
||||||
if isinstance(external, dict):
|
if isinstance(external, dict):
|
||||||
@ -265,10 +270,12 @@ def rec_subs(value, subs_dict):
|
|||||||
do bash-like substitution in value and if list of dictionary do that recursively
|
do bash-like substitution in value and if list of dictionary do that recursively
|
||||||
"""
|
"""
|
||||||
if isinstance(value, dict):
|
if isinstance(value, dict):
|
||||||
if 'environment' in value and isinstance(value['environment'], dict):
|
if "environment" in value and isinstance(value["environment"], dict):
|
||||||
# Load service's environment variables
|
# Load service's environment variables
|
||||||
subs_dict = subs_dict.copy()
|
subs_dict = subs_dict.copy()
|
||||||
svc_envs = {k: v for k, v in value['environment'].items() if k not in subs_dict}
|
svc_envs = {
|
||||||
|
k: v for k, v in value["environment"].items() if k not in subs_dict
|
||||||
|
}
|
||||||
# we need to add `svc_envs` to the `subs_dict` so that it can evaluate the
|
# we need to add `svc_envs` to the `subs_dict` so that it can evaluate the
|
||||||
# service environment that reference to another service environment.
|
# service environment that reference to another service environment.
|
||||||
svc_envs = rec_subs(svc_envs, subs_dict)
|
svc_envs = rec_subs(svc_envs, subs_dict)
|
||||||
@ -389,7 +396,9 @@ async def assert_volume(compose, mount_dict):
|
|||||||
if mount_dict["type"] == "bind":
|
if mount_dict["type"] == "bind":
|
||||||
basedir = os.path.realpath(compose.dirname)
|
basedir = os.path.realpath(compose.dirname)
|
||||||
mount_src = mount_dict["source"]
|
mount_src = mount_dict["source"]
|
||||||
mount_src = os.path.realpath(os.path.join(basedir, os.path.expanduser(mount_src)))
|
mount_src = os.path.realpath(
|
||||||
|
os.path.join(basedir, os.path.expanduser(mount_src))
|
||||||
|
)
|
||||||
if not os.path.exists(mount_src):
|
if not os.path.exists(mount_src):
|
||||||
try:
|
try:
|
||||||
os.makedirs(mount_src, exist_ok=True)
|
os.makedirs(mount_src, exist_ok=True)
|
||||||
@ -405,7 +414,9 @@ async def assert_volume(compose, mount_dict):
|
|||||||
# podman volume list --format '{{.Name}}\t{{.MountPoint}}' \
|
# podman volume list --format '{{.Name}}\t{{.MountPoint}}' \
|
||||||
# -f 'label=io.podman.compose.project=HERE'
|
# -f 'label=io.podman.compose.project=HERE'
|
||||||
try:
|
try:
|
||||||
_ = (await compose.podman.output([], "volume", ["inspect", vol_name])).decode("utf-8")
|
_ = (await compose.podman.output([], "volume", ["inspect", vol_name])).decode(
|
||||||
|
"utf-8"
|
||||||
|
)
|
||||||
except subprocess.CalledProcessError as e:
|
except subprocess.CalledProcessError as e:
|
||||||
if is_ext:
|
if is_ext:
|
||||||
raise RuntimeError(f"External volume [{vol_name}] does not exists") from e
|
raise RuntimeError(f"External volume [{vol_name}] does not exists") from e
|
||||||
@ -427,10 +438,14 @@ async def assert_volume(compose, mount_dict):
|
|||||||
args.extend(["--opt", f"{opt}={value}"])
|
args.extend(["--opt", f"{opt}={value}"])
|
||||||
args.append(vol_name)
|
args.append(vol_name)
|
||||||
await compose.podman.output([], "volume", args)
|
await compose.podman.output([], "volume", args)
|
||||||
_ = (await compose.podman.output([], "volume", ["inspect", vol_name])).decode("utf-8")
|
_ = (await compose.podman.output([], "volume", ["inspect", vol_name])).decode(
|
||||||
|
"utf-8"
|
||||||
|
)
|
||||||
|
|
||||||
|
|
||||||
def mount_desc_to_mount_args(compose, mount_desc, srv_name, cnt_name): # pylint: disable=unused-argument
|
def mount_desc_to_mount_args(
|
||||||
|
compose, mount_desc, srv_name, cnt_name
|
||||||
|
): # pylint: disable=unused-argument
|
||||||
mount_type = mount_desc.get("type")
|
mount_type = mount_desc.get("type")
|
||||||
vol = mount_desc.get("_vol") if mount_type == "volume" else None
|
vol = mount_desc.get("_vol") if mount_type == "volume" else None
|
||||||
source = vol["name"] if vol else mount_desc.get("source")
|
source = vol["name"] if vol else mount_desc.get("source")
|
||||||
@ -493,7 +508,9 @@ def container_to_ulimit_build_args(cnt, podman_args):
|
|||||||
ulimit_to_ulimit_args(build.get("ulimits", []), podman_args)
|
ulimit_to_ulimit_args(build.get("ulimits", []), podman_args)
|
||||||
|
|
||||||
|
|
||||||
def mount_desc_to_volume_args(compose, mount_desc, srv_name, cnt_name): # pylint: disable=unused-argument
|
def mount_desc_to_volume_args(
|
||||||
|
compose, mount_desc, srv_name, cnt_name
|
||||||
|
): # pylint: disable=unused-argument
|
||||||
mount_type = mount_desc["type"]
|
mount_type = mount_desc["type"]
|
||||||
if mount_type not in ("bind", "volume"):
|
if mount_type not in ("bind", "volume"):
|
||||||
raise ValueError("unknown mount type:" + mount_type)
|
raise ValueError("unknown mount type:" + mount_type)
|
||||||
@ -504,9 +521,13 @@ def mount_desc_to_volume_args(compose, mount_desc, srv_name, cnt_name): # pylin
|
|||||||
target = mount_desc["target"]
|
target = mount_desc["target"]
|
||||||
opts = []
|
opts = []
|
||||||
|
|
||||||
propagations = set(filteri(mount_desc.get(mount_type, {}).get("propagation", "").split(",")))
|
propagations = set(
|
||||||
|
filteri(mount_desc.get(mount_type, {}).get("propagation", "").split(","))
|
||||||
|
)
|
||||||
if mount_type != "bind":
|
if mount_type != "bind":
|
||||||
propagations.update(filteri(mount_desc.get("bind", {}).get("propagation", "").split(",")))
|
propagations.update(
|
||||||
|
filteri(mount_desc.get("bind", {}).get("propagation", "").split(","))
|
||||||
|
)
|
||||||
opts.extend(propagations)
|
opts.extend(propagations)
|
||||||
# --volume, -v[=[[SOURCE-VOLUME|HOST-DIR:]CONTAINER-DIR[:OPTIONS]]]
|
# --volume, -v[=[[SOURCE-VOLUME|HOST-DIR:]CONTAINER-DIR[:OPTIONS]]]
|
||||||
# [rw|ro]
|
# [rw|ro]
|
||||||
@ -574,7 +595,9 @@ def get_secret_args(compose, cnt, secret, podman_is_building=False):
|
|||||||
"""
|
"""
|
||||||
secret_name = secret if isinstance(secret, str) else secret.get("source")
|
secret_name = secret if isinstance(secret, str) else secret.get("source")
|
||||||
if not secret_name or secret_name not in compose.declared_secrets.keys():
|
if not secret_name or secret_name not in compose.declared_secrets.keys():
|
||||||
raise ValueError(f'ERROR: undeclared secret: "{secret}", service: {cnt["_service"]}')
|
raise ValueError(
|
||||||
|
f'ERROR: undeclared secret: "{secret}", service: {cnt["_service"]}'
|
||||||
|
)
|
||||||
declared_secret = compose.declared_secrets[secret_name]
|
declared_secret = compose.declared_secrets[secret_name]
|
||||||
|
|
||||||
source_file = declared_secret.get("file")
|
source_file = declared_secret.get("file")
|
||||||
@ -596,7 +619,9 @@ def get_secret_args(compose, cnt, secret, podman_is_building=False):
|
|||||||
if source_file:
|
if source_file:
|
||||||
# assemble path for source file first, because we need it for all cases
|
# assemble path for source file first, because we need it for all cases
|
||||||
basedir = compose.dirname
|
basedir = compose.dirname
|
||||||
source_file = os.path.realpath(os.path.join(basedir, os.path.expanduser(source_file)))
|
source_file = os.path.realpath(
|
||||||
|
os.path.join(basedir, os.path.expanduser(source_file))
|
||||||
|
)
|
||||||
|
|
||||||
if podman_is_building:
|
if podman_is_building:
|
||||||
# pass file secrets to "podman build" with param --secret
|
# pass file secrets to "podman build" with param --secret
|
||||||
@ -642,7 +667,9 @@ def get_secret_args(compose, cnt, secret, podman_is_building=False):
|
|||||||
secret_opts += f",gid={secret_gid}" if secret_gid else ""
|
secret_opts += f",gid={secret_gid}" if secret_gid else ""
|
||||||
secret_opts += f",mode={secret_mode}" if secret_mode else ""
|
secret_opts += f",mode={secret_mode}" if secret_mode else ""
|
||||||
secret_opts += f",type={secret_type}" if secret_type else ""
|
secret_opts += f",type={secret_type}" if secret_type else ""
|
||||||
secret_opts += f",target={secret_target}" if secret_target and secret_type == "env" else ""
|
secret_opts += (
|
||||||
|
f",target={secret_target}" if secret_target and secret_type == "env" else ""
|
||||||
|
)
|
||||||
# The target option is only valid for type=env,
|
# The target option is only valid for type=env,
|
||||||
# which in an ideal world would work
|
# which in an ideal world would work
|
||||||
# for type=mount as well.
|
# for type=mount as well.
|
||||||
@ -655,9 +682,9 @@ def get_secret_args(compose, cnt, secret, podman_is_building=False):
|
|||||||
)
|
)
|
||||||
if ext_name and ext_name != secret_name:
|
if ext_name and ext_name != secret_name:
|
||||||
raise ValueError(err_str.format(secret_name, ext_name))
|
raise ValueError(err_str.format(secret_name, ext_name))
|
||||||
if secret_target and secret_target != secret_name and secret_type != 'env':
|
if secret_target and secret_target != secret_name and secret_type != "env":
|
||||||
raise ValueError(err_str.format(secret_target, secret_name))
|
raise ValueError(err_str.format(secret_target, secret_name))
|
||||||
if secret_target and secret_type != 'env':
|
if secret_target and secret_type != "env":
|
||||||
log.warning(
|
log.warning(
|
||||||
'WARNING: Service "%s" uses target: "%s" for secret: "%s".'
|
'WARNING: Service "%s" uses target: "%s" for secret: "%s".'
|
||||||
+ " That is un-supported and a no-op and is ignored.",
|
+ " That is un-supported and a no-op and is ignored.",
|
||||||
@ -668,7 +695,9 @@ def get_secret_args(compose, cnt, secret, podman_is_building=False):
|
|||||||
return ["--secret", "{}{}".format(secret_name, secret_opts)]
|
return ["--secret", "{}{}".format(secret_name, secret_opts)]
|
||||||
|
|
||||||
raise ValueError(
|
raise ValueError(
|
||||||
'ERROR: unparsable secret: "{}", service: "{}"'.format(secret_name, cnt["_service"])
|
'ERROR: unparsable secret: "{}", service: "{}"'.format(
|
||||||
|
secret_name, cnt["_service"]
|
||||||
|
)
|
||||||
)
|
)
|
||||||
|
|
||||||
|
|
||||||
@ -702,26 +731,32 @@ def container_to_gpu_res_args(cnt, podman_args):
|
|||||||
device_ids = device.get("device_ids", "all")
|
device_ids = device.get("device_ids", "all")
|
||||||
if device_ids != "all" and len(device_ids) > 0:
|
if device_ids != "all" and len(device_ids) > 0:
|
||||||
for device_id in device_ids:
|
for device_id in device_ids:
|
||||||
podman_args.extend((
|
podman_args.extend(
|
||||||
"--device",
|
(
|
||||||
f"nvidia.com/gpu={device_id}",
|
"--device",
|
||||||
))
|
f"nvidia.com/gpu={device_id}",
|
||||||
|
)
|
||||||
|
)
|
||||||
gpu_on = True
|
gpu_on = True
|
||||||
continue
|
continue
|
||||||
|
|
||||||
if count != "all":
|
if count != "all":
|
||||||
for device_id in range(count):
|
for device_id in range(count):
|
||||||
podman_args.extend((
|
podman_args.extend(
|
||||||
"--device",
|
(
|
||||||
f"nvidia.com/gpu={device_id}",
|
"--device",
|
||||||
))
|
f"nvidia.com/gpu={device_id}",
|
||||||
|
)
|
||||||
|
)
|
||||||
gpu_on = True
|
gpu_on = True
|
||||||
continue
|
continue
|
||||||
|
|
||||||
podman_args.extend((
|
podman_args.extend(
|
||||||
"--device",
|
(
|
||||||
"nvidia.com/gpu=all",
|
"--device",
|
||||||
))
|
"nvidia.com/gpu=all",
|
||||||
|
)
|
||||||
|
)
|
||||||
gpu_on = True
|
gpu_on = True
|
||||||
|
|
||||||
if gpu_on:
|
if gpu_on:
|
||||||
@ -749,27 +784,35 @@ def container_to_cpu_res_args(cnt, podman_args):
|
|||||||
# add args
|
# add args
|
||||||
cpus = cpus_limit_v3 or cpus_limit_v2
|
cpus = cpus_limit_v3 or cpus_limit_v2
|
||||||
if cpus:
|
if cpus:
|
||||||
podman_args.extend((
|
podman_args.extend(
|
||||||
"--cpus",
|
(
|
||||||
str(cpus),
|
"--cpus",
|
||||||
))
|
str(cpus),
|
||||||
|
)
|
||||||
|
)
|
||||||
if cpu_shares_v2:
|
if cpu_shares_v2:
|
||||||
podman_args.extend((
|
podman_args.extend(
|
||||||
"--cpu-shares",
|
(
|
||||||
str(cpu_shares_v2),
|
"--cpu-shares",
|
||||||
))
|
str(cpu_shares_v2),
|
||||||
|
)
|
||||||
|
)
|
||||||
mem = mem_limit_v3 or mem_limit_v2
|
mem = mem_limit_v3 or mem_limit_v2
|
||||||
if mem:
|
if mem:
|
||||||
podman_args.extend((
|
podman_args.extend(
|
||||||
"-m",
|
(
|
||||||
str(mem).lower(),
|
"-m",
|
||||||
))
|
str(mem).lower(),
|
||||||
|
)
|
||||||
|
)
|
||||||
mem_res = mem_res_v3 or mem_res_v2
|
mem_res = mem_res_v3 or mem_res_v2
|
||||||
if mem_res:
|
if mem_res:
|
||||||
podman_args.extend((
|
podman_args.extend(
|
||||||
"--memory-reservation",
|
(
|
||||||
str(mem_res).lower(),
|
"--memory-reservation",
|
||||||
))
|
str(mem_res).lower(),
|
||||||
|
)
|
||||||
|
)
|
||||||
|
|
||||||
|
|
||||||
def port_dict_to_str(port_desc):
|
def port_dict_to_str(port_desc):
|
||||||
@ -836,10 +879,12 @@ def get_network_create_args(net_desc, proj_name, net_name):
|
|||||||
if net_desc.get("x-podman.disable_dns"):
|
if net_desc.get("x-podman.disable_dns"):
|
||||||
args.append("--disable-dns")
|
args.append("--disable-dns")
|
||||||
if net_desc.get("x-podman.dns"):
|
if net_desc.get("x-podman.dns"):
|
||||||
args.extend((
|
args.extend(
|
||||||
"--dns",
|
(
|
||||||
",".join(norm_as_list(net_desc.get("x-podman.dns"))),
|
"--dns",
|
||||||
))
|
",".join(norm_as_list(net_desc.get("x-podman.dns"))),
|
||||||
|
)
|
||||||
|
)
|
||||||
|
|
||||||
if isinstance(ipam_config_ls, dict):
|
if isinstance(ipam_config_ls, dict):
|
||||||
ipam_config_ls = [ipam_config_ls]
|
ipam_config_ls = [ipam_config_ls]
|
||||||
@ -880,7 +925,9 @@ async def assert_cnt_nets(compose, cnt):
|
|||||||
await compose.podman.output([], "network", ["exists", net_name])
|
await compose.podman.output([], "network", ["exists", net_name])
|
||||||
except subprocess.CalledProcessError as e:
|
except subprocess.CalledProcessError as e:
|
||||||
if is_ext:
|
if is_ext:
|
||||||
raise RuntimeError(f"External network [{net_name}] does not exists") from e
|
raise RuntimeError(
|
||||||
|
f"External network [{net_name}] does not exists"
|
||||||
|
) from e
|
||||||
args = get_network_create_args(net_desc, compose.project_name, net_name)
|
args = get_network_create_args(net_desc, compose.project_name, net_name)
|
||||||
await compose.podman.output([], "network", args)
|
await compose.podman.output([], "network", args)
|
||||||
await compose.podman.output([], "network", ["exists", net_name])
|
await compose.podman.output([], "network", ["exists", net_name])
|
||||||
@ -968,13 +1015,17 @@ def get_net_args_from_networks(compose, cnt):
|
|||||||
if is_list(multiple_nets):
|
if is_list(multiple_nets):
|
||||||
multiple_nets = {net: {} for net in multiple_nets}
|
multiple_nets = {net: {} for net in multiple_nets}
|
||||||
else:
|
else:
|
||||||
multiple_nets = {net: net_config or {} for net, net_config in multiple_nets.items()}
|
multiple_nets = {
|
||||||
|
net: net_config or {} for net, net_config in multiple_nets.items()
|
||||||
|
}
|
||||||
|
|
||||||
# if a mac_address was specified on the container level, we need to check that it is not
|
# if a mac_address was specified on the container level, we need to check that it is not
|
||||||
# specified on the network level as well
|
# specified on the network level as well
|
||||||
if mac_address is not None:
|
if mac_address is not None:
|
||||||
for net_config in multiple_nets.values():
|
for net_config in multiple_nets.values():
|
||||||
network_mac = net_config.get("mac_address", net_config.get("x-podman.mac_address"))
|
network_mac = net_config.get(
|
||||||
|
"mac_address", net_config.get("x-podman.mac_address")
|
||||||
|
)
|
||||||
if network_mac is not None:
|
if network_mac is not None:
|
||||||
raise RuntimeError(
|
raise RuntimeError(
|
||||||
f"conflicting mac addresses {mac_address} and {network_mac}:"
|
f"conflicting mac addresses {mac_address} and {network_mac}:"
|
||||||
@ -1198,10 +1249,12 @@ async def container_to_args(compose, cnt, detached=True, no_deps=False):
|
|||||||
# If it's a string, it's equivalent to specifying CMD-SHELL
|
# If it's a string, it's equivalent to specifying CMD-SHELL
|
||||||
if isinstance(healthcheck_test, str):
|
if isinstance(healthcheck_test, str):
|
||||||
# podman does not add shell to handle command with whitespace
|
# podman does not add shell to handle command with whitespace
|
||||||
podman_args.extend([
|
podman_args.extend(
|
||||||
"--healthcheck-command",
|
[
|
||||||
"/bin/sh -c " + cmd_quote(healthcheck_test),
|
"--healthcheck-command",
|
||||||
])
|
"/bin/sh -c " + cmd_quote(healthcheck_test),
|
||||||
|
]
|
||||||
|
)
|
||||||
elif is_list(healthcheck_test):
|
elif is_list(healthcheck_test):
|
||||||
healthcheck_test = healthcheck_test.copy()
|
healthcheck_test = healthcheck_test.copy()
|
||||||
# If it's a list, first item is either NONE, CMD or CMD-SHELL.
|
# If it's a list, first item is either NONE, CMD or CMD-SHELL.
|
||||||
@ -1237,24 +1290,26 @@ async def container_to_args(compose, cnt, detached=True, no_deps=False):
|
|||||||
podman_args.extend(["--healthcheck-retries", str(healthcheck["retries"])])
|
podman_args.extend(["--healthcheck-retries", str(healthcheck["retries"])])
|
||||||
|
|
||||||
# handle podman extension
|
# handle podman extension
|
||||||
if 'x-podman' in cnt:
|
if "x-podman" in cnt:
|
||||||
raise ValueError(
|
raise ValueError(
|
||||||
'Configuration under x-podman has been migrated to x-podman.uidmaps and '
|
"Configuration under x-podman has been migrated to x-podman.uidmaps and "
|
||||||
'x-podman.gidmaps fields'
|
"x-podman.gidmaps fields"
|
||||||
)
|
)
|
||||||
|
|
||||||
rootfs_mode = False
|
rootfs_mode = False
|
||||||
for uidmap in cnt.get('x-podman.uidmaps', []):
|
for uidmap in cnt.get("x-podman.uidmaps", []):
|
||||||
podman_args.extend(["--uidmap", uidmap])
|
podman_args.extend(["--uidmap", uidmap])
|
||||||
for gidmap in cnt.get('x-podman.gidmaps', []):
|
for gidmap in cnt.get("x-podman.gidmaps", []):
|
||||||
podman_args.extend(["--gidmap", gidmap])
|
podman_args.extend(["--gidmap", gidmap])
|
||||||
if cnt.get("x-podman.no_hosts", False):
|
if cnt.get("x-podman.no_hosts", False):
|
||||||
podman_args.extend(["--no-hosts"])
|
podman_args.extend(["--no-hosts"])
|
||||||
rootfs = cnt.get('x-podman.rootfs')
|
rootfs = cnt.get("x-podman.rootfs")
|
||||||
if rootfs is not None:
|
if rootfs is not None:
|
||||||
rootfs_mode = True
|
rootfs_mode = True
|
||||||
podman_args.extend(["--rootfs", rootfs])
|
podman_args.extend(["--rootfs", rootfs])
|
||||||
log.warning("WARNING: x-podman.rootfs and image both specified, image field ignored")
|
log.warning(
|
||||||
|
"WARNING: x-podman.rootfs and image both specified, image field ignored"
|
||||||
|
)
|
||||||
|
|
||||||
if not rootfs_mode:
|
if not rootfs_mode:
|
||||||
podman_args.append(cnt["image"]) # command, ..etc.
|
podman_args.append(cnt["image"]) # command, ..etc.
|
||||||
@ -1296,7 +1351,10 @@ class ServiceDependencyCondition(Enum):
|
|||||||
try:
|
try:
|
||||||
return docker_to_podman_cond[value]
|
return docker_to_podman_cond[value]
|
||||||
except KeyError:
|
except KeyError:
|
||||||
raise ValueError(f"Value '{value}' is not a valid condition for a service dependency") # pylint: disable=raise-missing-from
|
# pylint: disable-next=raise-missing-from
|
||||||
|
raise ValueError(
|
||||||
|
f"Value '{value}' is not a valid condition for a service dependency"
|
||||||
|
)
|
||||||
|
|
||||||
|
|
||||||
class ServiceDependency:
|
class ServiceDependency:
|
||||||
@ -1314,7 +1372,7 @@ class ServiceDependency:
|
|||||||
|
|
||||||
def __hash__(self):
|
def __hash__(self):
|
||||||
# Compute hash based on the frozenset of items to ensure order does not matter
|
# Compute hash based on the frozenset of items to ensure order does not matter
|
||||||
return hash(('name', self._name) + ('condition', self._condition))
|
return hash(("name", self._name) + ("condition", self._condition))
|
||||||
|
|
||||||
def __eq__(self, other):
|
def __eq__(self, other):
|
||||||
# Compare equality based on dictionary content
|
# Compare equality based on dictionary content
|
||||||
@ -1370,7 +1428,9 @@ def flat_deps(services, with_extends=False):
|
|||||||
links_ls = srv.get("links", [])
|
links_ls = srv.get("links", [])
|
||||||
if not is_list(links_ls):
|
if not is_list(links_ls):
|
||||||
links_ls = [links_ls]
|
links_ls = [links_ls]
|
||||||
deps.update([ServiceDependency(c.split(":")[0], "service_started") for c in links_ls])
|
deps.update(
|
||||||
|
[ServiceDependency(c.split(":")[0], "service_started") for c in links_ls]
|
||||||
|
)
|
||||||
for c in links_ls:
|
for c in links_ls:
|
||||||
if ":" in c:
|
if ":" in c:
|
||||||
dep_name, dep_alias = c.split(":")
|
dep_name, dep_alias = c.split(":")
|
||||||
@ -1432,7 +1492,9 @@ class Podman:
|
|||||||
if p.returncode == 0:
|
if p.returncode == 0:
|
||||||
return stdout_data
|
return stdout_data
|
||||||
|
|
||||||
raise subprocess.CalledProcessError(p.returncode, " ".join(cmd_ls), stderr_data)
|
raise subprocess.CalledProcessError(
|
||||||
|
p.returncode, " ".join(cmd_ls), stderr_data
|
||||||
|
)
|
||||||
|
|
||||||
async def _readchunk(self, reader):
|
async def _readchunk(self, reader):
|
||||||
try:
|
try:
|
||||||
@ -1527,7 +1589,9 @@ class Podman:
|
|||||||
err_t.add_done_callback(task_reference.discard)
|
err_t.add_done_callback(task_reference.discard)
|
||||||
|
|
||||||
else:
|
else:
|
||||||
p = await asyncio.create_subprocess_exec(*cmd_ls, close_fds=False) # pylint: disable=consider-using-with
|
p = await asyncio.create_subprocess_exec(
|
||||||
|
*cmd_ls, close_fds=False
|
||||||
|
) # pylint: disable=consider-using-with
|
||||||
|
|
||||||
try:
|
try:
|
||||||
exit_code = await p.wait()
|
exit_code = await p.wait()
|
||||||
@ -1638,7 +1702,7 @@ def normalize_service(service, sub_dir=""):
|
|||||||
# the dependency service_started is set by default
|
# the dependency service_started is set by default
|
||||||
# unless requested otherwise.
|
# unless requested otherwise.
|
||||||
for k, v in deps.items():
|
for k, v in deps.items():
|
||||||
v.setdefault('condition', 'service_started')
|
v.setdefault("condition", "service_started")
|
||||||
service["depends_on"] = deps
|
service["depends_on"] = deps
|
||||||
return service
|
return service
|
||||||
|
|
||||||
@ -1697,13 +1761,17 @@ def rec_merge_one(target, source):
|
|||||||
if not isinstance(value2, type(value)):
|
if not isinstance(value2, type(value)):
|
||||||
value_type = type(value)
|
value_type = type(value)
|
||||||
value2_type = type(value2)
|
value2_type = type(value2)
|
||||||
raise ValueError(f"can't merge value of [{key}] of type {value_type} and {value2_type}")
|
raise ValueError(
|
||||||
|
f"can't merge value of [{key}] of type {value_type} and {value2_type}"
|
||||||
|
)
|
||||||
if is_list(value2):
|
if is_list(value2):
|
||||||
if key == "volumes":
|
if key == "volumes":
|
||||||
# clean duplicate mount targets
|
# clean duplicate mount targets
|
||||||
pts = {v.split(":", 2)[1] for v in value2 if ":" in v}
|
pts = {v.split(":", 2)[1] for v in value2 if ":" in v}
|
||||||
del_ls = [
|
del_ls = [
|
||||||
ix for (ix, v) in enumerate(value) if ":" in v and v.split(":", 2)[1] in pts
|
ix
|
||||||
|
for (ix, v) in enumerate(value)
|
||||||
|
if ":" in v and v.split(":", 2)[1] in pts
|
||||||
]
|
]
|
||||||
for ix in reversed(del_ls):
|
for ix in reversed(del_ls):
|
||||||
del value[ix]
|
del value[ix]
|
||||||
@ -1846,14 +1914,16 @@ class PodmanCompose:
|
|||||||
if args.dry_run is False:
|
if args.dry_run is False:
|
||||||
log.fatal("Binary %s has not been found.", podman_path)
|
log.fatal("Binary %s has not been found.", podman_path)
|
||||||
sys.exit(1)
|
sys.exit(1)
|
||||||
self.podman = Podman(self, podman_path, args.dry_run, asyncio.Semaphore(args.parallel))
|
self.podman = Podman(
|
||||||
|
self, podman_path, args.dry_run, asyncio.Semaphore(args.parallel)
|
||||||
|
)
|
||||||
|
|
||||||
if not args.dry_run:
|
if not args.dry_run:
|
||||||
# just to make sure podman is running
|
# just to make sure podman is running
|
||||||
try:
|
try:
|
||||||
self.podman_version = (await self.podman.output(["--version"], "", [])).decode(
|
self.podman_version = (
|
||||||
"utf-8"
|
await self.podman.output(["--version"], "", [])
|
||||||
).strip() or ""
|
).decode("utf-8").strip() or ""
|
||||||
self.podman_version = (self.podman_version.split() or [""])[-1]
|
self.podman_version = (self.podman_version.split() or [""])[-1]
|
||||||
except subprocess.CalledProcessError:
|
except subprocess.CalledProcessError:
|
||||||
self.podman_version = None
|
self.podman_version = None
|
||||||
@ -1908,7 +1978,7 @@ class PodmanCompose:
|
|||||||
"pass files with -f"
|
"pass files with -f"
|
||||||
)
|
)
|
||||||
sys.exit(-1)
|
sys.exit(-1)
|
||||||
ex = map(lambda x: x == '-' or os.path.exists(x), files)
|
ex = map(lambda x: x == "-" or os.path.exists(x), files)
|
||||||
missing = [fn0 for ex0, fn0 in zip(ex, files) if not ex0]
|
missing = [fn0 for ex0, fn0 in zip(ex, files) if not ex0]
|
||||||
if missing:
|
if missing:
|
||||||
log.fatal("missing files: %s", missing)
|
log.fatal("missing files: %s", missing)
|
||||||
@ -1937,20 +2007,26 @@ class PodmanCompose:
|
|||||||
dotenv_path = os.path.realpath(args.env_file)
|
dotenv_path = os.path.realpath(args.env_file)
|
||||||
dotenv_dict.update(dotenv_to_dict(dotenv_path))
|
dotenv_dict.update(dotenv_to_dict(dotenv_path))
|
||||||
|
|
||||||
os.environ.update({
|
os.environ.update(
|
||||||
key: value for key, value in dotenv_dict.items() if key.startswith("PODMAN_")
|
{
|
||||||
})
|
key: value
|
||||||
|
for key, value in dotenv_dict.items()
|
||||||
|
if key.startswith("PODMAN_")
|
||||||
|
}
|
||||||
|
)
|
||||||
self.environ = dotenv_dict
|
self.environ = dotenv_dict
|
||||||
self.environ.update(dict(os.environ))
|
self.environ.update(dict(os.environ))
|
||||||
# see: https://docs.docker.com/compose/reference/envvars/
|
# see: https://docs.docker.com/compose/reference/envvars/
|
||||||
# see: https://docs.docker.com/compose/env-file/
|
# see: https://docs.docker.com/compose/env-file/
|
||||||
self.environ.update({
|
self.environ.update(
|
||||||
"COMPOSE_PROJECT_DIR": dirname,
|
{
|
||||||
"COMPOSE_FILE": pathsep.join(relative_files),
|
"COMPOSE_PROJECT_DIR": dirname,
|
||||||
"COMPOSE_PATH_SEPARATOR": pathsep,
|
"COMPOSE_FILE": pathsep.join(relative_files),
|
||||||
})
|
"COMPOSE_PATH_SEPARATOR": pathsep,
|
||||||
|
}
|
||||||
|
)
|
||||||
|
|
||||||
if args and 'env' in args and args.env:
|
if args and "env" in args and args.env:
|
||||||
env_vars = norm_as_dict(args.env)
|
env_vars = norm_as_dict(args.env)
|
||||||
self.environ.update(env_vars)
|
self.environ.update(env_vars)
|
||||||
|
|
||||||
@ -1964,7 +2040,7 @@ class PodmanCompose:
|
|||||||
except StopIteration:
|
except StopIteration:
|
||||||
break
|
break
|
||||||
|
|
||||||
if filename.strip().split('/')[-1] == '-':
|
if filename.strip().split("/")[-1] == "-":
|
||||||
content = yaml.safe_load(sys.stdin)
|
content = yaml.safe_load(sys.stdin)
|
||||||
else:
|
else:
|
||||||
with open(filename, "r", encoding="utf-8") as f:
|
with open(filename, "r", encoding="utf-8") as f:
|
||||||
@ -1978,10 +2054,12 @@ class PodmanCompose:
|
|||||||
content = normalize(content)
|
content = normalize(content)
|
||||||
# log(filename, json.dumps(content, indent = 2))
|
# log(filename, json.dumps(content, indent = 2))
|
||||||
content = rec_subs(content, self.environ)
|
content = rec_subs(content, self.environ)
|
||||||
if isinstance(services := content.get('services'), dict):
|
if isinstance(services := content.get("services"), dict):
|
||||||
for service in services.values():
|
for service in services.values():
|
||||||
if 'extends' in service and (service_file := service['extends'].get('file')):
|
if "extends" in service and (
|
||||||
service['extends']['file'] = os.path.join(
|
service_file := service["extends"].get("file")
|
||||||
|
):
|
||||||
|
service["extends"]["file"] = os.path.join(
|
||||||
os.path.dirname(filename), service_file
|
os.path.dirname(filename), service_file
|
||||||
)
|
)
|
||||||
|
|
||||||
@ -1989,14 +2067,18 @@ class PodmanCompose:
|
|||||||
# If `include` is used, append included files to files
|
# If `include` is used, append included files to files
|
||||||
include = compose.get("include")
|
include = compose.get("include")
|
||||||
if include:
|
if include:
|
||||||
files.extend([os.path.join(os.path.dirname(filename), i) for i in include])
|
files.extend(
|
||||||
|
[os.path.join(os.path.dirname(filename), i) for i in include]
|
||||||
|
)
|
||||||
# As compose obj is updated and tested with every loop, not deleting `include`
|
# As compose obj is updated and tested with every loop, not deleting `include`
|
||||||
# from it, results in it being tested again and again, original values for
|
# from it, results in it being tested again and again, original values for
|
||||||
# `include` be appended to `files`, and, included files be processed for ever.
|
# `include` be appended to `files`, and, included files be processed for ever.
|
||||||
# Solution is to remove 'include' key from compose obj. This doesn't break
|
# Solution is to remove 'include' key from compose obj. This doesn't break
|
||||||
# having `include` present and correctly processed in included files
|
# having `include` present and correctly processed in included files
|
||||||
del compose["include"]
|
del compose["include"]
|
||||||
resolved_services = self._resolve_profiles(compose.get("services", {}), set(args.profile))
|
resolved_services = self._resolve_profiles(
|
||||||
|
compose.get("services", {}), set(args.profile)
|
||||||
|
)
|
||||||
compose["services"] = resolved_services
|
compose["services"] = resolved_services
|
||||||
if not getattr(args, "no_normalize", None):
|
if not getattr(args, "no_normalize", None):
|
||||||
compose = normalize_final(compose, self.dirname)
|
compose = normalize_final(compose, self.dirname)
|
||||||
@ -2014,10 +2096,14 @@ class PodmanCompose:
|
|||||||
if project_name is None:
|
if project_name is None:
|
||||||
# More strict then actually needed for simplicity:
|
# More strict then actually needed for simplicity:
|
||||||
# podman requires [a-zA-Z0-9][a-zA-Z0-9_.-]*
|
# podman requires [a-zA-Z0-9][a-zA-Z0-9_.-]*
|
||||||
project_name = self.environ.get("COMPOSE_PROJECT_NAME", dir_basename.lower())
|
project_name = self.environ.get(
|
||||||
|
"COMPOSE_PROJECT_NAME", dir_basename.lower()
|
||||||
|
)
|
||||||
project_name = norm_re.sub("", project_name)
|
project_name = norm_re.sub("", project_name)
|
||||||
if not project_name:
|
if not project_name:
|
||||||
raise RuntimeError(f"Project name [{dir_basename}] normalized to empty")
|
raise RuntimeError(
|
||||||
|
f"Project name [{dir_basename}] normalized to empty"
|
||||||
|
)
|
||||||
|
|
||||||
self.project_name = project_name
|
self.project_name = project_name
|
||||||
self.environ.update({"COMPOSE_PROJECT_NAME": self.project_name})
|
self.environ.update({"COMPOSE_PROJECT_NAME": self.project_name})
|
||||||
@ -2031,11 +2117,15 @@ class PodmanCompose:
|
|||||||
|
|
||||||
# NOTE: maybe add "extends.service" to _deps at this stage
|
# NOTE: maybe add "extends.service" to _deps at this stage
|
||||||
flat_deps(services, with_extends=True)
|
flat_deps(services, with_extends=True)
|
||||||
service_names = sorted([(len(srv["_deps"]), name) for name, srv in services.items()])
|
service_names = sorted(
|
||||||
|
[(len(srv["_deps"]), name) for name, srv in services.items()]
|
||||||
|
)
|
||||||
service_names = [name for _, name in service_names]
|
service_names = [name for _, name in service_names]
|
||||||
resolve_extends(services, service_names, self.environ)
|
resolve_extends(services, service_names, self.environ)
|
||||||
flat_deps(services)
|
flat_deps(services)
|
||||||
service_names = sorted([(len(srv["_deps"]), name) for name, srv in services.items()])
|
service_names = sorted(
|
||||||
|
[(len(srv["_deps"]), name) for name, srv in services.items()]
|
||||||
|
)
|
||||||
service_names = [name for _, name in service_names]
|
service_names = [name for _, name in service_names]
|
||||||
nets = compose.get("networks", {})
|
nets = compose.get("networks", {})
|
||||||
if not nets:
|
if not nets:
|
||||||
@ -2061,7 +2151,9 @@ class PodmanCompose:
|
|||||||
for name, srv in services.items():
|
for name, srv in services.items():
|
||||||
srv_nets = srv.get("networks", self.default_net)
|
srv_nets = srv.get("networks", self.default_net)
|
||||||
srv_nets = (
|
srv_nets = (
|
||||||
list(srv_nets.keys()) if isinstance(srv_nets, dict) else norm_as_list(srv_nets)
|
list(srv_nets.keys())
|
||||||
|
if isinstance(srv_nets, dict)
|
||||||
|
else norm_as_list(srv_nets)
|
||||||
)
|
)
|
||||||
allnets.update(srv_nets)
|
allnets.update(srv_nets)
|
||||||
given_nets = set(nets.keys())
|
given_nets = set(nets.keys())
|
||||||
@ -2092,7 +2184,9 @@ class PodmanCompose:
|
|||||||
container_names_by_service = {}
|
container_names_by_service = {}
|
||||||
self.services = services
|
self.services = services
|
||||||
for service_name, service_desc in services.items():
|
for service_name, service_desc in services.items():
|
||||||
replicas = try_int(service_desc.get("deploy", {}).get("replicas"), fallback=1)
|
replicas = try_int(
|
||||||
|
service_desc.get("deploy", {}).get("replicas"), fallback=1
|
||||||
|
)
|
||||||
|
|
||||||
container_names_by_service[service_name] = []
|
container_names_by_service[service_name] = []
|
||||||
for num in range(1, replicas + 1):
|
for num in range(1, replicas + 1):
|
||||||
@ -2110,16 +2204,20 @@ class PodmanCompose:
|
|||||||
**service_desc,
|
**service_desc,
|
||||||
}
|
}
|
||||||
x_podman = service_desc.get("x-podman")
|
x_podman = service_desc.get("x-podman")
|
||||||
rootfs_mode = x_podman is not None and x_podman.get("rootfs") is not None
|
rootfs_mode = (
|
||||||
|
x_podman is not None and x_podman.get("rootfs") is not None
|
||||||
|
)
|
||||||
if "image" not in cnt and not rootfs_mode:
|
if "image" not in cnt and not rootfs_mode:
|
||||||
cnt["image"] = f"{project_name}_{service_name}"
|
cnt["image"] = f"{project_name}_{service_name}"
|
||||||
labels = norm_as_list(cnt.get("labels"))
|
labels = norm_as_list(cnt.get("labels"))
|
||||||
cnt["ports"] = norm_ports(cnt.get("ports"))
|
cnt["ports"] = norm_ports(cnt.get("ports"))
|
||||||
labels.extend(podman_compose_labels)
|
labels.extend(podman_compose_labels)
|
||||||
labels.extend([
|
labels.extend(
|
||||||
f"com.docker.compose.container-number={num}",
|
[
|
||||||
"com.docker.compose.service=" + service_name,
|
f"com.docker.compose.container-number={num}",
|
||||||
])
|
"com.docker.compose.service=" + service_name,
|
||||||
|
]
|
||||||
|
)
|
||||||
cnt["labels"] = labels
|
cnt["labels"] = labels
|
||||||
cnt["_service"] = service_name
|
cnt["_service"] = service_name
|
||||||
cnt["_project"] = project_name
|
cnt["_project"] = project_name
|
||||||
@ -2133,7 +2231,9 @@ class PodmanCompose:
|
|||||||
and mnt_dict["source"] not in self.vols
|
and mnt_dict["source"] not in self.vols
|
||||||
):
|
):
|
||||||
vol_name = mnt_dict["source"]
|
vol_name = mnt_dict["source"]
|
||||||
raise RuntimeError(f"volume [{vol_name}] not defined in top level")
|
raise RuntimeError(
|
||||||
|
f"volume [{vol_name}] not defined in top level"
|
||||||
|
)
|
||||||
self.container_names_by_service = container_names_by_service
|
self.container_names_by_service = container_names_by_service
|
||||||
self.all_services = set(container_names_by_service.keys())
|
self.all_services = set(container_names_by_service.keys())
|
||||||
container_by_name = {c["name"]: c for c in given_containers}
|
container_by_name = {c["name"]: c for c in given_containers}
|
||||||
@ -2169,7 +2269,9 @@ class PodmanCompose:
|
|||||||
|
|
||||||
for name, config in defined_services.items():
|
for name, config in defined_services.items():
|
||||||
service_profiles = set(config.get("profiles", []))
|
service_profiles = set(config.get("profiles", []))
|
||||||
if not service_profiles or requested_profiles.intersection(service_profiles):
|
if not service_profiles or requested_profiles.intersection(
|
||||||
|
service_profiles
|
||||||
|
):
|
||||||
services[name] = config
|
services[name] = config
|
||||||
return services
|
return services
|
||||||
|
|
||||||
@ -2179,26 +2281,35 @@ class PodmanCompose:
|
|||||||
subparsers = parser.add_subparsers(title="command", dest="command")
|
subparsers = parser.add_subparsers(title="command", dest="command")
|
||||||
subparser = subparsers.add_parser("help", help="show help")
|
subparser = subparsers.add_parser("help", help="show help")
|
||||||
for cmd_name, cmd in self.commands.items():
|
for cmd_name, cmd in self.commands.items():
|
||||||
subparser = subparsers.add_parser(cmd_name, help=cmd.desc) # pylint: disable=protected-access
|
subparser = subparsers.add_parser(
|
||||||
|
cmd_name, help=cmd.desc
|
||||||
|
) # pylint: disable=protected-access
|
||||||
for cmd_parser in cmd._parse_args: # pylint: disable=protected-access
|
for cmd_parser in cmd._parse_args: # pylint: disable=protected-access
|
||||||
cmd_parser(subparser)
|
cmd_parser(subparser)
|
||||||
self.global_args = parser.parse_args(argv)
|
self.global_args = parser.parse_args(argv)
|
||||||
if self.global_args.in_pod is not None and self.global_args.in_pod.lower() not in (
|
if (
|
||||||
'',
|
self.global_args.in_pod is not None
|
||||||
'true',
|
and self.global_args.in_pod.lower()
|
||||||
'1',
|
not in (
|
||||||
'false',
|
"",
|
||||||
'0',
|
"true",
|
||||||
|
"1",
|
||||||
|
"false",
|
||||||
|
"0",
|
||||||
|
)
|
||||||
):
|
):
|
||||||
raise ValueError(
|
raise ValueError(
|
||||||
f'Invalid --in-pod value: \'{self.global_args.in_pod}\'. '
|
f"Invalid --in-pod value: '{self.global_args.in_pod}'. "
|
||||||
'It must be set to either of: empty value, true, 1, false, 0'
|
"It must be set to either of: empty value, true, 1, false, 0"
|
||||||
)
|
)
|
||||||
|
|
||||||
if self.global_args.in_pod == '' or self.global_args.in_pod is None:
|
if self.global_args.in_pod == "" or self.global_args.in_pod is None:
|
||||||
self.global_args.in_pod_bool = None
|
self.global_args.in_pod_bool = None
|
||||||
else:
|
else:
|
||||||
self.global_args.in_pod_bool = self.global_args.in_pod.lower() in ('true', '1')
|
self.global_args.in_pod_bool = self.global_args.in_pod.lower() in (
|
||||||
|
"true",
|
||||||
|
"1",
|
||||||
|
)
|
||||||
|
|
||||||
if self.global_args.version:
|
if self.global_args.version:
|
||||||
self.global_args.command = "version"
|
self.global_args.command = "version"
|
||||||
@ -2292,7 +2403,9 @@ class PodmanCompose:
|
|||||||
action="store_true",
|
action="store_true",
|
||||||
)
|
)
|
||||||
parser.add_argument(
|
parser.add_argument(
|
||||||
"--parallel", type=int, default=os.environ.get("COMPOSE_PARALLEL_LIMIT", sys.maxsize)
|
"--parallel",
|
||||||
|
type=int,
|
||||||
|
default=os.environ.get("COMPOSE_PARALLEL_LIMIT", sys.maxsize),
|
||||||
)
|
)
|
||||||
parser.add_argument(
|
parser.add_argument(
|
||||||
"--verbose",
|
"--verbose",
|
||||||
@ -2500,7 +2613,9 @@ def container_to_build_args(compose, cnt, args, path_exists, cleanup_callbacks=N
|
|||||||
dockerfile_inline = str(dockerfile_inline)
|
dockerfile_inline = str(dockerfile_inline)
|
||||||
# Error if both `dockerfile_inline` and `dockerfile` are set
|
# Error if both `dockerfile_inline` and `dockerfile` are set
|
||||||
if dockerfile and dockerfile_inline:
|
if dockerfile and dockerfile_inline:
|
||||||
raise OSError("dockerfile_inline and dockerfile can't be used simultaneously")
|
raise OSError(
|
||||||
|
"dockerfile_inline and dockerfile can't be used simultaneously"
|
||||||
|
)
|
||||||
dockerfile = tempfile.NamedTemporaryFile(delete=False, suffix=".containerfile")
|
dockerfile = tempfile.NamedTemporaryFile(delete=False, suffix=".containerfile")
|
||||||
dockerfile.write(dockerfile_inline.encode())
|
dockerfile.write(dockerfile_inline.encode())
|
||||||
dockerfile.close()
|
dockerfile.close()
|
||||||
@ -2534,7 +2649,9 @@ def container_to_build_args(compose, cnt, args, path_exists, cleanup_callbacks=N
|
|||||||
if "platform" in cnt:
|
if "platform" in cnt:
|
||||||
build_args.extend(["--platform", cnt["platform"]])
|
build_args.extend(["--platform", cnt["platform"]])
|
||||||
for secret in build_desc.get("secrets", []):
|
for secret in build_desc.get("secrets", []):
|
||||||
build_args.extend(get_secret_args(compose, cnt, secret, podman_is_building=True))
|
build_args.extend(
|
||||||
|
get_secret_args(compose, cnt, secret, podman_is_building=True)
|
||||||
|
)
|
||||||
for tag in build_desc.get("tags", []):
|
for tag in build_desc.get("tags", []):
|
||||||
build_args.extend(["-t", tag])
|
build_args.extend(["-t", tag])
|
||||||
labels = build_desc.get("labels", [])
|
labels = build_desc.get("labels", [])
|
||||||
@ -2557,10 +2674,12 @@ def container_to_build_args(compose, cnt, args, path_exists, cleanup_callbacks=N
|
|||||||
build_args.append("--pull")
|
build_args.append("--pull")
|
||||||
args_list = norm_as_list(build_desc.get("args", {}))
|
args_list = norm_as_list(build_desc.get("args", {}))
|
||||||
for build_arg in args_list + args.build_arg:
|
for build_arg in args_list + args.build_arg:
|
||||||
build_args.extend((
|
build_args.extend(
|
||||||
"--build-arg",
|
(
|
||||||
build_arg,
|
"--build-arg",
|
||||||
))
|
build_arg,
|
||||||
|
)
|
||||||
|
)
|
||||||
for cache_img in build_desc.get("cache_from", []):
|
for cache_img in build_desc.get("cache_from", []):
|
||||||
build_args.extend(["--cache-from", cache_img])
|
build_args.extend(["--cache-from", cache_img])
|
||||||
for cache_img in build_desc.get("cache_to", []):
|
for cache_img in build_desc.get("cache_to", []):
|
||||||
@ -2673,12 +2792,14 @@ async def check_dep_conditions(compose: PodmanCompose, deps: set) -> None:
|
|||||||
log.debug(
|
log.debug(
|
||||||
"dependencies for condition %s have been fulfilled on containers %s",
|
"dependencies for condition %s have been fulfilled on containers %s",
|
||||||
condition.value,
|
condition.value,
|
||||||
', '.join(deps_cd),
|
", ".join(deps_cd),
|
||||||
)
|
)
|
||||||
break
|
break
|
||||||
except subprocess.CalledProcessError as _exc:
|
except subprocess.CalledProcessError as _exc:
|
||||||
output = list(
|
output = list(
|
||||||
((_exc.stdout or b"") + (_exc.stderr or b"")).decode().split('\n')
|
((_exc.stdout or b"") + (_exc.stderr or b""))
|
||||||
|
.decode()
|
||||||
|
.split("\n")
|
||||||
)
|
)
|
||||||
log.debug(
|
log.debug(
|
||||||
'Podman wait returned an error (%d) when executing "%s": %s',
|
'Podman wait returned an error (%d) when executing "%s": %s',
|
||||||
@ -2690,7 +2811,11 @@ async def check_dep_conditions(compose: PodmanCompose, deps: set) -> None:
|
|||||||
|
|
||||||
|
|
||||||
async def run_container(
|
async def run_container(
|
||||||
compose: PodmanCompose, name: str, deps: set, command: tuple, log_formatter: str = None
|
compose: PodmanCompose,
|
||||||
|
name: str,
|
||||||
|
deps: set,
|
||||||
|
command: tuple,
|
||||||
|
log_formatter: str = None,
|
||||||
):
|
):
|
||||||
"""runs a container after waiting for its dependencies to be fulfilled"""
|
"""runs a container after waiting for its dependencies to be fulfilled"""
|
||||||
|
|
||||||
@ -2707,10 +2832,12 @@ async def run_container(
|
|||||||
def deps_from_container(args, cnt):
|
def deps_from_container(args, cnt):
|
||||||
if args.no_deps:
|
if args.no_deps:
|
||||||
return set()
|
return set()
|
||||||
return cnt['_deps']
|
return cnt["_deps"]
|
||||||
|
|
||||||
|
|
||||||
@cmd_run(podman_compose, "up", "Create and start the entire stack or some of its services")
|
@cmd_run(
|
||||||
|
podman_compose, "up", "Create and start the entire stack or some of its services"
|
||||||
|
)
|
||||||
async def compose_up(compose: PodmanCompose, args):
|
async def compose_up(compose: PodmanCompose, args):
|
||||||
excluded = get_excluded(compose, args)
|
excluded = get_excluded(compose, args)
|
||||||
if not args.no_build:
|
if not args.no_build:
|
||||||
@ -2757,7 +2884,10 @@ async def compose_up(compose: PodmanCompose, args):
|
|||||||
subproc = await compose.podman.run([], podman_command, podman_args)
|
subproc = await compose.podman.run([], podman_command, podman_args)
|
||||||
if podman_command == "run" and subproc is not None:
|
if podman_command == "run" and subproc is not None:
|
||||||
await run_container(
|
await run_container(
|
||||||
compose, cnt["name"], deps_from_container(args, cnt), ([], "start", [cnt["name"]])
|
compose,
|
||||||
|
cnt["name"],
|
||||||
|
deps_from_container(args, cnt),
|
||||||
|
([], "start", [cnt["name"]]),
|
||||||
)
|
)
|
||||||
if args.no_start or args.detach or args.dry_run:
|
if args.no_start or args.detach or args.dry_run:
|
||||||
return
|
return
|
||||||
@ -2771,12 +2901,16 @@ async def compose_up(compose: PodmanCompose, args):
|
|||||||
max_service_length = 0
|
max_service_length = 0
|
||||||
for cnt in compose.containers:
|
for cnt in compose.containers:
|
||||||
curr_length = len(cnt["_service"])
|
curr_length = len(cnt["_service"])
|
||||||
max_service_length = curr_length if curr_length > max_service_length else max_service_length
|
max_service_length = (
|
||||||
|
curr_length if curr_length > max_service_length else max_service_length
|
||||||
|
)
|
||||||
|
|
||||||
tasks = set()
|
tasks = set()
|
||||||
|
|
||||||
loop = asyncio.get_event_loop()
|
loop = asyncio.get_event_loop()
|
||||||
loop.add_signal_handler(signal.SIGINT, lambda: [t.cancel("User exit") for t in tasks])
|
loop.add_signal_handler(
|
||||||
|
signal.SIGINT, lambda: [t.cancel("User exit") for t in tasks]
|
||||||
|
)
|
||||||
|
|
||||||
for i, cnt in enumerate(compose.containers):
|
for i, cnt in enumerate(compose.containers):
|
||||||
# Add colored service prefix to output by piping output through sed
|
# Add colored service prefix to output by piping output through sed
|
||||||
@ -2868,7 +3002,8 @@ async def compose_down(compose: PodmanCompose, args):
|
|||||||
podman_stop_args.extend(["-t", str(timeout)])
|
podman_stop_args.extend(["-t", str(timeout)])
|
||||||
down_tasks.append(
|
down_tasks.append(
|
||||||
asyncio.create_task(
|
asyncio.create_task(
|
||||||
compose.podman.run([], "stop", [*podman_stop_args, cnt["name"]]), name=cnt["name"]
|
compose.podman.run([], "stop", [*podman_stop_args, cnt["name"]]),
|
||||||
|
name=cnt["name"],
|
||||||
)
|
)
|
||||||
)
|
)
|
||||||
await asyncio.gather(*down_tasks)
|
await asyncio.gather(*down_tasks)
|
||||||
@ -2920,7 +3055,11 @@ async def compose_down(compose: PodmanCompose, args):
|
|||||||
|
|
||||||
@cmd_run(podman_compose, "ps", "show status of containers")
|
@cmd_run(podman_compose, "ps", "show status of containers")
|
||||||
async def compose_ps(compose, args):
|
async def compose_ps(compose, args):
|
||||||
ps_args = ["-a", "--filter", f"label=io.podman.compose.project={compose.project_name}"]
|
ps_args = [
|
||||||
|
"-a",
|
||||||
|
"--filter",
|
||||||
|
f"label=io.podman.compose.project={compose.project_name}",
|
||||||
|
]
|
||||||
if args.quiet is True:
|
if args.quiet is True:
|
||||||
ps_args.extend(["--format", "{{.ID}}"])
|
ps_args.extend(["--format", "{{.ID}}"])
|
||||||
elif args.format:
|
elif args.format:
|
||||||
@ -2965,7 +3104,10 @@ async def compose_run(compose, args):
|
|||||||
await compose.commands["up"](compose, up_args)
|
await compose.commands["up"](compose, up_args)
|
||||||
|
|
||||||
build_args = argparse.Namespace(
|
build_args = argparse.Namespace(
|
||||||
services=[args.service], if_not_exists=(not args.build), build_arg=[], **args.__dict__
|
services=[args.service],
|
||||||
|
if_not_exists=(not args.build),
|
||||||
|
build_arg=[],
|
||||||
|
**args.__dict__,
|
||||||
)
|
)
|
||||||
await compose.commands["build"](compose, build_args)
|
await compose.commands["build"](compose, build_args)
|
||||||
|
|
||||||
@ -2982,7 +3124,9 @@ async def compose_run(compose, args):
|
|||||||
|
|
||||||
def compose_run_update_container_from_args(compose, cnt, args):
|
def compose_run_update_container_from_args(compose, cnt, args):
|
||||||
# adjust one-off container options
|
# adjust one-off container options
|
||||||
name0 = "{}_{}_tmp{}".format(compose.project_name, args.service, random.randrange(0, 65536))
|
name0 = "{}_{}_tmp{}".format(
|
||||||
|
compose.project_name, args.service, random.randrange(0, 65536)
|
||||||
|
)
|
||||||
cnt["name"] = args.name or name0
|
cnt["name"] = args.name or name0
|
||||||
if args.entrypoint:
|
if args.entrypoint:
|
||||||
cnt["entrypoint"] = args.entrypoint
|
cnt["entrypoint"] = args.entrypoint
|
||||||
@ -2992,7 +3136,9 @@ def compose_run_update_container_from_args(compose, cnt, args):
|
|||||||
cnt["working_dir"] = args.workdir
|
cnt["working_dir"] = args.workdir
|
||||||
env = dict(cnt.get("environment", {}))
|
env = dict(cnt.get("environment", {}))
|
||||||
if args.env:
|
if args.env:
|
||||||
additional_env_vars = dict(map(lambda each: each.split("=", maxsplit=1), args.env))
|
additional_env_vars = dict(
|
||||||
|
map(lambda each: each.split("=", maxsplit=1), args.env)
|
||||||
|
)
|
||||||
env.update(additional_env_vars)
|
env.update(additional_env_vars)
|
||||||
cnt["environment"] = env
|
cnt["environment"] = env
|
||||||
if not args.service_ports:
|
if not args.service_ports:
|
||||||
@ -3042,7 +3188,12 @@ def compose_exec_args(cnt, container_name, args):
|
|||||||
env = dict(cnt.get("environment", {}))
|
env = dict(cnt.get("environment", {}))
|
||||||
if args.env:
|
if args.env:
|
||||||
additional_env_vars = dict(
|
additional_env_vars = dict(
|
||||||
map(lambda each: each.split("=", maxsplit=1) if "=" in each else (each, None), args.env)
|
map(
|
||||||
|
lambda each: (
|
||||||
|
each.split("=", maxsplit=1) if "=" in each else (each, None)
|
||||||
|
),
|
||||||
|
args.env,
|
||||||
|
)
|
||||||
)
|
)
|
||||||
env.update(additional_env_vars)
|
env.update(additional_env_vars)
|
||||||
for name, value in env.items():
|
for name, value in env.items():
|
||||||
@ -3079,7 +3230,9 @@ async def transfer_service_status(compose, args, action):
|
|||||||
timeout = str_to_seconds(timeout_str)
|
timeout = str_to_seconds(timeout_str)
|
||||||
if timeout is not None:
|
if timeout is not None:
|
||||||
podman_args.extend(["-t", str(timeout)])
|
podman_args.extend(["-t", str(timeout)])
|
||||||
tasks.append(asyncio.create_task(compose.podman.run([], action, podman_args + [target])))
|
tasks.append(
|
||||||
|
asyncio.create_task(compose.podman.run([], action, podman_args + [target]))
|
||||||
|
)
|
||||||
await asyncio.gather(*tasks)
|
await asyncio.gather(*tasks)
|
||||||
|
|
||||||
|
|
||||||
@ -3188,7 +3341,9 @@ async def compose_unpause(compose, args):
|
|||||||
await compose.podman.run([], "unpause", targets)
|
await compose.podman.run([], "unpause", targets)
|
||||||
|
|
||||||
|
|
||||||
@cmd_run(podman_compose, "kill", "Kill one or more running containers with a specific signal")
|
@cmd_run(
|
||||||
|
podman_compose, "kill", "Kill one or more running containers with a specific signal"
|
||||||
|
)
|
||||||
async def compose_kill(compose, args):
|
async def compose_kill(compose, args):
|
||||||
# to ensure that the user did not execute the command by mistake
|
# to ensure that the user did not execute the command by mistake
|
||||||
if not args.services and not args.all:
|
if not args.services and not args.all:
|
||||||
@ -3258,7 +3413,9 @@ async def compose_images(compose, args):
|
|||||||
if args.quiet is True:
|
if args.quiet is True:
|
||||||
for img in img_containers:
|
for img in img_containers:
|
||||||
name = img["name"]
|
name = img["name"]
|
||||||
output = await compose.podman.output([], "images", ["--quiet", img["image"]])
|
output = await compose.podman.output(
|
||||||
|
[], "images", ["--quiet", img["image"]]
|
||||||
|
)
|
||||||
data.append(output.decode("utf-8").split())
|
data.append(output.decode("utf-8").split())
|
||||||
else:
|
else:
|
||||||
data.append(["CONTAINER", "REPOSITORY", "TAG", "IMAGE ID", "SIZE", ""])
|
data.append(["CONTAINER", "REPOSITORY", "TAG", "IMAGE ID", "SIZE", ""])
|
||||||
@ -3317,13 +3474,17 @@ def compose_up_parse(parser):
|
|||||||
help="Detached mode: Run container in the background, print new container name. \
|
help="Detached mode: Run container in the background, print new container name. \
|
||||||
Incompatible with --abort-on-container-exit.",
|
Incompatible with --abort-on-container-exit.",
|
||||||
)
|
)
|
||||||
parser.add_argument("--no-color", action="store_true", help="Produce monochrome output.")
|
parser.add_argument(
|
||||||
|
"--no-color", action="store_true", help="Produce monochrome output."
|
||||||
|
)
|
||||||
parser.add_argument(
|
parser.add_argument(
|
||||||
"--quiet-pull",
|
"--quiet-pull",
|
||||||
action="store_true",
|
action="store_true",
|
||||||
help="Pull without printing progress information.",
|
help="Pull without printing progress information.",
|
||||||
)
|
)
|
||||||
parser.add_argument("--no-deps", action="store_true", help="Don't start linked services.")
|
parser.add_argument(
|
||||||
|
"--no-deps", action="store_true", help="Don't start linked services."
|
||||||
|
)
|
||||||
parser.add_argument(
|
parser.add_argument(
|
||||||
"--force-recreate",
|
"--force-recreate",
|
||||||
action="store_true",
|
action="store_true",
|
||||||
@ -3422,7 +3583,9 @@ def compose_run_parse(parser):
|
|||||||
action="store_true",
|
action="store_true",
|
||||||
help="Detached mode: Run container in the background, print new container name.",
|
help="Detached mode: Run container in the background, print new container name.",
|
||||||
)
|
)
|
||||||
parser.add_argument("--name", type=str, default=None, help="Assign a name to the container")
|
parser.add_argument(
|
||||||
|
"--name", type=str, default=None, help="Assign a name to the container"
|
||||||
|
)
|
||||||
parser.add_argument(
|
parser.add_argument(
|
||||||
"--entrypoint",
|
"--entrypoint",
|
||||||
type=str,
|
type=str,
|
||||||
@ -3446,7 +3609,9 @@ def compose_run_parse(parser):
|
|||||||
parser.add_argument(
|
parser.add_argument(
|
||||||
"-u", "--user", type=str, default=None, help="Run as specified username or uid"
|
"-u", "--user", type=str, default=None, help="Run as specified username or uid"
|
||||||
)
|
)
|
||||||
parser.add_argument("--no-deps", action="store_true", help="Don't start linked services")
|
parser.add_argument(
|
||||||
|
"--no-deps", action="store_true", help="Don't start linked services"
|
||||||
|
)
|
||||||
parser.add_argument(
|
parser.add_argument(
|
||||||
"--rm",
|
"--rm",
|
||||||
action="store_true",
|
action="store_true",
|
||||||
@ -3572,15 +3737,21 @@ def compose_logs_parse(parser):
|
|||||||
action="store_true",
|
action="store_true",
|
||||||
help="Output the container name in the log",
|
help="Output the container name in the log",
|
||||||
)
|
)
|
||||||
parser.add_argument("--since", help="Show logs since TIMESTAMP", type=str, default=None)
|
parser.add_argument(
|
||||||
parser.add_argument("-t", "--timestamps", action="store_true", help="Show timestamps.")
|
"--since", help="Show logs since TIMESTAMP", type=str, default=None
|
||||||
|
)
|
||||||
|
parser.add_argument(
|
||||||
|
"-t", "--timestamps", action="store_true", help="Show timestamps."
|
||||||
|
)
|
||||||
parser.add_argument(
|
parser.add_argument(
|
||||||
"--tail",
|
"--tail",
|
||||||
help="Number of lines to show from the end of the logs for each container.",
|
help="Number of lines to show from the end of the logs for each container.",
|
||||||
type=str,
|
type=str,
|
||||||
default="all",
|
default="all",
|
||||||
)
|
)
|
||||||
parser.add_argument("--until", help="Show logs until TIMESTAMP", type=str, default=None)
|
parser.add_argument(
|
||||||
|
"--until", help="Show logs until TIMESTAMP", type=str, default=None
|
||||||
|
)
|
||||||
parser.add_argument(
|
parser.add_argument(
|
||||||
"services", metavar="services", nargs="*", default=None, help="service names"
|
"services", metavar="services", nargs="*", default=None, help="service names"
|
||||||
)
|
)
|
||||||
@ -3605,7 +3776,9 @@ def compose_pull_parse(parser):
|
|||||||
default=False,
|
default=False,
|
||||||
help="Also pull unprefixed images for services which have a build section",
|
help="Also pull unprefixed images for services which have a build section",
|
||||||
)
|
)
|
||||||
parser.add_argument("services", metavar="services", nargs="*", help="services to pull")
|
parser.add_argument(
|
||||||
|
"services", metavar="services", nargs="*", help="services to pull"
|
||||||
|
)
|
||||||
|
|
||||||
|
|
||||||
@cmd_parse(podman_compose, "push")
|
@cmd_parse(podman_compose, "push")
|
||||||
@ -3615,12 +3788,16 @@ def compose_push_parse(parser):
|
|||||||
action="store_true",
|
action="store_true",
|
||||||
help="Push what it can and ignores images with push failures. (not implemented)",
|
help="Push what it can and ignores images with push failures. (not implemented)",
|
||||||
)
|
)
|
||||||
parser.add_argument("services", metavar="services", nargs="*", help="services to push")
|
parser.add_argument(
|
||||||
|
"services", metavar="services", nargs="*", help="services to push"
|
||||||
|
)
|
||||||
|
|
||||||
|
|
||||||
@cmd_parse(podman_compose, "ps")
|
@cmd_parse(podman_compose, "ps")
|
||||||
def compose_ps_parse(parser):
|
def compose_ps_parse(parser):
|
||||||
parser.add_argument("-q", "--quiet", help="Only display container IDs", action="store_true")
|
parser.add_argument(
|
||||||
|
"-q", "--quiet", help="Only display container IDs", action="store_true"
|
||||||
|
)
|
||||||
|
|
||||||
|
|
||||||
@cmd_parse(podman_compose, ["build", "up"])
|
@cmd_parse(podman_compose, ["build", "up"])
|
||||||
@ -3723,7 +3900,9 @@ def compose_kill_parse(parser):
|
|||||||
|
|
||||||
@cmd_parse(podman_compose, "images")
|
@cmd_parse(podman_compose, "images")
|
||||||
def compose_images_parse(parser):
|
def compose_images_parse(parser):
|
||||||
parser.add_argument("-q", "--quiet", help="Only display images IDs", action="store_true")
|
parser.add_argument(
|
||||||
|
"-q", "--quiet", help="Only display images IDs", action="store_true"
|
||||||
|
)
|
||||||
|
|
||||||
|
|
||||||
@cmd_parse(podman_compose, ["stats"])
|
@cmd_parse(podman_compose, ["stats"])
|
||||||
|
Loading…
Reference in New Issue
Block a user