Compare commits

...

5 Commits

Author SHA1 Message Date
8e0fe77274 test if a source is loaded 2021-09-13 14:58:11 +02:00
8dca850683 build image is now in risotto 2021-08-28 07:34:31 +02:00
1def6e4e4c add preprocessors function 2021-05-26 20:19:08 +02:00
2ea04e708d fix 2021-05-24 22:24:15 +02:00
4853bb47f0 risotto is now a lib 2021-05-24 20:41:04 +02:00
7 changed files with 267 additions and 139 deletions

View File

@ -113,12 +113,12 @@ if 'PASSWORD_ADMIN_EMAIL' in environ:
PASSWORD_ADMIN_EMAIL = environ['PASSWORD_ADMIN_EMAIL'] PASSWORD_ADMIN_EMAIL = environ['PASSWORD_ADMIN_EMAIL']
else: else:
# this parameter is mandatory # this parameter is mandatory
PASSWORD_ADMIN_EMAIL = config['PASSWORD_ADMIN_EMAIL'] PASSWORD_ADMIN_EMAIL = config.get('PASSWORD_ADMIN_EMAIL', 'XXX')
if 'PASSWORD_ADMIN_PASSWORD' in environ: if 'PASSWORD_ADMIN_PASSWORD' in environ:
PASSWORD_ADMIN_PASSWORD = environ['PASSWORD_ADMIN_PASSWORD'] PASSWORD_ADMIN_PASSWORD = environ['PASSWORD_ADMIN_PASSWORD']
else: else:
# this parameter is mandatory # this parameter is mandatory
PASSWORD_ADMIN_PASSWORD = config['PASSWORD_ADMIN_PASSWORD'] PASSWORD_ADMIN_PASSWORD = config.get('PASSWORD_ADMIN_PASSWORD', 'XXX')
if 'PASSWORD_DEVICE_IDENTIFIER' in environ: if 'PASSWORD_DEVICE_IDENTIFIER' in environ:
PASSWORD_DEVICE_IDENTIFIER = environ['PASSWORD_DEVICE_IDENTIFIER'] PASSWORD_DEVICE_IDENTIFIER = environ['PASSWORD_DEVICE_IDENTIFIER']
else: else:
@ -135,11 +135,11 @@ else:
if 'PKI_ADMIN_PASSWORD' in environ: if 'PKI_ADMIN_PASSWORD' in environ:
PKI_ADMIN_PASSWORD = environ['PKI_ADMIN_PASSWORD'] PKI_ADMIN_PASSWORD = environ['PKI_ADMIN_PASSWORD']
else: else:
PKI_ADMIN_PASSWORD = config['PKI_ADMIN_PASSWORD'] PKI_ADMIN_PASSWORD = config.get('PKI_ADMIN_PASSWORD', 'XXX')
if 'PKI_ADMIN_EMAIL' in environ: if 'PKI_ADMIN_EMAIL' in environ:
PKI_ADMIN_EMAIL = environ['PKI_ADMIN_EMAIL'] PKI_ADMIN_EMAIL = environ['PKI_ADMIN_EMAIL']
else: else:
PKI_ADMIN_EMAIL = config['PKI_ADMIN_EMAIL'] PKI_ADMIN_EMAIL = config.get('PKI_ADMIN_EMAIL', 'XXX')
if 'PKI_URL' in environ: if 'PKI_URL' in environ:
PKI_URL = environ['PKI_URL'] PKI_URL = environ['PKI_URL']
else: else:

View File

@ -12,7 +12,7 @@ except:
from .config import get_config from .config import get_config
from .utils import _, tiramisu_display_name from .utils import _, tiramisu_display_name
from .logger import log from .logger import log
from .dispatcher import dispatcher from .dispatcher import get_dispatcher
from .context import Context from .context import Context
@ -25,7 +25,7 @@ class Controller:
def __init__(self, def __init__(self,
test: bool, test: bool,
) -> None: ) -> None:
pass self.dispatcher = get_dispatcher()
async def call(self, async def call(self,
uri: str, uri: str,
@ -42,11 +42,11 @@ class Controller:
module = message.split('.', 1)[0] module = message.split('.', 1)[0]
if current_module != module: if current_module != module:
raise ValueError(_(f'cannot call to external module ("{module}") to the URI "{uri}" from "{current_module}"')) raise ValueError(_(f'cannot call to external module ("{module}") to the URI "{uri}" from "{current_module}"'))
return await dispatcher.call(version, return await self.dispatcher.call(version,
message, message,
risotto_context, risotto_context,
**kwargs, **kwargs,
) )
async def publish(self, async def publish(self,
uri: str, uri: str,
@ -58,19 +58,19 @@ class Controller:
if args: if args:
raise ValueError(_(f'the URI "{uri}" can only be published with keyword arguments')) raise ValueError(_(f'the URI "{uri}" can only be published with keyword arguments'))
version, message = uri.split('.', 1) version, message = uri.split('.', 1)
await dispatcher.publish(version, await self.dispatcher.publish(version,
message, message,
risotto_context, risotto_context,
**kwargs, **kwargs,
) )
@staticmethod async def check_role(self,
async def check_role(uri: str, uri: str,
username: str, username: str,
**kwargs: dict, **kwargs: dict,
) -> None: ) -> None:
# create a new config # create a new config
async with await Config(dispatcher.option) as config: async with await Config(self.dispatcher.option) as config:
await config.property.read_write() await config.property.read_write()
await config.option('message').value.set(uri) await config.option('message').value.set(uri)
subconfig = config.option(uri) subconfig = config.option(uri)
@ -83,10 +83,10 @@ class Controller:
raise ValueError(_(f'unknown parameter in "{uri}": "{key}"')) raise ValueError(_(f'unknown parameter in "{uri}": "{key}"'))
except ValueOptionError as err: except ValueOptionError as err:
raise ValueError(_(f'invalid parameter in "{uri}": {err}')) raise ValueError(_(f'invalid parameter in "{uri}": {err}'))
await dispatcher.check_role(subconfig, await self.dispatcher.check_role(subconfig,
username, username,
uri, uri,
) )
async def on_join(self, async def on_join(self,
risotto_context, risotto_context,
@ -98,20 +98,23 @@ class TiramisuController(Controller):
def __init__(self, def __init__(self,
test: bool, test: bool,
) -> None: ) -> None:
self.source_imported = None
if not 'dataset_name' in vars(self): if not 'dataset_name' in vars(self):
raise Exception(f'please specify "dataset_name" to "{self.__class__.__name__}"') raise Exception(f'please specify "dataset_name" to "{self.__class__.__name__}"')
self.tiramisu_cache_root_path = join(get_config()['cache']['root_path'], self.dataset_name) self.tiramisu_cache_root_path = join(get_config()['cache']['root_path'], self.dataset_name)
super().__init__(test)
self.internal_source_name = get_config()['servermodel']['internal_source']
if not test: if not test:
db_conf = get_config()['database']['tiramisu_dsn'] db_conf = get_config()['database']['tiramisu_dsn']
self.save_storage = Storage(engine='postgres') self.save_storage = Storage(engine='postgres')
self.save_storage.setting(dsn=db_conf) self.save_storage.setting(dsn=db_conf)
if self.dataset_name != 'servermodel': if self.dataset_name != 'servermodel':
self.optiondescription = None self.optiondescription = None
dispatcher.set_function('v1.setting.dataset.updated', self.dispatcher.set_function('v1.setting.dataset.updated',
None, None,
TiramisuController.dataset_updated, TiramisuController.dataset_updated,
self.__class__.__module__, self.__class__.__module__,
) )
async def on_join(self, async def on_join(self,
risotto_context: Context, risotto_context: Context,
@ -129,6 +132,12 @@ class TiramisuController(Controller):
risotto_context: Context, risotto_context: Context,
) -> None: ) -> None:
sources = await self.get_sources(risotto_context) sources = await self.get_sources(risotto_context)
source_imported = sources != [self.internal_source_name]
if source_imported and self.source_imported is False:
await self.load_datas(risotto_context)
self.source_imported = source_imported
if not self.source_imported:
return
self._aggregate_tiramisu_funcs(sources) self._aggregate_tiramisu_funcs(sources)
self._convert_dictionaries_to_tiramisu(sources) self._convert_dictionaries_to_tiramisu(sources)
@ -174,7 +183,9 @@ except:
funcs.write(fh.read()) funcs.write(fh.read())
funcs.write(b'\n') funcs.write(b'\n')
def _convert_dictionaries_to_tiramisu(self, sources: list) -> None: def _convert_dictionaries_to_tiramisu(self,
sources: list,
) -> None:
funcs_file = join(self.tiramisu_cache_root_path, 'funcs.py') funcs_file = join(self.tiramisu_cache_root_path, 'funcs.py')
tiramisu_file = join(self.tiramisu_cache_root_path, 'tiramisu.py') tiramisu_file = join(self.tiramisu_cache_root_path, 'tiramisu.py')
dictionaries_dir = join(self.tiramisu_cache_root_path, 'dictionaries') dictionaries_dir = join(self.tiramisu_cache_root_path, 'dictionaries')

View File

@ -18,6 +18,9 @@ from .context import Context
from . import register from . import register
DISPATCHER = None
class CallDispatcher: class CallDispatcher:
async def valid_call_returns(self, async def valid_call_returns(self,
risotto_context: Context, risotto_context: Context,
@ -492,5 +495,9 @@ class Dispatcher(register.RegisterDispatcher,
return returns return returns
dispatcher = Dispatcher() def get_dispatcher():
register.dispatcher = dispatcher global DISPATCHER
if DISPATCHER is None:
DISPATCHER = Dispatcher()
register.dispatcher = DISPATCHER
return DISPATCHER

View File

@ -7,7 +7,7 @@ except:
from tiramisu import Config, default_storage from tiramisu import Config, default_storage
from .dispatcher import dispatcher from .dispatcher import get_dispatcher
from .utils import _ from .utils import _
from .context import Context from .context import Context
from .error import CallError, NotAllowedError, RegistrationError from .error import CallError, NotAllowedError, RegistrationError
@ -70,6 +70,7 @@ class extra_route_handler:
if function_name != 'risotto.http': if function_name != 'risotto.http':
risotto_module_name, submodule_name = function_name.split('.', 2)[:-1] risotto_module_name, submodule_name = function_name.split('.', 2)[:-1]
module_name = risotto_module_name.split('_')[-1] module_name = risotto_module_name.split('_')[-1]
dispatcher = get_dispatcher()
kwargs['self'] = dispatcher.injected_self[module_name + '.' + submodule_name] kwargs['self'] = dispatcher.injected_self[module_name + '.' + submodule_name]
try: try:
returns = await cls.function(**kwargs) returns = await cls.function(**kwargs)
@ -93,6 +94,7 @@ async def handle(request):
risotto_context = create_context(request) risotto_context = create_context(request)
kwargs = await request.json() kwargs = await request.json()
try: try:
dispatcher = get_dispatcher()
pattern = dispatcher.messages[version][message]['pattern'] pattern = dispatcher.messages[version][message]['pattern']
if pattern == 'rpc': if pattern == 'rpc':
method = dispatcher.call method = dispatcher.call
@ -142,6 +144,7 @@ async def api(request,
# check all URI that have an associated role # check all URI that have an associated role
# all URI without role is concidered has a private URI # all URI without role is concidered has a private URI
uris = [] uris = []
dispatcher = get_dispatcher()
async with dispatcher.pool.acquire() as connection: async with dispatcher.pool.acquire() as connection:
async with connection.transaction(): async with connection.transaction():
# Check role with ACL # Check role with ACL
@ -167,6 +170,7 @@ async def get_app(loop):
""" build all routes """ build all routes
""" """
global extra_routes, extra_statics global extra_routes, extra_statics
dispatcher = get_dispatcher()
services.link_to_dispatcher(dispatcher) services.link_to_dispatcher(dispatcher)
app = Application(loop=loop) app = Application(loop=loop)
routes = [] routes = []

View File

@ -66,12 +66,23 @@ def zone_information(**kwargs):
""" """
class Image: class Images:
def __init__(self): def __init__(self,
self.parse_applications() image_dir: str=None,
tmp_dir: str=None,
):
if image_dir is None:
image_dir = IMAGES_DIRECTORY
self.image_dir = image_dir
if isdir(self.image_dir):
rmtree(self.image_dir)
if tmp_dir is None:
tmp_dir = PACKER_TMP_DIRECTORY
self.tmp_dir = tmp_dir
self.load_applications()
def parse_applications(self) -> None: def load_applications(self) -> None:
self.builds = [] self.build_images = []
self.applications = {} self.applications = {}
for distrib in listdir(join(DATASET_PATH, 'seed')): for distrib in listdir(join(DATASET_PATH, 'seed')):
distrib_dir = join(DATASET_PATH, 'seed', distrib, 'applicationservice') distrib_dir = join(DATASET_PATH, 'seed', distrib, 'applicationservice')
@ -93,48 +104,101 @@ class Image:
'yml': app, 'yml': app,
} }
if 'service' in app and app['service']: if 'service' in app and app['service']:
self.builds.append(applicationservice) self.build_images.append(applicationservice)
def calc_depends(self, def calc_depends(self,
dependencies: list, dependencies: list,
appname, appname,
key_is_name=False,
): ):
app = self.applications[appname]['yml'] app = self.applications[appname]['yml']
if not 'depends' in app or not app['depends']: if not 'depends' in app or not app['depends']:
return return
for dependency in app['depends']: for dependency in app['depends']:
dependency_path = self.applications[dependency]['path'] if key_is_name:
if dependency_path not in dependencies: key = appname
dependencies.insert(0, dependency_path) else:
self.calc_depends(dependencies, dependency) key = self.applications[dependency]['path']
if key not in dependencies:
dependencies.insert(0, key)
self.calc_depends(dependencies, dependency, key_is_name)
def list_oses(self):
oses = set()
for build in self.build_images:
dependencies = [build]
self.calc_depends(dependencies, build, True)
for dependency in dependencies:
if isdir(join(self.applications[dependency]['path'], 'packer', 'os')):
oses.add(dependency)
break
for os in oses:
dependencies = [self.applications[os]['path']]
self.calc_depends(dependencies, os)
yield os, dependencies
def list_images(self): def list_images(self):
for build in self.builds: for build in self.build_images:
dependencies = [self.applications[build]['path']] dependencies = [self.applications[build]['path']]
self.calc_depends(dependencies, build) self.calc_depends(dependencies, build)
yield build, dependencies yield build, dependencies
def copy_files(self, async def build(self) -> None:
src_path: str, if isdir(self.tmp_dir):
dst_path: str, rmtree(self.tmp_dir)
) -> None: image = Image(self.image_dir,
root_len = len(src_path) + 1 self.tmp_dir,
for dir_name, subdir_names, filenames in walk(src_path): )
subdir = join(dst_path, dir_name[root_len:]) print(_('Build OSes'))
if not isdir(subdir): if not isdir(join(self.image_dir, 'os')):
makedirs(subdir) makedirs(join(self.image_dir, 'os'))
for filename in filenames: for application, dependencies_path in self.list_oses():
path = join(dir_name, filename) print(_(f'Build OS {application}'))
sub_dst_path = join(subdir, filename) await image.build_os(application,
if isfile(sub_dst_path): dependencies_path,
raise Exception(_(f'Try to copy {sub_dst_path} which is already exists')) )
copy2(path, sub_dst_path) print(_('Build images'))
for application, dependencies_path in self.list_images():
print(_(f'Build image {application}'))
await image.build_image(application,
dependencies_path,
)
def load_configuration(self,
dependencies_path: list, class Image:
packer_tmp_directory: str, def __init__(self,
) -> dict: image_dir: str,
tmp_dir: str,
):
self.image_dir = image_dir
self.tmp_dir = tmp_dir
@staticmethod
def copy_files(dependencies_path: list,
dst_path: str,
element: str,
) -> None:
for dependency_path in dependencies_path:
src_path = join(dependency_path,
'packer',
element,
)
root_len = len(src_path) + 1
for dir_name, subdir_names, filenames in walk(src_path):
subdir = join(dst_path, dir_name[root_len:])
if not isdir(subdir):
makedirs(subdir)
for filename in filenames:
path = join(dir_name, filename)
sub_dst_path = join(subdir, filename)
if isfile(sub_dst_path):
raise Exception(_(f'Try to copy {sub_dst_path} which is already exists'))
copy2(path, sub_dst_path)
async def load_configuration(self,
dependencies_path: list,
packer_tmp_directory: str,
) -> dict:
config = RougailConfig.copy() config = RougailConfig.copy()
dictionaries = [join(dependency_path, 'dictionaries') for dependency_path in dependencies_path if isdir(join(dependency_path, 'dictionaries'))] dictionaries = [join(dependency_path, 'dictionaries') for dependency_path in dependencies_path if isdir(join(dependency_path, 'dictionaries'))]
upgrade = RougailUpgrade() upgrade = RougailUpgrade()
@ -157,11 +221,12 @@ class Image:
) )
config['dictionaries_dir'] = [dest_dictionaries] config['dictionaries_dir'] = [dest_dictionaries]
config['extra_dictionaries'] = {'packer': [dest_dictionaries_extras]} config['extra_dictionaries'] = {'packer': [dest_dictionaries_extras]}
return config self.merge_funcs(config, dependencies_path, packer_tmp_directory)
packer_configuration = await self.get_packer_information(config, packer_tmp_directory)
return packer_configuration
@staticmethod
def merge_funcs(self, def merge_funcs(config: RougailConfig,
config: RougailConfig,
dependencies_path: list, dependencies_path: list,
packer_tmp_directory: str, packer_tmp_directory: str,
): ):
@ -178,8 +243,8 @@ class Image:
fh.write(functions) fh.write(functions)
config['functions_file'] = func_name config['functions_file'] = func_name
async def get_packer_information(self, @staticmethod
config: RougailConfig, async def get_packer_information(config: RougailConfig,
packer_tmp_directory: str, packer_tmp_directory: str,
) -> dict: ) -> dict:
eolobj = RougailConvert(config) eolobj = RougailConvert(config)
@ -187,83 +252,116 @@ class Image:
optiondescription = {} optiondescription = {}
exec(xml, None, optiondescription) exec(xml, None, optiondescription)
config = await Config(optiondescription['option_0']) config = await Config(optiondescription['option_0'])
return await config.option('packer').value.dict(flatten=True) return await config.option('packer').value.dict(leader_to_list=True, flatten=True)
def do_recipe_checksum(self, @staticmethod
path: str, def do_recipe_checksum(path: str,
) -> str: ) -> str:
files = [] files = []
root_len = len(path) + 1 root_len = len(path) + 1
for dir_name, subdir_names, filenames in walk(path): for dir_name, subdir_names, filenames in walk(path):
subpath = dir_name[root_len:] subpath = dir_name[root_len:]
for filename in filenames: for filename in filenames:
with open(join(dir_name, filename), 'rb') as fh: with open(join(dir_name, filename), 'rb') as fh:
ctl_sum = sha512(fh.read()).hexdigest() ctl_sum = sha512(fh.read()).hexdigest()
files.append(f'{subpath}/{filename}/ctl_sum') abs_path = join(subpath, filename)
files.append(f'{abs_path}/{ctl_sum}')
files.sort()
print(files, sha512('\n'.join(files).encode()).hexdigest())
return sha512('\n'.join(files).encode()).hexdigest() return sha512('\n'.join(files).encode()).hexdigest()
async def build(self) -> None: def get_tmp_directory(self,
if isdir(PACKER_TMP_DIRECTORY): application: str,
rmtree(PACKER_TMP_DIRECTORY) ) -> str:
for application, dependencies_path in self.list_images(): return join(self.tmp_dir,
packer_tmp_directory = join(PACKER_TMP_DIRECTORY, application + '_' + str(time()),
application + '_' + str(time()), )
)
makedirs(packer_tmp_directory)
packer_tmp_os_directory = join(packer_tmp_directory, 'os')
makedirs(packer_tmp_os_directory)
packer_tmp_img_directory = join(packer_tmp_directory, 'image')
makedirs(packer_tmp_img_directory)
config = self.load_configuration(dependencies_path, packer_tmp_directory)
self.merge_funcs(config, dependencies_path, packer_tmp_directory)
packer_configuration = await self.get_packer_information(config, packer_tmp_directory)
# OS image needed ?
packer_dst_os_filename = join(IMAGES_DIRECTORY,
'os',
packer_configuration['os_name'] + '_' + packer_configuration['os_version'] + '.img',
)
for dependency_path in dependencies_path:
packer_directory = join(dependency_path,
'packer',
'os',
)
self.copy_files(packer_directory,
packer_tmp_os_directory,
)
packer_directory = join(dependency_path,
'packer',
'image',
)
self.copy_files(packer_directory,
packer_tmp_img_directory,
)
if not isfile(packer_dst_os_filename):
self.build_image(packer_dst_os_filename,
packer_tmp_os_directory,
packer_configuration,
)
recipe_checksum = self.do_recipe_checksum(packer_tmp_img_directory)
packer_dst_filename = join(IMAGES_DIRECTORY,
f'{recipe_checksum}.img',
)
self.build_image(packer_dst_filename,
packer_tmp_img_directory,
packer_configuration,
)
def build_image(self, def get_os_filename(self,
packer_dst_filename: str, packer_configuration: dict,
tmp_directory: str, ) -> str:
packer_configuration: dict, return join(self.image_dir,
) -> None: 'os',
packer_configuration['tmp_directory'] = tmp_directory packer_configuration['os_name'] + '_' + packer_configuration['os_version'] + '.img',
)
def get_image_filename(self,
recipe_checksum: str,
) -> str:
return join(self.image_dir,
f'{recipe_checksum}.img',
)
async def build_os(self,
application: str,
dependencies_path: list,
) -> None:
packer_tmp_directory = self.get_tmp_directory(application)
packer_configuration = await self.load_configuration(dependencies_path, packer_tmp_directory)
packer_dst_os_filename = self.get_os_filename(packer_configuration)
self.copy_files(dependencies_path,
packer_tmp_directory,
'os',
)
packer_configuration['tmp_directory'] = packer_tmp_directory
recipe = {'variables': packer_configuration} recipe = {'variables': packer_configuration}
self.build(packer_dst_os_filename,
packer_tmp_directory,
recipe,
)
async def build_image(self,
application: str,
dependencies_path: list,
) -> None:
packer_tmp_directory = self.get_tmp_directory(application)
makedirs(packer_tmp_directory)
self.copy_files(dependencies_path,
packer_tmp_directory,
'image',
)
recipe_checksum = self.do_recipe_checksum(packer_tmp_directory)
packer_dst_filename = self.get_image_filename(recipe_checksum)
packer_configuration = await self.load_configuration(dependencies_path, packer_tmp_directory)
packer_dst_os_filename = join(self.image_dir,
'os',
packer_configuration['os_name'] + '_' + packer_configuration['os_version'] + '.img',
)
packer_configuration['tmp_directory'] = packer_tmp_directory
recipe = {'variables': packer_configuration}
recipe['variables']['iso_url'] = packer_dst_os_filename
self.build(packer_dst_filename,
packer_tmp_directory,
recipe,
f'{packer_dst_os_filename}.sha256',
)
@staticmethod
def build(packer_dst_filename: str,
tmp_directory: str,
recipe: dict,
sha_file: str=None,
) -> None:
packer_filename = join(tmp_directory, PACKER_FILE_NAME) packer_filename = join(tmp_directory, PACKER_FILE_NAME)
if sha_file is not None:
with open(sha_file, 'r') as fh:
sha256 = fh.read().split(' ', 1)[0]
recipe['variables']['iso_checksum'] = sha256
with open(packer_filename, 'r') as recipe_fd: with open(packer_filename, 'r') as recipe_fd:
for key, value in jload(recipe_fd).items(): for key, value in jload(recipe_fd).items():
recipe[key] = value recipe[key] = value
with open(packer_filename, 'w') as recipe_fd: with open(packer_filename, 'w') as recipe_fd:
jdump(recipe, recipe_fd, indent=2) jdump(recipe, recipe_fd, indent=2)
preprocessors = join(tmp_directory, 'preprocessors')
if isfile(preprocessors):
proc = Popen([preprocessors],
#stdout=PIPE,
#stderr=PIPE,
cwd=tmp_directory,
)
proc.wait()
if proc.returncode:
raise Exception(_(f'error when executing {preprocessors}'))
proc = Popen(['packer', 'build', packer_filename], proc = Popen(['packer', 'build', packer_filename],
#stdout=PIPE, #stdout=PIPE,
#stderr=PIPE, #stderr=PIPE,
@ -272,7 +370,9 @@ class Image:
proc.wait() proc.wait()
if proc.returncode: if proc.returncode:
raise Exception(_(f'cannot build {packer_dst_filename} with {packer_filename}')) raise Exception(_(f'cannot build {packer_dst_filename} with {packer_filename}'))
if not isdir(dirname(packer_dst_filename)):
makedirs(dirname(packer_dst_filename))
move(join(tmp_directory, 'image.img'), packer_dst_filename) move(join(tmp_directory, 'image.img'), packer_dst_filename)
move(join(tmp_directory, 'image.sha256'), f'{packer_dst_filename}.sha256') move(join(tmp_directory, 'image.sha256'), f'{packer_dst_filename}.sha256')
rmtree(tmp_directory)
print(_(f'Image {packer_dst_filename} created')) print(_(f'Image {packer_dst_filename} created'))
rmtree(tmp_directory)

View File

@ -19,8 +19,8 @@ from .utils import _
MESSAGE_ROOT_PATH = get_config()['global']['message_root_path'] MESSAGE_ROOT_PATH = get_config()['global']['message_root_path']
groups.addgroup('message') groups.addgroup('message')
MESSAGE_TRANSLATION = translation('risotto-message', join(MESSAGE_ROOT_PATH, '..', 'locale')).gettext CUSTOMTYPES = None
MESSAGE_TRANSLATION = None
class DictOption(Option): class DictOption(Option):
@ -593,6 +593,11 @@ def get_messages(current_module_names,
): ):
"""generate description from yml files """generate description from yml files
""" """
global MESSAGE_TRANSLATION, CUSTOMTYPES
if MESSAGE_TRANSLATION is None:
MESSAGE_TRANSLATION = translation('risotto-message', join(MESSAGE_ROOT_PATH, '..', 'locale')).gettext
if CUSTOMTYPES is None:
CUSTOMTYPES = load_customtypes()
optiondescriptions = {} optiondescriptions = {}
optiondescriptions_info = {} optiondescriptions_info = {}
messages = list(list_messages(uris, messages = list(list_messages(uris,
@ -632,6 +637,3 @@ def get_messages(current_module_names,
optiondescriptions, optiondescriptions,
) )
return optiondescriptions_info, root return optiondescriptions_info, root
CUSTOMTYPES = load_customtypes()

View File

@ -83,12 +83,16 @@ def register(uris: str,
uris = [uris] uris = [uris]
def decorator(function): def decorator(function):
for uri in uris: try:
dispatcher.set_function(uri, for uri in uris:
notification, dispatcher.set_function(uri,
function, notification,
function.__module__ function,
) function.__module__
)
except NameError:
# if you when register uri, please use get_dispatcher before registered uri
pass
return decorator return decorator