diff sat_frontends/jp/common.py @ 2562:26edcf3a30eb

core, setup: huge cleaning: - moved directories from src and frontends/src to sat and sat_frontends, which is the recommanded naming convention - move twisted directory to root - removed all hacks from setup.py, and added missing dependencies, it is now clean - use https URL for website in setup.py - removed "Environment :: X11 Applications :: GTK", as wix is deprecated and removed - renamed sat.sh to sat and fixed its installation - added python_requires to specify Python version needed - replaced glib2reactor which use deprecated code by gtk3reactor sat can now be installed directly from virtualenv without using --system-site-packages anymore \o/
author Goffi <goffi@goffi.org>
date Mon, 02 Apr 2018 19:44:50 +0200
parents frontends/src/jp/common.py@501b0f827f63
children c9dddf691d7b
line wrap: on
line diff
--- /dev/null	Thu Jan 01 00:00:00 1970 +0000
+++ b/sat_frontends/jp/common.py	Mon Apr 02 19:44:50 2018 +0200
@@ -0,0 +1,746 @@
+#!/usr/bin/env python2
+# -*- coding: utf-8 -*-
+
+# jp: a SàT command line tool
+# Copyright (C) 2009-2018 Jérôme Poisson (goffi@goffi.org)
+
+# This program is free software: you can redistribute it and/or modify
+# it under the terms of the GNU Affero General Public License as published by
+# the Free Software Foundation, either version 3 of the License, or
+# (at your option) any later version.
+
+# This program is distributed in the hope that it will be useful,
+# but WITHOUT ANY WARRANTY; without even the implied warranty of
+# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the
+# GNU Affero General Public License for more details.
+
+# You should have received a copy of the GNU Affero General Public License
+# along with this program.  If not, see <http://www.gnu.org/licenses/>.
+
+from sat_frontends.jp.constants import Const as C
+from sat.core.i18n import _
+from sat.core import exceptions
+from sat.tools.common import regex
+from sat.tools.common.ansi import ANSI as A
+from sat.tools.common import uri as xmpp_uri
+from sat.tools import config
+from ConfigParser import NoSectionError, NoOptionError
+from collections import namedtuple
+from functools import partial
+import json
+import os
+import os.path
+import time
+import tempfile
+import subprocess
+import glob
+import shlex
+
+# defaut arguments used for some known editors (editing with metadata)
+VIM_SPLIT_ARGS = "-c 'set nospr|vsplit|wincmd w|next|wincmd w'"
+EMACS_SPLIT_ARGS = '--eval "(split-window-horizontally)"'
+EDITOR_ARGS_MAGIC = {
+    'vim': VIM_SPLIT_ARGS + ' {content_file} {metadata_file}',
+    'gvim': VIM_SPLIT_ARGS + ' --nofork {content_file} {metadata_file}',
+    'emacs': EMACS_SPLIT_ARGS + ' {content_file} {metadata_file}',
+    'xemacs': EMACS_SPLIT_ARGS + ' {content_file} {metadata_file}',
+    'nano': ' -F {content_file} {metadata_file}',
+    }
+
+SECURE_UNLINK_MAX = 10
+SECURE_UNLINK_DIR = ".backup"
+METADATA_SUFF = '_metadata.json'
+
+
+def ansi_ljust(s, width):
+    """ljust method handling ANSI escape codes"""
+    cleaned = regex.ansiRemove(s)
+    return s + u' ' * (width - len(cleaned))
+
+
+def ansi_center(s, width):
+    """ljust method handling ANSI escape codes"""
+    cleaned = regex.ansiRemove(s)
+    diff = width - len(cleaned)
+    half = diff/2
+    return half * u' ' + s + (half + diff % 2) * u' '
+
+
+def ansi_rjust(s, width):
+    """ljust method handling ANSI escape codes"""
+    cleaned = regex.ansiRemove(s)
+    return u' ' * (width - len(cleaned)) + s
+
+
+def getTmpDir(sat_conf, cat_dir, sub_dir=None):
+    """Return directory used to store temporary files
+
+    @param sat_conf(ConfigParser.ConfigParser): instance opened on sat configuration
+    @param cat_dir(unicode): directory of the category (e.g. "blog")
+    @param sub_dir(str): sub directory where data need to be put
+        profile can be used here, or special directory name
+        sub_dir will be escaped to be usable in path (use regex.pathUnescape to find
+        initial str)
+    @return (str): path to the dir
+    """
+    local_dir = config.getConfig(sat_conf, '', 'local_dir', Exception)
+    path = [local_dir.encode('utf-8'), cat_dir.encode('utf-8')]
+    if sub_dir is not None:
+        path.append(regex.pathEscape(sub_dir))
+    return os.path.join(*path)
+
+
+def parse_args(host, cmd_line, **format_kw):
+    """Parse command arguments
+
+    @param cmd_line(unicode): command line as found in sat.conf
+    @param format_kw: keywords used for formating
+    @return (list(unicode)): list of arguments to pass to subprocess function
+    """
+    try:
+        # we split the arguments and add the known fields
+        # we split arguments first to avoid escaping issues in file names
+        return [a.format(**format_kw) for a in shlex.split(cmd_line)]
+    except ValueError as e:
+        host.disp(u"Couldn't parse editor cmd [{cmd}]: {reason}".format(cmd=cmd_line, reason=e))
+        return []
+
+
+class BaseEdit(object):
+    u"""base class for editing commands
+
+    This class allows to edit file for PubSub or something else.
+    It works with temporary files in SàT local_dir, in a "cat_dir" subdir
+    """
+
+    def __init__(self, host, cat_dir, use_metadata=False):
+        """
+        @param sat_conf(ConfigParser.ConfigParser): instance opened on sat configuration
+        @param cat_dir(unicode): directory to use for drafts
+            this will be a sub-directory of SàT's local_dir
+        @param use_metadata(bool): True is edition need a second file for metadata
+            most of signature change with use_metadata with an additional metadata argument.
+            This is done to raise error if a command needs metadata but forget the flag, and vice versa
+        """
+        self.host = host
+        self.sat_conf = config.parseMainConf()
+        self.cat_dir_str = cat_dir.encode('utf-8')
+        self.use_metadata = use_metadata
+
+    def secureUnlink(self, path):
+        """Unlink given path after keeping it for a while
+
+        This method is used to prevent accidental deletion of a draft
+        If there are more file in SECURE_UNLINK_DIR than SECURE_UNLINK_MAX,
+        older file are deleted
+        @param path(str): file to unlink
+        """
+        if not os.path.isfile(path):
+            raise OSError(u"path must link to a regular file")
+        if not path.startswith(getTmpDir(self.sat_conf, self.cat_dir_str)):
+            self.disp(u"File {} is not in SàT temporary hierarchy, we do not remove it".format(path.decode('utf-8')), 2)
+            return
+        # we have 2 files per draft with use_metadata, so we double max
+        unlink_max = SECURE_UNLINK_MAX * 2 if self.use_metadata else SECURE_UNLINK_MAX
+        backup_dir = getTmpDir(self.sat_conf, self.cat_dir_str, SECURE_UNLINK_DIR)
+        if not os.path.exists(backup_dir):
+            os.makedirs(backup_dir)
+        filename = os.path.basename(path)
+        backup_path = os.path.join(backup_dir, filename)
+        # we move file to backup dir
+        self.host.disp(u"Backuping file {src} to {dst}".format(
+            src=path.decode('utf-8'), dst=backup_path.decode('utf-8')), 1)
+        os.rename(path, backup_path)
+        # and if we exceeded the limit, we remove older file
+        backup_files = [os.path.join(backup_dir, f) for f in os.listdir(backup_dir)]
+        if len(backup_files) > unlink_max:
+            backup_files.sort(key=lambda path: os.stat(path).st_mtime)
+            for path in backup_files[:len(backup_files) - unlink_max]:
+                self.host.disp(u"Purging backup file {}".format(path.decode('utf-8')), 2)
+                os.unlink(path)
+
+    def runEditor(self, editor_args_opt, content_file_path,
+                 content_file_obj, meta_file_path=None, meta_ori=None):
+        """run editor to edit content and metadata
+
+        @param editor_args_opt(unicode): option in [jp] section in configuration for
+            specific args
+        @param content_file_path(str): path to the content file
+        @param content_file_obj(file): opened file instance
+        @param meta_file_path(str, None): metadata file path
+            if None metadata will not be used
+        @param meta_ori(dict, None): original cotent of metadata
+            can't be used if use_metadata is False
+        """
+        if not self.use_metadata:
+            assert meta_file_path is None
+            assert meta_ori is None
+
+        # we calculate hashes to check for modifications
+        import hashlib
+        content_file_obj.seek(0)
+        tmp_ori_hash = hashlib.sha1(content_file_obj.read()).digest()
+        content_file_obj.close()
+
+        # we prepare arguments
+        editor = config.getConfig(self.sat_conf, 'jp', 'editor') or os.getenv('EDITOR', 'vi')
+        try:
+            # is there custom arguments in sat.conf ?
+            editor_args = config.getConfig(self.sat_conf, 'jp', editor_args_opt, Exception)
+        except (NoOptionError, NoSectionError):
+            # no, we check if we know the editor and have special arguments
+            if self.use_metadata:
+                editor_args = EDITOR_ARGS_MAGIC.get(os.path.basename(editor), '')
+            else:
+                editor_args = ''
+        parse_kwargs = {'content_file': content_file_path}
+        if self.use_metadata:
+            parse_kwargs['metadata_file'] = meta_file_path
+        args = parse_args(self.host, editor_args, **parse_kwargs)
+        if not args:
+            args = [content_file_path]
+
+        # actual editing
+        editor_exit = subprocess.call([editor] + args)
+
+        # edition will now be checked, and data will be sent if it was a success
+        if editor_exit != 0:
+            self.disp(u"Editor exited with an error code, so temporary file has not be deleted, and item is not published.\nYou can find temporary file at {path}".format(
+                path=content_file_path), error=True)
+        else:
+            # main content
+            try:
+                with open(content_file_path, 'rb') as f:
+                    content = f.read()
+            except (OSError, IOError):
+                self.disp(u"Can read file at {content_path}, have it been deleted?\nCancelling edition".format(
+                    content_path=content_file_path), error=True)
+                self.host.quit(C.EXIT_NOT_FOUND)
+
+            # metadata
+            if self.use_metadata:
+                try:
+                    with open(meta_file_path, 'rb') as f:
+                        metadata = json.load(f)
+                except (OSError, IOError):
+                    self.disp(u"Can read file at {meta_file_path}, have it been deleted?\nCancelling edition".format(
+                        content_path=content_file_path, meta_path=meta_file_path), error=True)
+                    self.host.quit(C.EXIT_NOT_FOUND)
+                except ValueError:
+                    self.disp(u"Can't parse metadata, please check it is correct JSON format. Cancelling edition.\n" +
+                        "You can find tmp file at {content_path} and temporary meta file at {meta_path}.".format(
+                        content_path=content_file_path,
+                        meta_path=meta_file_path), error=True)
+                    self.host.quit(C.EXIT_DATA_ERROR)
+
+            if self.use_metadata and not C.bool(metadata.get('publish', "true")):
+                self.disp(u'Publication blocked by "publish" key in metadata, cancelling edition.\n\n' +
+                    "temporary file path:\t{content_path}\nmetadata file path:\t{meta_path}".format(
+                    content_path=content_file_path, meta_path=meta_file_path), error=True)
+                self.host.quit()
+
+            if len(content) == 0:
+                self.disp(u"Content is empty, cancelling the edition")
+                if not content_file_path.startswith(getTmpDir(self.sat_conf, self.cat_dir_str)):
+                    self.disp(u"File are not in SàT temporary hierarchy, we do not remove them", 2)
+                    self.host.quit()
+                self.disp(u"Deletion of {}".format(content_file_path.decode('utf-8')), 2)
+                os.unlink(content_file_path)
+                if self.use_metadata:
+                    self.disp(u"Deletion of {}".format(meta_file_path.decode('utf-8')), 2)
+                    os.unlink(meta_file_path)
+                self.host.quit()
+
+            # time to re-check the hash
+            elif (tmp_ori_hash == hashlib.sha1(content).digest() and
+                  (not self.use_metadata or meta_ori == metadata)):
+                self.disp(u"The content has not been modified, cancelling the edition")
+                self.host.quit()
+
+            else:
+                # we can now send the item
+                content = content.decode('utf-8-sig') # we use utf-8-sig to avoid BOM
+                try:
+                    if self.use_metadata:
+                        self.publish(content, metadata)
+                    else:
+                        self.publish(content)
+                except Exception as e:
+                    if self.use_metadata:
+                        self.disp(u"Error while sending your item, the temporary files have been kept at {content_path} and {meta_path}: {reason}".format(
+                            content_path=content_file_path, meta_path=meta_file_path, reason=e), error=True)
+                    else:
+                        self.disp(u"Error while sending your item, the temporary file has been kept at {content_path}: {reason}".format(
+                            content_path=content_file_path, reason=e), error=True)
+                    self.host.quit(1)
+
+            self.secureUnlink(content_file_path)
+            if self.use_metadata:
+                self.secureUnlink(meta_file_path)
+
+    def publish(self, content):
+        # if metadata is needed, publish will be called with it last argument
+        raise NotImplementedError
+
+    def getTmpFile(self):
+        """Create a temporary file
+
+        @param suff (str): suffix to use for the filename
+        @return (tuple(file, str)): opened (w+b) file object and file path
+        """
+        suff = '.' + self.getTmpSuff()
+        cat_dir_str = self.cat_dir_str
+        tmp_dir = getTmpDir(self.sat_conf, self.cat_dir_str, self.profile.encode('utf-8'))
+        if not os.path.exists(tmp_dir):
+            try:
+                os.makedirs(tmp_dir)
+            except OSError as e:
+                self.disp(u"Can't create {path} directory: {reason}".format(
+                    path=tmp_dir, reason=e), error=True)
+                self.host.quit(1)
+        try:
+            fd, path = tempfile.mkstemp(suffix=suff.encode('utf-8'),
+                prefix=time.strftime(cat_dir_str + '_%Y-%m-%d_%H:%M:%S_'),
+                dir=tmp_dir, text=True)
+            return os.fdopen(fd, 'w+b'), path
+        except OSError as e:
+            self.disp(u"Can't create temporary file: {reason}".format(reason=e), error=True)
+            self.host.quit(1)
+
+    def getCurrentFile(self, profile):
+        """Get most recently edited file
+
+        @param profile(unicode): profile linked to the draft
+        @return(str): full path of current file
+        """
+        # we guess the item currently edited by choosing
+        # the most recent file corresponding to temp file pattern
+        # in tmp_dir, excluding metadata files
+        cat_dir_str = self.cat_dir_str
+        tmp_dir = getTmpDir(self.sat_conf, self.cat_dir_str, profile.encode('utf-8'))
+        available = [path for path in glob.glob(os.path.join(tmp_dir, cat_dir_str + '_*')) if not path.endswith(METADATA_SUFF)]
+        if not available:
+            self.disp(u"Could not find any content draft in {path}".format(path=tmp_dir), error=True)
+            self.host.quit(1)
+        return max(available, key=lambda path: os.stat(path).st_mtime)
+
+    def getItemData(self, service, node, item):
+        """return formatted content, metadata (or not if use_metadata is false), and item id"""
+        raise NotImplementedError
+
+    def getTmpSuff(self):
+        """return suffix used for content file"""
+        return u'xml'
+
+    def getItemPath(self):
+        """retrieve item path (i.e. service and node) from item argument
+
+        This method is obviously only useful for edition of PubSub based features
+        """
+        service = self.args.service
+        node = self.args.node
+        item = self.args.item
+        last_item = self.args.last_item
+
+        if self.args.current:
+            # user wants to continue current draft
+            content_file_path = self.getCurrentFile(self.profile)
+            self.disp(u'Continuing edition of current draft', 2)
+            content_file_obj = open(content_file_path, 'r+b')
+            # we seek at the end of file in case of an item already exist
+            # this will write content of the existing item at the end of the draft.
+            # This way no data should be lost.
+            content_file_obj.seek(0, os.SEEK_END)
+        elif self.args.draft_path:
+            # there is an existing draft that we use
+            content_file_path = os.path.expanduser(self.args.item)
+            content_file_obj = open(content_file_path, 'r+b')
+            # we seek at the end for the same reason as above
+            content_file_obj.seek(0, os.SEEK_END)
+        else:
+            # we need a temporary file
+            content_file_obj, content_file_path = self.getTmpFile()
+
+        if item or last_item:
+            self.disp(u'Editing requested published item', 2)
+            try:
+                if self.use_metadata:
+                    content, metadata, item = self.getItemData(service, node, item)
+                else:
+                    content, item = self.getItemData(service, node, item)
+            except Exception as e:
+                # FIXME: ugly but we have not good may to check errors in bridge
+                if u'item-not-found' in unicode(e):
+                    # item doesn't exist, we create a new one with requested id
+                    metadata = None
+                    if last_item:
+                        self.disp(_(u'no item found at all, we create a new one'), 2)
+                    else:
+                        self.disp(_(u'item "{item_id}" not found, we create a new item with this id').format(item_id=item), 2)
+                    content_file_obj.seek(0)
+                else:
+                    self.disp(u"Error while retrieving item: {}".format(e))
+                    self.host.quit(C.EXIT_ERROR)
+            else:
+                # item exists, we write content
+                if content_file_obj.tell() != 0:
+                    # we already have a draft,
+                    # we copy item content after it and add an indicator
+                    content_file_obj.write('\n*****\n')
+                content_file_obj.write(content.encode('utf-8'))
+                content_file_obj.seek(0)
+                self.disp(_(u'item "{item_id}" found, we edit it').format(item_id=item), 2)
+        else:
+            self.disp(u'Editing a new item', 2)
+            if self.use_metadata:
+                metadata = None
+
+        if self.use_metadata:
+            return service, node, item, content_file_path, content_file_obj, metadata
+        else:
+            return service, node, item, content_file_path, content_file_obj
+
+
+class Table(object):
+
+    def __init__(self, host, data, headers=None, filters=None, use_buffer=False):
+        """
+        @param data(iterable[list]): table data
+            all lines must have the same number of columns
+        @param headers(iterable[unicode], None): names/titles of the columns
+            if not None, must have same number of columns as data
+        @param filters(iterable[(callable, unicode)], None): values filters
+            the callable will get col value as argument and must return a string
+            if it's unicode, it will be used with .format and must countain u'{}' which will be replaced with the string
+            if not None, must have same number of columns as data
+        @param use_buffer(bool): if True, bufferise output instead of printing it directly
+        """
+        self.host = host
+        self._buffer = [] if use_buffer else None
+        # headers are columns names/titles, can be None
+        self.headers = headers
+        # sizes fof columns without headers,
+        # headers may be larger
+        self.sizes = []
+        # rows countains one list per row with columns values
+        self.rows = []
+
+        size = None
+        if headers:
+            row_cls = namedtuple('RowData', headers)
+        else:
+            row_cls = tuple
+
+        for row_data in data:
+            new_row = []
+            row_data_list = list(row_data)
+            for idx, value in enumerate(row_data_list):
+                if filters is not None and filters[idx] is not None:
+                    filter_ = filters[idx]
+                    if isinstance(filter_, basestring):
+                        col_value = filter_.format(value)
+                    else:
+                        col_value = filter_(value, row_cls(*row_data_list))
+                    # we count size without ANSI code as they will change length of the string
+                    # when it's mostly style/color changes.
+                    col_size = len(regex.ansiRemove(col_value))
+                else:
+                    col_value = unicode(value)
+                    col_size = len(col_value)
+                new_row.append(col_value)
+                if size is None:
+                    self.sizes.append(col_size)
+                else:
+                    self.sizes[idx] = max(self.sizes[idx], col_size)
+            if size is None:
+                size = len(new_row)
+                if headers is not None and len(headers) != size:
+                    raise exceptions.DataError(u'headers size is not coherent with rows')
+            else:
+                if len(new_row) != size:
+                    raise exceptions.DataError(u'rows size is not coherent')
+            self.rows.append(new_row)
+
+        if not data and headers is not None:
+            # the table is empty, we print headers at their lenght
+            self.sizes = [len(h) for h in headers]
+
+    @property
+    def string(self):
+        if self._buffer is None:
+            raise exceptions.InternalError(u'buffer must be used to get a string')
+        return u'\n'.join(self._buffer)
+
+    @staticmethod
+    def readDictValues(data, keys, defaults=None):
+        if defaults is None:
+            defaults = {}
+        for key in keys:
+            try:
+                yield data[key]
+            except KeyError as e:
+                default = defaults.get(key)
+                if default is not None:
+                    yield default
+                else:
+                    raise e
+
+    @classmethod
+    def fromDict(cls, host, data, keys=None, headers=None, filters=None, defaults=None):
+        """Prepare a table to display it
+
+        the whole data will be read and kept into memory,
+        to be printed
+        @param data(list[dict[unicode, unicode]]): data to create the table from
+        @param keys(iterable[unicode], None): keys to get
+            if None, all keys will be used
+        @param headers(iterable[unicode], None): name of the columns
+            names must be in same order as keys
+        @param filters(dict[unicode, (callable,unicode)), None): filter to use on values
+            keys correspond to keys to filter, and value is a callable or unicode which
+            will get the value as argument and must return a string
+        @param defaults(dict[unicode, unicode]): default value to use
+            if None, an exception will be raised if not value is found
+        """
+        if keys is None and headers is not None:
+            # FIXME: keys are not needed with OrderedDict,
+            raise exceptions.DataError(u'You must specify keys order to used headers')
+        if keys is None:
+            keys = data[0].keys()
+        if headers is None:
+            headers = keys
+        filters = [filters.get(k) for k in keys]
+        return cls(host, (cls.readDictValues(d, keys, defaults) for d in data), headers, filters)
+
+    def _headers(self, head_sep, headers, sizes, alignment=u'left', style=None):
+        """Render headers
+
+        @param head_sep(unicode): sequence to use as separator
+        @param alignment(unicode): how to align, can be left, center or right
+        @param style(unicode, iterable[unicode], None): ANSI escape sequences to apply
+        @param headers(list[unicode]): headers to show
+        @param sizes(list[int]): sizes of columns
+        """
+        rendered_headers = []
+        if isinstance(style, basestring):
+            style = [style]
+        for idx, header in enumerate(headers):
+            size = sizes[idx]
+            if alignment == u'left':
+                rendered = header[:size].ljust(size)
+            elif alignment == u'center':
+                rendered = header[:size].center(size)
+            elif alignment == u'right':
+                rendered = header[:size].rjust(size)
+            else:
+                raise exceptions.InternalError(u'bad alignment argument')
+            if style:
+                args = style + [rendered]
+                rendered = A.color(*args)
+            rendered_headers.append(rendered)
+        return head_sep.join(rendered_headers)
+
+    def _disp(self, data):
+        """output data (can be either bufferised or printed)"""
+        if self._buffer is not None:
+            self._buffer.append(data)
+        else:
+            self.host.disp(data)
+
+    def display(self,
+                head_alignment = u'left',
+                columns_alignment = u'left',
+                head_style = None,
+                show_header=True,
+                show_borders=True,
+                hide_cols=None,
+                col_sep=u' │ ',
+                top_left=u'┌',
+                top=u'─',
+                top_sep=u'─┬─',
+                top_right=u'┐',
+                left=u'│',
+                right=None,
+                head_sep=None,
+                head_line=u'┄',
+                head_line_left=u'├',
+                head_line_sep=u'┄┼┄',
+                head_line_right=u'┤',
+                bottom_left=u'└',
+                bottom=None,
+                bottom_sep=u'─┴─',
+                bottom_right=u'┘',
+                ):
+        """Print the table
+
+        @param show_header(bool): True if header need no be shown
+        @param show_borders(bool): True if borders need no be shown
+        @param hide_cols(None, iterable(unicode)): columns which should not be displayed
+        @param head_alignment(unicode): how to align headers, can be left, center or right
+        @param columns_alignment(unicode): how to align columns, can be left, center or right
+        @param col_sep(unicode): separator betweens columns
+        @param head_line(unicode): character to use to make line under head
+        @param disp(callable, None): method to use to display the table
+            None to use self.host.disp
+        """
+        if not self.sizes:
+            # the table is empty
+            return
+        col_sep_size = len(regex.ansiRemove(col_sep))
+
+        # if we have columns to hide, we remove them from headers and size
+        if not hide_cols:
+            headers = self.headers
+            sizes = self.sizes
+        else:
+            headers = list(self.headers)
+            sizes = self.sizes[:]
+            ignore_idx = [headers.index(to_hide) for to_hide in hide_cols]
+            for to_hide in hide_cols:
+                hide_idx = headers.index(to_hide)
+                del headers[hide_idx]
+                del sizes[hide_idx]
+
+        if right is None:
+            right = left
+        if top_sep is None:
+            top_sep = col_sep_size * top
+        if head_sep is None:
+            head_sep = col_sep
+        if bottom is None:
+            bottom = top
+        if bottom_sep is None:
+            bottom_sep = col_sep_size * bottom
+        if not show_borders:
+            left = right = head_line_left = head_line_right = u''
+        # top border
+        if show_borders:
+            self._disp(
+                top_left
+                + top_sep.join([top*size for size in sizes])
+                + top_right
+                )
+
+        # headers
+        if show_header:
+            self._disp(
+                left
+                + self._headers(head_sep, headers, sizes, head_alignment, head_style)
+                + right
+                )
+            # header line
+            self._disp(
+                head_line_left
+                + head_line_sep.join([head_line*size for size in sizes])
+                + head_line_right
+                )
+
+        # content
+        if columns_alignment == u'left':
+            alignment = lambda idx, s: ansi_ljust(s, sizes[idx])
+        elif columns_alignment == u'center':
+            alignment = lambda idx, s: ansi_center(s, sizes[idx])
+        elif columns_alignment == u'right':
+            alignment = lambda idx, s: ansi_rjust(s, sizes[idx])
+        else:
+            raise exceptions.InternalError(u'bad columns alignment argument')
+
+        for row in self.rows:
+            if hide_cols:
+                row = [v for idx,v in enumerate(row) if idx not in ignore_idx]
+            self._disp(left + col_sep.join([alignment(idx,c) for idx,c in enumerate(row)]) + right)
+
+        if show_borders:
+            # bottom border
+            self._disp(
+                bottom_left
+                + bottom_sep.join([bottom*size for size in sizes])
+                + bottom_right
+                )
+        # we return self so string can be used after display (table.display().string)
+        return self
+
+    def display_blank(self, **kwargs):
+        """Display table without visible borders"""
+        kwargs_ = {'col_sep':u' ', 'head_line_sep':u' ', 'show_borders':False}
+        kwargs_.update(kwargs)
+        return self.display(**kwargs_)
+
+
+class URIFinder(object):
+    """Helper class to find URIs in well-known locations"""
+
+    def __init__(self, command, path, key, callback, meta_map=None):
+        """
+        @param command(CommandBase): command instance
+            args of this instance will be updated with found values
+        @param path(unicode): absolute path to use as a starting point to look for URIs
+        @param key(unicode): key to look for
+        @param callback(callable): method to call once URIs are found (or not)
+        @param meta_map(dict, None): if not None, map metadata to arg name
+            key is metadata used attribute name
+            value is name to actually use, or None to ignore
+            use empty dict to only retrieve URI
+            possible keys are currently:
+                - labels
+        """
+        if not command.args.service and not command.args.node:
+            self.host = command.host
+            self.args = command.args
+            self.key = key
+            self.callback = callback
+            self.meta_map = meta_map
+            self.host.bridge.URIFind(path,
+                                     [key],
+                                     callback=self.URIFindCb,
+                                     errback=partial(command.errback,
+                                                     msg=_(u"can't find " + key + u" URI: {}"),
+                                                     exit_code=C.EXIT_BRIDGE_ERRBACK))
+        else:
+            callback()
+
+    def setMetadataList(self, uri_data, key):
+        """Helper method to set list of values from metadata
+
+        @param uri_data(dict): data of the found URI
+        @param key(unicode): key of the value to retrieve
+        """
+        new_values_json = uri_data.get(key)
+        if uri_data is not None:
+            if self.meta_map is None:
+                dest = key
+            else:
+                dest = self.meta_map.get(key)
+                if dest is None:
+                    return
+
+            try:
+                values = getattr(self.args, key)
+            except AttributeError:
+                raise exceptions.InternalError(u'there is no "{key}" arguments'.format(
+                    key=key))
+            else:
+                if values is None:
+                    values = []
+                values.extend(json.loads(new_values_json))
+                setattr(self.args, dest, values)
+
+
+    def URIFindCb(self, uris_data):
+        try:
+            uri_data = uris_data[self.key]
+        except KeyError:
+            self.host.disp(_(u"No {key} URI specified for this project, please specify service and node").format(key=self.key), error=True)
+            self.host.quit(C.EXIT_NOT_FOUND)
+        else:
+            uri = uri_data[u'uri']
+
+        self.setMetadataList(uri_data, u'labels')
+        parsed_uri = xmpp_uri.parseXMPPUri(uri)
+        try:
+            self.args.service = parsed_uri[u'path']
+            self.args.node = parsed_uri[u'node']
+        except KeyError:
+            self.host.disp(_(u"Invalid URI found: {uri}").format(uri=uri), error=True)
+            self.host.quit(C.EXIT_DATA_ERROR)
+        self.callback()