Mercurial > libervia-backend
view libervia/backend/plugins/plugin_misc_merge_requests.py @ 4270:0d7bb4df2343
Reformatted code base using black.
author | Goffi <goffi@goffi.org> |
---|---|
date | Wed, 19 Jun 2024 18:44:57 +0200 |
parents | 4b842c1fb686 |
children |
line wrap: on
line source
#!/usr/bin/env python3 # SAT plugin for Pubsub Schemas # Copyright (C) 2009-2021 Jérôme Poisson (goffi@goffi.org) # This program is free software: you can redistribute it and/or modify # it under the terms of the GNU Affero General Public License as published by # the Free Software Foundation, either version 3 of the License, or # (at your option) any later version. # This program is distributed in the hope that it will be useful, # but WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the # GNU Affero General Public License for more details. # You should have received a copy of the GNU Affero General Public License # along with this program. If not, see <http://www.gnu.org/licenses/>. from collections import namedtuple from twisted.internet import defer from twisted.words.protocols.jabber import jid from libervia.backend.core.i18n import _ from libervia.backend.core.constants import Const as C from libervia.backend.core import exceptions from libervia.backend.tools.common import data_format from libervia.backend.core.log import getLogger log = getLogger(__name__) APP_NS_MERGE_REQUESTS = "org.salut-a-toi.merge_requests:0" PLUGIN_INFO = { C.PI_NAME: _("Merge requests management"), C.PI_IMPORT_NAME: "MERGE_REQUESTS", C.PI_TYPE: "EXP", C.PI_PROTOCOLS: [], C.PI_DEPENDENCIES: ["XEP-0060", "XEP-0346", "LISTS", "TEXT_SYNTAXES"], C.PI_MAIN: "MergeRequests", C.PI_HANDLER: "no", C.PI_DESCRIPTION: _("""Merge requests management plugin"""), } FIELD_DATA_TYPE = "type" FIELD_DATA = "request_data" MergeRequestHandler = namedtuple( "MergeRequestHandler", ["name", "handler", "data_types", "short_desc", "priority"] ) class MergeRequests(object): META_AUTHOR = "author" META_EMAIL = "email" META_TIMESTAMP = "timestamp" META_HASH = "hash" META_PARENT_HASH = "parent_hash" META_COMMIT_MSG = "commit_msg" META_DIFF = "diff" # index of the diff in the whole data # needed to retrieve comments location META_DIFF_IDX = "diff_idx" def __init__(self, host): log.info(_("Merge requests plugin initialization")) self.host = host self._s = self.host.plugins["XEP-0346"] self.namespace = self._s.get_submitted_ns(APP_NS_MERGE_REQUESTS) host.register_namespace("merge_requests", self.namespace) self._p = self.host.plugins["XEP-0060"] self._t = self.host.plugins["LISTS"] self._handlers = {} self._handlers_list = [] # handlers sorted by priority self._type_handlers = {} # data type => handler map host.bridge.add_method( "merge_requests_get", ".plugin", in_sign="ssiassss", out_sign="s", method=self._get, async_=True, ) host.bridge.add_method( "merge_request_set", ".plugin", in_sign="ssssa{sas}ssss", out_sign="s", method=self._set, async_=True, ) host.bridge.add_method( "merge_requests_schema_get", ".plugin", in_sign="sss", out_sign="s", method=lambda service, nodeIdentifier, profile_key: self._s._get_ui_schema( service, nodeIdentifier, default_node=self.namespace, profile_key=profile_key, ), async_=True, ) host.bridge.add_method( "merge_request_parse_data", ".plugin", in_sign="ss", out_sign="aa{ss}", method=self._parse_data, async_=True, ) host.bridge.add_method( "merge_requests_import", ".plugin", in_sign="ssssa{ss}s", out_sign="", method=self._import, async_=True, ) def register(self, name, handler, data_types, short_desc, priority=0): """register an merge request handler @param name(unicode): name of the handler @param handler(object): instance of the handler. It must have the following methods, which may all return a Deferred: - check(repository)->bool: True if repository can be handled - export(repository)->str: return export data, i.e. the patches - parse(export_data): parse report data and return a list of dict (1 per patch) with: - title: title of the commit message (first line) - body: body of the commit message @aram data_types(list[unicode]): data types that his handler can generate or parse """ if name in self._handlers: raise exceptions.ConflictError( _("a handler with name {name} already " "exists!").format(name=name) ) self._handlers[name] = MergeRequestHandler( name, handler, data_types, short_desc, priority ) self._handlers_list.append(name) self._handlers_list.sort(key=lambda name: self._handlers[name].priority) if isinstance(data_types, str): data_types = [data_types] for data_type in data_types: if data_type in self._type_handlers: log.warning( _( "merge requests of type {type} are already handled by " "{old_handler}, ignoring {new_handler}" ).format( type=data_type, old_handler=self._type_handlers[data_type].name, new_handler=name, ) ) continue self._type_handlers[data_type] = self._handlers[name] def serialise(self, get_data): tickets_xmlui, metadata, items_patches = get_data tickets_xmlui_s, metadata = self._p.trans_items_data((tickets_xmlui, metadata)) return data_format.serialise( { "items": tickets_xmlui_s, "metadata": metadata, "items_patches": items_patches, } ) def _get( self, service="", node="", max_items=10, item_ids=None, sub_id=None, extra="", profile_key=C.PROF_KEY_NONE, ): extra = data_format.deserialise(extra) client, service, node, max_items, extra, sub_id = self._s.prepare_bridge_get( service, node, max_items, sub_id, extra, profile_key ) d = self.get( client, service, node or None, max_items, item_ids, sub_id or None, extra.rsm_request, extra.extra, ) d.addCallback(self.serialise) return d @defer.inlineCallbacks def get( self, client, service=None, node=None, max_items=None, item_ids=None, sub_id=None, rsm_request=None, extra=None, ): """Retrieve merge requests and convert them to XMLUI @param extra(XEP-0060.parse, None): can have following keys: - update(bool): if True, will return list of parsed request data other params are the same as for [TICKETS._get] @return (tuple[list[unicode], list[dict[unicode, unicode]])): tuple with - XMLUI of the tickets, like [TICKETS._get] - node metadata - list of parsed request data (if extra['parse'] is set, else empty list) """ if not node: node = self.namespace if extra is None: extra = {} # XXX: Q&D way to get list for labels when displaying them, but text when we # have to modify them if C.bool(extra.get("labels_as_list", C.BOOL_FALSE)): filters = {"labels": self._s.textbox_2_list_filter} else: filters = {} tickets_xmlui, metadata = yield defer.ensureDeferred( self._s.get_data_form_items( client, service, node, max_items=max_items, item_ids=item_ids, sub_id=sub_id, rsm_request=rsm_request, extra=extra, form_ns=APP_NS_MERGE_REQUESTS, filters=filters, ) ) parsed_patches = [] if extra.get("parse", False): for ticket in tickets_xmlui: request_type = ticket.named_widgets[FIELD_DATA_TYPE].value request_data = ticket.named_widgets[FIELD_DATA].value parsed_data = yield self.parse_data(request_type, request_data) parsed_patches.append(parsed_data) defer.returnValue((tickets_xmlui, metadata, parsed_patches)) def _set( self, service, node, repository, method, values, schema=None, item_id=None, extra="", profile_key=C.PROF_KEY_NONE, ): client, service, node, schema, item_id, extra = self._s.prepare_bridge_set( service, node, schema, item_id, extra, profile_key ) d = defer.ensureDeferred( self.set( client, service, node, repository, method, values, schema, item_id or None, extra, deserialise=True, ) ) d.addCallback(lambda ret: ret or "") return d async def set( self, client, service, node, repository, method="auto", values=None, schema=None, item_id=None, extra=None, deserialise=False, ): """Publish a tickets @param service(None, jid.JID): Pubsub service to use @param node(unicode, None): Pubsub node to use None to use default tickets node @param repository(unicode): path to the repository where the code stands @param method(unicode): name of one of the registered handler, or "auto" to try autodetection. other arguments are same as for [TICKETS.set] @return (unicode): id of the created item """ if not node: node = self.namespace if values is None: values = {} update = extra.get("update", False) if not repository and not update: # in case of update, we may re-user former patches data # so repository is not mandatory raise exceptions.DataError(_("repository must be specified")) if FIELD_DATA in values: raise exceptions.DataError( _("{field} is set by backend, you must not set " "it in frontend").format( field=FIELD_DATA ) ) if repository: if method == "auto": for name in self._handlers_list: handler = self._handlers[name].handler can_handle = await handler.check(repository) if can_handle: log.info(_("{name} handler will be used").format(name=name)) break else: log.warning( _( "repository {path} can't be handled by any installed " "handler" ).format(path=repository) ) raise exceptions.NotFound( _("no handler for this repository has " "been found") ) else: try: handler = self._handlers[name].handler except KeyError: raise exceptions.NotFound(_("No handler of this name found")) data = await handler.export(repository) if not data.strip(): raise exceptions.DataError( _("export data is empty, do you have any " "change to send?") ) if not values.get("title") or not values.get("body"): patches = handler.parse(data, values.get(FIELD_DATA_TYPE)) commits_msg = patches[-1][self.META_COMMIT_MSG] msg_lines = commits_msg.splitlines() if not values.get("title"): values["title"] = msg_lines[0] if not values.get("body"): ts = self.host.plugins["TEXT_SYNTAXES"] xhtml = await ts.convert( "\n".join(msg_lines[1:]), syntax_from=ts.SYNTAX_TEXT, syntax_to=ts.SYNTAX_XHTML, profile=client.profile, ) values["body"] = '<div xmlns="{ns}">{xhtml}</div>'.format( ns=C.NS_XHTML, xhtml=xhtml ) values[FIELD_DATA] = data item_id = await self._t.set( client, service, node, values, schema, item_id, extra, deserialise, form_ns=APP_NS_MERGE_REQUESTS, ) return item_id def _parse_data(self, data_type, data): d = self.parse_data(data_type, data) d.addCallback( lambda parsed_patches: { key: str(value) for key, value in parsed_patches.items() } ) return d def parse_data(self, data_type, data): """Parse a merge request data according to type @param data_type(unicode): type of the data to parse @param data(unicode): data to parse @return(list[dict[unicode, unicode]]): parsed data key of dictionary are self.META_* or keys specifics to handler @raise NotFound: no handler can parse this data_type """ try: handler = self._type_handlers[data_type] except KeyError: raise exceptions.NotFound( _('No handler can handle data type "{type}"').format(type=data_type) ) return defer.maybeDeferred(handler.handler.parse, data, data_type) def _import( self, repository, item_id, service=None, node=None, extra=None, profile_key=C.PROF_KEY_NONE, ): client = self.host.get_client(profile_key) service = jid.JID(service) if service else None d = self.import_request( client, repository, item_id, service, node or None, extra=extra or None ) return d @defer.inlineCallbacks def import_request( self, client, repository, item, service=None, node=None, extra=None ): """import a merge request in specified directory @param repository(unicode): path to the repository where the code stands """ if not node: node = self.namespace tickets_xmlui, metadata = yield defer.ensureDeferred( self._s.get_data_form_items( client, service, node, max_items=1, item_ids=[item], form_ns=APP_NS_MERGE_REQUESTS, ) ) ticket_xmlui = tickets_xmlui[0] data = ticket_xmlui.named_widgets[FIELD_DATA].value data_type = ticket_xmlui.named_widgets[FIELD_DATA_TYPE].value try: handler = self._type_handlers[data_type] except KeyError: raise exceptions.NotFound( _("No handler found to import {data_type}").format(data_type=data_type) ) log.info( _("Importing patch [{item_id}] using {name} handler").format( item_id=item, name=handler.name ) ) yield handler.handler.import_( repository, data, data_type, item, service, node, extra )