Mercurial > libervia-backend
view sat/plugins/plugin_blog_import.py @ 3981:acc9dfc8ba8d
component AP gateway: parse body immediately on `POST` request:
the body is parsed immediately during a `POST` request: this avoids duplication of code,
and allows to check the body data before continuing (will be used to filter some requests
in a future patch).
author | Goffi <goffi@goffi.org> |
---|---|
date | Tue, 15 Nov 2022 18:07:34 +0100 |
parents | 0ff265725489 |
children | 524856bd7b19 |
line wrap: on
line source
#!/usr/bin/env python3 # SàT plugin for import external blogs # Copyright (C) 2009-2021 Jérôme Poisson (goffi@goffi.org) # This program is free software: you can redistribute it and/or modify # it under the terms of the GNU Affero General Public License as published by # the Free Software Foundation, either version 3 of the License, or # (at your option) any later version. # This program is distributed in the hope that it will be useful, # but WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the # GNU Affero General Public License for more details. # You should have received a copy of the GNU Affero General Public License # along with this program. If not, see <http://www.gnu.org/licenses/>. from sat.core.i18n import _ from sat.core.constants import Const as C from sat.core.log import getLogger log = getLogger(__name__) from twisted.internet import defer from twisted.web import client as web_client from twisted.words.xish import domish from sat.core import exceptions from sat.tools import xml_tools import os import os.path import tempfile import urllib.parse import shortuuid PLUGIN_INFO = { C.PI_NAME: "blog import", C.PI_IMPORT_NAME: "BLOG_IMPORT", C.PI_TYPE: (C.PLUG_TYPE_BLOG, C.PLUG_TYPE_IMPORT), C.PI_DEPENDENCIES: ["IMPORT", "XEP-0060", "XEP-0277", "TEXT_SYNTAXES", "UPLOAD"], C.PI_MAIN: "BlogImportPlugin", C.PI_HANDLER: "no", C.PI_DESCRIPTION: _( """Blog import management: This plugin manage the different blog importers which can register to it, and handle generic importing tasks.""" ), } OPT_HOST = "host" OPT_UPLOAD_IMAGES = "upload_images" OPT_UPLOAD_IGNORE_HOST = "upload_ignore_host" OPT_IGNORE_TLS = "ignore_tls_errors" URL_REDIRECT_PREFIX = "url_redirect_" class BlogImportPlugin(object): BOOL_OPTIONS = (OPT_UPLOAD_IMAGES, OPT_IGNORE_TLS) JSON_OPTIONS = () OPT_DEFAULTS = {OPT_UPLOAD_IMAGES: True, OPT_IGNORE_TLS: False} def __init__(self, host): log.info(_("plugin Blog Import initialization")) self.host = host self._u = host.plugins["UPLOAD"] self._p = host.plugins["XEP-0060"] self._m = host.plugins["XEP-0277"] self._s = self.host.plugins["TEXT_SYNTAXES"] host.plugins["IMPORT"].initialize(self, "blog") def importItem( self, client, item_import_data, session, options, return_data, service, node ): """importItem specialized for blog import @param item_import_data(dict): * mandatory keys: 'blog' (dict): microblog data of the blog post (cf. http://wiki.goffi.org/wiki/Bridge_API_-_Microblogging/en) the importer MUST NOT create node or call XEP-0277 plugin itself 'comments*' key MUST NOT be used in this microblog_data, see bellow for comments It is recommanded to use a unique id in the "id" key which is constant per blog item, so if the import fail, a new import will overwrite the failed items and avoid duplicates. 'comments' (list[list[dict]],None): Dictionaries must have the same keys as main item (i.e. 'blog' and 'comments') a list of list is used because XEP-0277 can handler several comments nodes, but in most cases, there will we only one item it the first list (something like [[{comment1_data},{comment2_data}, ...]]) blog['allow_comments'] must be True if there is any comment, and False (or not present) if comments are not allowed. If allow_comments is False and some comments are present, an exceptions.DataError will be raised * optional keys: 'url' (unicode): former url of the post (only the path, without host part) if present the association to the new path will be displayed to user, so it can make redirections if necessary @param options(dict, None): Below are the generic options, blog importer can have specific ones. All options have unicode values generic options: - OPT_HOST (unicode): original host - OPT_UPLOAD_IMAGES (bool): upload images to XMPP server if True see OPT_UPLOAD_IGNORE_HOST. Default: True - OPT_UPLOAD_IGNORE_HOST (unicode): don't upload images from this host - OPT_IGNORE_TLS (bool): ignore TLS error for image upload. Default: False @param return_data(dict): will contain link between former posts and new items """ mb_data = item_import_data["blog"] try: item_id = mb_data["id"] except KeyError: item_id = mb_data["id"] = str(shortuuid.uuid()) try: # we keep the link between old url and new blog item # so the user can redirect its former blog urls old_uri = item_import_data["url"] except KeyError: pass else: new_uri = return_data[URL_REDIRECT_PREFIX + old_uri] = self._p.getNodeURI( service if service is not None else client.jid.userhostJID(), node or self._m.namespace, item_id, ) log.info("url link from {old} to {new}".format(old=old_uri, new=new_uri)) return mb_data @defer.inlineCallbacks def importSubItems(self, client, item_import_data, mb_data, session, options): # comments data if len(item_import_data["comments"]) != 1: raise NotImplementedError("can't manage multiple comment links") allow_comments = C.bool(mb_data.get("allow_comments", C.BOOL_FALSE)) if allow_comments: comments_service = yield self._m.getCommentsService(client) comments_node = self._m.getCommentsNode(mb_data["id"]) mb_data["comments_service"] = comments_service.full() mb_data["comments_node"] = comments_node recurse_kwargs = { "items_import_data": item_import_data["comments"][0], "service": comments_service, "node": comments_node, } defer.returnValue(recurse_kwargs) else: if item_import_data["comments"][0]: raise exceptions.DataError( "allow_comments set to False, but comments are there" ) defer.returnValue(None) def publishItem(self, client, mb_data, service, node, session): log.debug( "uploading item [{id}]: {title}".format( id=mb_data["id"], title=mb_data.get("title", "") ) ) return self._m.send(client, mb_data, service, node) @defer.inlineCallbacks def itemFilters(self, client, mb_data, session, options): """Apply filters according to options modify mb_data in place @param posts_data(list[dict]): data as returned by importer callback @param options(dict): dict as given in [blogImport] """ # FIXME: blog filters don't work on text content # TODO: text => XHTML conversion should handler links with <a/> # filters can then be used by converting text to XHTML if not options: return # we want only XHTML content for prefix in ( "content", ): # a tuple is use, if title need to be added in the future try: rich = mb_data["{}_rich".format(prefix)] except KeyError: pass else: if "{}_xhtml".format(prefix) in mb_data: raise exceptions.DataError( "importer gave {prefix}_rich and {prefix}_xhtml at the same time, this is not allowed".format( prefix=prefix ) ) # we convert rich syntax to XHTML here, so we can handle filters easily converted = yield self._s.convert( rich, self._s.getCurrentSyntax(client.profile), safe=False ) mb_data["{}_xhtml".format(prefix)] = converted del mb_data["{}_rich".format(prefix)] try: mb_data["txt"] except KeyError: pass else: if "{}_xhtml".format(prefix) in mb_data: log.warning( "{prefix}_text will be replaced by converted {prefix}_xhtml, so filters can be handled".format( prefix=prefix ) ) del mb_data["{}_text".format(prefix)] else: log.warning( "importer gave a text {prefix}, blog filters don't work on text {prefix}".format( prefix=prefix ) ) return # at this point, we have only XHTML version of content try: top_elt = xml_tools.ElementParser()( mb_data["content_xhtml"], namespace=C.NS_XHTML ) except domish.ParserError: # we clean the xml and try again our luck cleaned = yield self._s.cleanXHTML(mb_data["content_xhtml"]) top_elt = xml_tools.ElementParser()(cleaned, namespace=C.NS_XHTML) opt_host = options.get(OPT_HOST) if opt_host: # we normalise the domain parsed_host = urllib.parse.urlsplit(opt_host) opt_host = urllib.parse.urlunsplit( ( parsed_host.scheme or "http", parsed_host.netloc or parsed_host.path, "", "", "", ) ) tmp_dir = tempfile.mkdtemp() try: # TODO: would be nice to also update the hyperlinks to these images, e.g. when you have <a href="{url}"><img src="{url}"></a> for img_elt in xml_tools.findAll(top_elt, names=["img"]): yield self.imgFilters(client, img_elt, options, opt_host, tmp_dir) finally: os.rmdir(tmp_dir) # XXX: tmp_dir should be empty, or something went wrong # we now replace the content with filtered one mb_data["content_xhtml"] = top_elt.toXml() @defer.inlineCallbacks def imgFilters(self, client, img_elt, options, opt_host, tmp_dir): """Filters handling images url without host are fixed (if possible) according to options, images are uploaded to XMPP server @param img_elt(domish.Element): <img/> element to handle @param options(dict): filters options @param opt_host(unicode): normalised host given in options @param tmp_dir(str): path to temp directory """ try: url = img_elt["src"] if url[0] == "/": if not opt_host: log.warning( "host was not specified, we can't deal with src without host ({url}) and have to ignore the following <img/>:\n{xml}".format( url=url, xml=img_elt.toXml() ) ) return else: url = urllib.parse.urljoin(opt_host, url) filename = url.rsplit("/", 1)[-1].strip() if not filename: raise KeyError except (KeyError, IndexError): log.warning("ignoring invalid img element: {}".format(img_elt.toXml())) return # we change the url for the normalized one img_elt["src"] = url if options.get(OPT_UPLOAD_IMAGES, False): # upload is requested try: ignore_host = options[OPT_UPLOAD_IGNORE_HOST] except KeyError: pass else: # host is the ignored one, we skip parsed_url = urllib.parse.urlsplit(url) if ignore_host in parsed_url.hostname: log.info( "Don't upload image at {url} because of {opt} option".format( url=url, opt=OPT_UPLOAD_IGNORE_HOST ) ) return # we download images and re-upload them via XMPP tmp_file = os.path.join(tmp_dir, filename).encode("utf-8") upload_options = {"ignore_tls_errors": options.get(OPT_IGNORE_TLS, False)} try: yield web_client.downloadPage(url.encode("utf-8"), tmp_file) filename = filename.replace( "%", "_" ) # FIXME: tmp workaround for a bug in prosody http upload __, download_d = yield self._u.upload( client, tmp_file, filename, extra=upload_options ) download_url = yield download_d except Exception as e: log.warning( "can't download image at {url}: {reason}".format(url=url, reason=e) ) else: img_elt["src"] = download_url try: os.unlink(tmp_file) except OSError: pass