# -*- coding: utf-8 -*- # Copyright 2015-2019 Mike Fährmann # # This program is free software; you can redistribute it and/or modify # it under the terms of the GNU General Public License version 2 as # published by the Free Software Foundation. import sys import time import logging from . import extractor, downloader, postprocessor from . import config, text, util, output, exception from .extractor.message import Message class Job(): """Base class for Job-types""" ulog = None def __init__(self, extr, parent=None): if isinstance(extr, str): extr = extractor.find(extr) if not extr: raise exception.NoExtractorError() self.extractor = extr extr.log.extractor = extr extr.log.job = self extr.log.debug("Using %s for '%s'", extr.__class__.__name__, extr.url) self.pred_url = self._prepare_predicates("image", True) self.pred_queue = self._prepare_predicates("chapter", False) if parent and parent.extractor.config( "category-transfer", parent.extractor.categorytransfer): self.extractor.category = parent.extractor.category self.extractor.subcategory = parent.extractor.subcategory # user-supplied metadata self.userkwds = self.extractor.config("keywords") def run(self): """Execute or run the job""" try: log = self.extractor.log for msg in self.extractor: self.dispatch(msg) except exception.StopExtraction as exc: if exc.message: log.error("%s", exc.message) return exc.code except exception.GalleryDLException as exc: log.error("%s: %s", exc.__class__.__name__, exc) return exc.code except OSError as exc: log.error("Unable to download data: %s: %s", exc.__class__.__name__, exc) log.debug("", exc_info=True) return 128 except Exception as exc: log.error(("An unexpected error occurred: %s - %s. " "Please run gallery-dl again with the --verbose flag, " "copy its output and report this issue on " "https://github.com/mikf/gallery-dl/issues ."), exc.__class__.__name__, exc) log.debug("", exc_info=True) return 1 else: return 0 finally: self.handle_finalize() def dispatch(self, msg): """Call the appropriate message handler""" if msg[0] == Message.Url: _, url, kwds = msg if self.pred_url(url, kwds): self.update_kwdict(kwds) self.handle_url(url, kwds) elif msg[0] == Message.Directory: self.update_kwdict(msg[1]) self.handle_directory(msg[1]) elif msg[0] == Message.Queue: _, url, kwds = msg if self.pred_queue(url, kwds): self.handle_queue(url, kwds) elif msg[0] == Message.Urllist: _, urls, kwds = msg if self.pred_url(urls[0], kwds): self.update_kwdict(kwds) self.handle_urllist(urls, kwds) elif msg[0] == Message.Version: if msg[1] != 1: raise "unsupported message-version ({}, {})".format( self.extractor.category, msg[1] ) # TODO: support for multiple message versions def handle_url(self, url, keywords): """Handle Message.Url""" def handle_urllist(self, urls, keywords): """Handle Message.Urllist""" self.handle_url(urls[0], keywords) def handle_directory(self, keywords): """Handle Message.Directory""" def handle_queue(self, url, keywords): """Handle Message.Queue""" def handle_finalize(self): """Handle job finalization""" def update_kwdict(self, kwdict): """Update 'kwdict' with additional metadata""" kwdict["category"] = self.extractor.category kwdict["subcategory"] = self.extractor.subcategory if self.userkwds: kwdict.update(self.userkwds) def _prepare_predicates(self, target, skip=True): predicates = [] if self.extractor.config(target + "-unique"): predicates.append(util.UniquePredicate()) pfilter = self.extractor.config(target + "-filter") if pfilter: try: pred = util.FilterPredicate(pfilter, target) except (SyntaxError, ValueError, TypeError) as exc: self.extractor.log.warning(exc) else: predicates.append(pred) prange = self.extractor.config(target + "-range") if prange: try: pred = util.RangePredicate(prange) except ValueError as exc: self.extractor.log.warning( "invalid %s range: %s", target, exc) else: if skip and pred.lower > 1 and not pfilter: pred.index += self.extractor.skip(pred.lower - 1) predicates.append(pred) return util.build_predicate(predicates) def _write_unsupported(self, url): if self.ulog: self.ulog.info(url) @staticmethod def _filter(kwdict): """Return a copy of 'kwdict' without "private" entries""" return {k: v for k, v in kwdict.items() if k[0] != "_"} class DownloadJob(Job): """Download images into appropriate directory/filename locations""" def __init__(self, url, parent=None): Job.__init__(self, url, parent) self.log = logging.getLogger("download") self.pathfmt = None self.archive = None self.sleep = None self.downloaders = {} self.postprocessors = None self.out = output.select() def handle_url(self, url, keywords, fallback=None): """Download the resource specified in 'url'""" postprocessors = self.postprocessors pathfmt = self.pathfmt archive = self.archive # prepare download pathfmt.set_filename(keywords) if postprocessors: for pp in postprocessors: pp.prepare(pathfmt) if pathfmt.exists(archive): self.handle_skip() return if self.sleep: time.sleep(self.sleep) # download from URL if not self.download(url): # use fallback URLs if available for num, url in enumerate(fallback or (), 1): self.log.info("Trying fallback URL #%d", num) if self.download(url): break else: # download failed self.log.error("Failed to download %s", pathfmt.filename or url) return if not pathfmt.temppath: self.handle_skip() return # run post processors if postprocessors: for pp in postprocessors: pp.run(pathfmt) # download succeeded pathfmt.finalize() self.out.success(pathfmt.path, 0) if archive: archive.add(keywords) if postprocessors: for pp in postprocessors: pp.run_after(pathfmt) self._skipcnt = 0 def handle_urllist(self, urls, keywords): """Download the resource specified in 'url'""" fallback = iter(urls) url = next(fallback) self.handle_url(url, keywords, fallback) def handle_directory(self, keywords): """Set and create the target directory for downloads""" if not self.pathfmt: self.initialize(keywords) else: self.pathfmt.set_directory(keywords) def handle_queue(self, url, keywords): if "_extractor" in keywords: extr = keywords["_extractor"].from_url(url) else: extr = extractor.find(url) if extr: self.__class__(extr, self).run() else: self._write_unsupported(url) def handle_finalize(self): if self.postprocessors: for pp in self.postprocessors: pp.finalize() if self.archive: self.archive.close() if self.pathfmt: self.extractor._store_cookies() def handle_skip(self): self.out.skip(self.pathfmt.path) if self._skipexc: self._skipcnt += 1 if self._skipcnt >= self._skipmax: raise self._skipexc() def download(self, url): """Download 'url'""" scheme = url.partition(":")[0] downloader = self.get_downloader(scheme) if downloader: return downloader.download(url, self.pathfmt) self._write_unsupported(url) return False def get_downloader(self, scheme): """Return a downloader suitable for 'scheme'""" try: return self.downloaders[scheme] except KeyError: pass klass = downloader.find(scheme) if klass and config.get(("downloader", klass.scheme, "enabled"), True): instance = klass(self.extractor, self.out) else: instance = None self.log.error("'%s:' URLs are not supported/enabled", scheme) if klass and klass.scheme == "http": self.downloaders["http"] = self.downloaders["https"] = instance else: self.downloaders[scheme] = instance return instance def initialize(self, keywords=None): """Delayed initialization of PathFormat, etc.""" self.pathfmt = util.PathFormat(self.extractor) if keywords: self.pathfmt.set_directory(keywords) self.sleep = self.extractor.config("sleep") if not self.extractor.config("download", True): self.download = self.pathfmt.fix_extension skip = self.extractor.config("skip", True) if skip: self._skipexc = None if skip == "enumerate": self.pathfmt.check_file = self.pathfmt._enum_file elif isinstance(skip, str): skip, _, smax = skip.partition(":") if skip == "abort": self._skipexc = exception.StopExtraction elif skip == "exit": self._skipexc = sys.exit self._skipcnt = 0 self._skipmax = text.parse_int(smax) else: self.pathfmt.exists = lambda x=None: False archive = self.extractor.config("archive") if archive: path = util.expand_path(archive) try: self.archive = util.DownloadArchive(path, self.extractor) except Exception as exc: self.extractor.log.warning( "Failed to open download archive at '%s' ('%s: %s')", path, exc.__class__.__name__, exc) else: self.extractor.log.debug("Using download archive '%s'", path) postprocessors = self.extractor.config("postprocessors") if postprocessors: pp_list = [] for pp_dict in postprocessors: whitelist = pp_dict.get("whitelist") blacklist = pp_dict.get("blacklist") if (whitelist and self.extractor.category not in whitelist or blacklist and self.extractor.category in blacklist): continue name = pp_dict.get("name") pp_cls = postprocessor.find(name) if not pp_cls: postprocessor.log.warning("module '%s' not found", name) continue try: pp_obj = pp_cls(self.pathfmt, pp_dict) except Exception as exc: postprocessor.log.error( "'%s' initialization failed: %s: %s", name, exc.__class__.__name__, exc) else: pp_list.append(pp_obj) if pp_list: self.postprocessors = pp_list self.extractor.log.debug( "Active postprocessor modules: %s", pp_list) class SimulationJob(DownloadJob): """Simulate the extraction process without downloading anything""" def handle_url(self, url, keywords, fallback=None): self.pathfmt.set_filename(keywords) self.out.skip(self.pathfmt.path) if self.sleep: time.sleep(self.sleep) if self.archive: self.archive.add(keywords) def handle_directory(self, keywords): if not self.pathfmt: self.initialize() class KeywordJob(Job): """Print available keywords""" def handle_url(self, url, keywords): print("\nKeywords for filenames and --filter:") print("------------------------------------") self.print_keywords(keywords) raise exception.StopExtraction() def handle_directory(self, keywords): print("Keywords for directory names:") print("-----------------------------") self.print_keywords(keywords) def handle_queue(self, url, keywords): if not keywords: self.extractor.log.info( "This extractor delegates work to other extractors " "and does not provide any keywords on its own. Try " "'gallery-dl -K \"%s\"' instead.", url) else: print("Keywords for --chapter-filter:") print("------------------------------") self.print_keywords(keywords) if self.extractor.categorytransfer: print() KeywordJob(url, self).run() raise exception.StopExtraction() @staticmethod def print_keywords(keywords, prefix=""): """Print key-value pairs with formatting""" suffix = "]" if prefix else "" for key, value in sorted(keywords.items()): if key[0] == "_": continue key = prefix + key + suffix if isinstance(value, dict): KeywordJob.print_keywords(value, key + "[") elif isinstance(value, list): if value and isinstance(value[0], dict): KeywordJob.print_keywords(value[0], key + "[][") else: print(key, "[]", sep="") for val in value: print(" -", val) else: # string or number print(key, "\n ", value, sep="") class UrlJob(Job): """Print download urls""" maxdepth = 1 def __init__(self, url, parent=None, depth=1): Job.__init__(self, url, parent) self.depth = depth if depth >= self.maxdepth: self.handle_queue = self.handle_url @staticmethod def handle_url(url, _): print(url) @staticmethod def handle_urllist(urls, _): prefix = "" for url in urls: print(prefix, url, sep="") prefix = "| " def handle_queue(self, url, _): try: UrlJob(url, self, self.depth + 1).run() except exception.NoExtractorError: self._write_unsupported(url) class DataJob(Job): """Collect extractor results and dump them""" def __init__(self, url, parent=None, file=sys.stdout, ensure_ascii=True): Job.__init__(self, url, parent) self.file = file self.data = [] self.ascii = config.get(("output", "ascii"), ensure_ascii) def run(self): # collect data try: for msg in self.extractor: self.dispatch(msg) except exception.StopExtraction: pass except Exception as exc: self.data.append((exc.__class__.__name__, str(exc))) except BaseException: pass # convert numbers to string if config.get(("output", "num-to-str"), False): for msg in self.data: util.transform_dict(msg[-1], util.number_to_string) # dump to 'file' util.dump_json(self.data, self.file, self.ascii, 2) return 0 def handle_url(self, url, kwdict): self.data.append((Message.Url, url, self._filter(kwdict))) def handle_urllist(self, urls, kwdict): self.data.append((Message.Urllist, list(urls), self._filter(kwdict))) def handle_directory(self, kwdict): self.data.append((Message.Directory, self._filter(kwdict))) def handle_queue(self, url, kwdict): self.data.append((Message.Queue, url, self._filter(kwdict))) def handle_finalize(self): self.file.close()