From ad12180027fb01b7031255a2f233b43b4d67020e Mon Sep 17 00:00:00 2001 From: =?UTF-8?q?Mike=20F=C3=A4hrmann?= Date: Sat, 21 Nov 2015 04:21:04 +0100 Subject: [PATCH] [redhawkscans] remove module --- gallery_dl/extractor/__init__.py | 1 - gallery_dl/extractor/redhawkscans.py | 68 ---------------------------- 2 files changed, 69 deletions(-) delete mode 100644 gallery_dl/extractor/redhawkscans.py diff --git a/gallery_dl/extractor/__init__.py b/gallery_dl/extractor/__init__.py index 066cd18a..2898ffe5 100644 --- a/gallery_dl/extractor/__init__.py +++ b/gallery_dl/extractor/__init__.py @@ -38,7 +38,6 @@ modules = [ "nhentai", "nijie", "powermanga", - "redhawkscans", "safebooru", "sankaku", "spectrumnexus", diff --git a/gallery_dl/extractor/redhawkscans.py b/gallery_dl/extractor/redhawkscans.py deleted file mode 100644 index 2a7c30b4..00000000 --- a/gallery_dl/extractor/redhawkscans.py +++ /dev/null @@ -1,68 +0,0 @@ -# -*- coding: utf-8 -*- - -# Copyright 2015 Mike Fährmann -# -# This program is free software; you can redistribute it and/or modify -# it under the terms of the GNU General Public License version 2 as -# published by the Free Software Foundation. - -"""Extract manga pages from http://manga.redhawkscans.com/""" - -from .common import Extractor, Message -from .. import text, iso639_1 -import os.path -import json -import re - -info = { - "category": "redhawkscans", - "extractor": "RedHawkScansExtractor", - "directory": ["{category}", "{manga}", "c{chapter:>03}{chapter-minor} - {title}"], - "filename": "{manga}_c{chapter:>03}{chapter-minor}_{page:>03}.{extension}", - "pattern": [ - (r"(?:https?://)?manga\.redhawkscans\.com/reader/read/" - r"(.+/([a-z]{2})/\d+/\d+)(?:/page)?"), - ], -} - -class RedHawkScansExtractor(Extractor): - - url_base = "https://manga.redhawkscans.com/reader/read/" - - def __init__(self, match): - Extractor.__init__(self) - self.category = info["category"] - self.part = match.group(1) - self.lang = match.group(2) - - def items(self): - yield Message.Version, 1 - data, pages = self.get_job_metadata() - yield Message.Directory, data - for page_index, page_data in enumerate(pages, 1): - name, ext = os.path.splitext(page_data["filename"]) - page_data.update(data) - page_data["page"] = page_index - page_data["name"] = name - page_data["extension"] = ext[1:] - yield Message.Url, "https" + page_data["url"][4:], page_data - - def get_job_metadata(self): - """Collect metadata for extractor-job""" - response = self.request(self.url_base + self.part) - response.encoding = "utf-8" - page = response.text - _ , pos = text.extract(page, '

', '') - manga , pos = text.extract(page, 'title="', '"', pos) - chapter , pos = text.extract(page, '">', '', pos) - json_data, pos = text.extract(page, 'var pages = ', ';', pos) - match = re.match(r"(\w+ (\d+)([^:+]*)(?:: (.*))?|[^:]+)", chapter) - return { - "category": self.category, - "manga": text.unescape(manga), - "chapter": match.group(2) or match.group(1), - "chapter-minor": match.group(3) or "", - "lang": self.lang, - "language": iso639_1.code_to_language(self.lang), - "title": text.unescape(match.group(4) or ""), - }, json.loads(json_data)