# -*- coding: utf-8 -*- # Copyright 2015-2017 Mike Fährmann # # This program is free software; you can redistribute it and/or modify # it under the terms of the GNU General Public License version 2 as # published by the Free Software Foundation. """Extract manga-chapters and entire manga from http://www.mangareader.net/""" from .common import AsynchronousExtractor, MangaExtractor, Message from .. import text class MangareaderBase(): """Base class for mangareader extractors""" category = "mangareader" directory_fmt = ["{category}", "{manga}", "c{chapter:>03} - {title}"] filename_fmt = "{manga}_c{chapter:>03}_{page:>03}.{extension}" root = "http://www.mangareader.net" class MangareaderMangaExtractor(MangareaderBase, MangaExtractor): """Extractor for manga from mangareader.net""" pattern = [r"(?:https?://)?((?:www\.)?mangareader\.net/[^/]+)$"] reverse = False test = [("http://www.mangareader.net/mushishi", { "url": "249042420b67a07b32e7f6be4c7410b6d810b808", })] def chapter_paths(self, page): needle = '
\n') return text.extract_iter(page, needle, '"', pos) class MangareaderChapterExtractor(MangareaderBase, AsynchronousExtractor): """Extractor for manga-chapters from mangareader.net""" subcategory = "chapter" pattern = [ (r"(?:https?://)?(?:www\.)?mangareader\.net((/[^/]+)/(\d+))"), (r"(?:https?://)?(?:www\.)?mangareader\.net(/\d+-\d+-\d+(/[^/]+)/" r"chapter-(\d+).html)"), ] test = [(("http://www.mangareader.net/" "karate-shoukoushi-kohinata-minoru/11"), { "url": "84ffaab4c027ef9022695c53163c3aeabd07ca58", "keyword": "09b4ad57a082eb371dec027ccfc8ed1157c6eac6", })] def __init__(self, match): AsynchronousExtractor.__init__(self) self.part, self.url_title, self.chapter = match.groups() def items(self): page = self.request(self.root + self.part).text data = self.get_job_metadata(page) yield Message.Version, 1 yield Message.Directory, data for i in range(1, int(data["count"])+1): next_url, image_url, image_data = self.get_page_metadata(page) image_data.update(data) image_data["page"] = i yield Message.Url, image_url, image_data if next_url: page = self.request(next_url).text def get_job_metadata(self, chapter_page): """Collect metadata for extractor-job""" page = self.request(self.root + self.url_title).text data = { "chapter": self.chapter, "lang": "en", "language": "English", } data, _ = text.extract_all(page, ( (None, 'Name:', ''), ("manga", '

', '

'), (None, 'Year of Release:', ''), ('manga-release', '', ''), (None, 'Author:', ''), ('author', '', ''), (None, 'Artist:', ''), ('artist', '', ''), (None, '
', ''), ('title', ' ' + self.chapter + ' : ', ''), ('chapter-date', '', ''), ), values=data) data, _ = text.extract_all(chapter_page, ( (None, '