You can not select more than 25 topics Topics must start with a letter or number, can include dashes ('-') and can be up to 35 characters long.
gallery-dl/gallery_dl/extractor/mangasee.py

118 lines
4.1 KiB

# -*- coding: utf-8 -*-
# Copyright 2021-2023 Mike Fährmann
#
# This program is free software; you can redistribute it and/or modify
# it under the terms of the GNU General Public License version 2 as
# published by the Free Software Foundation.
"""Extractors for https://mangasee123.com/"""
from .common import ChapterExtractor, MangaExtractor
from .. import text, util
class MangaseeBase():
category = "mangasee"
browser = "firefox"
root = "https://mangasee123.com"
@staticmethod
def _transform_chapter(data):
chapter = data["Chapter"]
return {
"title" : data["ChapterName"] or "",
"index" : chapter[0],
"chapter" : int(chapter[1:-1]),
"chapter_minor": "" if chapter[-1] == "0" else "." + chapter[-1],
"chapter_string": chapter,
"lang" : "en",
"language": "English",
"date" : text.parse_datetime(
data["Date"], "%Y-%m-%d %H:%M:%S"),
}
class MangaseeChapterExtractor(MangaseeBase, ChapterExtractor):
pattern = (r"(?:https?://)?(mangasee123|manga4life)\.com"
r"(/read-online/[^/?#]+\.html)")
example = "https://mangasee123.com/read-online/MANGA-chapter-1-page-1.html"
def __init__(self, match):
if match.group(1) == "manga4life":
self.category = "mangalife"
self.root = "https://manga4life.com"
ChapterExtractor.__init__(self, match, self.root + match.group(2))
def _init(self):
self.session.headers["Referer"] = self.gallery_url
domain = self.root.rpartition("/")[2]
cookies = self.cookies
if not cookies.get("PHPSESSID", domain=domain):
cookies.set("PHPSESSID", util.generate_token(13), domain=domain)
def metadata(self, page):
extr = text.extract_from(page)
author = util.json_loads(extr('"author":', '],') + "]")
genre = util.json_loads(extr('"genre":', '],') + "]")
self.chapter = data = util.json_loads(extr("vm.CurChapter =", ";\r\n"))
self.domain = extr('vm.CurPathName = "', '"')
self.slug = extr('vm.IndexName = "', '"')
data = self._transform_chapter(data)
data["manga"] = text.unescape(extr('vm.SeriesName = "', '"'))
data["author"] = author
data["genre"] = genre
return data
def images(self, page):
chapter = self.chapter["Chapter"][1:]
if chapter[-1] == "0":
chapter = chapter[:-1]
else:
chapter = chapter[:-1] + "." + chapter[-1]
base = "https://{}/manga/{}/".format(self.domain, self.slug)
if self.chapter["Directory"]:
base += self.chapter["Directory"] + "/"
base += chapter + "-"
return [
("{}{:>03}.png".format(base, i), None)
for i in range(1, int(self.chapter["Page"]) + 1)
]
class MangaseeMangaExtractor(MangaseeBase, MangaExtractor):
chapterclass = MangaseeChapterExtractor
pattern = r"(?:https?://)?(mangasee123|manga4life)\.com(/manga/[^/?#]+)"
example = "https://mangasee123.com/manga/MANGA"
def __init__(self, match):
if match.group(1) == "manga4life":
self.category = "mangalife"
self.root = "https://manga4life.com"
MangaExtractor.__init__(self, match, self.root + match.group(2))
def chapters(self, page):
extr = text.extract_from(page)
author = util.json_loads(extr('"author":', '],') + "]")
genre = util.json_loads(extr('"genre":', '],') + "]")
slug = extr('vm.IndexName = "', '"')
chapters = util.json_loads(extr("vm.Chapters = ", ";\r\n"))
result = []
for data in map(self._transform_chapter, chapters):
url = "{}/read-online/{}-chapter-{}{}".format(
self.root, slug, data["chapter"], data["chapter_minor"])
if data["index"] != "1":
url += "-index-" + data["index"]
url += "-page-1.html"
data["manga"] = slug
data["author"] = author
data["genre"] = genre
result.append((url, data))
return result