# -*- coding: utf-8 -*- # Copyright 2014-2023 Mike Fährmann # # This program is free software; you can redistribute it and/or modify # it under the terms of the GNU General Public License version 2 as # published by the Free Software Foundation. """Extractors for https://e621.net/ and other e621 instances""" from .common import Message from . import danbooru from ..cache import memcache from .. import text, util class E621Extractor(danbooru.DanbooruExtractor): """Base class for e621 extractors""" basecategory = "E621" page_limit = 750 page_start = None per_page = 320 request_interval_min = 1.0 def items(self): self.session.headers["User-Agent"] = util.USERAGENT + " (by mikf)" includes = self.config("metadata") or () if includes: if isinstance(includes, str): includes = includes.split(",") elif not isinstance(includes, (list, tuple)): includes = ("notes", "pools") notes = ("notes" in includes) pools = ("pools" in includes) data = self.metadata() for post in self.posts(): file = post["file"] if not file["url"]: md5 = file["md5"] file["url"] = "https://static1.{}/data/{}/{}/{}.{}".format( self.root[8:], md5[0:2], md5[2:4], md5, file["ext"]) if notes and post.get("has_notes"): post["notes"] = self._get_notes(post["id"]) if pools and post["pools"]: post["pools"] = self._get_pools( ",".join(map(str, post["pools"]))) post["filename"] = file["md5"] post["extension"] = file["ext"] post["date"] = text.parse_datetime( post["created_at"], "%Y-%m-%dT%H:%M:%S.%f%z") post.update(data) yield Message.Directory, post yield Message.Url, file["url"], post def _get_notes(self, id): return self.request( "{}/notes.json?search[post_id]={}".format(self.root, id)).json() @memcache(keyarg=1) def _get_pools(self, ids): pools = self.request( "{}/pools.json?search[id]={}".format(self.root, ids)).json() for pool in pools: pool["name"] = pool["name"].replace("_", " ") return pools BASE_PATTERN = E621Extractor.update({ "e621": { "root": "https://e621.net", "pattern": r"e621\.net", }, "e926": { "root": "https://e926.net", "pattern": r"e926\.net", }, "e6ai": { "root": "https://e6ai.net", "pattern": r"e6ai\.net", }, }) class E621TagExtractor(E621Extractor, danbooru.DanbooruTagExtractor): """Extractor for e621 posts from tag searches""" pattern = BASE_PATTERN + r"/posts?(?:\?.*?tags=|/index/\d+/)([^&#]+)" example = "https://e621.net/posts?tags=TAG" class E621PoolExtractor(E621Extractor, danbooru.DanbooruPoolExtractor): """Extractor for e621 pools""" pattern = BASE_PATTERN + r"/pool(?:s|/show)/(\d+)" example = "https://e621.net/pools/12345" def posts(self): self.log.info("Fetching posts of pool %s", self.pool_id) id_to_post = { post["id"]: post for post in self._pagination( "/posts.json", {"tags": "pool:" + self.pool_id}) } posts = [] append = posts.append for num, pid in enumerate(self.post_ids, 1): if pid in id_to_post: post = id_to_post[pid] post["num"] = num append(post) else: self.log.warning("Post %s is unavailable", pid) return posts class E621PostExtractor(E621Extractor, danbooru.DanbooruPostExtractor): """Extractor for single e621 posts""" pattern = BASE_PATTERN + r"/post(?:s|/show)/(\d+)" example = "https://e621.net/posts/12345" def posts(self): url = "{}/posts/{}.json".format(self.root, self.post_id) return (self.request(url).json()["post"],) class E621PopularExtractor(E621Extractor, danbooru.DanbooruPopularExtractor): """Extractor for popular images from e621""" pattern = BASE_PATTERN + r"/explore/posts/popular(?:\?([^#]*))?" example = "https://e621.net/explore/posts/popular" def posts(self): return self._pagination("/popular.json", self.params) class E621FavoriteExtractor(E621Extractor): """Extractor for e621 favorites""" subcategory = "favorite" directory_fmt = ("{category}", "Favorites", "{user_id}") archive_fmt = "f_{user_id}_{id}" pattern = BASE_PATTERN + r"/favorites(?:\?([^#]*))?" example = "https://e621.net/favorites" def __init__(self, match): E621Extractor.__init__(self, match) self.query = text.parse_query(match.group(match.lastindex)) def metadata(self): return {"user_id": self.query.get("user_id", "")} def posts(self): return self._pagination("/favorites.json", self.query)