[adultempire] add gallery extractor (closes #340)
parent
b3851e01d9
commit
3fe5ccdfa6
@ -0,0 +1,58 @@
|
||||
# -*- coding: utf-8 -*-
|
||||
|
||||
# Copyright 2019 Mike Fährmann
|
||||
#
|
||||
# This program is free software; you can redistribute it and/or modify
|
||||
# it under the terms of the GNU General Public License version 2 as
|
||||
# published by the Free Software Foundation.
|
||||
|
||||
"""Extractors for https://www.adultempire.com/"""
|
||||
|
||||
from .common import GalleryExtractor
|
||||
from .. import text
|
||||
|
||||
|
||||
class AdultempireGalleryExtractor(GalleryExtractor):
|
||||
"""Extractor for image galleries from www.adultempire.com"""
|
||||
category = "adultempire"
|
||||
root = "https://www.adultempire.com"
|
||||
pattern = (r"(?:https?://)?(?:www\.)?adult(?:dvd)?empire\.com"
|
||||
r"(/(\d+)/gallery\.html)")
|
||||
test = (
|
||||
("https://www.adultempire.com/5998/gallery.html", {
|
||||
"range": "1",
|
||||
"keyword": "9057a960a7224873431612e4c95b88b2e8b5c97b",
|
||||
"content": "5c6beb31e5e3cdc90ee5910d5c30f9aaec977b9e",
|
||||
}),
|
||||
("https://www.adultdvdempire.com/5683/gallery.html", {
|
||||
"url": "b12cd1a65cae8019d837505adb4d6a2c1ed4d70d",
|
||||
"keyword": "59fe5d95929efc5040a819a5f77aba7a022bb85a",
|
||||
}),
|
||||
)
|
||||
|
||||
def __init__(self, match):
|
||||
GalleryExtractor.__init__(self, match)
|
||||
self.gallery_id = match.group(2)
|
||||
|
||||
def metadata(self, page):
|
||||
extr = text.extract_from(page, page.index('<div id="content">'))
|
||||
return {
|
||||
"gallery_id": text.parse_int(self.gallery_id),
|
||||
"title" : text.unescape(extr('title="', '"')),
|
||||
"studio" : extr(">studio</small>", "<").strip(),
|
||||
"date" : text.parse_datetime(extr(
|
||||
">released</small>", "<").strip(), "%m/%d/%Y"),
|
||||
"actors" : text.split_html(extr(
|
||||
'<ul class="item-details item-cast-list ', '</ul>'))[1:],
|
||||
}
|
||||
|
||||
def images(self, page):
|
||||
params = {"page": 1}
|
||||
while True:
|
||||
urls = list(text.extract_iter(page, 'rel="L"><img src="', '"'))
|
||||
for url in urls:
|
||||
yield url.replace("_200.", "_9600."), None
|
||||
if len(urls) < 24:
|
||||
return
|
||||
params["page"] += 1
|
||||
page = self.request(self.chapter_url, params=params).text
|
Loading…
Reference in new issue