You can not select more than 25 topics
Topics must start with a letter or number, can include dashes ('-') and can be up to 35 characters long.
Command-Line Options
General Options
‑h |
‑‑help |
|
Print this help message and exit |
|
‑‑version |
|
Print program version and exit |
‑i |
‑‑input‑file |
FILE |
Download URLs found in FILE (- for stdin). More than one --input-file can be specified |
‑d |
‑‑destination |
PATH |
Target location for file downloads |
‑D |
‑‑directory |
PATH |
Exact location for file downloads |
‑f |
‑‑filename |
FORMAT |
Filename format string for downloaded files (/O for "original" filenames) |
|
‑‑proxy |
URL |
Use the specified proxy |
|
‑‑source‑address |
IP |
Client-side IP address to bind to |
|
‑‑user‑agent |
UA |
User-Agent request header |
|
‑‑clear‑cache |
MODULE |
Delete cached login sessions, cookies, etc. for MODULE (ALL to delete everything) |
|
‑‑cookies |
FILE |
File to load additional cookies from |
|
‑‑cookies‑from‑browser |
BROWSER |
Name of the browser to load cookies from, with optional keyring name prefixed with + , profile prefixed with : , and container prefixed with :: (none for no container) |
Output Options
‑q |
‑‑quiet |
|
Activate quiet mode |
‑v |
‑‑verbose |
|
Print various debugging information |
‑g |
‑‑get‑urls |
|
Print URLs instead of downloading |
‑G |
‑‑resolve‑urls |
|
Print URLs instead of downloading; resolve intermediary URLs |
‑j |
‑‑dump‑json |
|
Print JSON information |
‑s |
‑‑simulate |
|
Simulate data extraction; do not download anything |
‑E |
‑‑extractor‑info |
|
Print extractor defaults and settings |
‑K |
‑‑list‑keywords |
|
Print a list of available keywords and example values for the given URLs |
|
‑‑list‑modules |
|
Print a list of available extractor modules |
|
‑‑list‑extractors |
|
Print a list of extractor classes with description, (sub)category and example URL |
|
‑‑write‑log |
FILE |
Write logging output to FILE |
|
‑‑write‑unsupported |
FILE |
Write URLs, which get emitted by other extractors but cannot be handled, to FILE |
|
‑‑write‑pages |
|
Write downloaded intermediary pages to files in the current directory to debug problems |
Downloader Options
‑r |
‑‑limit‑rate |
RATE |
Maximum download rate (e.g. 500k or 2.5M) |
‑R |
‑‑retries |
N |
Maximum number of retries for failed HTTP requests or -1 for infinite retries (default: 4) |
|
‑‑http‑timeout |
SECONDS |
Timeout for HTTP connections (default: 30.0) |
|
‑‑sleep |
SECONDS |
Number of seconds to wait before each download. This can be either a constant value or a range (e.g. 2.7 or 2.0-3.5) |
|
‑‑sleep‑request |
SECONDS |
Number of seconds to wait between HTTP requests during data extraction |
|
‑‑sleep‑extractor |
SECONDS |
Number of seconds to wait before starting data extraction for an input URL |
|
‑‑filesize‑min |
SIZE |
Do not download files smaller than SIZE (e.g. 500k or 2.5M) |
|
‑‑filesize‑max |
SIZE |
Do not download files larger than SIZE (e.g. 500k or 2.5M) |
|
‑‑chunk‑size |
SIZE |
Size of in-memory data chunks (default: 32k) |
|
‑‑no‑part |
|
Do not use .part files |
|
‑‑no‑skip |
|
Do not skip downloads; overwrite existing files |
|
‑‑no‑mtime |
|
Do not set file modification times according to Last-Modified HTTP response headers |
|
‑‑no‑download |
|
Do not download any files |
|
‑‑no‑postprocessors |
|
Do not run any post processors |
|
‑‑no‑check‑certificate |
|
Disable HTTPS certificate validation |
Configuration Options
‑c |
‑‑config |
FILE |
Additional configuration files |
‑o |
‑‑option |
OPT |
Additional <key>=<value> option values |
|
‑‑ignore‑config |
|
Do not read default configuration files |
Authentication Options
‑u |
‑‑username |
USER |
Username to login with |
‑p |
‑‑password |
PASS |
Password belonging to the given username |
|
‑‑netrc |
|
Enable .netrc authentication data |
Selection Options
|
‑‑download‑archive |
FILE |
Record all downloaded or skipped files in FILE and skip downloading any file already in it |
‑A |
‑‑abort |
N |
Stop current extractor run after N consecutive file downloads were skipped |
‑T |
‑‑terminate |
N |
Stop current and parent extractor runs after N consecutive file downloads were skipped |
|
‑‑range |
RANGE |
Index range(s) specifying which files to download. These can be either a constant value, range, or slice (e.g. 5 , 8-20 , or 1:24:3 ) |
|
‑‑chapter‑range |
RANGE |
Like --range , but applies to manga chapters and other delegated URLs |
|
‑‑filter |
EXPR |
Python expression controlling which files to download. Files for which the expression evaluates to False are ignored. Available keys are the filename-specific ones listed by -K . Example: --filter "image_width >= 1000 and rating in ('s', 'q')" |
|
‑‑chapter‑filter |
EXPR |
Like --filter , but applies to manga chapters and other delegated URLs |
Post-processing Options
|
‑‑zip |
|
Store downloaded files in a ZIP archive |
|
‑‑ugoira‑conv |
|
Convert Pixiv Ugoira to WebM (requires FFmpeg) |
|
‑‑ugoira‑conv‑lossless |
|
Convert Pixiv Ugoira to WebM in VP9 lossless mode |
|
‑‑ugoira‑conv‑copy |
|
Convert Pixiv Ugoira to MKV without re-encoding any frames |
|
‑‑write‑metadata |
|
Write metadata to separate JSON files |
|
‑‑write‑info‑json |
|
Write gallery metadata to a info.json file |
|
‑‑write‑tags |
|
Write image tags to separate text files |
|
‑‑mtime‑from‑date |
|
Set file modification times according to date metadata |
|
‑‑exec |
CMD |
Execute CMD for each downloaded file. Example: --exec "convert {} {}.png && rm {}" |
|
‑‑exec‑after |
CMD |
Execute CMD after all files were downloaded successfully. Example: --exec-after "cd {} && convert * ../doc.pdf" |
‑P |
‑‑postprocessor |
NAME |
Activate the specified post processor |