From 9c7d0ecfbfe16a6e6e4f78139dfc70995f889030 Mon Sep 17 00:00:00 2001 From: Tonda39 <84375208+Tonda39@users.noreply.github.com> Date: Sat, 2 Nov 2024 16:12:58 +0100 Subject: [PATCH] bdfr: add page (#14535) Co-authored-by: Wiktor Perskawiec Co-authored-by: Sebastiaan Speck <12570668+sebastiaanspeck@users.noreply.github.com> --- pages/common/bdfr.md | 28 ++++++++++++++++++++++++++++ 1 file changed, 28 insertions(+) create mode 100644 pages/common/bdfr.md diff --git a/pages/common/bdfr.md b/pages/common/bdfr.md new file mode 100644 index 000000000..abe24d002 --- /dev/null +++ b/pages/common/bdfr.md @@ -0,0 +1,28 @@ +# bdfr + +> Bulk downloader for Reddit. +> More information: . + +- Download videos/images from the specified [l]inks to URL or ID's of posts: + +`bdfr download {{path/to/output_directory}} -l {{post_url}}` + +- Download the maximum possible number (roughly 1000) of videos/images from a specified [u]ser: + +`bdfr download {{path/to/output_directory}} -u {{reddit_user}} --submitted` + +- Download submission data (text, upvotes, comments, etc.) [L]imited to 10 submissions for each [s]ubreddit (30 total): + +`bdfr archive {{path/to/output_directory}} -s '{{Python, all, mindustry}}' -L 10` + +- Download videos/images from the [s]ubreddit r/Python [S]orted by top (default is hot) using [t]ime filter all, [L]imited to 10 submissions: + +`bdfr download {{path/to/output_directory}} -s Python -S top -t all -L 10` + +- Download the maximum possible number of both submission data and videos/images from [s]ubreddit r/Python skipping over submissions with mp4 or gif file extensions and creating hard links for duplicate files: + +`bdfr clone {{path/to/output_directory}} -s Python --skip mp4 --skip gif --make-hard-links` + +- Download saved posts of the authenticated user, naming each file according to a specified format. Avoid downloading duplicates and posts already present in the output directory: + +`bdfr download {{path/to/output_directory}} --user me --saved --authenticate --file-scheme '{{ {POSTID}_{TITLE}_{UPVOTES} }}' --no-dupes --search-existing`