Question about --log, --no-dupes and --skip #585
-
First of all, amazing tool, after you told me about ffmpeg missing, It worked like a charm! Kudos and thanks to you! I am using : python -m bdfr download ./lolo --subreddit "MemeVideos" --no-dupes --time "day" --skip "jpeg, gif, png" -L 50 --file-scheme "{POSTID}" What I am doing though is I manually delete some videos but I need a way to prevent them from re-downloading when deleted by me. Is there a way to use --log to prevent re-downloading items which I downloaded at some point? because if I have to keep files in order to prevent them from re-downloading my HDD would be full quite quickly :(
Appreciate you, thanks! |
Beta Was this translation helpful? Give feedback.
Replies: 1 comment 9 replies
-
So The The What you do want is For just such an occasion, we have a folder of scripts that will read the logfile generated by each run and extract all IDs from that run, either all successfully downloaded IDs (so they aren't retried) or the failed ones (so you don't waste time reattempting those, if that's what you want). This is the best way to prevent duplicates. The scripts are available here and come in bash and Powershell so they'll work for whatever system you have. Use them with the |
Beta Was this translation helpful? Give feedback.
So
--log
only specifies where the logfile should be saved. This is only really useful if you're running multiple instances of the BDFR at the same time, since multiple processes accessing the same file will crash on Windows, and write gibberish on Linux and Mac.The
--no-dupes
option is only there to prevent saving the same image on the same run. What happens is that we calculate the hash of the image once it's been downloaded, and check to see if the hash exists in memory. This only gets hashes from the current run, so it's of limited use by itself. It was originally paired with the--search-existing
option, which will scan the destination and compute all the hashes for every file, but t…