mirror of
https://github.com/Radiquum/furaffinity-dl.git
synced 2025-04-06 00:04:38 +00:00
179 lines
5.8 KiB
Bash
Executable file
179 lines
5.8 KiB
Bash
Executable file
#!/bin/bash
|
|
set -e
|
|
|
|
# Detect installed metadata injectors
|
|
eyed3=true
|
|
if [ -z $(command -v eyeD3) ]; then
|
|
eyed3=false
|
|
echo "INFO: eyed3 is not installed, no metadata will be injected into music files."
|
|
fi
|
|
|
|
exiftool=true
|
|
if [ -z $(command -v exiftool) ]; then
|
|
exiftool=false
|
|
echo "INFO: exiftool is not installed, no metadata will be injected into pictures."
|
|
fi
|
|
|
|
# Helper functions
|
|
help() {
|
|
echo "Usage: $0 [ARGUMENTS] SECTION/USER
|
|
Downloads the entire gallery/scraps/favorites of any furaffinity user.
|
|
|
|
Arguments:
|
|
-h (H)elp screen
|
|
-i Use an (I)nsecure connection when downloading
|
|
-o The (O)utput directory to put files in
|
|
-c If you need to download restricted content
|
|
you can provide a path to a (C)ookie file
|
|
-p (P)lain file without any additional metadata
|
|
-r Don't (R)ename files, just give them the same
|
|
filename as on facdn
|
|
-n (N)unmber of images to download, starting from
|
|
the most recent submission
|
|
|
|
Examples:
|
|
$0 gallery/mylafox
|
|
$0 -o mylasArt gallery/mylafox
|
|
$0 -o koulsFavs favorites/koul
|
|
|
|
You can also log in to FurAffinity to download restricted content, like this:
|
|
$0 -c /path/to/your/cookies.txt gallery/gonnaneedabiggerboat
|
|
|
|
DISCLAIMER: It is your own responsibility to check whether batch downloading is allowed by FurAffinity terms of service and to abide by them."
|
|
exit 1
|
|
}
|
|
cleanup() {
|
|
rm -r "$tempfile"
|
|
}
|
|
|
|
# Arguments
|
|
[[ $# -eq 0 ]] && help
|
|
|
|
outdir="."
|
|
prefix="https:"
|
|
metadata=true
|
|
rename=true
|
|
maxsavefiles="0"
|
|
while getopts 'o:c:iphrn:' flag; do
|
|
case "${flag}" in
|
|
o) outdir=${OPTARG};;
|
|
c) cookiefile=${OPTARG};;
|
|
i) prefix="http:";;
|
|
p) metadata=false;;
|
|
r) rename=false;;
|
|
n) maxsavefiles=${OPTARG};;
|
|
h) help;;
|
|
*) help;;
|
|
esac
|
|
done
|
|
|
|
# Attempt to create the output directory
|
|
mkdir -p "$outdir"
|
|
|
|
# Setup runtime directory
|
|
runtime_dir="$HOME"'/.cache/furaffinity-dl'
|
|
mkdir -p "$runtime_dir"
|
|
tempfile="$(umask u=rwx,g=,o= && mktemp $runtime_dir/fa-dl.XXXXXXXXXX)"
|
|
|
|
# Call cleanup function on exit
|
|
trap cleanup EXIT
|
|
|
|
if [ -z "$cookiefile" ]; then
|
|
# Set wget with a custom user agent
|
|
fwget() {
|
|
wget -nv --user-agent="Mozilla/5.0 furaffinity-dl (https://github.com/Xerbo/furaffinity-dl)" $*
|
|
}
|
|
else
|
|
# Set wget with a custom user agent and cookies
|
|
fwget() {
|
|
wget -nv --user-agent="Mozilla/5.0 furaffinity-dl (https://github.com/Xerbo/furaffinity-dl)" --load-cookies "$cookiefile" $*
|
|
}
|
|
fi
|
|
|
|
url="https://www.furaffinity.net/${@: -1}"
|
|
download_count="0"
|
|
|
|
# Iterate over the gallery pages with thumbnails and links to artwork view pages
|
|
while true; do
|
|
fwget "$url" -O "$tempfile"
|
|
if [ -n "$cookiefile" ] && grep -q 'furaffinity.net/login/' "$tempfile"; then
|
|
echo "ERROR: You have provided a cookies file, but it does not contain valid cookies.
|
|
|
|
If this file used to work, this means that the cookies have expired;
|
|
you will have to log in to FurAffinity from your web browser and export the cookies again.
|
|
|
|
If this is the first time you're trying to use cookies, make sure you have exported them
|
|
in Netscape format (this is normally done through \"cookie export\" browser extensions)
|
|
and supplied the correct path to the cookies.txt file to this script.
|
|
|
|
If that doesn't resolve the issue, please report the problem at
|
|
https://github.com/Xerbo/furaffinity-dl/issues" >&2
|
|
exit 1
|
|
fi
|
|
|
|
# Get URL for next page out of "Next" button. Required for favorites, pages of which are not numbered
|
|
next_page_url="$(grep '<a class="button-link right" href="' "$tempfile" | grep '">Next ❯❯</a>' | cut -d '"' -f 4 | sort -u)"
|
|
|
|
# Extract links to pages with individual artworks and iterate over them
|
|
artwork_pages=$(grep '<a href="/view/' "$tempfile" | grep -E --only-matching '/view/[[:digit:]]+/' | uniq)
|
|
for page in $artwork_pages; do
|
|
# Download the submission page
|
|
fwget -O "$tempfile" 'https://www.furaffinity.net'"$page"
|
|
|
|
if grep -q "System Message" "$tempfile"; then
|
|
echo "WARNING: $page seems to be inaccessible, skipping."
|
|
continue
|
|
fi
|
|
|
|
# Get the full size image URL.
|
|
# This will be a facdn.net link, we will default to HTTPS
|
|
# but this can be disabled with -i or --http for specific reasons
|
|
image_url=$prefix$(grep --only-matching --max-count=1 ' href="//d.facdn.net/art/.\+">Download' "$tempfile" | cut -d '"' -f 2)
|
|
|
|
# Get metadata
|
|
description=$(grep 'og:description" content="' "$tempfile" | cut -d '"' -f4)
|
|
title=$(grep 'og:title" content="' "$tempfile" | cut -d '"' -f4)
|
|
title="${title%" by"*}" # Remove the " by Artist" bit
|
|
title="$(echo $title | tr -d "/")"
|
|
file_type=${image_url##*.}
|
|
file="$outdir/$title.$file_type"
|
|
|
|
# Download the image
|
|
if [ $rename = true ]; then
|
|
wget "$image_url" -O "$file"
|
|
else
|
|
wget "$image_url" -P "$outdir"
|
|
fi
|
|
|
|
# Add metadata
|
|
if [ $file_type == "mp3" ] || [ $file_type == "wav" ] || [ $file_type == "wmv" ] || [ $file_type == "ogg" ] || [ $file_type == "flac" ]; then
|
|
# Use eyeD3 for injecting metadata into audio files (if it's installed)
|
|
if [ $eyed3 = true ] && [ $metadata = true ]; then
|
|
if [ -z "$description" ]; then
|
|
eyeD3 -t "$title" "$file"
|
|
else
|
|
# HACK: eyeD3 throws an error if a description containing a ":"
|
|
eyeD3 -t "$title" --add-comment "${description//:/\\:}" "$file"
|
|
fi
|
|
fi
|
|
elif [ $file_type == "jpg" ] || [ $file_type == "jpeg" ]; then
|
|
# Use exiftool for injecting metadata into pictures (if it's installed)
|
|
if [ $exiftool = true ] && [ $metadata = true ]; then
|
|
exiftool "$file" -description="$description" -title="$title" -overwrite_original
|
|
fi
|
|
fi
|
|
|
|
# If there is a file download limit then keep track of it
|
|
if [ "$maxsavefiles" -ne "0" ]; then
|
|
download_count="$((download_count + 1))"
|
|
|
|
if [ "$download_count" -ge "$maxsavefiles" ]; then
|
|
echo "Reached set file download limit."
|
|
exit 0
|
|
fi
|
|
fi
|
|
done
|
|
|
|
[ -z "$next_page_url" ] && break
|
|
url='https://www.furaffinity.net'"$next_page_url"
|
|
done
|