mirror of
https://github.com/Radiquum/furaffinity-dl.git
synced 2025-04-05 07:44:37 +00:00
changelog:
- workaround for NTFS filesystem when username ends with dot. - ReadMe changes - other minor changes
This commit is contained in:
parent
007f00b8ba
commit
d610cd350e
8 changed files with 141 additions and 162 deletions
6
.gitignore
vendored
6
.gitignore
vendored
|
@ -11,7 +11,9 @@ cookies.txt
|
||||||
# Default download folder
|
# Default download folder
|
||||||
Submissions/
|
Submissions/
|
||||||
|
|
||||||
# vscode stuff
|
#Dev stuff
|
||||||
.vscode
|
|
||||||
list.txt
|
list.txt
|
||||||
|
.vscode
|
||||||
|
.idea
|
||||||
|
venv
|
||||||
__pycache__
|
__pycache__
|
||||||
|
|
|
@ -1,4 +1,5 @@
|
||||||
import argparse
|
import argparse
|
||||||
|
import os
|
||||||
|
|
||||||
parser = argparse.ArgumentParser(
|
parser = argparse.ArgumentParser(
|
||||||
formatter_class=argparse.RawTextHelpFormatter,
|
formatter_class=argparse.RawTextHelpFormatter,
|
||||||
|
@ -28,7 +29,8 @@ parser.add_argument(
|
||||||
"username",
|
"username",
|
||||||
nargs="?",
|
nargs="?",
|
||||||
help="username of the furaffinity \
|
help="username of the furaffinity \
|
||||||
user",
|
user (if username is starting with '-' or '--' \
|
||||||
|
provide them through a file instead)",
|
||||||
)
|
)
|
||||||
parser.add_argument(
|
parser.add_argument(
|
||||||
"category",
|
"category",
|
||||||
|
@ -36,14 +38,16 @@ parser.add_argument(
|
||||||
help="the category to download, gallery/scraps/favorites \
|
help="the category to download, gallery/scraps/favorites \
|
||||||
[default: gallery]",
|
[default: gallery]",
|
||||||
default="gallery",
|
default="gallery",
|
||||||
|
type=str,
|
||||||
)
|
)
|
||||||
parser.add_argument("-c", "--cookies", help="path to a NetScape cookies file")
|
parser.add_argument("--cookies", "-c", help="path to a NetScape cookies file", type=str)
|
||||||
parser.add_argument(
|
parser.add_argument(
|
||||||
"--output",
|
"--output",
|
||||||
"-o",
|
"-o",
|
||||||
dest="output_folder",
|
dest="output_folder",
|
||||||
default="Submissions",
|
default="Submissions",
|
||||||
help="set a custom output folder",
|
help="set a custom output folder",
|
||||||
|
type=str,
|
||||||
)
|
)
|
||||||
parser.add_argument(
|
parser.add_argument(
|
||||||
"--check",
|
"--check",
|
||||||
|
@ -51,56 +55,53 @@ parser.add_argument(
|
||||||
help="check and download latest submissions of a user",
|
help="check and download latest submissions of a user",
|
||||||
)
|
)
|
||||||
parser.add_argument(
|
parser.add_argument(
|
||||||
"-ua",
|
|
||||||
"--user-agent",
|
"--user-agent",
|
||||||
|
"-ua",
|
||||||
dest="user_agent",
|
dest="user_agent",
|
||||||
default="Mozilla/5.0 (Windows NT 10.0; Win64; x64; rv:101.0) Gecko/20100101 \
|
default="Mozilla/5.0 (Windows NT 10.0; Win64; x64; rv:101.0) Gecko/20100101 \
|
||||||
Firefox/101.0",
|
Firefox/101.0",
|
||||||
help="Your browser's useragent, may be required, depending on your luck",
|
help="Your browser's user agent, may be required, depending on your luck",
|
||||||
|
type=str,
|
||||||
)
|
)
|
||||||
parser.add_argument(
|
parser.add_argument(
|
||||||
"-sub",
|
|
||||||
"--submissions",
|
"--submissions",
|
||||||
|
"-sub",
|
||||||
action="store_true",
|
action="store_true",
|
||||||
help="download your \
|
help="download your \
|
||||||
submissions",
|
submissions",
|
||||||
)
|
)
|
||||||
parser.add_argument(
|
parser.add_argument(
|
||||||
"-f",
|
|
||||||
"--folder",
|
"--folder",
|
||||||
|
"-f",
|
||||||
help="full path of the furaffinity gallery folder. for instance 123456/\
|
help="full path of the furaffinity gallery folder. for instance 123456/\
|
||||||
Folder-Name-Here",
|
Folder-Name-Here",
|
||||||
|
type=str,
|
||||||
)
|
)
|
||||||
|
parser.add_argument("--start", default=1, help="page number to start from", type=str)
|
||||||
parser.add_argument(
|
parser.add_argument(
|
||||||
"-s",
|
|
||||||
"--start",
|
|
||||||
default=1,
|
|
||||||
help="page number to start from",
|
|
||||||
)
|
|
||||||
parser.add_argument(
|
|
||||||
"-S",
|
|
||||||
"--stop",
|
"--stop",
|
||||||
default=0,
|
default=0,
|
||||||
help="Page number to stop on. Specify the full URL after the username: for \
|
help="Page number to stop on. Specify the full URL after the username: for \
|
||||||
favorites pages (1234567890/next) or for submissions pages: \
|
favorites pages (1234567890/next) or for submissions pages: \
|
||||||
(new~123456789@48)",
|
(new~123456789@48)",
|
||||||
|
type=str,
|
||||||
)
|
)
|
||||||
parser.add_argument(
|
parser.add_argument(
|
||||||
"-rd",
|
|
||||||
"--redownload",
|
"--redownload",
|
||||||
|
"-rd",
|
||||||
action="store_false",
|
action="store_false",
|
||||||
help="Redownload files that have been downloaded already",
|
help="Redownload files that have been downloaded already",
|
||||||
)
|
)
|
||||||
parser.add_argument(
|
parser.add_argument(
|
||||||
"-i",
|
|
||||||
"--interval",
|
"--interval",
|
||||||
type=int,
|
"-i",
|
||||||
default=0,
|
default=0,
|
||||||
help="delay between downloading pages in seconds [default: 0]",
|
help="delay between downloading pages in seconds [default: 0]",
|
||||||
|
type=int,
|
||||||
)
|
)
|
||||||
parser.add_argument(
|
parser.add_argument(
|
||||||
"-r",
|
|
||||||
"--rating",
|
"--rating",
|
||||||
|
"-r",
|
||||||
action="store_false",
|
action="store_false",
|
||||||
help="disable rating separation",
|
help="disable rating separation",
|
||||||
)
|
)
|
||||||
|
@ -111,18 +112,17 @@ parser.add_argument(
|
||||||
help="enable submission filter",
|
help="enable submission filter",
|
||||||
)
|
)
|
||||||
parser.add_argument(
|
parser.add_argument(
|
||||||
"-m",
|
|
||||||
"--metadata",
|
"--metadata",
|
||||||
|
"-m",
|
||||||
action="store_true",
|
action="store_true",
|
||||||
help="enable metadata saving",
|
help="enable metadata saving",
|
||||||
)
|
)
|
||||||
parser.add_argument(
|
parser.add_argument(
|
||||||
"--download",
|
"--download", help="download a specific submission by providing its id", type=str
|
||||||
help="download a specific submission by providing its id",
|
|
||||||
)
|
)
|
||||||
parser.add_argument(
|
parser.add_argument(
|
||||||
"-jd",
|
|
||||||
"--json-description",
|
"--json-description",
|
||||||
|
"-jd",
|
||||||
dest="json_description",
|
dest="json_description",
|
||||||
action="store_true",
|
action="store_true",
|
||||||
help="download description as a JSON list",
|
help="download description as a JSON list",
|
||||||
|
@ -147,6 +147,10 @@ category = args.category
|
||||||
if username is not None:
|
if username is not None:
|
||||||
username = username.split(" ")
|
username = username.split(" ")
|
||||||
|
|
||||||
|
if os.path.exists(username[0]):
|
||||||
|
data = open(username[0]).read()
|
||||||
|
username = filter(None, data.split("\n"))
|
||||||
|
|
||||||
# Custom input
|
# Custom input
|
||||||
cookies = args.cookies
|
cookies = args.cookies
|
||||||
output_folder = args.output_folder
|
output_folder = args.output_folder
|
||||||
|
@ -199,4 +203,4 @@ search = 'YCH[a-z $-/:-?{-~!"^_`\\[\\]]*OPEN\
|
||||||
|TELEGRAM[a-z $-/:-?{-~!"^_`\\[\\]]*STICK\
|
|TELEGRAM[a-z $-/:-?{-~!"^_`\\[\\]]*STICK\
|
||||||
|TG[a-z $-/:-?{-~!"^_`\\[\\]]*STICK\
|
|TG[a-z $-/:-?{-~!"^_`\\[\\]]*STICK\
|
||||||
|REM[insder]*\\b\
|
|REM[insder]*\\b\
|
||||||
|\\bREF|\\bSale|auction|multislot|stream|adopt'
|
|\\bREF|\\bSale|auction|multislot|multi slot|stream|adopt'
|
||||||
|
|
|
@ -1,28 +1,18 @@
|
||||||
import http.cookiejar as cookielib
|
|
||||||
import json
|
import json
|
||||||
import os
|
import os
|
||||||
|
|
||||||
import requests
|
|
||||||
from bs4 import BeautifulSoup
|
from bs4 import BeautifulSoup
|
||||||
from pathvalidate import sanitize_filename
|
from pathvalidate import sanitize_filename
|
||||||
from tqdm import tqdm
|
from tqdm import tqdm
|
||||||
|
|
||||||
import Modules.config as config
|
import Modules.config as config
|
||||||
from Modules.functions import download_complete
|
from Modules.functions import DownloadComplete
|
||||||
from Modules.functions import requests_retry_session
|
from Modules.functions import requests_retry_session
|
||||||
from Modules.functions import system_message_handler
|
from Modules.functions import system_message_handler
|
||||||
|
|
||||||
session = requests.session()
|
|
||||||
if config.cookies is not None: # add cookies if present
|
|
||||||
cookies = cookielib.MozillaCookieJar(config.cookies)
|
|
||||||
cookies.load()
|
|
||||||
session.cookies = cookies
|
|
||||||
|
|
||||||
|
|
||||||
def download(path):
|
def download(path):
|
||||||
response = requests_retry_session(session=session).get(
|
response = requests_retry_session().get(f"{config.BASE_URL}{path}")
|
||||||
f"{config.BASE_URL}{path}"
|
|
||||||
)
|
|
||||||
s = BeautifulSoup(response.text, "html.parser")
|
s = BeautifulSoup(response.text, "html.parser")
|
||||||
|
|
||||||
# System messages
|
# System messages
|
||||||
|
@ -32,7 +22,7 @@ def download(path):
|
||||||
image = s.find(class_="download").find("a").attrs.get("href")
|
image = s.find(class_="download").find("a").attrs.get("href")
|
||||||
except AttributeError:
|
except AttributeError:
|
||||||
print(
|
print(
|
||||||
f"{config.ERROR_COLOR}uncessesful download of {config.BASE_URL}{path}{config.END}"
|
f"{config.ERROR_COLOR}unsuccessful download of {config.BASE_URL}{path}{config.END}"
|
||||||
)
|
)
|
||||||
download(path)
|
download(path)
|
||||||
return True
|
return True
|
||||||
|
@ -40,10 +30,14 @@ def download(path):
|
||||||
filename = sanitize_filename(image.split("/")[-1:][0])
|
filename = sanitize_filename(image.split("/")[-1:][0])
|
||||||
|
|
||||||
author = (
|
author = (
|
||||||
s.find(class_="submission-id-sub-container").find("a").find("strong").text
|
s.find(class_="submission-id-sub-container")
|
||||||
|
.find("a")
|
||||||
|
.find("strong")
|
||||||
|
.text.replace(".", "._")
|
||||||
)
|
)
|
||||||
|
|
||||||
title = sanitize_filename(
|
title = sanitize_filename(
|
||||||
s.find(class_="submission-title").find("p").contents[0]
|
str(s.find(class_="submission-title").find("p").contents[0])
|
||||||
)
|
)
|
||||||
view_id = int(path.split("/")[-2:-1][0])
|
view_id = int(path.split("/")[-2:-1][0])
|
||||||
|
|
||||||
|
@ -70,18 +64,19 @@ def download(path):
|
||||||
|
|
||||||
image_url = f"https:{image}"
|
image_url = f"https:{image}"
|
||||||
|
|
||||||
if download_file(image_url, output_path, f"{title} - [{rating}]") is True:
|
if (
|
||||||
|
download_file(
|
||||||
|
image_url, f"{config.BASE_URL}{path}", output_path, f"{title} - [{rating}]"
|
||||||
|
)
|
||||||
|
is True
|
||||||
|
):
|
||||||
with open(
|
with open(
|
||||||
f"{config.output_folder}/index.idx", encoding="utf-8", mode="a+"
|
f"{config.output_folder}/index.idx", encoding="utf-8", mode="a+"
|
||||||
) as idx:
|
) as idx:
|
||||||
idx.write(f"({view_id})\n")
|
idx.write(f"({view_id})\n")
|
||||||
|
|
||||||
if config.metadata is True:
|
if config.metadata is True:
|
||||||
dsc = (
|
dsc = s.find(class_="submission-description").text.strip().replace("\r\n", "\n")
|
||||||
s.find(class_="submission-description")
|
|
||||||
.text.strip()
|
|
||||||
.replace("\r\n", "\n")
|
|
||||||
)
|
|
||||||
if config.json_description is True:
|
if config.json_description is True:
|
||||||
dsc = []
|
dsc = []
|
||||||
data = {
|
data = {
|
||||||
|
@ -98,9 +93,7 @@ def download(path):
|
||||||
"species": s.find(class_="info").findAll("div")[2].find("span").text,
|
"species": s.find(class_="info").findAll("div")[2].find("span").text,
|
||||||
"gender": s.find(class_="info").findAll("div")[3].find("span").text,
|
"gender": s.find(class_="info").findAll("div")[3].find("span").text,
|
||||||
"views": int(s.find(class_="views").find(class_="font-large").text),
|
"views": int(s.find(class_="views").find(class_="font-large").text),
|
||||||
"favorites": int(
|
"favorites": int(s.find(class_="favorites").find(class_="font-large").text),
|
||||||
s.find(class_="favorites").find(class_="font-large").text
|
|
||||||
),
|
|
||||||
"rating": rating,
|
"rating": rating,
|
||||||
"comments": [],
|
"comments": [],
|
||||||
}
|
}
|
||||||
|
@ -114,17 +107,17 @@ def download(path):
|
||||||
return True
|
return True
|
||||||
|
|
||||||
|
|
||||||
def download_file(url, fname, desc):
|
def download_file(url, view_url, file_name, desc):
|
||||||
try:
|
try:
|
||||||
r = session.get(url, stream=True)
|
r = requests_retry_session().get(url, stream=True)
|
||||||
if r.status_code != 200:
|
if r.status_code != 200:
|
||||||
print(
|
print(
|
||||||
f'{config.ERROR_COLOR}Got a HTTP {r.status_code} while downloading \
|
f'{config.ERROR_COLOR}Got a HTTP {r.status_code} while downloading \
|
||||||
"{fname}". URL {url} ...skipping{config.END}'
|
"{file_name}" ({view_url}) ...skipping{config.END}'
|
||||||
)
|
)
|
||||||
return False
|
return False
|
||||||
total = int(r.headers.get("Content-Length", 0))
|
total = int(r.headers.get("Content-Length", 0))
|
||||||
with open(fname, "wb") as file, tqdm(
|
with open(file_name, "wb") as file, tqdm(
|
||||||
desc=desc.ljust(40),
|
desc=desc.ljust(40),
|
||||||
total=total,
|
total=total,
|
||||||
miniters=100,
|
miniters=100,
|
||||||
|
@ -137,7 +130,7 @@ def download_file(url, fname, desc):
|
||||||
bar.update(size)
|
bar.update(size)
|
||||||
except KeyboardInterrupt:
|
except KeyboardInterrupt:
|
||||||
print(f"{config.SUCCESS_COLOR}Finished downloading{config.END}")
|
print(f"{config.SUCCESS_COLOR}Finished downloading{config.END}")
|
||||||
os.remove(fname)
|
os.remove(file_name)
|
||||||
exit()
|
exit()
|
||||||
return True
|
return True
|
||||||
|
|
||||||
|
@ -155,7 +148,7 @@ def create_metadata(output, data, s, title, filename):
|
||||||
for desc in s.find("div", class_="submission-description").stripped_strings:
|
for desc in s.find("div", class_="submission-description").stripped_strings:
|
||||||
data["description"].append(desc)
|
data["description"].append(desc)
|
||||||
|
|
||||||
# Extact tags
|
# Extract tags
|
||||||
|
|
||||||
try:
|
try:
|
||||||
for tag in s.find(class_="tags-row").findAll(class_="tags"):
|
for tag in s.find(class_="tags-row").findAll(class_="tags"):
|
||||||
|
@ -194,7 +187,7 @@ def file_exists_fallback(author, title, view_id):
|
||||||
f'fallback: {config.SUCCESS_COLOR}Downloaded all recent files of \
|
f'fallback: {config.SUCCESS_COLOR}Downloaded all recent files of \
|
||||||
"{author}"{config.END}'
|
"{author}"{config.END}'
|
||||||
)
|
)
|
||||||
raise download_complete
|
raise DownloadComplete
|
||||||
print(
|
print(
|
||||||
f'fallback: {config.WARN_COLOR}Skipping "{title}" since \
|
f'fallback: {config.WARN_COLOR}Skipping "{title}" since \
|
||||||
it\'s already downloaded{config.END}'
|
it\'s already downloaded{config.END}'
|
||||||
|
|
|
@ -9,14 +9,6 @@ from urllib3.util import Retry
|
||||||
|
|
||||||
import Modules.config as config
|
import Modules.config as config
|
||||||
|
|
||||||
session = requests.session()
|
|
||||||
if config.cookies is not None: # add cookies if present
|
|
||||||
cookies = cookielib.MozillaCookieJar(config.cookies)
|
|
||||||
cookies.load()
|
|
||||||
session.cookies = cookies
|
|
||||||
|
|
||||||
session.headers.update({"User-Agent": config.user_agent})
|
|
||||||
|
|
||||||
|
|
||||||
def requests_retry_session(
|
def requests_retry_session(
|
||||||
retries=3,
|
retries=3,
|
||||||
|
@ -24,7 +16,13 @@ def requests_retry_session(
|
||||||
status_forcelist=(500, 502, 504, 104),
|
status_forcelist=(500, 502, 504, 104),
|
||||||
session=None,
|
session=None,
|
||||||
):
|
):
|
||||||
|
"""Get a session, and retry in case of an error"""
|
||||||
session = session or requests.Session()
|
session = session or requests.Session()
|
||||||
|
if config.cookies is not None: # add cookies if present
|
||||||
|
cookies = cookielib.MozillaCookieJar(config.cookies)
|
||||||
|
cookies.load()
|
||||||
|
session.cookies = cookies
|
||||||
|
session.headers.update({"User-Agent": config.user_agent})
|
||||||
retry = Retry(
|
retry = Retry(
|
||||||
total=retries,
|
total=retries,
|
||||||
read=retries,
|
read=retries,
|
||||||
|
@ -38,11 +36,12 @@ def requests_retry_session(
|
||||||
return session
|
return session
|
||||||
|
|
||||||
|
|
||||||
class download_complete(Exception):
|
class DownloadComplete(Exception):
|
||||||
pass
|
pass
|
||||||
|
|
||||||
|
|
||||||
def check_filter(title):
|
def check_filter(title):
|
||||||
|
"""Compare post title and search string, then return 'True' if match found"""
|
||||||
|
|
||||||
match = re.search(
|
match = re.search(
|
||||||
config.search,
|
config.search,
|
||||||
|
@ -56,6 +55,7 @@ def check_filter(title):
|
||||||
|
|
||||||
|
|
||||||
def system_message_handler(s):
|
def system_message_handler(s):
|
||||||
|
"""Parse and return system message text"""
|
||||||
try:
|
try:
|
||||||
message = {
|
message = {
|
||||||
s.find(class_="notice-message")
|
s.find(class_="notice-message")
|
||||||
|
@ -78,18 +78,19 @@ def system_message_handler(s):
|
||||||
.text.strip()
|
.text.strip()
|
||||||
)
|
)
|
||||||
print(f"{config.WARN_COLOR}System Message: {message}{config.END}")
|
print(f"{config.WARN_COLOR}System Message: {message}{config.END}")
|
||||||
raise download_complete
|
raise DownloadComplete
|
||||||
|
|
||||||
|
|
||||||
def login():
|
def login():
|
||||||
|
"""Get cookies from any browser with logged in furaffinity and save them to file"""
|
||||||
|
session = requests.Session()
|
||||||
|
cj = browser_cookie3.load()
|
||||||
|
|
||||||
CJ = browser_cookie3.load()
|
response = session.get(config.BASE_URL, cookies=cj)
|
||||||
|
fa_cookies = cj._cookies[".furaffinity.net"]["/"]
|
||||||
|
|
||||||
response = session.get(config.BASE_URL, cookies=CJ)
|
cookie_a = fa_cookies["a"]
|
||||||
FA_COOKIES = CJ._cookies[".furaffinity.net"]["/"]
|
cookie_b = fa_cookies["b"]
|
||||||
|
|
||||||
cookie_a = FA_COOKIES["a"]
|
|
||||||
cookie_b = FA_COOKIES["b"]
|
|
||||||
|
|
||||||
s = BeautifulSoup(response.text, "html.parser")
|
s = BeautifulSoup(response.text, "html.parser")
|
||||||
try:
|
try:
|
||||||
|
@ -116,48 +117,51 @@ furaffinity in your browser, or you can export cookies.txt manually{config.END}"
|
||||||
|
|
||||||
|
|
||||||
def next_button(page_url):
|
def next_button(page_url):
|
||||||
response = session.get(page_url)
|
"""Parse Next button and get next page url"""
|
||||||
|
response = requests_retry_session().get(page_url)
|
||||||
s = BeautifulSoup(response.text, "html.parser")
|
s = BeautifulSoup(response.text, "html.parser")
|
||||||
if config.submissions is True:
|
if config.submissions is True:
|
||||||
# unlike galleries that are sequentially numbered, submissions use a different scheme.
|
# unlike galleries that are sequentially numbered, submissions use a different scheme.
|
||||||
# the "page_num" is instead: new~[set of numbers]@(12 or 48 or 72) if sorting by new
|
# the "page_num" is instead: new~[set of numbers]@(12 or 48 or 72) if sorting by new
|
||||||
try:
|
try:
|
||||||
next_button = s.find("a", class_="button standard more").attrs.get("href")
|
parse_next_button = s.find("a", class_="button standard more").attrs.get(
|
||||||
|
"href"
|
||||||
|
)
|
||||||
except AttributeError:
|
except AttributeError:
|
||||||
try:
|
try:
|
||||||
next_button = s.find("a", class_="button standard more-half").attrs.get(
|
parse_next_button = s.find(
|
||||||
"href"
|
"a", class_="button standard more-half"
|
||||||
)
|
).attrs.get("href")
|
||||||
except AttributeError as e:
|
except AttributeError as e:
|
||||||
print(f"{config.WARN_COLOR}Unable to find next button{config.END}")
|
print(f"{config.WARN_COLOR}Unable to find next button{config.END}")
|
||||||
raise download_complete from e
|
raise DownloadComplete from e
|
||||||
page_num = next_button.split("/")[-2]
|
page_num = parse_next_button.split("/")[-2]
|
||||||
elif config.category != "favorites":
|
elif config.category != "favorites":
|
||||||
next_button = s.find("button", class_="button standard", text="Next")
|
parse_next_button = s.find("button", class_="button standard", text="Next")
|
||||||
if next_button is None or next_button.parent is None:
|
if parse_next_button is None or parse_next_button.parent is None:
|
||||||
print(f"{config.WARN_COLOR}Unable to find next button{config.END}")
|
print(f"{config.WARN_COLOR}Unable to find next button{config.END}")
|
||||||
raise download_complete
|
raise DownloadComplete
|
||||||
page_num = next_button.parent.attrs["action"].split("/")[-2]
|
page_num = parse_next_button.parent.attrs["action"].split("/")[-2]
|
||||||
else:
|
else:
|
||||||
next_button = s.find("a", class_="button standard right", text="Next")
|
parse_next_button = s.find("a", class_="button standard right", text="Next")
|
||||||
page_num = fav_next_button(s)
|
page_num = fav_next_button(parse_next_button)
|
||||||
print(
|
print(
|
||||||
f"Downloading page {page_num} - {config.BASE_URL}{next_button.parent.attrs['action']}"
|
f"Downloading page {page_num} - {config.BASE_URL}{parse_next_button.parent.attrs['action']}"
|
||||||
)
|
)
|
||||||
return page_num
|
return page_num
|
||||||
|
|
||||||
|
|
||||||
def fav_next_button():
|
def fav_next_button(parse_next_button):
|
||||||
# unlike galleries that are sequentially numbered, favorites use a different scheme.
|
# unlike galleries that are sequentially numbered, favorites use a different scheme.
|
||||||
# the "page_num" is instead: [set of numbers]/next (the trailing /next is required)
|
# the "page_num" is instead: [set of numbers]/next (the trailing /next is required)
|
||||||
if next_button is None:
|
if parse_next_button is None:
|
||||||
print(f"{config.WARN_COLOR}Unable to find next button{config.END}")
|
print(f"{config.WARN_COLOR}Unable to find next button{config.END}")
|
||||||
raise download_complete
|
raise DownloadComplete
|
||||||
next_page_link = next_button.attrs["href"]
|
next_page_link = parse_next_button.attrs["href"]
|
||||||
next_fav_num = re.search(r"\d+", next_page_link)
|
next_fav_num = re.search(r"\d+", next_page_link)
|
||||||
|
|
||||||
if next_fav_num is None:
|
if next_fav_num is None:
|
||||||
print(f"{config.WARN_COLOR}Failed to parse next favorite link{config.END}")
|
print(f"{config.WARN_COLOR}Failed to parse next favorite link{config.END}")
|
||||||
raise download_complete
|
raise DownloadComplete
|
||||||
|
|
||||||
return f"{next_fav_num[0]}/next"
|
return f"{next_fav_num[0]}/next"
|
||||||
|
|
|
@ -9,8 +9,8 @@ import Modules.config as config
|
||||||
@lru_cache(maxsize=None)
|
@lru_cache(maxsize=None)
|
||||||
def start_indexing(path, layer=0):
|
def start_indexing(path, layer=0):
|
||||||
|
|
||||||
"""Recursively iterate over each item in path
|
"""Recursively iterate over each item in path, then
|
||||||
and print item's name.
|
save and print item's name.
|
||||||
"""
|
"""
|
||||||
|
|
||||||
# make Path object from input string
|
# make Path object from input string
|
||||||
|
@ -23,7 +23,7 @@ def start_indexing(path, layer=0):
|
||||||
if p.is_file():
|
if p.is_file():
|
||||||
name = p.stem
|
name = p.stem
|
||||||
ext = p.suffix
|
ext = p.suffix
|
||||||
match = re.search(r"\([0-9]{5,}\)", name)
|
match = re.search(r"\(\d{5,}\)", name)
|
||||||
if match is None and ext not in [".txt", ".idx"]:
|
if match is None and ext not in [".txt", ".idx"]:
|
||||||
return
|
return
|
||||||
|
|
||||||
|
@ -39,6 +39,7 @@ def start_indexing(path, layer=0):
|
||||||
|
|
||||||
@lru_cache(maxsize=None)
|
@lru_cache(maxsize=None)
|
||||||
def check_file(path):
|
def check_file(path):
|
||||||
|
"""compare file view id with index list"""
|
||||||
view_id = path.split("/")[-2:-1][0]
|
view_id = path.split("/")[-2:-1][0]
|
||||||
with contextlib.suppress(FileNotFoundError):
|
with contextlib.suppress(FileNotFoundError):
|
||||||
with open(f"{config.output_folder}/index.idx", encoding="utf-8") as idx:
|
with open(f"{config.output_folder}/index.idx", encoding="utf-8") as idx:
|
||||||
|
|
61
README.md
61
README.md
|
@ -1,22 +1,20 @@
|
||||||
This branch is the development version of furaffinity-dl rewritten in python.
|
|
||||||
|
|
||||||
# FurAffinity Downloader
|
# FurAffinity Downloader
|
||||||
|
|
||||||
**furaffinity-dl** is a python script for batch downloading of galleries (and scraps/favourites) from furaffinity users users or your submissons!
|
**furaffinity-dl** is a python script for batch downloading of galleries (and scraps/favorites) from furaffinity users users or your submission notifications!
|
||||||
It was written for preservation of culture, to counter the people nuking their galleries every once a while.
|
Mainly it was written for preservation of culture, to counter the people nuking their galleries every once a while.
|
||||||
and then modified for confinience.
|
But no-one is restricting you from just using is for convenience.
|
||||||
|
|
||||||
Supports all known submission types: images, text, flash and audio.
|
Supports all known submission types: images, text, flash and audio.
|
||||||
|
|
||||||
## Requirements
|
## Requirements
|
||||||
|
|
||||||
`python 3`
|
`python3` (Recommended version is 3.10.x and above)
|
||||||
|
|
||||||
`pip3 install -r requirements.txt`
|
`pip3 install -r requirements.txt`
|
||||||
|
|
||||||
**The script currently only works with the "Modern" theme**
|
furaffinity-dl has been tested on Linux and Windows OSs, however it should also work on Mac or any other platform that supports python.
|
||||||
|
|
||||||
furaffinity-dl has only been tested only on Linux, however it should also work on Mac, Windows or any other platform that supports python.
|
***The script currently only works with the "Modern" theme***
|
||||||
|
|
||||||
## Usage
|
## Usage
|
||||||
|
|
||||||
|
@ -24,46 +22,47 @@ When downloading a folder make sure to put everything after **/folder/**, for ex
|
||||||
|
|
||||||
```help
|
```help
|
||||||
|
|
||||||
usage: furaffinity-dl.py [-h] [-c COOKIES] [--output OUTPUT_FOLDER] [--check] [-ua USER_AGENT] [-sub] [-f FOLDER] [-s START [START ...]]
|
usage: furaffinity-dl.py [-h] [--cookies COOKIES] [--output OUTPUT_FOLDER] [--check] [--user-agent USER_AGENT] [--submissions] [--folder FOLDER] [--start START]
|
||||||
[-S STOP] [-rd] [-i INTERVAL] [-r] [--filter] [-m] [--download DOWNLOAD] [-jd] [--login]
|
[--stop STOP] [--redownload] [--interval INTERVAL] [--rating] [--filter] [--metadata] [--download DOWNLOAD] [--json-description] [--login]
|
||||||
|
[--index]
|
||||||
[username] [category]
|
[username] [category]
|
||||||
|
|
||||||
Downloads the entire gallery/scraps/folder/favorites of a furaffinity user, or your submissions notifications
|
Downloads the entire gallery/scraps/folder/favorites of a furaffinity user, or your submissions notifications
|
||||||
|
|
||||||
positional arguments:
|
positional arguments:
|
||||||
username username of the furaffinity user
|
username username of the furaffinity user (if username is starting with '-' or '--' provide them through a file instead)
|
||||||
category the category to download, gallery/scraps/favorites [default: gallery]
|
category the category to download, gallery/scraps/favorites [default: gallery]
|
||||||
|
|
||||||
options:
|
options:
|
||||||
-h, --help show this help message and exit
|
-h, --help show this help message and exit
|
||||||
-c COOKIES, --cookies COOKIES
|
--cookies COOKIES, -c COOKIES
|
||||||
path to a NetScape cookies file
|
path to a NetScape cookies file
|
||||||
--output OUTPUT_FOLDER, -o OUTPUT_FOLDER
|
--output OUTPUT_FOLDER, -o OUTPUT_FOLDER
|
||||||
set a custom output folder
|
set a custom output folder
|
||||||
--check check and download latest submissions of a user
|
--check check and download latest submissions of a user
|
||||||
-ua USER_AGENT, --user-agent USER_AGENT
|
--user-agent USER_AGENT, -ua USER_AGENT
|
||||||
Your browser's useragent, may be required, depending on your luck
|
Your browser's user agent, may be required, depending on your luck
|
||||||
-sub, --submissions download your submissions
|
--submissions, -sub download your submissions
|
||||||
-f FOLDER, --folder FOLDER
|
--folder FOLDER, -f FOLDER
|
||||||
full path of the furaffinity gallery folder. for instance 123456/Folder-Name-Here
|
full path of the furaffinity gallery folder. for instance 123456/Folder-Name-Here
|
||||||
-s START [START ...], --start START [START ...]
|
--start START page number to start from
|
||||||
page number to start from
|
--stop STOP Page number to stop on. Specify the full URL after the username: for favorites pages (1234567890/next) or for submissions pages: (new~123456789@48)
|
||||||
-S STOP, --stop STOP Page number to stop on. Specify the full URL after the username: for favorites pages (1234567890/next) or for submissions pages: (new~123456789@48)
|
--redownload, -rd Redownload files that have been downloaded already
|
||||||
-rd, --redownload Redownload files that have been downloaded already
|
--interval INTERVAL, -i INTERVAL
|
||||||
-i INTERVAL, --interval INTERVAL
|
|
||||||
delay between downloading pages in seconds [default: 0]
|
delay between downloading pages in seconds [default: 0]
|
||||||
-r, --rating disable rating separation
|
--rating, -r disable rating separation
|
||||||
--filter enable submission filter
|
--filter enable submission filter
|
||||||
-m, --metadata enable metadata saving
|
--metadata, -m enable metadata saving
|
||||||
--download DOWNLOAD download a specific submission /view/12345678/
|
--download DOWNLOAD download a specific submission by providing its id
|
||||||
-jd, --json-description
|
--json-description, -jd
|
||||||
download description as a JSON list
|
download description as a JSON list
|
||||||
--login extract furaffinity cookies directly from your browser
|
--login extract furaffinity cookies directly from your browser
|
||||||
|
--index create an index of downloaded files in an output folder
|
||||||
|
|
||||||
Examples:
|
Examples:
|
||||||
python3 furaffinity-dl.py koul -> will download gallery of user koul
|
python3 furaffinity-dl.py koul -> will download gallery of user koul
|
||||||
python3 furaffinity-dl.py koul scraps -> will download scraps of user koul
|
python3 furaffinity-dl.py koul scraps -> will download scraps of user koul
|
||||||
python3 furaffinity-dl.py mylafox favorites -> will download favorites of user mylafox
|
python3 furaffinity-dl.py mylafox favorites -> will download favorites of user mylafox
|
||||||
|
|
||||||
You can also download a several users in one go like this:
|
You can also download a several users in one go like this:
|
||||||
python3 furaffinity-dl.py "koul radiquum mylafox" -> will download gallery of users koul -> radiquum -> mylafox
|
python3 furaffinity-dl.py "koul radiquum mylafox" -> will download gallery of users koul -> radiquum -> mylafox
|
||||||
|
@ -71,21 +70,21 @@ You can also provide a file with user names that are separated by a new line
|
||||||
|
|
||||||
You can also log in to FurAffinity in a web browser and load cookies to download age restricted content or submissions:
|
You can also log in to FurAffinity in a web browser and load cookies to download age restricted content or submissions:
|
||||||
python3 furaffinity-dl.py letodoesart -c cookies.txt -> will download gallery of user letodoesart including Mature and Adult submissions
|
python3 furaffinity-dl.py letodoesart -c cookies.txt -> will download gallery of user letodoesart including Mature and Adult submissions
|
||||||
python3 furaffinity-dl.py --submissions -c cookies.txt -> will download your submissions notifications
|
python3 furaffinity-dl.py --submissions -c cookies.txt -> will download your submissions notifications
|
||||||
|
|
||||||
DISCLAIMER: It is your own responsibility to check whether batch downloading is allowed by FurAffinity terms of service and to abide by them.
|
DISCLAIMER: It is your own responsibility to check whether batch downloading is allowed by FurAffinity terms of service and to abide by them.
|
||||||
|
|
||||||
```
|
```
|
||||||
|
|
||||||
You can also log in to download restricted content. To do that, log in to FurAffinity in your web browser, and use `python3 furaffinity-dl.py --login` to export furaffinity cookies from your web browser in Netscape format directly in file `cookies.txt` or export them manually with extensions: [for Firefox](https://addons.mozilla.org/en-US/firefox/addon/ganbo/) and [for Chrome based browsers](https://chrome.google.com/webstore/detail/get-cookiestxt/bgaddhkoddajcdgocldbbfleckgcbcid?hl=en), then you can then pass them to the script with the `-c` flag, like this (you may also have to provide your user agent):
|
You can also log in to download restricted content. To do that, log in to FurAffinity in your web browser, and use `python3 furaffinity-dl.py --login` to export furaffinity cookies from your web browser in Netscape format directly in to the file `cookies.txt` or export them manually with extensions: [for Firefox](https://addons.mozilla.org/en-US/firefox/addon/ganbo/) and [for Chrome based browsers](https://chrome.google.com/webstore/detail/get-cookiestxt/bgaddhkoddajcdgocldbbfleckgcbcid?hl=en), then you can then pass them to the script with the `-c` flag, like this (you may also have to provide your user agent):
|
||||||
|
|
||||||
`python3 furaffinity-dl.py letodoesart -c cookies.txt --user_agent 'Mozilla/5.0 ....'`
|
`python3 furaffinity-dl.py letodoesart -c cookies.txt --user-agent 'Mozilla/5.0 ....'`
|
||||||
|
|
||||||
## TODO
|
<!-- ## TODO
|
||||||
|
|
||||||
- Download user profile information.
|
- Download user profile information.
|
||||||
- "Classic" theme support
|
- "Classic" theme support
|
||||||
- Login without having to export cookies
|
- Login without having to export cookies -->
|
||||||
|
|
||||||
## Disclaimer
|
## Disclaimer
|
||||||
|
|
||||||
|
|
|
@ -1,16 +1,14 @@
|
||||||
#!/usr/bin/python3
|
#!/usr/bin/python3
|
||||||
import contextlib
|
import contextlib
|
||||||
import http.cookiejar as cookielib
|
|
||||||
import os
|
import os
|
||||||
from time import sleep
|
from time import sleep
|
||||||
|
|
||||||
import requests
|
|
||||||
from bs4 import BeautifulSoup
|
from bs4 import BeautifulSoup
|
||||||
|
|
||||||
import Modules.config as config
|
import Modules.config as config
|
||||||
from Modules.download import download
|
from Modules.download import download
|
||||||
from Modules.functions import check_filter
|
from Modules.functions import check_filter
|
||||||
from Modules.functions import download_complete
|
from Modules.functions import DownloadComplete
|
||||||
from Modules.functions import login
|
from Modules.functions import login
|
||||||
from Modules.functions import next_button
|
from Modules.functions import next_button
|
||||||
from Modules.functions import requests_retry_session
|
from Modules.functions import requests_retry_session
|
||||||
|
@ -18,20 +16,11 @@ from Modules.functions import system_message_handler
|
||||||
from Modules.index import check_file
|
from Modules.index import check_file
|
||||||
from Modules.index import start_indexing
|
from Modules.index import start_indexing
|
||||||
|
|
||||||
# get session
|
|
||||||
session = requests.session()
|
|
||||||
session.headers.update({"User-Agent": config.user_agent})
|
|
||||||
|
|
||||||
if config.cookies is not None: # add cookies if present
|
|
||||||
cookies = cookielib.MozillaCookieJar(config.cookies)
|
|
||||||
cookies.load()
|
|
||||||
session.cookies = cookies
|
|
||||||
|
|
||||||
|
|
||||||
def main():
|
def main():
|
||||||
# download loop
|
"""loop over and download all images on the page(s)"""
|
||||||
page_num = config.start
|
page_num = config.start
|
||||||
with contextlib.suppress(download_complete):
|
with contextlib.suppress(DownloadComplete):
|
||||||
while True:
|
while True:
|
||||||
if config.stop == page_num:
|
if config.stop == page_num:
|
||||||
print(
|
print(
|
||||||
|
@ -41,7 +30,7 @@ stopping.{config.END}'
|
||||||
break
|
break
|
||||||
|
|
||||||
page_url = f"{download_url}/{page_num}"
|
page_url = f"{download_url}/{page_num}"
|
||||||
response = requests_retry_session(session=session).get(page_url)
|
response = requests_retry_session().get(page_url)
|
||||||
s = BeautifulSoup(response.text, "html.parser")
|
s = BeautifulSoup(response.text, "html.parser")
|
||||||
|
|
||||||
# System messages
|
# System messages
|
||||||
|
@ -71,7 +60,7 @@ downloaded - {config.BASE_URL}{img_url}{config.END}'
|
||||||
f'{config.SUCCESS_COLOR}Downloaded all recent files of \
|
f'{config.SUCCESS_COLOR}Downloaded all recent files of \
|
||||||
"{username}"{config.END}'
|
"{username}"{config.END}'
|
||||||
)
|
)
|
||||||
raise download_complete
|
raise DownloadComplete
|
||||||
print(
|
print(
|
||||||
f'{config.WARN_COLOR}Skipping "{title}" since \
|
f'{config.WARN_COLOR}Skipping "{title}" since \
|
||||||
it\'s already downloaded{config.END}'
|
it\'s already downloaded{config.END}'
|
||||||
|
@ -96,15 +85,12 @@ if __name__ == "__main__":
|
||||||
print(f"{config.SUCCESS_COLOR}indexing finished{config.END}")
|
print(f"{config.SUCCESS_COLOR}indexing finished{config.END}")
|
||||||
exit()
|
exit()
|
||||||
|
|
||||||
try:
|
one_time_response = requests_retry_session().get(config.BASE_URL)
|
||||||
response = requests_retry_session(session=session).get(config.BASE_URL)
|
one_time_s = BeautifulSoup(one_time_response.text, "html.parser")
|
||||||
except KeyboardInterrupt:
|
if one_time_s.find(class_="loggedin_user_avatar") is not None:
|
||||||
print(f"{config.WARN_COLOR}Aborted by user{config.END}")
|
account_username = one_time_s.find(class_="loggedin_user_avatar").attrs.get(
|
||||||
exit()
|
"alt"
|
||||||
|
)
|
||||||
s = BeautifulSoup(response.text, "html.parser")
|
|
||||||
if s.find(class_="loggedin_user_avatar") is not None:
|
|
||||||
account_username = s.find(class_="loggedin_user_avatar").attrs.get("alt")
|
|
||||||
print(
|
print(
|
||||||
f'{config.SUCCESS_COLOR}Logged in as \
|
f'{config.SUCCESS_COLOR}Logged in as \
|
||||||
"{account_username}"{config.END}'
|
"{account_username}"{config.END}'
|
||||||
|
@ -146,17 +132,6 @@ downloading "{config.folder[1]}"{config.END}'
|
||||||
)
|
)
|
||||||
exit()
|
exit()
|
||||||
|
|
||||||
try:
|
|
||||||
if os.path.exists(config.username[0]):
|
|
||||||
data = open(config.username[0]).read()
|
|
||||||
config.username = filter(None, data.split("\n"))
|
|
||||||
except TypeError or AttributeError:
|
|
||||||
print(
|
|
||||||
f"{config.ERROR_COLOR}Please enter a username \
|
|
||||||
or provide a file with usernames (1 username per line){config.END}"
|
|
||||||
)
|
|
||||||
exit()
|
|
||||||
|
|
||||||
for username in config.username:
|
for username in config.username:
|
||||||
username = username.split("#")[0].translate(
|
username = username.split("#")[0].translate(
|
||||||
str.maketrans(config.username_replace_chars)
|
str.maketrans(config.username_replace_chars)
|
||||||
|
|
|
@ -1,6 +1,7 @@
|
||||||
beautifulsoup4
|
urllib3
|
||||||
requests
|
requests
|
||||||
|
beautifulsoup4
|
||||||
tqdm
|
tqdm
|
||||||
browser-cookie3
|
|
||||||
pathvalidate
|
pathvalidate
|
||||||
pre-commit
|
pre-commit
|
||||||
|
browser-cookie3
|
||||||
|
|
Loading…
Add table
Reference in a new issue