mirror of
https://github.com/Radiquum/furaffinity-dl.git
synced 2025-04-05 15:54:38 +00:00
some cosmetic changes:
- add shorter arguments - metadata file saving will now be in same folder as an image (it's same as it was before) - print output path when downloaded a specific file
This commit is contained in:
parent
7b17023597
commit
c4400d4f78
2 changed files with 41 additions and 35 deletions
27
README.md
27
README.md
|
@ -24,10 +24,9 @@ When downloading a folder make sure to put everything after **/folder/**, for ex
|
||||||
|
|
||||||
```help
|
```help
|
||||||
|
|
||||||
usage: furaffinity-dl.py [-h] [--submissions] [--folder FOLDER [FOLDER ...]] [--cookies COOKIES [COOKIES ...]]
|
usage: furaffinity-dl.py [-h] [-sub] [-f FOLDER [FOLDER ...]] [-c COOKIES [COOKIES ...]] [-ua USER_AGENT [USER_AGENT ...]] [--start START [START ...]]
|
||||||
[--user-agent USER_AGENT [USER_AGENT ...]] [--start START [START ...]] [--stop STOP [STOP ...]] [--dont-redownload]
|
[--stop STOP [STOP ...]] [--redownload] [--interval INTERVAL [INTERVAL ...]] [--rating] [--filter] [--metadata] [--download DOWNLOAD]
|
||||||
[--interval INTERVAL [INTERVAL ...]] [--rating] [--filter] [--metadata] [--download DOWNLOAD] [--json-description]
|
[-jd] [--login]
|
||||||
[--login]
|
|
||||||
[username] [category]
|
[username] [category]
|
||||||
|
|
||||||
Downloads the entire gallery/scraps/folder/favorites of a furaffinity user, or your submissions notifications
|
Downloads the entire gallery/scraps/folder/favorites of a furaffinity user, or your submissions notifications
|
||||||
|
@ -38,36 +37,36 @@ positional arguments:
|
||||||
|
|
||||||
options:
|
options:
|
||||||
-h, --help show this help message and exit
|
-h, --help show this help message and exit
|
||||||
--submissions download your submissions
|
-sub, --submissions download your submissions
|
||||||
--folder FOLDER [FOLDER ...]
|
-f FOLDER [FOLDER ...], --folder FOLDER [FOLDER ...]
|
||||||
full path of the furaffinity gallery folder. for instance 123456/Folder-Name-Here
|
full path of the furaffinity gallery folder. for instance 123456/Folder-Name-Here
|
||||||
--cookies COOKIES [COOKIES ...], -c COOKIES [COOKIES ...]
|
-c COOKIES [COOKIES ...], --cookies COOKIES [COOKIES ...]
|
||||||
path to a NetScape cookies file
|
path to a NetScape cookies file
|
||||||
--user-agent USER_AGENT [USER_AGENT ...]
|
-ua USER_AGENT [USER_AGENT ...], --user-agent USER_AGENT [USER_AGENT ...]
|
||||||
Your browser's useragent, may be required, depending on your luck
|
Your browser's useragent, may be required, depending on your luck
|
||||||
--start START [START ...], -s START [START ...]
|
--start START [START ...], -s START [START ...]
|
||||||
page number to start from
|
page number to start from
|
||||||
--stop STOP [STOP ...], -S STOP [STOP ...]
|
--stop STOP [STOP ...], -S STOP [STOP ...]
|
||||||
Page number to stop on. Specify the full URL after the username: for favorites pages (1234567890/next) or for submissions pages: (new~123456789@48)
|
Page number to stop on. Specify the full URL after the username: for favorites pages (1234567890/next) or for submissions pages: (new~123456789@48)
|
||||||
--dont-redownload, -d
|
--redownload, -rd Redownload files that have been downloaded already
|
||||||
Allow to redownload files that have been downloaded already
|
|
||||||
--interval INTERVAL [INTERVAL ...], -i INTERVAL [INTERVAL ...]
|
--interval INTERVAL [INTERVAL ...], -i INTERVAL [INTERVAL ...]
|
||||||
delay between downloading pages in seconds [default: 0]
|
delay between downloading pages in seconds [default: 0]
|
||||||
--rating, -r disable rating separation
|
--rating, -r disable rating separation
|
||||||
--filter disable submission filter
|
--filter disable submission filter
|
||||||
--metadata, -m enable downloading of metadata
|
--metadata, -m enable metadata saving
|
||||||
--download DOWNLOAD download a specific submission /view/12345678/
|
--download DOWNLOAD download a specific submission /view/12345678/
|
||||||
--json-description download description as a JSON list
|
-jd, --json-description
|
||||||
|
download description as a JSON list
|
||||||
--login extract furaffinity cookies directly from your browser
|
--login extract furaffinity cookies directly from your browser
|
||||||
|
|
||||||
Examples:
|
Examples:
|
||||||
python3 furaffinity-dl.py koul -> will download gallery of user koul
|
python3 furaffinity-dl.py koul -> will download gallery of user koul
|
||||||
python3 furaffinity-dl.py koul scraps -> will download scraps of user koul
|
python3 furaffinity-dl.py koul scraps -> will download scraps of user koul
|
||||||
python3 furaffinity-dl.py mylafox favorites -> will download favorites of user mylafox
|
python3 furaffinity-dl.py mylafox favorites -> will download favorites of user mylafox
|
||||||
|
|
||||||
You can also log in to FurAffinity in a web browser and load cookies to download age restricted content or submissions:
|
You can also log in to FurAffinity in a web browser and load cookies to download age restricted content or submissions:
|
||||||
python3 furaffinity-dl.py letodoesart -c cookies.txt -> will download gallery of user letodoesart including Mature and Adult submissions
|
python3 furaffinity-dl.py letodoesart -c cookies.txt -> will download gallery of user letodoesart including Mature and Adult submissions
|
||||||
python3 furaffinity-dl.py --submissions -c cookies.txt -> will download your submissions notifications
|
python3 furaffinity-dl.py --submissions -c cookies.txt -> will download your submissions notifications
|
||||||
|
|
||||||
DISCLAIMER: It is your own responsibility to check whether batch downloading is allowed by FurAffinity terms of service and to abide by them.
|
DISCLAIMER: It is your own responsibility to check whether batch downloading is allowed by FurAffinity terms of service and to abide by them.
|
||||||
|
|
||||||
|
|
|
@ -40,17 +40,19 @@ parser.add_argument(
|
||||||
default="gallery",
|
default="gallery",
|
||||||
)
|
)
|
||||||
parser.add_argument(
|
parser.add_argument(
|
||||||
"--submissions", action="store_true", help="download your submissions"
|
"-sub", "--submissions", action="store_true", help="download your submissions"
|
||||||
)
|
)
|
||||||
parser.add_argument(
|
parser.add_argument(
|
||||||
|
"-f",
|
||||||
"--folder",
|
"--folder",
|
||||||
nargs="+",
|
nargs="+",
|
||||||
help="full path of the furaffinity gallery folder. for instance 123456/Folder-Name-Here",
|
help="full path of the furaffinity gallery folder. for instance 123456/Folder-Name-Here",
|
||||||
)
|
)
|
||||||
parser.add_argument(
|
parser.add_argument(
|
||||||
"--cookies", "-c", nargs="+", help="path to a NetScape cookies file"
|
"-c", "--cookies", nargs="+", help="path to a NetScape cookies file"
|
||||||
)
|
)
|
||||||
parser.add_argument(
|
parser.add_argument(
|
||||||
|
"-ua",
|
||||||
"--user-agent",
|
"--user-agent",
|
||||||
dest="user_agent",
|
dest="user_agent",
|
||||||
nargs="+",
|
nargs="+",
|
||||||
|
@ -70,10 +72,11 @@ parser.add_argument(
|
||||||
help="Page number to stop on. Specify the full URL after the username: for favorites pages (1234567890/next) or for submissions pages: (new~123456789@48)",
|
help="Page number to stop on. Specify the full URL after the username: for favorites pages (1234567890/next) or for submissions pages: (new~123456789@48)",
|
||||||
)
|
)
|
||||||
parser.add_argument(
|
parser.add_argument(
|
||||||
"--dont-redownload",
|
"--redownload",
|
||||||
"-d",
|
"-rd",
|
||||||
|
dest="dont_redownload",
|
||||||
action="store_false",
|
action="store_false",
|
||||||
help="Allow to redownload files that have been downloaded already",
|
help="Redownload files that have been downloaded already",
|
||||||
)
|
)
|
||||||
parser.add_argument(
|
parser.add_argument(
|
||||||
"--interval",
|
"--interval",
|
||||||
|
@ -98,13 +101,14 @@ parser.add_argument(
|
||||||
"--metadata",
|
"--metadata",
|
||||||
"-m",
|
"-m",
|
||||||
action="store_true",
|
action="store_true",
|
||||||
help="enable downloading of metadata",
|
help="enable metadata saving",
|
||||||
)
|
)
|
||||||
parser.add_argument(
|
parser.add_argument(
|
||||||
"--download",
|
"--download",
|
||||||
help="download a specific submission /view/12345678/",
|
help="download a specific submission /view/12345678/",
|
||||||
)
|
)
|
||||||
parser.add_argument(
|
parser.add_argument(
|
||||||
|
"-jd",
|
||||||
"--json-description",
|
"--json-description",
|
||||||
dest="json_description",
|
dest="json_description",
|
||||||
action="store_true",
|
action="store_true",
|
||||||
|
@ -163,7 +167,7 @@ def download_file(url, fname, desc):
|
||||||
|
|
||||||
total = int(r.headers.get("Content-Length", 0))
|
total = int(r.headers.get("Content-Length", 0))
|
||||||
with open(fname, "wb") as file, tqdm(
|
with open(fname, "wb") as file, tqdm(
|
||||||
desc=desc.ljust(60)[:60],
|
desc=desc.ljust(40)[:40],
|
||||||
total=total,
|
total=total,
|
||||||
miniters=100,
|
miniters=100,
|
||||||
unit="b",
|
unit="b",
|
||||||
|
@ -216,7 +220,7 @@ def download(path):
|
||||||
exit()
|
exit()
|
||||||
|
|
||||||
image = s.find(class_="download").find("a").attrs.get("href")
|
image = s.find(class_="download").find("a").attrs.get("href")
|
||||||
title = s.find(class_="submission-title").find("p").contents[0] + " "
|
title = f' {s.find(class_="submission-title").find("p").contents[0]} '
|
||||||
description = (
|
description = (
|
||||||
s.find(class_="submission-description").text.strip().replace("\r\n", "\n")
|
s.find(class_="submission-description").text.strip().replace("\r\n", "\n")
|
||||||
)
|
)
|
||||||
|
@ -246,7 +250,7 @@ def download(path):
|
||||||
"comments": [],
|
"comments": [],
|
||||||
}
|
}
|
||||||
|
|
||||||
if args.submissions is True:
|
if args.submissions is True or args.download is not None:
|
||||||
global output
|
global output
|
||||||
output = f"furaffinity-dl/gallery/{data.get('author')}"
|
output = f"furaffinity-dl/gallery/{data.get('author')}"
|
||||||
|
|
||||||
|
@ -258,33 +262,40 @@ def download(path):
|
||||||
)
|
)
|
||||||
if match is not None and title == match.string:
|
if match is not None and title == match.string:
|
||||||
print(
|
print(
|
||||||
f"{YELLOW}<i> post {title} was filtered and will not be downloaded - {data.get('url')}{END}"
|
f"{YELLOW}<i> post:{title}was filtered and will not be downloaded - {data.get('url')}{END}"
|
||||||
)
|
)
|
||||||
return True
|
return True
|
||||||
|
|
||||||
image_url = f"https:{image}"
|
image_url = f"https:{image}"
|
||||||
|
|
||||||
os.makedirs(output, exist_ok=True)
|
os.makedirs(output, exist_ok=True)
|
||||||
|
global output_path
|
||||||
output_path = f"{output}/{filename}"
|
output_path = f"{output}/{filename}"
|
||||||
if args.rating is True:
|
if args.rating is True:
|
||||||
os.makedirs(f'{output}/{data.get("rating")}', exist_ok=True)
|
os.makedirs(f'{output}/{data.get("rating")}', exist_ok=True)
|
||||||
output_path = f'{output}/{data.get("rating")}/{filename}'
|
output_path = f'{output}/{data.get("rating")}/{filename}'
|
||||||
|
|
||||||
if args.dont_redownload is True and os.path.isfile(output_path):
|
if args.dont_redownload is True and os.path.isfile(output_path):
|
||||||
print(f'{YELLOW}<i> Skipping "{title}", since it\'s already downloaded{END}')
|
print(f"{YELLOW}<i> Skipping:{title} since it's already downloaded{END}")
|
||||||
else:
|
else:
|
||||||
download_file(image_url, output_path, title)
|
download_file(image_url, output_path, title)
|
||||||
|
|
||||||
if args.metadata is True:
|
if args.metadata is True:
|
||||||
|
|
||||||
metadata = f"{output}/metadata"
|
metadata = output_path
|
||||||
|
|
||||||
# Extract description as list
|
# Extract description as list
|
||||||
if args.json_description is True:
|
if args.json_description is True:
|
||||||
for desc in s.find("div", class_="submission-description").stripped_strings:
|
for desc in s.find("div", class_="submission-description").stripped_strings:
|
||||||
|
|
||||||
if re.search("[<>/]", desc) is True:
|
if re.search("<", desc) is True:
|
||||||
desc = desc.replace("<", "").replace(">", "").replace("/", "")
|
desc = desc.replace("<", "")
|
||||||
|
|
||||||
|
if re.search(">", desc) is True:
|
||||||
|
desc = desc.replace(">", "")
|
||||||
|
|
||||||
|
if re.search("/", desc) is True:
|
||||||
|
desc = desc.replace("/", "")
|
||||||
|
|
||||||
data["description"].append(desc)
|
data["description"].append(desc)
|
||||||
|
|
||||||
|
@ -294,7 +305,7 @@ def download(path):
|
||||||
for tag in s.find(class_="tags-row").findAll(class_="tags"):
|
for tag in s.find(class_="tags-row").findAll(class_="tags"):
|
||||||
data["tags"].append(tag.find("a").text)
|
data["tags"].append(tag.find("a").text)
|
||||||
except AttributeError:
|
except AttributeError:
|
||||||
print(f'{YELLOW}<i> post: "{title}", has no tags{END}')
|
print(f"{YELLOW}<i> post:{title} has no tags{END}")
|
||||||
|
|
||||||
# Extract comments
|
# Extract comments
|
||||||
for comment in s.findAll(class_="comment_container"):
|
for comment in s.findAll(class_="comment_container"):
|
||||||
|
@ -319,19 +330,15 @@ def download(path):
|
||||||
)
|
)
|
||||||
|
|
||||||
# Write a UTF-8 encoded JSON file for metadata
|
# Write a UTF-8 encoded JSON file for metadata
|
||||||
os.makedirs(metadata, exist_ok=True)
|
with open(f"{metadata}.json", "w", encoding="utf-8") as f:
|
||||||
with open(
|
|
||||||
os.path.join(metadata, f"{filename}.json"), "w", encoding="utf-8"
|
|
||||||
) as f:
|
|
||||||
json.dump(data, f, ensure_ascii=False, indent=4)
|
json.dump(data, f, ensure_ascii=False, indent=4)
|
||||||
|
|
||||||
return True
|
return True
|
||||||
|
|
||||||
|
|
||||||
if args.download is not None:
|
if args.download is not None:
|
||||||
output = "furaffinity-dl/downloaded/"
|
|
||||||
download(args.download)
|
download(args.download)
|
||||||
print(f"{GREEN}<i> File downloaded{END}")
|
print(f"{GREEN}<i> File saved as {output_path} {END}")
|
||||||
exit()
|
exit()
|
||||||
|
|
||||||
# Main function
|
# Main function
|
||||||
|
|
Loading…
Add table
Reference in a new issue