mirror of
https://github.com/ClaytonWWilson/Scraper-for-theTVDB.com.git
synced 2025-12-15 17:28:46 +00:00
Refactoring - Added download support for images.
This commit is contained in:
parent
772581c6ec
commit
e4cd55501b
156
main.py
156
main.py
@ -33,75 +33,123 @@ def download(series):
|
||||
|
||||
api_con = APIConnector()
|
||||
|
||||
# # Download series text data to info.json
|
||||
# api_path = "https://api.thetvdb.com/series/{}".format(series.id)
|
||||
# res = api_con.send_http_req(api_path)
|
||||
# Download series text data to info.json
|
||||
api_path = "https://api.thetvdb.com/series/{}".format(series.id)
|
||||
res = api_con.send_http_req(api_path)
|
||||
|
||||
# info_path = os.path.join("downloads", series.folder_name, "info.json")
|
||||
info_path = os.path.join("downloads", series.folder_name, "info.json")
|
||||
|
||||
# with open(info_path, 'wb') as f:
|
||||
# f.write(res.content)
|
||||
with open(info_path, 'wb') as f:
|
||||
f.write(res.content)
|
||||
|
||||
# # Make a folder for actors
|
||||
# actors_folder_path = os.path.join("downloads", series.folder_name, "actors")
|
||||
# os.makedirs(actors_folder_path)
|
||||
# Make a folder for actors
|
||||
actors_folder_path = os.path.join("downloads", series.folder_name, "actors")
|
||||
os.makedirs(actors_folder_path)
|
||||
|
||||
# # Download actors to actors.json
|
||||
# api_path = "https://api.thetvdb.com/series/{}/actors".format(series.id)
|
||||
# res = api_con.send_http_req(api_path)
|
||||
# Download actors to actors.json
|
||||
api_path = "https://api.thetvdb.com/series/{}/actors".format(series.id)
|
||||
res = api_con.send_http_req(api_path)
|
||||
|
||||
# actors_path = os.path.join("downloads", series.folder_name, "actors", "actors.json")
|
||||
actors_path = os.path.join("downloads", series.folder_name, "actors", "actors.json")
|
||||
|
||||
# with open(actors_path, 'wb') as f:
|
||||
# f.write(res.content)
|
||||
with open(actors_path, 'wb') as f:
|
||||
f.write(res.content)
|
||||
|
||||
# # Make folder for actor profile images
|
||||
# actors_profile_folder_path = os.path.join("downloads", series.folder_name, "actors", "profiles")
|
||||
# os.makedirs(actors_profile_folder_path)
|
||||
# Make folder for actor profile images
|
||||
actors_profile_folder_path = os.path.join("downloads", series.folder_name, "actors", "profiles")
|
||||
os.makedirs(actors_profile_folder_path)
|
||||
|
||||
# # Download each actor's profile picture and save it as their name
|
||||
# for actor in json.loads(res.content)["data"]:
|
||||
# name = create_file_name(actor["name"])
|
||||
# Download each actor's profile picture and save it as their name
|
||||
for actor in json.loads(res.content)["data"]:
|
||||
name = create_file_name(actor["name"])
|
||||
|
||||
# # Check if there is an image for the actor
|
||||
# if actor["image"] != "":
|
||||
# print("downloading " + actor["image"])
|
||||
# img_res = requests.get("https://www.thetvdb.com/banners/" + actor["image"])
|
||||
# with open(os.path.join(actors_profile_folder_path, name + '_' + str(actor["id"]) + ".jpg"), 'wb') as f:
|
||||
# f.write(img_res.content)
|
||||
# else:
|
||||
# # Use a default image if one does not exist on theTVDB.com
|
||||
# shutil.copyfile(os.path.join("resources", "default_person.jpg"), os.path.join(actors_profile_folder_path, name + '_' + str(actor["id"]) + ".jpg"))
|
||||
# Check if there is an image for the actor
|
||||
if actor["image"] != "":
|
||||
# print("downloading " + actor["image"])
|
||||
img_res = requests.get("https://www.thetvdb.com/banners/" + actor["image"])
|
||||
with open(os.path.join(actors_profile_folder_path, name + '_' + str(actor["id"]) + ".jpg"), 'wb') as f:
|
||||
f.write(img_res.content)
|
||||
else:
|
||||
# Use a default image if one does not exist on theTVDB.com
|
||||
shutil.copyfile(os.path.join("resources", "default_person.jpg"), os.path.join(actors_profile_folder_path, name + '_' + str(actor["id"]) + ".jpg"))
|
||||
|
||||
# # Make a folder for episodes
|
||||
# episodes_folder_path = os.path.join("downloads", series.folder_name, "episodes")
|
||||
# os.makedirs(episodes_folder_path)
|
||||
# Make a folder for episodes
|
||||
episodes_folder_path = os.path.join("downloads", series.folder_name, "episodes")
|
||||
os.makedirs(episodes_folder_path)
|
||||
|
||||
|
||||
# # Get number of seasons
|
||||
# api_path = "https://api.thetvdb.com/series/{}/episodes/summary".format(series.id)
|
||||
# res = api_con.send_http_req(api_path)
|
||||
# seasons = json.loads(res.content)["data"]["airedSeasons"]
|
||||
# Get number of seasons
|
||||
api_path = "https://api.thetvdb.com/series/{}/episodes/summary".format(series.id)
|
||||
res = api_con.send_http_req(api_path)
|
||||
seasons = json.loads(res.content)["data"]["airedSeasons"]
|
||||
|
||||
# # Create a folder for each season
|
||||
# for season in seasons:
|
||||
# season_folder_path = os.path.join(episodes_folder_path, "Season " + season)
|
||||
# os.makedirs(season_folder_path)
|
||||
# Create a folder for each season
|
||||
for season in seasons:
|
||||
season_folder_path = os.path.join(episodes_folder_path, "Season " + season)
|
||||
os.makedirs(season_folder_path)
|
||||
|
||||
# # Download episode info to episodes.json
|
||||
# api_path = "https://api.thetvdb.com/series/{}/episodes".format(series.id)
|
||||
# res = api_con.send_http_req(api_path)
|
||||
# with open(os.path.join(episodes_folder_path, "episodes.json"), 'wb') as f:
|
||||
# f.write(res.content)
|
||||
# Download episode info to episodes.json
|
||||
api_path = "https://api.thetvdb.com/series/{}/episodes".format(series.id)
|
||||
res = api_con.send_http_req(api_path)
|
||||
with open(os.path.join(episodes_folder_path, "episodes.json"), 'wb') as f:
|
||||
f.write(res.content)
|
||||
|
||||
# # Seperate episode data into individual json files for each episode and download episode thumbnails
|
||||
# for episode in json.loads(res.content)["data"]:
|
||||
# episode_path = os.path.join(episodes_folder_path, "Season " + str(episode["airedSeason"]), "Episode {} - {}".format(str(episode["airedEpisodeNumber"]), episode["episodeName"]))
|
||||
# img_res = requests.get("https://www.thetvdb.com/banners/" + episode["filename"])
|
||||
# with open(episode_path + ".json", 'w') as f:
|
||||
# f.write(json.dumps(episode))
|
||||
# with open(episode_path + ".jpg", 'wb') as f:
|
||||
# f.write(img_res.content)
|
||||
# Seperate episode data into individual json files for each episode and download episode thumbnails
|
||||
for episode in json.loads(res.content)["data"]:
|
||||
episode_path = os.path.join(episodes_folder_path, "Season " + str(episode["airedSeason"]), "Episode {} - {}".format(str(episode["airedEpisodeNumber"]), episode["episodeName"]))
|
||||
img_res = requests.get("https://www.thetvdb.com/banners/" + episode["filename"])
|
||||
with open(episode_path + ".json", 'w') as f:
|
||||
f.write(json.dumps(episode))
|
||||
with open(episode_path + ".jpg", 'wb') as f:
|
||||
f.write(img_res.content)
|
||||
|
||||
# Make a folder for images
|
||||
images_folder_path = os.path.join("downloads", series.folder_name, "images")
|
||||
os.makedirs(images_folder_path)
|
||||
|
||||
# Make a folder for each image type
|
||||
banners_folder_path = os.path.join(images_folder_path, "banners")
|
||||
os.makedirs(banners_folder_path)
|
||||
fanart_folder_path = os.path.join(images_folder_path, "fanart")
|
||||
os.makedirs(fanart_folder_path)
|
||||
posters_folder_path = os.path.join(images_folder_path, "posters")
|
||||
os.makedirs(posters_folder_path)
|
||||
|
||||
# The API doesn't like to send links to all of the images hosted on the website,
|
||||
# so the best option to get every images is to scrape the website directly
|
||||
|
||||
# Download banners
|
||||
banners_page = requests.get("{}/artwork/banners".format(series.url))
|
||||
banners_soup = BeautifulSoup(banners_page.content, "html.parser")
|
||||
|
||||
counter = 0
|
||||
for image in banners_soup.find_all("img", attrs={"class":"media-object img-responsive"}):
|
||||
image_res = requests.get(image["src"])
|
||||
with open(os.path.join(banners_folder_path, "{:03d}.jpg".format(counter)), 'wb') as f:
|
||||
f.write(image_res.content)
|
||||
counter+=1
|
||||
|
||||
# Download fanart
|
||||
fanart_page = requests.get("{}/artwork/fanart".format(series.url))
|
||||
fanart_soup = BeautifulSoup(fanart_page.content, "html.parser")
|
||||
|
||||
counter = 0
|
||||
for image in fanart_soup.find_all("img", attrs={"class":"media-object img-responsive"}):
|
||||
image_res = requests.get(image["src"])
|
||||
with open(os.path.join(fanart_folder_path, "{:03d}.jpg".format(counter)), 'wb') as f:
|
||||
f.write(image_res.content)
|
||||
counter+=1
|
||||
|
||||
# Download posters
|
||||
posters_page = requests.get("{}/artwork/poster".format(series.url))
|
||||
posters_soup = BeautifulSoup(posters_page.content, "html.parser")
|
||||
|
||||
counter = 0
|
||||
for image in posters_soup.find_all("img", attrs={"class":"media-object img-responsive"}):
|
||||
image_res = requests.get(image["src"])
|
||||
with open(os.path.join(posters_folder_path, "{:03d}.jpg".format(counter)), 'wb') as f:
|
||||
f.write(image_res.content)
|
||||
counter+=1
|
||||
|
||||
|
||||
|
||||
|
||||
Loading…
Reference in New Issue
Block a user