Add custom headers in Image Download func

This commit is contained in:
yoshiko2
2022-05-21 23:42:30 +08:00
parent 47110c567c
commit daf431b9f5
2 changed files with 42 additions and 30 deletions

View File

@@ -29,7 +29,7 @@ def getXpathSingle(htmlcode, xpath):
G_USER_AGENT = r'Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/100.0.4896.133 Safari/537.36'
def get_html(url, cookies: dict = None, ua: str = None, return_type: str = None, encoding: str = None):
def get_html(url, cookies: dict = None, ua: str = None, return_type: str = None, encoding: str = None, json_headers = None):
"""
网页请求核心函数
"""
@@ -38,6 +38,8 @@ def get_html(url, cookies: dict = None, ua: str = None, return_type: str = None,
errors = ""
headers = {"User-Agent": ua or G_USER_AGENT} # noqa
if json_headers != None:
headers.update(json_headers)
for i in range(configProxy.retry):
try:
@@ -518,15 +520,15 @@ def download_one_file(args) -> str:
wrapped for map function
"""
(url, save_path) = args
filebytes = get_html(url, return_type='content')
(url, save_path, json_data) = args
filebytes = get_html(url, return_type='content', json_headers=json_data['headers'])
if isinstance(filebytes, bytes) and len(filebytes):
with save_path.open('wb') as fpbyte:
if len(filebytes) == fpbyte.write(filebytes):
return str(save_path)
def parallel_download_files(dn_list: typing.Iterable[typing.Sequence], parallel: int = 0):
def parallel_download_files(dn_list: typing.Iterable[typing.Sequence], parallel: int = 0, json_data=None):
"""
download files in parallel 多线程下载文件
@@ -545,7 +547,7 @@ def parallel_download_files(dn_list: typing.Iterable[typing.Sequence], parallel:
and fullpath and isinstance(fullpath, (str, Path)) and len(str(fullpath)):
fullpath = Path(fullpath)
fullpath.parent.mkdir(parents=True, exist_ok=True)
mp_args.append((url, fullpath))
mp_args.append((url, fullpath, json_data))
if not len(mp_args):
return []
if not isinstance(parallel, int) or parallel not in range(1, 200):

60
core.py
View File

@@ -71,11 +71,11 @@ def get_info(json_data): # 返回json里的数据
return title, studio, year, outline, runtime, director, actor_photo, release, number, cover, trailer, website, series, label
def small_cover_check(path, filename, cover_small, movie_path):
def small_cover_check(path, filename, cover_small, movie_path, json_data=None):
full_filepath = Path(path) / filename
if config.getInstance().download_only_missing_images() and not file_not_exist_or_empty(str(full_filepath)):
return
download_file_with_filename(cover_small, filename, path, movie_path)
download_file_with_filename(cover_small, filename, path, movie_path, json_data)
print('[+]Image Downloaded! ' + full_filepath.name)
@@ -113,7 +113,7 @@ def create_folder(json_data): # 创建文件夹
# =====================资源下载部分===========================
# path = examle:photo , video.in the Project Folder!
def download_file_with_filename(url, filename, path, filepath):
def download_file_with_filename(url, filename, path, filepath, json_headers=None):
conf = config.getInstance()
configProxy = conf.proxy()
@@ -127,8 +127,9 @@ def download_file_with_filename(url, filename, path, filepath):
print(f"[-]Fatal error! Can not make folder '{path}'")
os._exit(0)
proxies = configProxy.proxies()
headers = {
'User-Agent': G_USER_AGENT}
headers = {'User-Agent': G_USER_AGENT}
if json_headers != None:
headers.update(json_headers)
r = requests.get(url, headers=headers, timeout=configProxy.timeout, proxies=proxies)
if r == '':
print('[-]Movie Download Data not found!')
@@ -143,8 +144,9 @@ def download_file_with_filename(url, filename, path, filepath):
except:
print(f"[-]Fatal error! Can not make folder '{path}'")
os._exit(0)
headers = {
'User-Agent': G_USER_AGENT}
headers = {'User-Agent': G_USER_AGENT}
if json_headers != None:
headers.update(json_headers)
r = requests.get(url, timeout=configProxy.timeout, headers=headers)
if r == '':
print('[-]Movie Download Data not found!')
@@ -226,13 +228,13 @@ def actor_photo_download(actors, save_dir, number):
# 剧照下载成功否则移动到failed
def extrafanart_download(data, path, number, filepath):
def extrafanart_download(data, path, number, filepath, json_data=None):
if config.getInstance().extrafanart_thread_pool_download():
return extrafanart_download_threadpool(data, path, number)
extrafanart_download_one_by_one(data, path, filepath)
return extrafanart_download_threadpool(data, path, number, json_data)
extrafanart_download_one_by_one(data, path, filepath, json_data)
def extrafanart_download_one_by_one(data, path, filepath):
def extrafanart_download_one_by_one(data, path, filepath, json_data=None):
tm_start = time.perf_counter()
j = 1
conf = config.getInstance()
@@ -244,13 +246,13 @@ def extrafanart_download_one_by_one(data, path, filepath):
jpg_fullpath = os.path.join(path, jpg_filename)
if download_only_missing_images and not file_not_exist_or_empty(jpg_fullpath):
continue
if download_file_with_filename(url, jpg_filename, path, filepath) == 'failed':
if download_file_with_filename(url, jpg_filename, path, filepath, json_data) == 'failed':
moveFailedFolder(filepath)
return
for i in range(configProxy.retry):
if file_not_exist_or_empty(jpg_fullpath):
print('[!]Image Download Failed! Trying again. [{}/3]', i + 1)
download_file_with_filename(url, jpg_filename, path, filepath)
download_file_with_filename(url, jpg_filename, path, filepath, json_data)
continue
else:
break
@@ -262,7 +264,7 @@ def extrafanart_download_one_by_one(data, path, filepath):
print(f'[!]Extrafanart download one by one mode runtime {time.perf_counter() - tm_start:.3f}s')
def extrafanart_download_threadpool(url_list, save_dir, number):
def extrafanart_download_threadpool(url_list, save_dir, number,json_data=None):
tm_start = time.perf_counter()
conf = config.getInstance()
extrafanart_dir = Path(save_dir) / conf.get_extrafanart()
@@ -278,7 +280,7 @@ def extrafanart_download_threadpool(url_list, save_dir, number):
parallel = min(len(dn_list), conf.extrafanart_thread_pool_download())
if parallel > 100:
print('[!]Warrning: Parallel download thread too large may cause website ban IP!')
result = parallel_download_files(dn_list, parallel)
result = parallel_download_files(dn_list, parallel, json_data)
failed = 0
for i, r in enumerate(result, start=1):
if not r:
@@ -300,19 +302,27 @@ def image_ext(url):
# 封面是否下载成功否则移动到failed
def image_download(cover, fanart_path, thumb_path, path, filepath):
def image_download(cover, fanart_path, thumb_path, path, filepath, json_data):
full_filepath = os.path.join(path, fanart_path)
if config.getInstance().download_only_missing_images() and not file_not_exist_or_empty(full_filepath):
return
if download_file_with_filename(cover, fanart_path, path, filepath) == 'failed':
moveFailedFolder(filepath)
return
if "headers" in json_data:
if download_file_with_filename(cover, fanart_path, path, filepath, json_data['headers']) == 'failed':
moveFailedFolder(filepath)
return
else:
if download_file_with_filename(cover, fanart_path, path, filepath) == 'failed':
moveFailedFolder(filepath)
return
configProxy = config.getInstance().proxy()
for i in range(configProxy.retry):
if file_not_exist_or_empty(full_filepath):
print('[!]Image Download Failed! Trying again. [{}/3]', i + 1)
download_file_with_filename(cover, fanart_path, path, filepath)
if "headers" in json_data:
download_file_with_filename(cover, fanart_path, path, filepath, json_data['headers'])
else:
download_file_with_filename(cover, fanart_path, path, filepath)
continue
else:
break
@@ -853,10 +863,10 @@ def core_main(movie_path, number_th, oCC):
# 检查小封面, 如果image cut为3则下载小封面
if imagecut == 3:
small_cover_check(path, poster_path, json_data.get('cover_small'), movie_path)
small_cover_check(path, poster_path, json_data.get('cover_small'), movie_path, json_data)
# creatFolder会返回番号路径
image_download( cover, fanart_path,thumb_path, path, movie_path)
image_download( cover, fanart_path,thumb_path, path, movie_path, json_data)
if not multi_part or part.lower() == '-cd1':
try:
@@ -866,7 +876,7 @@ def core_main(movie_path, number_th, oCC):
# 下载剧照 data, path, filepath
if conf.is_extrafanart() and json_data.get('extrafanart'):
extrafanart_download(json_data.get('extrafanart'), path, number, movie_path)
extrafanart_download(json_data.get('extrafanart'), path, number, movie_path, json_data)
# 下载演员头像 KODI .actors 目录位置
if conf.download_actor_photo_for_kodi():
@@ -907,7 +917,7 @@ def core_main(movie_path, number_th, oCC):
# 检查小封面, 如果image cut为3则下载小封面
if imagecut == 3:
small_cover_check(path, poster_path, json_data.get('cover_small'), movie_path)
small_cover_check(path, poster_path, json_data.get('cover_small'), movie_path, json_data)
# creatFolder会返回番号路径
image_download( cover, fanart_path, thumb_path, path, movie_path)
@@ -920,7 +930,7 @@ def core_main(movie_path, number_th, oCC):
# 下载剧照 data, path, filepath
if conf.is_extrafanart() and json_data.get('extrafanart'):
extrafanart_download(json_data.get('extrafanart'), path, number, movie_path)
extrafanart_download(json_data.get('extrafanart'), path, number, movie_path, json_data)
# 下载演员头像 KODI .actors 目录位置
if conf.download_actor_photo_for_kodi():