diff --git a/ADC_function.py b/ADC_function.py index bfa4bb8..2219219 100644 --- a/ADC_function.py +++ b/ADC_function.py @@ -470,15 +470,12 @@ def download_file_with_filename(url: str, filename: str, path: str) -> None: except: print(f"[-]Fatal error! Can not make folder '{path}'") os._exit(0) - proxies = configProxy.proxies() - headers = { - 'User-Agent': G_USER_AGENT} - r = requests.get(url, headers=headers, timeout=configProxy.timeout, proxies=proxies) + r = get_html(url=url, return_type='content') if r == '': print('[-]Movie Download Data not found!') return with open(os.path.join(path, filename), "wb") as code: - code.write(r.content) + code.write(r) return else: if not os.path.exists(path): @@ -487,14 +484,12 @@ def download_file_with_filename(url: str, filename: str, path: str) -> None: except: print(f"[-]Fatal error! Can not make folder '{path}'") os._exit(0) - headers = { - 'User-Agent': G_USER_AGENT} - r = requests.get(url, timeout=configProxy.timeout, headers=headers) + r = get_html(url=url, return_type='content') if r == '': print('[-]Movie Download Data not found!') return with open(os.path.join(path, filename), "wb") as code: - code.write(r.content) + code.write(r) return except requests.exceptions.RequestException: i += 1 @@ -522,15 +517,18 @@ def download_one_file(args) -> str: wrapped for map function """ - (url, save_path, json_data) = args - filebytes = get_html(url, return_type='content', json_headers=json_data['headers']) + (url, save_path, json_headers) = args + if json_headers != None: + filebytes = get_html(url, return_type='content', json_headers=json_headers['headers']) + else: + filebytes = get_html(url, return_type='content') if isinstance(filebytes, bytes) and len(filebytes): with save_path.open('wb') as fpbyte: if len(filebytes) == fpbyte.write(filebytes): return str(save_path) -def parallel_download_files(dn_list: typing.Iterable[typing.Sequence], parallel: int = 0, json_data=None): +def parallel_download_files(dn_list: typing.Iterable[typing.Sequence], parallel: int = 0, json_headers=None): """ download files in parallel 多线程下载文件 @@ -549,7 +547,7 @@ def parallel_download_files(dn_list: typing.Iterable[typing.Sequence], parallel: and fullpath and isinstance(fullpath, (str, Path)) and len(str(fullpath)): fullpath = Path(fullpath) fullpath.parent.mkdir(parents=True, exist_ok=True) - mp_args.append((url, fullpath, json_data)) + mp_args.append((url, fullpath, json_headers)) if not len(mp_args): return [] if not isinstance(parallel, int) or parallel not in range(1, 200): diff --git a/Movie_Data_Capture.py b/Movie_Data_Capture.py index 2b1b37a..10888a1 100644 --- a/Movie_Data_Capture.py +++ b/Movie_Data_Capture.py @@ -591,6 +591,7 @@ def main(args: tuple) -> Path: print("[!] " + "Mapping Table Download FAILED".center(47)) print("[!] " + "无法连接github".center(47)) print("[!] " + "请过几小时再试试".center(47)) + print("[!]", e) print("[-] " + "------ AUTO EXIT AFTER 30s !!! ------ ".center(47)) time.sleep(30) os._exit(-1) @@ -677,7 +678,7 @@ def period(delta, pattern): if __name__ == '__main__': - version = '6.2.1' + version = '6.2.2' urllib3.disable_warnings() # Ignore http proxy warning app_start = time.time() diff --git a/core.py b/core.py index 29df3a7..98c538d 100644 --- a/core.py +++ b/core.py @@ -72,11 +72,14 @@ def get_info(json_data): # 返回json里的数据 return title, studio, year, outline, runtime, director, actor_photo, release, number, cover, trailer, website, series, label -def small_cover_check(path, filename, cover_small, movie_path, json_data=None): +def small_cover_check(path, filename, cover_small, movie_path, json_headers=None): full_filepath = Path(path) / filename if config.getInstance().download_only_missing_images() and not file_not_exist_or_empty(str(full_filepath)): return - download_file_with_filename(cover_small, filename, path, movie_path, json_data) + if json_headers != None: + download_file_with_filename(cover_small, filename, path, movie_path, json_headers['headers']) + else: + download_file_with_filename(cover_small, filename, path, movie_path) print('[+]Image Downloaded! ' + full_filepath.name) @@ -120,57 +123,28 @@ def download_file_with_filename(url, filename, path, filepath, json_headers=None for i in range(configProxy.retry): try: - if configProxy.enable: - if not os.path.exists(path): - try: - os.makedirs(path) - except: - print(f"[-]Fatal error! Can not make folder '{path}'") - os._exit(0) - proxies = configProxy.proxies() - headers = {'User-Agent': G_USER_AGENT} - if json_headers != None: - headers.update(json_headers) - r = requests.get(url, headers=headers, timeout=configProxy.timeout, proxies=proxies) - if r == '': - print('[-]Movie Download Data not found!') - return - with open(os.path.join(path, filename), "wb") as code: - code.write(r.content) + if not os.path.exists(path): + try: + os.makedirs(path) + except: + print(f"[-]Fatal error! Can not make folder '{path}'") + os._exit(0) + r = get_html(url=url,return_type='content',json_headers=json_headers) + if r == '': + print('[-]Movie Download Data not found!') return - else: - if not os.path.exists(path): - try: - os.makedirs(path) - except: - print(f"[-]Fatal error! Can not make folder '{path}'") - os._exit(0) - headers = {'User-Agent': G_USER_AGENT} - if json_headers != None: - headers.update(json_headers) - r = requests.get(url, timeout=configProxy.timeout, headers=headers) - if r == '': - print('[-]Movie Download Data not found!') - return - with open(os.path.join(path, filename), "wb") as code: - code.write(r.content) - return - except requests.exceptions.RequestException: - i += 1 - print('[-]Image Download : Connect retry ' + str(i) + '/' + str(configProxy.retry)) - except requests.exceptions.ConnectionError: - i += 1 - print('[-]Image Download : Connect retry ' + str(i) + '/' + str(configProxy.retry)) + with open(os.path.join(path, filename), "wb") as code: + code.write(r) + return except requests.exceptions.ProxyError: i += 1 - print('[-]Image Download : Connect retry ' + str(i) + '/' + str(configProxy.retry)) - except requests.exceptions.ConnectTimeout: - i += 1 - print('[-]Image Download : Connect retry ' + str(i) + '/' + str(configProxy.retry)) - except IOError: - print(f"[-]Create Directory '{path}' failed!") - moveFailedFolder(filepath) - return + print('[-]Image Download : Proxy error ' + str(i) + '/' + str(configProxy.retry)) + # except IOError: + # print(f"[-]Create Directory '{path}' failed!") + # moveFailedFolder(filepath) + # return + except Exception as e: + print('[-]Image Download :Error',e) print('[-]Connect Failed! Please check your Proxy or Network!') moveFailedFolder(filepath) return @@ -303,12 +277,12 @@ def image_ext(url): # 封面是否下载成功,否则移动到failed -def image_download(cover, fanart_path, thumb_path, path, filepath, json_data): +def image_download(cover, fanart_path, thumb_path, path, filepath, json_headers=None): full_filepath = os.path.join(path, fanart_path) if config.getInstance().download_only_missing_images() and not file_not_exist_or_empty(full_filepath): return - if "headers" in json_data: - if download_file_with_filename(cover, fanart_path, path, filepath, json_data['headers']) == 'failed': + if json_headers != None: + if download_file_with_filename(cover, fanart_path, path, filepath, json_headers['headers']) == 'failed': moveFailedFolder(filepath) return else: @@ -320,8 +294,8 @@ def image_download(cover, fanart_path, thumb_path, path, filepath, json_data): for i in range(configProxy.retry): if file_not_exist_or_empty(full_filepath): print('[!]Image Download Failed! Trying again. [{}/3]', i + 1) - if "headers" in json_data: - download_file_with_filename(cover, fanart_path, path, filepath, json_data['headers']) + if json_headers != None: + download_file_with_filename(cover, fanart_path, path, filepath, json_headers['headers']) else: download_file_with_filename(cover, fanart_path, path, filepath) continue @@ -364,8 +338,10 @@ def print_files(path, leak_word, c_word, naming_rule, part, cn_sub, json_data, f print(" ", file=code) print(" JP-18+", file=code) print(" JP-18+", file=code) - print(" ", file=code) - print(" ", file=code) + try: + print(" " + series + "", file=code) + except: + print(" ", file=code) print(" " + studio + "", file=code) print(" " + year + "", file=code) print(" ", file=code) @@ -399,7 +375,7 @@ def print_files(path, leak_word, c_word, naming_rule, part, cn_sub, json_data, f try: for i in tag: print(" " + i + "", file=code) - print(" " + series + "", file=code) + # print(" " + series + "", file=code) except: aaaaa = '' if cn_sub == '1': @@ -413,7 +389,7 @@ def print_files(path, leak_word, c_word, naming_rule, part, cn_sub, json_data, f try: for i in tag: print(" " + i + "", file=code) - print(" " + series + "", file=code) + # print(" " + series + "", file=code) except: aaaaaaaa = '' print(" " + number + "", file=code) @@ -864,10 +840,16 @@ def core_main(movie_path, number_th, oCC): # 检查小封面, 如果image cut为3,则下载小封面 if imagecut == 3: - small_cover_check(path, poster_path, json_data.get('cover_small'), movie_path, json_data) + if 'headers' in json_data: + small_cover_check(path, poster_path, json_data.get('cover_small'), movie_path, json_data) + else: + small_cover_check(path, poster_path, json_data.get('cover_small'), movie_path) # creatFolder会返回番号路径 - image_download( cover, fanart_path,thumb_path, path, movie_path, json_data) + if 'headers' in json_data: + image_download(cover, fanart_path, thumb_path, path, movie_path, json_data) + else: + image_download(cover, fanart_path, thumb_path, path, movie_path) if not multi_part or part.lower() == '-cd1': try: @@ -877,7 +859,10 @@ def core_main(movie_path, number_th, oCC): # 下载剧照 data, path, filepath if conf.is_extrafanart() and json_data.get('extrafanart'): - extrafanart_download(json_data.get('extrafanart'), path, number, movie_path, json_data) + if 'headers' in json_data: + extrafanart_download(json_data.get('extrafanart'), path, number, movie_path, json_data) + else: + extrafanart_download(json_data.get('extrafanart'), path, number, movie_path) # 下载演员头像 KODI .actors 目录位置 if conf.download_actor_photo_for_kodi(): @@ -918,10 +903,16 @@ def core_main(movie_path, number_th, oCC): # 检查小封面, 如果image cut为3,则下载小封面 if imagecut == 3: - small_cover_check(path, poster_path, json_data.get('cover_small'), movie_path, json_data) + if 'headers' in json_data: + small_cover_check(path, poster_path, json_data.get('cover_small'), movie_path, json_data) + else: + small_cover_check(path, poster_path, json_data.get('cover_small'), movie_path) # creatFolder会返回番号路径 - image_download( cover, fanart_path, thumb_path, path, movie_path) + if 'headers' in json_data: + image_download(cover, fanart_path, thumb_path, path, movie_path, json_data) + else: + image_download(cover, fanart_path, thumb_path, path, movie_path) if not multi_part or part.lower() == '-cd1': try: @@ -931,7 +922,10 @@ def core_main(movie_path, number_th, oCC): # 下载剧照 data, path, filepath if conf.is_extrafanart() and json_data.get('extrafanart'): - extrafanart_download(json_data.get('extrafanart'), path, number, movie_path, json_data) + if 'headers' in json_data: + extrafanart_download(json_data.get('extrafanart'), path, number, movie_path, json_data) + else: + extrafanart_download(json_data.get('extrafanart'), path, number, movie_path) # 下载演员头像 KODI .actors 目录位置 if conf.download_actor_photo_for_kodi(): diff --git a/number_parser.py b/number_parser.py index e3e7245..cbe94fa 100755 --- a/number_parser.py +++ b/number_parser.py @@ -79,7 +79,33 @@ def get_number(debug: bool, file_path: str) -> str: if debug: print(f'[-]Number Parser exception: {e} [{file_path}]') return None + +# modou提取number +def md(filename): + m = re.search(r'(md[a-z]{0,2}-?)(\d{2,})(-ep\d*)*', filename, re.I) + return f'{m.group(1).replace("-","").upper()}{m.group(2).zfill(4)}{m.group(3) or ""}' +def mmz(filename): + m = re.search(r'(mmz-?)(\d{2,})(-ep\d*)*', filename, re.I) + return f'{m.group(1).replace("-","").upper()}{m.group(2).zfill(3)}{m.group(3) or ""}' + +def msd(filename): + m = re.search(r'(msd-?)(\d{2,})(-ep\d*)*', filename, re.I) + return f'{m.group(1).replace("-","").upper()}{m.group(2).zfill(3)}{m.group(3) or ""}' + +def mky(filename): + m = re.search(r'(mky-[a-z]{2,2}-?)(\d{2,})(-ep\d*)*', filename, re.I) + return f'{m.group(1).replace("-","").upper()}{m.group(2).zfill(3)}{m.group(3) or ""}' + +def yk(filename): + m = re.search(r'(yk-?)(\d{2,})(-ep\d*)*', filename, re.I) + return f'{m.group(1).replace("-","").upper()}{m.group(2).zfill(3)}{m.group(3) or ""}' + +def pm(filename): + m = re.search(r'(pm[a-z]?-?)(\d{2,})(-ep\d*)*', filename, re.I) + return f'{m.group(1).replace("-","").upper()}{m.group(2).zfill(3)}{m.group(3) or ""}' + + # 按javdb数据源的命名规范提取number G_TAKE_NUM_RULES = { @@ -90,7 +116,13 @@ G_TAKE_NUM_RULES = { 'x-art': lambda x: str(re.search(r'x-art\.\d{2}\.\d{2}\.\d{2}', x, re.I).group()), 'xxx-av': lambda x: ''.join(['xxx-av-', re.findall(r'xxx-av[^\d]*(\d{3,5})[^\d]*', x, re.I)[0]]), 'heydouga': lambda x: 'heydouga-' + '-'.join(re.findall(r'(\d{4})[\-_](\d{3,4})[^\d]*', x, re.I)[0]), - 'heyzo': lambda x: 'HEYZO-' + re.findall(r'heyzo[^\d]*(\d{4})', x, re.I)[0] + 'heyzo': lambda x: 'HEYZO-' + re.findall(r'heyzo[^\d]*(\d{4})', x, re.I)[0], + r'\bmd[a-z]{0,2}-\d{2,}': md, + r'\bmmz-\d{2,}':mmz, + r'\bmsd-\d{2,}':msd, + r'\bmky-[a-z]{2,2}-\d{2,}':mky, + r'\byk-\d{2,3}': yk, + r'\bpm[a-z]?-?\d{2,}':pm } @@ -176,6 +208,9 @@ if __name__ == "__main__": "rctd-461CH-CD2.mp4", # ch后可加CDn "rctd-461-Cd3-C.mp4", # CDn后可加-C "rctd-461-C-cD4.mp4", # cD1 Cd1 cd1 CD1 最终生成.nfo时统一为大写CD1 + "MD-123.ts", + "MDSR-0001-ep2.ts", + "MKY-NS-001.mp4" )