ADC_function.py:换装getInstance(),load_cookies()改用pathlib
This commit is contained in:
@@ -2,7 +2,7 @@ from os import replace
|
|||||||
import requests
|
import requests
|
||||||
import hashlib
|
import hashlib
|
||||||
from pathlib import Path
|
from pathlib import Path
|
||||||
import random
|
#import secrets
|
||||||
import os.path
|
import os.path
|
||||||
import uuid
|
import uuid
|
||||||
import json
|
import json
|
||||||
@@ -24,8 +24,8 @@ G_USER_AGENT = r'Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36 (K
|
|||||||
|
|
||||||
# 网页请求核心
|
# 网页请求核心
|
||||||
def get_html(url, cookies: dict = None, ua: str = None, return_type: str = None):
|
def get_html(url, cookies: dict = None, ua: str = None, return_type: str = None):
|
||||||
verify = config.Config().cacert_file()
|
verify = config.getInstance().cacert_file()
|
||||||
configProxy = config.Config().proxy()
|
configProxy = config.getInstance().proxy()
|
||||||
errors = ""
|
errors = ""
|
||||||
|
|
||||||
if ua is None:
|
if ua is None:
|
||||||
@@ -61,7 +61,7 @@ def get_html(url, cookies: dict = None, ua: str = None, return_type: str = None)
|
|||||||
|
|
||||||
|
|
||||||
def post_html(url: str, query: dict, headers: dict = None) -> requests.Response:
|
def post_html(url: str, query: dict, headers: dict = None) -> requests.Response:
|
||||||
configProxy = config.Config().proxy()
|
configProxy = config.getInstance().proxy()
|
||||||
errors = ""
|
errors = ""
|
||||||
headers_ua = {"User-Agent": G_USER_AGENT}
|
headers_ua = {"User-Agent": G_USER_AGENT}
|
||||||
if headers is None:
|
if headers is None:
|
||||||
@@ -86,7 +86,7 @@ def post_html(url: str, query: dict, headers: dict = None) -> requests.Response:
|
|||||||
|
|
||||||
def get_html_by_browser(url, cookies: dict = None, ua: str = None, return_type: str = None):
|
def get_html_by_browser(url, cookies: dict = None, ua: str = None, return_type: str = None):
|
||||||
browser = mechanicalsoup.StatefulBrowser(user_agent=G_USER_AGENT if ua is None else ua)
|
browser = mechanicalsoup.StatefulBrowser(user_agent=G_USER_AGENT if ua is None else ua)
|
||||||
configProxy = config.Config().proxy()
|
configProxy = config.getInstance().proxy()
|
||||||
if configProxy.enable:
|
if configProxy.enable:
|
||||||
browser.session.proxies = configProxy.proxies()
|
browser.session.proxies = configProxy.proxies()
|
||||||
result = browser.open(url)
|
result = browser.open(url)
|
||||||
@@ -107,7 +107,7 @@ def get_html_by_form(url, form_name: str = None, fields: dict = None, cookies: d
|
|||||||
browser = mechanicalsoup.StatefulBrowser(user_agent=G_USER_AGENT if ua is None else ua)
|
browser = mechanicalsoup.StatefulBrowser(user_agent=G_USER_AGENT if ua is None else ua)
|
||||||
if isinstance(cookies, dict):
|
if isinstance(cookies, dict):
|
||||||
requests.utils.add_dict_to_cookiejar(browser.session.cookies, cookies)
|
requests.utils.add_dict_to_cookiejar(browser.session.cookies, cookies)
|
||||||
configProxy = config.Config().proxy()
|
configProxy = config.getInstance().proxy()
|
||||||
if configProxy.enable:
|
if configProxy.enable:
|
||||||
browser.session.proxies = configProxy.proxies()
|
browser.session.proxies = configProxy.proxies()
|
||||||
result = browser.open(url)
|
result = browser.open(url)
|
||||||
@@ -131,7 +131,7 @@ def get_html_by_form(url, form_name: str = None, fields: dict = None, cookies: d
|
|||||||
|
|
||||||
# def get_javlib_cookie() -> [dict, str]:
|
# def get_javlib_cookie() -> [dict, str]:
|
||||||
# import cloudscraper
|
# import cloudscraper
|
||||||
# switch, proxy, timeout, retry_count, proxytype = config.Config().proxy()
|
# switch, proxy, timeout, retry_count, proxytype = config.getInstance().proxy()
|
||||||
# proxies = get_proxy(proxy, proxytype)
|
# proxies = get_proxy(proxy, proxytype)
|
||||||
#
|
#
|
||||||
# raw_cookie = {}
|
# raw_cookie = {}
|
||||||
@@ -158,7 +158,7 @@ def get_html_by_form(url, form_name: str = None, fields: dict = None, cookies: d
|
|||||||
|
|
||||||
|
|
||||||
def translateTag_to_sc(tag):
|
def translateTag_to_sc(tag):
|
||||||
tranlate_to_sc = config.Config().transalte_to_sc()
|
tranlate_to_sc = config.getInstance().transalte_to_sc()
|
||||||
if tranlate_to_sc:
|
if tranlate_to_sc:
|
||||||
dict_gen = {'中文字幕': '中文字幕',
|
dict_gen = {'中文字幕': '中文字幕',
|
||||||
'高清': 'XXXX', '字幕': 'XXXX', '推薦作品': '推荐作品', '通姦': '通奸', '淋浴': '淋浴', '舌頭': '舌头',
|
'高清': 'XXXX', '字幕': 'XXXX', '推薦作品': '推荐作品', '通姦': '通奸', '淋浴': '淋浴', '舌頭': '舌头',
|
||||||
@@ -506,7 +506,7 @@ def translate(
|
|||||||
):
|
):
|
||||||
trans_result = ""
|
trans_result = ""
|
||||||
if engine == "google-free":
|
if engine == "google-free":
|
||||||
gsite = config.Config().get_translate_service_site()
|
gsite = config.getInstance().get_translate_service_site()
|
||||||
if not re.match('^translate\.google\.(com|com\.\w{2}|\w{2})$', gsite):
|
if not re.match('^translate\.google\.(com|com\.\w{2}|\w{2})$', gsite):
|
||||||
gsite = 'translate.google.cn'
|
gsite = 'translate.google.cn'
|
||||||
url = (
|
url = (
|
||||||
@@ -521,7 +521,7 @@ f"https://{gsite}/translate_a/single?client=gtx&dt=t&dj=1&ie=UTF-8&sl=auto&tl={t
|
|||||||
trans_result = trans_result.join(translate_list)
|
trans_result = trans_result.join(translate_list)
|
||||||
# elif engine == "baidu":
|
# elif engine == "baidu":
|
||||||
# url = "https://fanyi-api.baidu.com/api/trans/vip/translate"
|
# url = "https://fanyi-api.baidu.com/api/trans/vip/translate"
|
||||||
# salt = random.randint(1, 1435660288)
|
# salt = secrets.randbelow(1435660287) + 1 # random.randint(1, 1435660288)
|
||||||
# sign = app_id + src + str(salt) + key
|
# sign = app_id + src + str(salt) + key
|
||||||
# sign = hashlib.md5(sign.encode()).hexdigest()
|
# sign = hashlib.md5(sign.encode()).hexdigest()
|
||||||
# url += (
|
# url += (
|
||||||
@@ -564,7 +564,7 @@ f"https://{gsite}/translate_a/single?client=gtx&dt=t&dj=1&ie=UTF-8&sl=auto&tl={t
|
|||||||
def is_uncensored(number):
|
def is_uncensored(number):
|
||||||
if re.match('^\d{4,}', number) or re.match('n\d{4}', number) or 'HEYZO' in number.upper():
|
if re.match('^\d{4,}', number) or re.match('n\d{4}', number) or 'HEYZO' in number.upper():
|
||||||
return True
|
return True
|
||||||
configs = config.Config().get_uncensored()
|
configs = config.getInstance().get_uncensored()
|
||||||
prefix_list = str(configs).split(',')
|
prefix_list = str(configs).split(',')
|
||||||
for pre in prefix_list:
|
for pre in prefix_list:
|
||||||
if pre.upper() in number.upper():
|
if pre.upper() in number.upper():
|
||||||
@@ -593,20 +593,20 @@ def load_cookies(filename):
|
|||||||
filename = os.path.basename(filename)
|
filename = os.path.basename(filename)
|
||||||
if not len(filename):
|
if not len(filename):
|
||||||
return None, None
|
return None, None
|
||||||
path_search_order = [
|
path_search_order = (
|
||||||
f"./{filename}",
|
Path.cwd() / filename,
|
||||||
os.path.join(Path.home(), filename),
|
Path.home() / filename,
|
||||||
os.path.join(Path.home(), f".avdc/{filename}"),
|
Path.home() / f".avdc/{filename}",
|
||||||
os.path.join(Path.home(), f".local/share/avdc/{filename}")
|
Path.home() / f".local/share/avdc/{filename}"
|
||||||
]
|
)
|
||||||
cookies_filename = None
|
cookies_filename = None
|
||||||
for p in path_search_order:
|
|
||||||
if os.path.exists(p):
|
|
||||||
cookies_filename = os.path.abspath(p)
|
|
||||||
break
|
|
||||||
if not cookies_filename:
|
|
||||||
return None, None
|
|
||||||
try:
|
try:
|
||||||
|
for p in path_search_order:
|
||||||
|
if p.is_file():
|
||||||
|
cookies_filename = str(p.resolve())
|
||||||
|
break
|
||||||
|
if not cookies_filename:
|
||||||
|
return None, None
|
||||||
return json.load(open(cookies_filename)), cookies_filename
|
return json.load(open(cookies_filename)), cookies_filename
|
||||||
except:
|
except:
|
||||||
return None, None
|
return None, None
|
||||||
|
|||||||
Reference in New Issue
Block a user