mirror of
https://github.com/baiqwerdvd/ArknightsUID.git
synced 2025-05-05 03:23:45 +08:00
✨支持群友 @谦虚 的资源站
This commit is contained in:
parent
b7ac18cebf
commit
a1e74d1129
@ -6,26 +6,28 @@ from gsuid_core.logger import logger
|
|||||||
from gsuid_core.models import Event
|
from gsuid_core.models import Event
|
||||||
from gsuid_core.sv import SV
|
from gsuid_core.sv import SV
|
||||||
|
|
||||||
from ..utils.resource.download_all_resource import download_all_resource
|
from ..utils.resource.download_from_cos import check_use
|
||||||
from .memoryStore import store
|
|
||||||
from .constants import Excel
|
from .constants import Excel
|
||||||
|
from .memoryStore import store
|
||||||
|
|
||||||
sv_download_config = SV("下载资源", pm=2)
|
sv_download_config = SV("下载资源", pm=2)
|
||||||
|
|
||||||
|
|
||||||
@sv_download_config.on_fullmatch(("下载全部资源")) # noqa: UP034
|
@sv_download_config.on_fullmatch(("ark下载全部资源")) # noqa: UP034
|
||||||
async def send_download_resource_msg(bot: Bot, ev: Event):
|
async def send_download_resource_msg(bot: Bot, ev: Event):
|
||||||
await bot.send("正在开始下载~可能需要较久的时间!")
|
await bot.send("正在开始下载~可能需要较久的时间!")
|
||||||
im = await download_all_resource()
|
im = await check_use()
|
||||||
await bot.send(im)
|
await bot.send(im)
|
||||||
|
|
||||||
|
|
||||||
async def startup():
|
async def startup():
|
||||||
logger.info("[资源文件下载] 正在检查与下载缺失的资源文件, 可能需要较长时间, 请稍等")
|
logger.info("[资源文件下载] 正在检查与下载缺失的资源文件, 可能需要较长时间, 请稍等")
|
||||||
await download_all_resource()
|
await check_use()
|
||||||
logger.info("[资源文件下载] 检查完毕, 正在加载 gamedata")
|
logger.info("[资源文件下载] 检查完毕, 正在加载 gamedata")
|
||||||
|
|
||||||
for file_path in Path(get_res_path(["ArknightsUID", "resource", "gamedata"])).rglob("*.json"):
|
for file_path in Path(
|
||||||
|
get_res_path(["ArknightsUID", "resource", "gamedata"])
|
||||||
|
).rglob("*.json"):
|
||||||
await store.get_file(file_path)
|
await store.get_file(file_path)
|
||||||
|
|
||||||
await Excel.preload_table()
|
await Excel.preload_table()
|
||||||
|
@ -2,12 +2,16 @@ import asyncio
|
|||||||
from pathlib import Path
|
from pathlib import Path
|
||||||
from typing import Dict, List, Tuple, Union
|
from typing import Dict, List, Tuple, Union
|
||||||
|
|
||||||
|
from aiohttp import ClientTimeout, TCPConnector
|
||||||
from aiohttp.client import ClientSession
|
from aiohttp.client import ClientSession
|
||||||
|
from bs4 import BeautifulSoup
|
||||||
from gsuid_core.logger import logger
|
from gsuid_core.logger import logger
|
||||||
|
from gsuid_core.utils.download_resource.download_core import check_url
|
||||||
|
from gsuid_core.utils.download_resource.download_file import download
|
||||||
from msgspec import json as msgjson
|
from msgspec import json as msgjson
|
||||||
|
|
||||||
from .download_url import download_file
|
from .download_url import download_file
|
||||||
from .RESOURCE_PATH import RESOURCE_PATH
|
from .RESOURCE_PATH import GAMEDATA_PATH, RESOURCE_PATH
|
||||||
|
|
||||||
MAX_DOWNLOAD = 10
|
MAX_DOWNLOAD = 10
|
||||||
|
|
||||||
@ -20,6 +24,66 @@ with Path.open(
|
|||||||
)
|
)
|
||||||
|
|
||||||
|
|
||||||
|
async def find_fastest_url(urls: Dict[str, str]):
|
||||||
|
tasks = []
|
||||||
|
for tag in urls:
|
||||||
|
tasks.append(asyncio.create_task(check_url(tag, urls[tag])))
|
||||||
|
|
||||||
|
results: list[tuple[str, str, float]] = await asyncio.gather(
|
||||||
|
*tasks, return_exceptions=True
|
||||||
|
)
|
||||||
|
fastest_tag = ''
|
||||||
|
fastest_url = None
|
||||||
|
fastest_time = float('inf')
|
||||||
|
|
||||||
|
for result in results:
|
||||||
|
if isinstance(result, Exception):
|
||||||
|
continue
|
||||||
|
tag, url, elapsed_time = result
|
||||||
|
if elapsed_time < fastest_time:
|
||||||
|
fastest_url = url
|
||||||
|
fastest_time = elapsed_time
|
||||||
|
fastest_tag = tag
|
||||||
|
|
||||||
|
return fastest_tag, fastest_url
|
||||||
|
|
||||||
|
|
||||||
|
async def check_speed():
|
||||||
|
logger.info('[GsCore资源下载]测速中...')
|
||||||
|
|
||||||
|
URL_LIB = {
|
||||||
|
'[qxqx]': 'https://kr-arm.qxqx.me',
|
||||||
|
'[cos]': 'http://182.43.43.40:8765',
|
||||||
|
'[JPFRP]': 'http://jp-2.lcf.icu:13643',
|
||||||
|
}
|
||||||
|
|
||||||
|
TAG, BASE_URL = await find_fastest_url(URL_LIB)
|
||||||
|
logger.info(f'最快资源站: {TAG} {BASE_URL}')
|
||||||
|
return TAG, BASE_URL
|
||||||
|
|
||||||
|
|
||||||
|
async def check_use():
|
||||||
|
tag, _ = await check_speed()
|
||||||
|
logger.info(tag, _)
|
||||||
|
if tag == '[qxqx]':
|
||||||
|
await download_all_file(
|
||||||
|
'https://kr-arm.qxqx.me',
|
||||||
|
'[qxqx]',
|
||||||
|
'ArknightsUID',
|
||||||
|
{'resource/gamedata': GAMEDATA_PATH},
|
||||||
|
)
|
||||||
|
if tag == '[JPFRP]':
|
||||||
|
await download_all_file(
|
||||||
|
'http://jp-2.lcf.icu:13643',
|
||||||
|
'[JPFRP]',
|
||||||
|
'ArknightsUID',
|
||||||
|
{'resource/gamedata': GAMEDATA_PATH},
|
||||||
|
)
|
||||||
|
if tag == '[cos]':
|
||||||
|
await download_all_file_from_cos()
|
||||||
|
return 'ark全部资源下载完成!'
|
||||||
|
|
||||||
|
|
||||||
async def download_all_file_from_cos():
|
async def download_all_file_from_cos():
|
||||||
async def _download(tasks: List[asyncio.Task]):
|
async def _download(tasks: List[asyncio.Task]):
|
||||||
failed_list.extend(
|
failed_list.extend(
|
||||||
@ -92,3 +156,71 @@ async def download_all_file_from_cos():
|
|||||||
await _download(TASKS)
|
await _download(TASKS)
|
||||||
if count := len(failed_list):
|
if count := len(failed_list):
|
||||||
logger.error(f'[cos]仍有{count}个文件未下载, 请使用命令 `下载全部资源` 重新下载')
|
logger.error(f'[cos]仍有{count}个文件未下载, 请使用命令 `下载全部资源` 重新下载')
|
||||||
|
|
||||||
|
|
||||||
|
async def _get_url(url: str, sess: ClientSession):
|
||||||
|
req = await sess.get(url=url)
|
||||||
|
return await req.read()
|
||||||
|
|
||||||
|
|
||||||
|
async def download_all_file(
|
||||||
|
BASE_URL: str, TAG: str, plugin_name: str, EPATH_MAP: Dict[str, Path]
|
||||||
|
):
|
||||||
|
PLUGIN_RES = f'{BASE_URL}/{plugin_name}'
|
||||||
|
|
||||||
|
TASKS = []
|
||||||
|
async with ClientSession(
|
||||||
|
connector=TCPConnector(verify_ssl=False),
|
||||||
|
timeout=ClientTimeout(total=None, sock_connect=20, sock_read=200),
|
||||||
|
) as sess:
|
||||||
|
for endpoint in EPATH_MAP:
|
||||||
|
url = f'{PLUGIN_RES}/{endpoint}/'
|
||||||
|
path = EPATH_MAP[endpoint]
|
||||||
|
|
||||||
|
base_data = await _get_url(url, sess)
|
||||||
|
content_bs = BeautifulSoup(base_data, 'lxml')
|
||||||
|
pre_data = content_bs.find_all('pre')[0]
|
||||||
|
data_list = pre_data.find_all('a')
|
||||||
|
size_list = list(content_bs.strings)
|
||||||
|
logger.info(f'{TAG} 数据库 {endpoint} 中存在 {len(data_list)} 个内容!')
|
||||||
|
|
||||||
|
temp_num = 0
|
||||||
|
for index, data in enumerate(data_list):
|
||||||
|
if data['href'] == '../':
|
||||||
|
continue
|
||||||
|
file_url = f'{url}{data["href"]}'
|
||||||
|
name: str = data.text
|
||||||
|
size = size_list[index * 2 + 6].split(' ')[-1]
|
||||||
|
size = size.replace('\r\n', '')
|
||||||
|
file_path = path / name
|
||||||
|
if file_path.exists():
|
||||||
|
is_diff = size == str(Path.stat(file_path).st_size)
|
||||||
|
else:
|
||||||
|
is_diff = True
|
||||||
|
if (
|
||||||
|
not file_path.exists()
|
||||||
|
or not Path.stat(file_path).st_size
|
||||||
|
or not is_diff
|
||||||
|
):
|
||||||
|
logger.info(
|
||||||
|
f'{TAG} {plugin_name} 开始下载 {endpoint}/{name} ...'
|
||||||
|
)
|
||||||
|
temp_num += 1
|
||||||
|
TASKS.append(
|
||||||
|
asyncio.wait_for(
|
||||||
|
download(file_url, path, name, sess, TAG),
|
||||||
|
timeout=600,
|
||||||
|
)
|
||||||
|
)
|
||||||
|
if len(TASKS) >= 10:
|
||||||
|
await asyncio.gather(*TASKS)
|
||||||
|
TASKS.clear()
|
||||||
|
await asyncio.gather(*TASKS)
|
||||||
|
TASKS.clear()
|
||||||
|
|
||||||
|
if temp_num == 0:
|
||||||
|
im = f'{TAG} 数据库 {endpoint} 无需下载!'
|
||||||
|
else:
|
||||||
|
im = f'{TAG}数据库 {endpoint} 已下载{temp_num}个内容!'
|
||||||
|
temp_num = 0
|
||||||
|
logger.info(im)
|
||||||
|
Loading…
x
Reference in New Issue
Block a user