支持群友 @谦虚 的资源站

This commit is contained in:
qwerdvd 2023-10-12 13:18:19 +08:00
parent b7ac18cebf
commit a1e74d1129
2 changed files with 141 additions and 7 deletions

View File

@ -6,26 +6,28 @@ from gsuid_core.logger import logger
from gsuid_core.models import Event
from gsuid_core.sv import SV
from ..utils.resource.download_all_resource import download_all_resource
from .memoryStore import store
from ..utils.resource.download_from_cos import check_use
from .constants import Excel
from .memoryStore import store
sv_download_config = SV("下载资源", pm=2)
@sv_download_config.on_fullmatch(("下载全部资源")) # noqa: UP034
@sv_download_config.on_fullmatch(("ark下载全部资源")) # noqa: UP034
async def send_download_resource_msg(bot: Bot, ev: Event):
await bot.send("正在开始下载~可能需要较久的时间!")
im = await download_all_resource()
im = await check_use()
await bot.send(im)
async def startup():
logger.info("[资源文件下载] 正在检查与下载缺失的资源文件, 可能需要较长时间, 请稍等")
await download_all_resource()
await check_use()
logger.info("[资源文件下载] 检查完毕, 正在加载 gamedata")
for file_path in Path(get_res_path(["ArknightsUID", "resource", "gamedata"])).rglob("*.json"):
for file_path in Path(
get_res_path(["ArknightsUID", "resource", "gamedata"])
).rglob("*.json"):
await store.get_file(file_path)
await Excel.preload_table()

View File

@ -2,12 +2,16 @@ import asyncio
from pathlib import Path
from typing import Dict, List, Tuple, Union
from aiohttp import ClientTimeout, TCPConnector
from aiohttp.client import ClientSession
from bs4 import BeautifulSoup
from gsuid_core.logger import logger
from gsuid_core.utils.download_resource.download_core import check_url
from gsuid_core.utils.download_resource.download_file import download
from msgspec import json as msgjson
from .download_url import download_file
from .RESOURCE_PATH import RESOURCE_PATH
from .RESOURCE_PATH import GAMEDATA_PATH, RESOURCE_PATH
MAX_DOWNLOAD = 10
@ -20,6 +24,66 @@ with Path.open(
)
async def find_fastest_url(urls: Dict[str, str]):
tasks = []
for tag in urls:
tasks.append(asyncio.create_task(check_url(tag, urls[tag])))
results: list[tuple[str, str, float]] = await asyncio.gather(
*tasks, return_exceptions=True
)
fastest_tag = ''
fastest_url = None
fastest_time = float('inf')
for result in results:
if isinstance(result, Exception):
continue
tag, url, elapsed_time = result
if elapsed_time < fastest_time:
fastest_url = url
fastest_time = elapsed_time
fastest_tag = tag
return fastest_tag, fastest_url
async def check_speed():
logger.info('[GsCore资源下载]测速中...')
URL_LIB = {
'[qxqx]': 'https://kr-arm.qxqx.me',
'[cos]': 'http://182.43.43.40:8765',
'[JPFRP]': 'http://jp-2.lcf.icu:13643',
}
TAG, BASE_URL = await find_fastest_url(URL_LIB)
logger.info(f'最快资源站: {TAG} {BASE_URL}')
return TAG, BASE_URL
async def check_use():
tag, _ = await check_speed()
logger.info(tag, _)
if tag == '[qxqx]':
await download_all_file(
'https://kr-arm.qxqx.me',
'[qxqx]',
'ArknightsUID',
{'resource/gamedata': GAMEDATA_PATH},
)
if tag == '[JPFRP]':
await download_all_file(
'http://jp-2.lcf.icu:13643',
'[JPFRP]',
'ArknightsUID',
{'resource/gamedata': GAMEDATA_PATH},
)
if tag == '[cos]':
await download_all_file_from_cos()
return 'ark全部资源下载完成!'
async def download_all_file_from_cos():
async def _download(tasks: List[asyncio.Task]):
failed_list.extend(
@ -92,3 +156,71 @@ async def download_all_file_from_cos():
await _download(TASKS)
if count := len(failed_list):
logger.error(f'[cos]仍有{count}个文件未下载, 请使用命令 `下载全部资源` 重新下载')
async def _get_url(url: str, sess: ClientSession):
req = await sess.get(url=url)
return await req.read()
async def download_all_file(
BASE_URL: str, TAG: str, plugin_name: str, EPATH_MAP: Dict[str, Path]
):
PLUGIN_RES = f'{BASE_URL}/{plugin_name}'
TASKS = []
async with ClientSession(
connector=TCPConnector(verify_ssl=False),
timeout=ClientTimeout(total=None, sock_connect=20, sock_read=200),
) as sess:
for endpoint in EPATH_MAP:
url = f'{PLUGIN_RES}/{endpoint}/'
path = EPATH_MAP[endpoint]
base_data = await _get_url(url, sess)
content_bs = BeautifulSoup(base_data, 'lxml')
pre_data = content_bs.find_all('pre')[0]
data_list = pre_data.find_all('a')
size_list = list(content_bs.strings)
logger.info(f'{TAG} 数据库 {endpoint} 中存在 {len(data_list)} 个内容!')
temp_num = 0
for index, data in enumerate(data_list):
if data['href'] == '../':
continue
file_url = f'{url}{data["href"]}'
name: str = data.text
size = size_list[index * 2 + 6].split(' ')[-1]
size = size.replace('\r\n', '')
file_path = path / name
if file_path.exists():
is_diff = size == str(Path.stat(file_path).st_size)
else:
is_diff = True
if (
not file_path.exists()
or not Path.stat(file_path).st_size
or not is_diff
):
logger.info(
f'{TAG} {plugin_name} 开始下载 {endpoint}/{name} ...'
)
temp_num += 1
TASKS.append(
asyncio.wait_for(
download(file_url, path, name, sess, TAG),
timeout=600,
)
)
if len(TASKS) >= 10:
await asyncio.gather(*TASKS)
TASKS.clear()
await asyncio.gather(*TASKS)
TASKS.clear()
if temp_num == 0:
im = f'{TAG} 数据库 {endpoint} 无需下载!'
else:
im = f'{TAG}数据库 {endpoint} 已下载{temp_num}个内容!'
temp_num = 0
logger.info(im)