rework redis, rework web for work with array of links

This commit is contained in:
nikili0n 2023-09-29 05:53:27 +03:00
parent 30ce9aee9a
commit e4fa00e1ce
10 changed files with 153 additions and 124 deletions

View File

@ -8,5 +8,5 @@ redis = RedisClient()
class AsyncQueue(asyncio.Queue): class AsyncQueue(asyncio.Queue):
async def put(self, item): async def put(self, item):
await redis.set_task_to_queue(item) await redis.set_task_to_queue(item["link"], item)
return await super().put(item) return await super().put(item)

View File

@ -12,12 +12,10 @@ from src.core.redis_client import RedisClient
from src.core.result import Result, ResultTypeEnum from src.core.result import Result, ResultTypeEnum
from src.exceptions.download_exceptions import FileAlreadyExistException, SiteNotImplementedException from src.exceptions.download_exceptions import FileAlreadyExistException, SiteNotImplementedException
from src.parsers.MyMail.my_mail_parser import MyMailParser from src.parsers.MyMail.my_mail_parser import MyMailParser
from src.parsers.Okru.ok_parser import OkParser
from src.parsers.Yappy.yappy_parser import YappyParser from src.parsers.Yappy.yappy_parser import YappyParser
from src.parsers.base_parser import BaseParser from src.parsers.base_parser import BaseParser
from loguru import logger from loguru import logger
from src.parsers.parser_mapping import get_parser from src.parsers.parser_mapping import get_parser
# TODO: добавить логгер с временными метками в yt-dlp
class MasterService: class MasterService:
@ -40,12 +38,17 @@ class MasterService:
await asyncio.gather(self.rabbit_consumer(self.queue), *tasks) await asyncio.gather(self.rabbit_consumer(self.queue), *tasks)
async def result_processing(self, result: Result | list, redis: RedisClient, video_params: dict):
await redis.del_task_from_tasks_and_add_to_task_done(task=result.value, link=video_params["link"])
await publish_message_with_task_done(task=result.value)
self.queue.task_done()
async def create_workers(self): async def create_workers(self):
while True: while True:
video_params = await self.queue.get() video_params = await self.queue.get()
redis = RedisClient() redis = RedisClient()
await redis.del_tasks_queue() await redis.del_tasks_queue()
await redis.del_task_from_queue_and_add_to_tasks(task=video_params) await redis.del_task_from_queue_and_add_to_tasks(link=video_params["link"], task=video_params)
self.currently_underway[video_params['link']] = video_params self.currently_underway[video_params['link']] = video_params
download_task = self.loop.run_in_executor(self.executor, partial( download_task = self.loop.run_in_executor(self.executor, partial(
@ -53,18 +56,7 @@ class MasterService:
)) ))
result: Result = await download_task result: Result = await download_task
if result.result_type in [ResultTypeEnum.DONE, ResultTypeEnum.EXIST]: await self.result_processing(result, redis, video_params)
await redis.del_task_from_tasks_and_add_to_task_done(task=result.value, working_task=video_params)
await publish_message_with_task_done(task=result.value)
self.queue.task_done()
else:
error_message = {
"link": video_params["link"],
"result": result.value,
"status": "error"
}
await redis.del_task_from_tasks_and_add_to_task_done(task=result.value, working_task=video_params)
await publish_message_with_task_done(task=error_message)
if video_params['link'] in self.currently_underway: if video_params['link'] in self.currently_underway:
del self.currently_underway[video_params['link']] del self.currently_underway[video_params['link']]
@ -79,7 +71,7 @@ class MasterService:
def get_parser(params: dict): def get_parser(params: dict):
try: try:
url_parse_result = urlparse(params["link"]) url_parse_result = urlparse(params["link"])
uri = f"{url_parse_result.netloc}{url_parse_result.path}" uri = f"{url_parse_result.netloc}{url_parse_result.path}"
logger.info(uri) logger.info(uri)
# # TODO: похоже нужно переделать на регулярки, т.к. добавлять каждую вариацию домена моветон, вероятно я сделаюне- # # TODO: похоже нужно переделать на регулярки, т.к. добавлять каждую вариацию домена моветон, вероятно я сделаюне-
# parser_mapping = { # parser_mapping = {
@ -106,7 +98,13 @@ class MasterService:
"link": video_params["link"], "link": video_params["link"],
"result": result, "result": result,
"status": "done" "status": "done"
}) }) if not isinstance(result, list)\
else Result(result_type=ResultTypeEnum.DONE, value={
"link": video_params["link"],
"result": [result_part for result_part in result],
"status": "done"
})
except FileAlreadyExistException as ex: except FileAlreadyExistException as ex:
return Result(result_type=ResultTypeEnum.EXIST, value={ return Result(result_type=ResultTypeEnum.EXIST, value={
"link": video_params["link"], "link": video_params["link"],
@ -120,16 +118,11 @@ class MasterService:
"status": "error" "status": "error"
}) })
except Exception as ex: except Exception as ex:
Result(result_type=ResultTypeEnum.EXCEPTION, value={ return Result(result_type=ResultTypeEnum.EXCEPTION, value={
"link": video_params["link"], "link": video_params["link"],
"result": traceback.format_exc(), "result": traceback.format_exc(),
"status": "error" "status": "error"
}) })
return Result(result_type=ResultTypeEnum.EXCEPTION, value={
"link": video_params["link"],
"result": "Unknown exception encountered",
"status": "error"
})
# TODO upload to server # TODO upload to server

View File

@ -28,7 +28,7 @@ async def get_messages(inner_queue) -> None:
await asyncio.Future() await asyncio.Future()
async def publish_message_with_task_done(task: dict) -> None: async def publish_message_with_task_done(task: dict | list) -> None:
queue_name = "tasks_done" queue_name = "tasks_done"
async with await connect("amqp://guest:guest@localhost/") as connection: async with await connect("amqp://guest:guest@localhost/") as connection:
# Creating channel # Creating channel

View File

@ -7,61 +7,57 @@ class RedisClient:
SET_NAME = "queue" SET_NAME = "queue"
TASKS_NAME = "tasks_working" TASKS_NAME = "tasks_working"
TASKS_DONE_NAME = "tasks_done" TASKS_DONE_NAME = "tasks_done"
# TODO: переписать всё вазимодействие редиса обратно на ключ-значение
def __init__(self): def __init__(self):
self.connection = redis.Redis(host="localhost", port=6379, db=0) self.connection = redis.Redis(host="localhost", port=6379, db=0)
async def _set_task(self, task: dict, queue_name) -> int: async def _set_task(self, queue_name: str, link: str, task: dict | list, ) -> int:
async with self.connection as connection: async with self.connection as connection:
res = await connection.sadd(f'{queue_name}:1', json.dumps(task, indent=4).encode('utf-8')) res = await connection.hset(queue_name, link, json.dumps(task, indent=4).encode('utf-8'))
return res return res
async def _del_task(self, task: dict, queue_name) -> int: async def _del_task(self, queue_name: str, link: str,) -> int:
async with self.connection as connection: async with self.connection as connection:
res = await connection.srem(f'{queue_name}:1', json.dumps(task, indent=4).encode('utf-8')) res = await connection.hdel(queue_name, link)
return res return res
async def set_task_to_queue(self, task: dict) -> int: async def set_task_to_queue(self, link: str, task: dict | list) -> int:
async with self.connection as connection: res = await self._set_task(queue_name=self.SET_NAME, link=link, task=task)
res = await connection.sadd(self.SET_NAME, json.dumps(task, indent=4).encode('utf-8'))
return res return res
async def get_queue(self) -> set: async def get_all_tasks_from_queue(self, queue_name: str) -> dict:
async with self.connection as connection: async with self.connection as connection:
res = await connection.smembers(self.SET_NAME + f":1") res = await connection.hgetall(queue_name)
return res return res
async def get_tasks(self) -> set: # async def get_tasks(self) -> set:
async with self.connection as connection: # async with self.connection as connection:
res = await connection.smembers(self.TASKS_NAME + f":1") # res = await connection.smembers(self.TASKS_NAME + f":1")
return res # return res
async def get_task_done_queue(self) -> set: # async def get_task_done_queue(self) -> set:
async with self.connection as connection: # async with self.connection as connection:
res = await connection.smembers(self.TASKS_DONE_NAME + f":1") # res = await connection.smembers(self.TASKS_DONE_NAME + f":1")
return res # return res
async def del_task_from_queue(self, task: dict) -> int: # async def del_task_from_queue(self, link, task: dict) -> int:
async with self.connection as connection: # async with self.connection as connection:
res = await connection.srem(self.SET_NAME, json.dumps(task, indent=4).encode('utf-8')) # res = await self._del_task(self.SET_NAME, json.dumps(task, indent=4).encode('utf-8'))
return res # return res
async def del_task_from_queue_and_add_to_tasks(self, task: dict) -> int: async def del_task_from_queue_and_add_to_tasks(self, link: str, task: dict | list) -> int:
await self._del_task(task, self.SET_NAME) await self._del_task(self.SET_NAME, link)
return await self._set_task(task, self.TASKS_NAME) return await self._set_task(self.TASKS_NAME, link, task)
async def del_task_from_tasks_and_add_to_task_done(self, task: dict, working_task: dict) -> int: async def del_task_from_tasks_and_add_to_task_done(self, task: dict | list, link: str) -> int:
await self._del_task(working_task, self.TASKS_NAME) await self._del_task(self.TASKS_NAME, link)
return await self._set_task(task, self.TASKS_DONE_NAME) return await self._set_task(self.TASKS_DONE_NAME, link, task)
async def del_task_from_task_done_queue(self, task) -> int: async def del_task_from_task_done_queue(self, task) -> int:
async with self.connection as connection: res = await self._del_task(self.TASKS_DONE_NAME, task["link"])
res = await connection.srem(self.TASKS_DONE_NAME + f":1", json.dumps(task, indent=4).encode('utf-8'))
return res return res
async def del_tasks_queue(self) -> int: async def del_tasks_queue(self) -> int:
async with self.connection as connection: async with self.connection as connection:
res = await connection.delete(self.TASKS_NAME + f":1") res = await connection.delete(self.TASKS_NAME)
return res return res

View File

@ -10,7 +10,7 @@ class ResultTypeEnum(Enum):
class Result: class Result:
def __init__(self, result_type: ResultTypeEnum, value: str | dict = None): def __init__(self, result_type: ResultTypeEnum, value: str | dict | list = None):
self.result_type = result_type self.result_type = result_type
self.value = value self.value = value

View File

@ -4,6 +4,7 @@ import requests
from bs4 import BeautifulSoup from bs4 import BeautifulSoup
from src.exceptions.download_exceptions import FileAlreadyExistException
from src.parsers.base_parser import BaseParser from src.parsers.base_parser import BaseParser
@ -16,12 +17,24 @@ class OkParser(BaseParser):
resp.encoding = self.BASE_ENCODING resp.encoding = self.BASE_ENCODING
soup = BeautifulSoup(resp.text, 'lxml') soup = BeautifulSoup(resp.text, 'lxml')
required_div = [div for div in soup.find_all('div', {'class': 'invisible'}) if len(div['class']) < 2][0] required_div = [div for div in soup.find_all('div', {'class': 'invisible'}) if len(div['class']) < 2][0]
link = required_div.find('span').find('span').find('a').get("href") video_tags = required_div.find('span').find_all_next('span', {'itemprop': "video"})
self.params["link"] = link links = [video_tag.find('a').get("href") for video_tag in video_tags]
return link return links
except Exception as ex: except Exception as ex:
raise raise
def video_download(self): def video_download(self):
self.get_video_link() base_link = self.params["link"]
super().video_download() links = self.get_video_link()
file_paths = []
for link in links:
try:
self.params["link"] = link
file_path = super().video_download()
file_paths.append(file_path)
except FileAlreadyExistException as ex:
file_paths.append(ex.message)
continue
self.params["link"] = base_link
return file_paths

View File

@ -20,11 +20,11 @@ class BaseParser:
"logger": logger, "logger": logger,
"merge_output_format": self.params["merge_output_format"], "merge_output_format": self.params["merge_output_format"],
'outtmpl': self.params["outtmpl"], 'outtmpl': self.params["outtmpl"],
"quiet": True # "quiet": True
} }
downloader = VideoDownloader(link=self.params["link"], ydl_opts=ydl_opts) downloader = VideoDownloader(link=self.params["link"], ydl_opts=ydl_opts)
downloader.get_info() downloader.get_info()
path_to_video = f"{downloader.info['extractor_key']}/{downloader.info['id']}_{downloader.info['width']}p.{downloader.info['ext']}" path_to_video = f"{downloader.info['extractor_key']}/{downloader.info['id']}_{downloader.info['resolution']}.{downloader.info['ext']}"
if os.path.exists(os.path.join(os.getcwd() + "/downloads/" + path_to_video)): if os.path.exists(os.path.join(os.getcwd() + "/downloads/" + path_to_video)):
raise FileAlreadyExistException(message=path_to_video) raise FileAlreadyExistException(message=path_to_video)
downloader.ydl_opts["quiet"] = False downloader.ydl_opts["quiet"] = False

View File

@ -6,12 +6,14 @@ from src.parsers.Okru.ok_parser import OkParser
from src.parsers.Yappy.yappy_parser import YappyParser from src.parsers.Yappy.yappy_parser import YappyParser
from src.parsers.base_parser import BaseParser from src.parsers.base_parser import BaseParser
def compile_regex(regex): def compile_regex(regex):
return re.compile(regex, re.IGNORECASE | re.DOTALL | re.MULTILINE) return re.compile(regex, re.IGNORECASE | re.DOTALL | re.MULTILINE)
parser_mapping = OrderedDict( parser_mapping = OrderedDict(
{ {
compile_regex(r"^my.mail.ru/") : MyMailParser, compile_regex(r"^my.mail.ru/"): MyMailParser,
compile_regex(r"^(?:www.)?(?:youtube.com|youtu.be)/"): BaseParser, compile_regex(r"^(?:www.)?(?:youtube.com|youtu.be)/"): BaseParser,
compile_regex(r"^vk.com/"): BaseParser, compile_regex(r"^vk.com/"): BaseParser,
compile_regex(r"^ok.ru/okvideo/topic"): OkParser, compile_regex(r"^ok.ru/okvideo/topic"): OkParser,
@ -24,7 +26,8 @@ parser_mapping = OrderedDict(
} }
) )
def get_parser(uri): def get_parser(uri):
for regex in parser_mapping: for regex in parser_mapping:
if regex.match(uri): if regex.match(uri):
return parser_mapping[regex] return parser_mapping[regex]

View File

@ -27,7 +27,6 @@ app.add_middleware(
allow_headers=["*"], allow_headers=["*"],
) )
''' '''
await self.app(scope, receive, send) await self.app(scope, receive, send)
File "/home/admin/video_downloader_service/.venv/lib/python3.10/site-packages/starlette/routing.py", line 66, in app File "/home/admin/video_downloader_service/.venv/lib/python3.10/site-packages/starlette/routing.py", line 66, in app
@ -61,32 +60,19 @@ queue_name -> {
async def is_task_already_done_or_exist(redis: RedisClient, link: str): async def is_task_already_done_or_exist(redis: RedisClient, link: str):
messages = await redis.get_task_done_queue() messages = await redis.get_all_tasks_from_queue(redis.TASKS_DONE_NAME)
temp = [json.loads(msg) for msg in messages] tasks = {k.decode("utf-8"): json.loads(v.decode("utf-8")) for k, v in messages.items()} if messages else None
tasks = [
msg for msg in temp
if msg["link"] == link
and msg["status"] in ["done", "exist"]
]
if len(tasks) > 0: if tasks and link in tasks and tasks[link]["status"] in ["done", "exist"]:
task = tasks[0] return tasks[link]
if os.path.exists(os.path.join(os.getcwd(), os.pardir, os.pardir + "/downloads/" + task["result"])):
return task
await redis.del_task_from_task_done_queue(task)
async def is_task_in_process(redis: RedisClient, link: str): async def is_task_in_process(redis: RedisClient, link: str):
messages = await redis.get_tasks() messages = await redis.get_all_tasks_from_queue(redis.TASKS_NAME)
tasks = {k.decode("utf-8"): json.loads(v.decode("utf-8")) for k, v in messages.items()} if messages else None
tasks = [ if tasks and link in tasks:
literal_eval(message.decode('utf-8')) for message in messages return tasks[link]
if literal_eval(message.decode('utf-8'))["link"] == link
]
if len(tasks) > 0:
task = tasks[0]
return task
@app.get("/") @app.get("/")
@ -101,10 +87,13 @@ async def get_url_for_download_video(request: Request, data: SubmitIn = Depends(
# TODO: где-то не обновился статус после выполнения\провала задачи # TODO: где-то не обновился статус после выполнения\провала задачи
task_in_process = await is_task_in_process(red, data.link) task_in_process = await is_task_in_process(red, data.link)
if task_in_process: if task_in_process:
return JSONResponse({"result": "Задача в работе. Ожидайте"}) return JSONResponse(status_code=202, content={"result": "Задача в работе. Ожидайте"})
if task_done: if task_done:
link_to_download_video = str(request.base_url) + "get/?file_path=" + task_done["result"] if isinstance(task_done["result"], str):
return JSONResponse({"result": link_to_download_video}) links_to_download_video = [str(request.base_url) + "get/?file_path=" + task_done["result"]]
else:
links_to_download_video = [str(request.base_url) + "get/?file_path=" + path for path in task_done["result"]]
return JSONResponse({"result": links_to_download_video})
# TODO: учесть, что если делать запрос CURL\urllib3\etc, в теле может быть несколько ссылок -> должно быть создано несколько задач # TODO: учесть, что если делать запрос CURL\urllib3\etc, в теле может быть несколько ссылок -> должно быть создано несколько задач
async with await connect("amqp://guest:guest@localhost/") as connection: async with await connect("amqp://guest:guest@localhost/") as connection:
@ -115,7 +104,7 @@ async def get_url_for_download_video(request: Request, data: SubmitIn = Depends(
"link": data.link, "link": data.link,
"format": f"bestvideo[ext={data.video_format.value}]+bestaudio[ext={data.audio_format.value}]/best[ext={data.video_format.value}]/best", "format": f"bestvideo[ext={data.video_format.value}]+bestaudio[ext={data.audio_format.value}]/best[ext={data.video_format.value}]/best",
"merge_output_format": data.merge_output_format.value, "merge_output_format": data.merge_output_format.value,
"outtmpl": f"downloads/%(extractor_key)s/%(id)s_%(width)sp.%(ext)s", "outtmpl": f"downloads/%(extractor_key)s/%(id)s_%(resolution)s.%(ext)s",
}, ] }, ]
# Sending the message # Sending the message
for link in body: for link in body:
@ -147,44 +136,41 @@ async def download_video(file_path):
with open(base_download_dir + f'/{file_path}', mode="rb") as file_like: with open(base_download_dir + f'/{file_path}', mode="rb") as file_like:
yield from file_like yield from file_like
return StreamingResponse(iterfile(), headers={'Content-Disposition': f'inline; filename="{file_path}"'}, media_type="video") return StreamingResponse(iterfile(), headers={'Content-Disposition': f'inline; filename="{file_path}"'},
media_type="video")
@app.post('/check/', response_class=FileResponse, status_code=200) @app.post('/check/', response_class=FileResponse, status_code=200)
async def download_video(data: CheckIn, request: Request): async def download_video(data: CheckIn, request: Request):
try: try:
red = RedisClient() red = RedisClient()
messages_task_done = await red.get_all_tasks_from_queue(red.TASKS_DONE_NAME)
messages_tasks = await red.get_all_tasks_from_queue(red.TASKS_NAME)
messages_task_done = await red.get_task_done_queue() tasks_done = {k.decode("utf-8"): json.loads(v.decode("utf-8")) for k, v in
messages_tasks = await red.get_tasks() messages_task_done.items()} if messages_task_done else None
tasks_done = [ tasks = {k.decode("utf-8"): json.loads(v.decode("utf-8")) for k, v in
literal_eval(message.decode('utf-8')) for message in messages_task_done messages_tasks.items()} if messages_tasks else None
if literal_eval(message.decode('utf-8'))["link"] == data.link
]
tasks = [
literal_eval(message.decode('utf-8')) for message in messages_tasks
if literal_eval(message.decode('utf-8'))["link"] == data.link
]
error_tasks = [ if tasks and data.link in tasks:
tasks_done.pop(tasks_done.index(error_task)) for error_task in tasks_done if error_task["status"] == "error"
] if tasks_done else None
if tasks and len(tasks) > 0:
task = tasks[0]
return JSONResponse( return JSONResponse(
status_code=202, status_code=202,
content={"result": f"Задача {task['link']} в данный момент в работе, выполняется"} content={"result": f"Задача {data.link} в данный момент в работе, выполняется"}
) )
# TODO: если уже была попытка сделать задачу и в редисе она с ошибкой, то переташить её в очередь на выполнение с очисткой состояние об ошибке # TODO: если уже была попытка сделать задачу и в редисе она с ошибкой, то переташить её в очередь на выполнение с очисткой состояние об ошибке
if error_tasks and len(error_tasks) > 0: if data.link in tasks_done and tasks_done[data.link]["status"] == "error":
error_task = error_tasks[0] await red.del_task_from_task_done_queue(tasks_done[data.link])
await red.del_task_from_task_done_queue(error_task)
return JSONResponse(status_code=510, return JSONResponse(status_code=510,
content={"result": f"Задача выполнена с ошибкой, попробуйте загрузить еще раз"}) content={"result": f"Задача выполнена с ошибкой, попробуйте загрузить еще раз"})
if len(tasks_done) > 0: if tasks_done and data.link in tasks_done:
link_to_download_video = str(request.base_url) + "get/?file_path=" + tasks_done[0]["result"] if isinstance(tasks_done[data.link]["result"], str):
return JSONResponse({"result": link_to_download_video}) links_to_download_video = [str(request.base_url) + "get/?file_path=" + tasks_done[data.link]["result"]]
else:
links_to_download_video = [str(request.base_url) + "get/?file_path=" + path for path in
tasks_done[data.link]["result"]]
return JSONResponse({"result": links_to_download_video})
return JSONResponse(status_code=404, content={"result": "Задача не найдена"})
except (AttributeError, IndexError): except (AttributeError, IndexError):
return JSONResponse(status_code=404, content={"result": "Задача не найдена"}) return JSONResponse(status_code=404, content={"result": "Задача не найдена"})

View File

@ -77,7 +77,7 @@
<input type="text" name="merge_output_format" placeholder="merge_output_format"> <input type="text" name="merge_output_format" placeholder="merge_output_format">
<button type="submit" class="custom-btn btn-1"><span class="submit-spinner submit-spinner_hide"></span> Download</button> <button type="submit" class="custom-btn btn-1"><span class="submit-spinner submit-spinner_hide"></span> Download</button>
</form> </form>
<div class="col"> <div id="linksList" class="col">
<p> Ссылка для скачивания:</p> <p> Ссылка для скачивания:</p>
<br> <br>
<a id="result" href="" download></a> <a id="result" href="" download></a>
@ -87,12 +87,32 @@
</div> </div>
</body> </body>
<script> <script>
function createLink(url) {
var link = document.createElement('a');
link.className = "addedLink";
link.setAttribute('download', '');
link.innerHTML = url;
link.href = url;
return link;
}
function createLinksList(urls) {
var mydiv = document.getElementById("linksList");
urls.forEach((element) => {
mydiv.appendChild(createLink(element));
var br = document.createElement('br');
br.className = "addedBr";
mydiv.appendChild(br);
}
);
};
function sendReq() { function sendReq() {
document.forms.download.querySelector('[type="submit"]').disabled = true; document.forms.download.querySelector('[type="submit"]').disabled = true;
document.forms.download.querySelector('.submit-spinner').classList.remove('submit-spinner_hide'); document.forms.download.querySelector('.submit-spinner').classList.remove('submit-spinner_hide');
const link = document.getElementById("link").value const link = document.getElementById("link").value
const xhr2 = new XMLHttpRequest(); const xhr2 = new XMLHttpRequest();
// TODO: скорректировать ссылку, она должна быть относительной
xhr2.open('POST', 'http://'+document.location.host+'/check/'); xhr2.open('POST', 'http://'+document.location.host+'/check/');
xhr2.responseType = 'json'; xhr2.responseType = 'json';
xhr2.setRequestHeader("Content-Type", "application/json;charset=UTF-8"); xhr2.setRequestHeader("Content-Type", "application/json;charset=UTF-8");
@ -113,8 +133,11 @@
document.forms.download.querySelector('.submit-spinner').classList.add('submit-spinner_hide'); document.forms.download.querySelector('.submit-spinner').classList.add('submit-spinner_hide');
}; };
} else if (xhr2.status === 200) { } else if (xhr2.status === 200) {
result.innerHTML = xhr2.response.result; console.log(xhr2.response)
result.href = xhr2.response.result; if (document.getElementById("result").texContent !== '') {
document.getElementById("result").textContent = ''
};
createLinksList(xhr2.response["result"]);
document.forms.download.querySelector('[type="submit"]').disabled = false; document.forms.download.querySelector('[type="submit"]').disabled = false;
document.forms.download.querySelector('.submit-spinner').classList.add('submit-spinner_hide'); document.forms.download.querySelector('.submit-spinner').classList.add('submit-spinner_hide');
}; };
@ -127,6 +150,16 @@
function sendForm() { function sendForm() {
const links = document.querySelectorAll('.addedLink');
links.forEach(link => {
link.remove();
}
);
const brs = document.querySelectorAll('.addedBr');
brs.forEach(br => {
br.remove();
}
);
const xhr = new XMLHttpRequest(); const xhr = new XMLHttpRequest();
xhr.open('POST', document.forms.download.action); xhr.open('POST', document.forms.download.action);
xhr.responseType = 'json'; xhr.responseType = 'json';
@ -140,12 +173,17 @@
}; };
const response = xhr.response; const response = xhr.response;
result.innerHTML = xhr.response.result;
result.href = xhr.response.result;
console.log(response);
if (xhr.status === 201) { if (xhr.status === 201) {
sendReq() sendReq()
} } else if (xhr.status === 202) {
result.innerHTML = xhr.response.result;
result.href = xhr.response.result;
} else {
console.log(response["result"]);
createLinksList(response["result"]);
};
} }
xhr.onerror = () => { xhr.onerror = () => {
document.forms.download.querySelector('[type="submit"]').disabled = false; document.forms.download.querySelector('[type="submit"]').disabled = false;