2020-12-06 02:48:10 +00:00
|
|
|
import os
|
2020-11-26 03:03:55 +00:00
|
|
|
import re
|
2020-12-06 08:10:00 +00:00
|
|
|
import math
|
2020-12-06 02:48:10 +00:00
|
|
|
from pathlib import Path
|
2020-12-06 01:22:16 +00:00
|
|
|
import requests
|
|
|
|
from PIL import Image
|
2020-12-06 02:48:10 +00:00
|
|
|
from django.conf import settings
|
2020-11-26 03:03:55 +00:00
|
|
|
from urllib.parse import urlsplit, parse_qs
|
|
|
|
from django.forms import ValidationError
|
|
|
|
|
|
|
|
|
|
|
|
def validate_url(url, validator):
|
|
|
|
'''
|
2020-11-26 05:01:47 +00:00
|
|
|
Validate a URL against a dict of validation requirements. Returns an extracted
|
|
|
|
part of the URL if the URL is valid, if invalid raises a ValidationError.
|
2020-11-26 03:03:55 +00:00
|
|
|
'''
|
2020-12-07 10:26:46 +00:00
|
|
|
valid_scheme, valid_netloc, valid_path, invalid_paths, valid_query, \
|
|
|
|
extract_parts = (
|
|
|
|
validator['scheme'], validator['domain'], validator['path_regex'],
|
|
|
|
validator['path_must_not_match'], validator['qs_args'],
|
|
|
|
validator['extract_key']
|
|
|
|
)
|
2020-11-26 03:03:55 +00:00
|
|
|
url_parts = urlsplit(str(url).strip())
|
|
|
|
url_scheme = str(url_parts.scheme).strip().lower()
|
|
|
|
if url_scheme != valid_scheme:
|
2020-12-07 10:26:46 +00:00
|
|
|
raise ValidationError(f'invalid scheme "{url_scheme}" must be "{valid_scheme}"')
|
2020-11-26 03:03:55 +00:00
|
|
|
url_netloc = str(url_parts.netloc).strip().lower()
|
|
|
|
if url_netloc != valid_netloc:
|
2020-12-07 10:26:46 +00:00
|
|
|
raise ValidationError(f'invalid domain "{url_netloc}" must be "{valid_netloc}"')
|
2020-11-26 03:03:55 +00:00
|
|
|
url_path = str(url_parts.path).strip()
|
2020-11-26 05:01:47 +00:00
|
|
|
matches = re.findall(valid_path, url_path)
|
|
|
|
if not matches:
|
2020-12-07 10:26:46 +00:00
|
|
|
raise ValidationError(f'invalid path "{url_path}" must match "{valid_path}"')
|
|
|
|
for invalid_path in invalid_paths:
|
|
|
|
if url_path.lower() == invalid_path.lower():
|
|
|
|
raise ValidationError(f'path "{url_path}" is not valid')
|
2020-11-26 03:03:55 +00:00
|
|
|
url_query = str(url_parts.query).strip()
|
|
|
|
url_query_parts = parse_qs(url_query)
|
|
|
|
for required_query in valid_query:
|
|
|
|
if required_query not in url_query_parts:
|
2020-12-07 10:26:46 +00:00
|
|
|
raise ValidationError(f'invalid query string "{url_query}" must '
|
2020-11-26 05:01:47 +00:00
|
|
|
f'contain the parameter "{required_query}"')
|
|
|
|
extract_from, extract_param = extract_parts
|
|
|
|
extract_value = ''
|
|
|
|
if extract_from == 'path_regex':
|
|
|
|
try:
|
|
|
|
submatches = matches[0]
|
|
|
|
try:
|
|
|
|
extract_value = submatches[extract_param]
|
|
|
|
except IndexError:
|
|
|
|
pass
|
|
|
|
except IndexError:
|
|
|
|
pass
|
|
|
|
elif extract_from == 'qs_args':
|
|
|
|
extract_value = url_query_parts[extract_param][0]
|
|
|
|
return extract_value
|
2020-12-06 01:22:16 +00:00
|
|
|
|
|
|
|
|
|
|
|
def get_remote_image(url):
|
|
|
|
headers = {
|
|
|
|
'user-agent': ('Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36 '
|
|
|
|
'(KHTML, like Gecko) Chrome/69.0.3497.64 Safari/537.36')
|
|
|
|
}
|
|
|
|
r = requests.get(url, headers=headers, stream=True, timeout=60)
|
|
|
|
r.raw.decode_content = True
|
|
|
|
return Image.open(r.raw)
|
2020-12-06 02:48:10 +00:00
|
|
|
|
|
|
|
|
2020-12-06 08:10:00 +00:00
|
|
|
def resize_image_to_height(image, width, height):
|
|
|
|
'''
|
|
|
|
Resizes an image to 'height' pixels keeping the ratio. If the resulting width
|
|
|
|
is larger than 'width' then crop it. If the resulting width is smaller than
|
|
|
|
'width' then stretch it.
|
|
|
|
'''
|
2020-12-06 14:11:48 +00:00
|
|
|
image = image.convert('RGB')
|
2020-12-06 08:10:00 +00:00
|
|
|
ratio = image.width / image.height
|
|
|
|
scaled_width = math.ceil(height * ratio)
|
|
|
|
if scaled_width < width:
|
|
|
|
# Width too small, stretch it
|
|
|
|
scaled_width = width
|
|
|
|
image = image.resize((scaled_width, height), Image.ANTIALIAS)
|
|
|
|
if scaled_width > width:
|
|
|
|
# Width too large, crop it
|
|
|
|
delta = scaled_width - width
|
2020-12-07 10:26:46 +00:00
|
|
|
left, upper = round(delta / 2), 0
|
2020-12-06 08:10:00 +00:00
|
|
|
right, lower = (left + width), height
|
|
|
|
image = image.crop((left, upper, right, lower))
|
|
|
|
return image
|
|
|
|
|
|
|
|
|
2020-12-06 02:48:10 +00:00
|
|
|
def file_is_editable(filepath):
|
|
|
|
'''
|
|
|
|
Checks that a file exists and the file is in an allowed predefined tuple of
|
|
|
|
directories we want to allow writing or deleting in.
|
|
|
|
'''
|
|
|
|
allowed_paths = (
|
|
|
|
# Media item thumbnails
|
|
|
|
os.path.commonpath([os.path.abspath(str(settings.MEDIA_ROOT))]),
|
|
|
|
# Downloaded video files
|
|
|
|
os.path.commonpath([os.path.abspath(str(settings.SYNC_VIDEO_ROOT))]),
|
|
|
|
# Downloaded audio files
|
|
|
|
os.path.commonpath([os.path.abspath(str(settings.SYNC_AUDIO_ROOT))]),
|
|
|
|
)
|
|
|
|
filepath = os.path.abspath(str(filepath))
|
|
|
|
if not os.path.isfile(filepath):
|
|
|
|
return False
|
|
|
|
for allowed_path in allowed_paths:
|
|
|
|
if allowed_path == os.path.commonpath([allowed_path, filepath]):
|
|
|
|
return True
|
|
|
|
return False
|
|
|
|
|
|
|
|
|
|
|
|
def delete_file(filepath):
|
|
|
|
if file_is_editable(filepath):
|
|
|
|
return os.remove(filepath)
|
|
|
|
return False
|
2020-12-06 09:09:48 +00:00
|
|
|
|
|
|
|
|
|
|
|
def seconds_to_timestr(seconds):
|
|
|
|
seconds = seconds % (24 * 3600)
|
|
|
|
hour = seconds // 3600
|
|
|
|
seconds %= 3600
|
|
|
|
minutes = seconds // 60
|
|
|
|
seconds %= 60
|
|
|
|
return '{:02d}:{:02d}:{:02d}'.format(hour, minutes, seconds)
|
2020-12-06 14:11:48 +00:00
|
|
|
|
|
|
|
|
|
|
|
def parse_media_format(format_dict):
|
2020-12-08 05:56:43 +00:00
|
|
|
'''
|
|
|
|
This parser primarily adapts the format dict returned by youtube-dl into a
|
|
|
|
standard form used by the matchers in matching.py. If youtube-dl changes
|
|
|
|
any internals, update it here.
|
|
|
|
'''
|
2020-12-06 14:11:48 +00:00
|
|
|
vcodec_full = format_dict.get('vcodec', '')
|
|
|
|
vcodec_parts = vcodec_full.split('.')
|
|
|
|
if len(vcodec_parts) > 0:
|
|
|
|
vcodec = vcodec_parts[0].strip().upper()
|
|
|
|
else:
|
|
|
|
vcodec = None
|
|
|
|
if vcodec == 'NONE':
|
|
|
|
vcodec = None
|
|
|
|
acodec_full = format_dict.get('acodec', '')
|
|
|
|
acodec_parts = acodec_full.split('.')
|
|
|
|
if len(acodec_parts) > 0:
|
|
|
|
acodec = acodec_parts[0].strip().upper()
|
|
|
|
else:
|
|
|
|
acodec = None
|
|
|
|
if acodec == 'NONE':
|
|
|
|
acodec = None
|
2020-12-08 05:19:19 +00:00
|
|
|
try:
|
|
|
|
fps = int(format_dict.get('fps', 0))
|
|
|
|
except (ValueError, TypeError):
|
|
|
|
fps = 0
|
|
|
|
format_full = format_dict.get('format_note', '').strip().upper()
|
|
|
|
format_str = format_full[:-2] if format_full.endswith('60') else format_full
|
2020-12-06 14:11:48 +00:00
|
|
|
return {
|
|
|
|
'id': format_dict.get('format_id', ''),
|
2020-12-08 05:19:19 +00:00
|
|
|
'format': format_str,
|
|
|
|
'format_verbose': format_dict.get('format', ''),
|
2020-12-06 14:11:48 +00:00
|
|
|
'height': format_dict.get('height', 0),
|
|
|
|
'vcodec': vcodec,
|
2020-12-08 05:19:19 +00:00
|
|
|
'vbr': format_dict.get('tbr', 0),
|
2020-12-06 14:11:48 +00:00
|
|
|
'acodec': acodec,
|
2020-12-08 05:19:19 +00:00
|
|
|
'abr': format_dict.get('abr', 0),
|
|
|
|
'is_60fps': fps > 50,
|
|
|
|
'is_hdr': 'HDR' in format_dict.get('format', '').upper(),
|
2020-12-06 14:11:48 +00:00
|
|
|
}
|