mirror of
https://github.com/yt-dlp/yt-dlp.git
synced 2024-11-03 08:19:57 +01:00
[cleanup] Misc
This commit is contained in:
parent
8aa0e7cd96
commit
7b2c3f47c6
@ -58,6 +58,7 @@
|
|||||||
from .update import detect_variant
|
from .update import detect_variant
|
||||||
from .utils import (
|
from .utils import (
|
||||||
DEFAULT_OUTTMPL,
|
DEFAULT_OUTTMPL,
|
||||||
|
IDENTITY,
|
||||||
LINK_TEMPLATES,
|
LINK_TEMPLATES,
|
||||||
NO_DEFAULT,
|
NO_DEFAULT,
|
||||||
NUMBER_RE,
|
NUMBER_RE,
|
||||||
@ -1002,7 +1003,7 @@ def parse_outtmpl(self):
|
|||||||
return self.params['outtmpl']
|
return self.params['outtmpl']
|
||||||
|
|
||||||
def _parse_outtmpl(self):
|
def _parse_outtmpl(self):
|
||||||
sanitize = lambda x: x
|
sanitize = IDENTITY
|
||||||
if self.params.get('restrictfilenames'): # Remove spaces in the default template
|
if self.params.get('restrictfilenames'): # Remove spaces in the default template
|
||||||
sanitize = lambda x: x.replace(' - ', ' ').replace(' ', '-')
|
sanitize = lambda x: x.replace(' - ', ' ').replace(' ', '-')
|
||||||
|
|
||||||
@ -2983,13 +2984,12 @@ def existing_video_file(*filepaths):
|
|||||||
info_dict['ext'] = os.path.splitext(file)[1][1:]
|
info_dict['ext'] = os.path.splitext(file)[1][1:]
|
||||||
return file
|
return file
|
||||||
|
|
||||||
success = True
|
fd, success = None, True
|
||||||
merger, fd = FFmpegMergerPP(self), None
|
|
||||||
if info_dict.get('protocol') or info_dict.get('url'):
|
if info_dict.get('protocol') or info_dict.get('url'):
|
||||||
fd = get_suitable_downloader(info_dict, self.params, to_stdout=temp_filename == '-')
|
fd = get_suitable_downloader(info_dict, self.params, to_stdout=temp_filename == '-')
|
||||||
if fd is not FFmpegFD and (
|
if fd is not FFmpegFD and (
|
||||||
info_dict.get('section_start') or info_dict.get('section_end')):
|
info_dict.get('section_start') or info_dict.get('section_end')):
|
||||||
msg = ('This format cannot be partially downloaded' if merger.available
|
msg = ('This format cannot be partially downloaded' if FFmpegFD.available()
|
||||||
else 'You have requested downloading the video partially, but ffmpeg is not installed')
|
else 'You have requested downloading the video partially, but ffmpeg is not installed')
|
||||||
self.report_error(f'{msg}. Aborting')
|
self.report_error(f'{msg}. Aborting')
|
||||||
return
|
return
|
||||||
@ -3048,6 +3048,7 @@ def correct_ext(filename, ext=new_ext):
|
|||||||
dl_filename = existing_video_file(full_filename, temp_filename)
|
dl_filename = existing_video_file(full_filename, temp_filename)
|
||||||
info_dict['__real_download'] = False
|
info_dict['__real_download'] = False
|
||||||
|
|
||||||
|
merger = FFmpegMergerPP(self)
|
||||||
downloaded = []
|
downloaded = []
|
||||||
if dl_filename is not None:
|
if dl_filename is not None:
|
||||||
self.report_file_already_downloaded(dl_filename)
|
self.report_file_already_downloaded(dl_filename)
|
||||||
|
@ -12,6 +12,7 @@
|
|||||||
from .compat import compat_getpass, compat_shlex_quote
|
from .compat import compat_getpass, compat_shlex_quote
|
||||||
from .cookies import SUPPORTED_BROWSERS, SUPPORTED_KEYRINGS
|
from .cookies import SUPPORTED_BROWSERS, SUPPORTED_KEYRINGS
|
||||||
from .downloader import FileDownloader
|
from .downloader import FileDownloader
|
||||||
|
from .downloader.external import get_external_downloader
|
||||||
from .extractor import list_extractor_classes
|
from .extractor import list_extractor_classes
|
||||||
from .extractor.adobepass import MSO_INFO
|
from .extractor.adobepass import MSO_INFO
|
||||||
from .extractor.common import InfoExtractor
|
from .extractor.common import InfoExtractor
|
||||||
@ -39,6 +40,7 @@
|
|||||||
download_range_func,
|
download_range_func,
|
||||||
expand_path,
|
expand_path,
|
||||||
float_or_none,
|
float_or_none,
|
||||||
|
format_field,
|
||||||
int_or_none,
|
int_or_none,
|
||||||
match_filter_func,
|
match_filter_func,
|
||||||
parse_duration,
|
parse_duration,
|
||||||
@ -399,6 +401,10 @@ def metadataparser_actions(f):
|
|||||||
if opts.no_sponsorblock:
|
if opts.no_sponsorblock:
|
||||||
opts.sponsorblock_mark = opts.sponsorblock_remove = set()
|
opts.sponsorblock_mark = opts.sponsorblock_remove = set()
|
||||||
|
|
||||||
|
for proto, path in opts.external_downloader.items():
|
||||||
|
if get_external_downloader(path) is None:
|
||||||
|
raise ValueError(
|
||||||
|
f'No such {format_field(proto, None, "%s ", ignore="default")}external downloader "{path}"')
|
||||||
warnings, deprecation_warnings = [], []
|
warnings, deprecation_warnings = [], []
|
||||||
|
|
||||||
# Common mistake: -f best
|
# Common mistake: -f best
|
||||||
|
@ -704,12 +704,10 @@
|
|||||||
IqIE,
|
IqIE,
|
||||||
IqAlbumIE
|
IqAlbumIE
|
||||||
)
|
)
|
||||||
|
|
||||||
from .itprotv import (
|
from .itprotv import (
|
||||||
ITProTVIE,
|
ITProTVIE,
|
||||||
ITProTVCourseIE
|
ITProTVCourseIE
|
||||||
)
|
)
|
||||||
|
|
||||||
from .itv import (
|
from .itv import (
|
||||||
ITVIE,
|
ITVIE,
|
||||||
ITVBTCCIE,
|
ITVBTCCIE,
|
||||||
|
@ -16,7 +16,7 @@
|
|||||||
from ..utils import (
|
from ..utils import (
|
||||||
ExtractorError,
|
ExtractorError,
|
||||||
bytes_to_intlist,
|
bytes_to_intlist,
|
||||||
decode_base,
|
decode_base_n,
|
||||||
int_or_none,
|
int_or_none,
|
||||||
intlist_to_bytes,
|
intlist_to_bytes,
|
||||||
request_to_url,
|
request_to_url,
|
||||||
@ -123,7 +123,7 @@ def _get_videokey_from_ticket(self, ticket):
|
|||||||
'Content-Type': 'application/json',
|
'Content-Type': 'application/json',
|
||||||
})
|
})
|
||||||
|
|
||||||
res = decode_base(license_response['k'], self.STRTABLE)
|
res = decode_base_n(license_response['k'], table=self.STRTABLE)
|
||||||
encvideokey = bytes_to_intlist(struct.pack('>QQ', res >> 64, res & 0xffffffffffffffff))
|
encvideokey = bytes_to_intlist(struct.pack('>QQ', res >> 64, res & 0xffffffffffffffff))
|
||||||
|
|
||||||
h = hmac.new(
|
h = hmac.new(
|
||||||
|
@ -2817,7 +2817,7 @@ def extract_Initialization(source):
|
|||||||
base_url = ''
|
base_url = ''
|
||||||
for element in (representation, adaptation_set, period, mpd_doc):
|
for element in (representation, adaptation_set, period, mpd_doc):
|
||||||
base_url_e = element.find(_add_ns('BaseURL'))
|
base_url_e = element.find(_add_ns('BaseURL'))
|
||||||
if base_url_e is not None:
|
if base_url_e and base_url_e.text:
|
||||||
base_url = base_url_e.text + base_url
|
base_url = base_url_e.text + base_url
|
||||||
if re.match(r'^https?://', base_url):
|
if re.match(r'^https?://', base_url):
|
||||||
break
|
break
|
||||||
|
@ -73,9 +73,7 @@ def _real_extract(self, url):
|
|||||||
'display_id': slug,
|
'display_id': slug,
|
||||||
'title': traverse_obj(episode_info, 'title', 'name'),
|
'title': traverse_obj(episode_info, 'title', 'name'),
|
||||||
'description': episode_info.get('description'),
|
'description': episode_info.get('description'),
|
||||||
'creator': join_nonempty(
|
'creator': join_nonempty(('createdBy', 'firstName'), ('createdBy', 'lastName'), from_dict=episode_info, delim=' '),
|
||||||
traverse_obj(episode_info, ('createdBy','firstName')), traverse_obj(episode_info, ('createdBy','lastName')),
|
|
||||||
delim=' '),
|
|
||||||
'duration': float_or_none(episode_info.get('duration')),
|
'duration': float_or_none(episode_info.get('duration')),
|
||||||
'is_live': episode_info.get('isLive'),
|
'is_live': episode_info.get('isLive'),
|
||||||
'thumbnail': traverse_obj(episode_info, 'thumbnail', 'image', expected_type=url_or_none),
|
'thumbnail': traverse_obj(episode_info, 'thumbnail', 'image', expected_type=url_or_none),
|
||||||
|
@ -3116,6 +3116,7 @@ def _real_extract(self, url):
|
|||||||
wistia_urls = WistiaIE._extract_urls(webpage)
|
wistia_urls = WistiaIE._extract_urls(webpage)
|
||||||
if wistia_urls:
|
if wistia_urls:
|
||||||
playlist = self.playlist_from_matches(wistia_urls, video_id, video_title, ie=WistiaIE.ie_key())
|
playlist = self.playlist_from_matches(wistia_urls, video_id, video_title, ie=WistiaIE.ie_key())
|
||||||
|
playlist['entries'] = list(playlist['entries'])
|
||||||
for entry in playlist['entries']:
|
for entry in playlist['entries']:
|
||||||
entry.update({
|
entry.update({
|
||||||
'_type': 'url_transparent',
|
'_type': 'url_transparent',
|
||||||
|
@ -37,7 +37,7 @@ def md5_text(text):
|
|||||||
return hashlib.md5(text.encode('utf-8')).hexdigest()
|
return hashlib.md5(text.encode('utf-8')).hexdigest()
|
||||||
|
|
||||||
|
|
||||||
class IqiyiSDK(object):
|
class IqiyiSDK:
|
||||||
def __init__(self, target, ip, timestamp):
|
def __init__(self, target, ip, timestamp):
|
||||||
self.target = target
|
self.target = target
|
||||||
self.ip = ip
|
self.ip = ip
|
||||||
@ -131,7 +131,7 @@ def split_time_ip_sum(self):
|
|||||||
self.target = self.digit_sum(self.timestamp) + chunks[0] + compat_str(sum(ip))
|
self.target = self.digit_sum(self.timestamp) + chunks[0] + compat_str(sum(ip))
|
||||||
|
|
||||||
|
|
||||||
class IqiyiSDKInterpreter(object):
|
class IqiyiSDKInterpreter:
|
||||||
def __init__(self, sdk_code):
|
def __init__(self, sdk_code):
|
||||||
self.sdk_code = sdk_code
|
self.sdk_code = sdk_code
|
||||||
|
|
||||||
|
@ -146,6 +146,7 @@ def random_user_agent():
|
|||||||
|
|
||||||
|
|
||||||
NO_DEFAULT = object()
|
NO_DEFAULT = object()
|
||||||
|
IDENTITY = lambda x: x
|
||||||
|
|
||||||
ENGLISH_MONTH_NAMES = [
|
ENGLISH_MONTH_NAMES = [
|
||||||
'January', 'February', 'March', 'April', 'May', 'June',
|
'January', 'February', 'March', 'April', 'May', 'June',
|
||||||
@ -4744,22 +4745,42 @@ def pkcs1pad(data, length):
|
|||||||
return [0, 2] + pseudo_random + [0] + data
|
return [0, 2] + pseudo_random + [0] + data
|
||||||
|
|
||||||
|
|
||||||
def encode_base_n(num, n, table=None):
|
def _base_n_table(n, table):
|
||||||
FULL_TABLE = '0123456789abcdefghijklmnopqrstuvwxyzABCDEFGHIJKLMNOPQRSTUVWXYZ'
|
if not table and not n:
|
||||||
if not table:
|
raise ValueError('Either table or n must be specified')
|
||||||
table = FULL_TABLE[:n]
|
elif not table:
|
||||||
|
table = '0123456789abcdefghijklmnopqrstuvwxyzABCDEFGHIJKLMNOPQRSTUVWXYZ'[:n]
|
||||||
|
elif not n or n == len(table):
|
||||||
|
return table
|
||||||
|
raise ValueError(f'base {n} exceeds table length {len(table)}')
|
||||||
|
|
||||||
if n > len(table):
|
|
||||||
raise ValueError('base %d exceeds table length %d' % (n, len(table)))
|
|
||||||
|
|
||||||
if num == 0:
|
def encode_base_n(num, n=None, table=None):
|
||||||
|
"""Convert given int to a base-n string"""
|
||||||
|
table = _base_n_table(n)
|
||||||
|
if not num:
|
||||||
return table[0]
|
return table[0]
|
||||||
|
|
||||||
ret = ''
|
result, base = '', len(table)
|
||||||
while num:
|
while num:
|
||||||
ret = table[num % n] + ret
|
result = table[num % base] + result
|
||||||
num = num // n
|
num = num // result
|
||||||
return ret
|
return result
|
||||||
|
|
||||||
|
|
||||||
|
def decode_base_n(string, n=None, table=None):
|
||||||
|
"""Convert given base-n string to int"""
|
||||||
|
table = {char: index for index, char in enumerate(_base_n_table(n, table))}
|
||||||
|
result, base = 0, len(table)
|
||||||
|
for char in string:
|
||||||
|
result = result * base + table[char]
|
||||||
|
return result
|
||||||
|
|
||||||
|
|
||||||
|
def decode_base(value, digits):
|
||||||
|
write_string('DeprecationWarning: yt_dlp.utils.decode_base is deprecated '
|
||||||
|
'and may be removed in a future version. Use yt_dlp.decode_base_n instead')
|
||||||
|
return decode_base_n(value, table=digits)
|
||||||
|
|
||||||
|
|
||||||
def decode_packed_codes(code):
|
def decode_packed_codes(code):
|
||||||
@ -5062,11 +5083,11 @@ def to_high_limit_path(path):
|
|||||||
return path
|
return path
|
||||||
|
|
||||||
|
|
||||||
def format_field(obj, field=None, template='%s', ignore=NO_DEFAULT, default='', func=None):
|
def format_field(obj, field=None, template='%s', ignore=NO_DEFAULT, default='', func=IDENTITY):
|
||||||
val = traverse_obj(obj, *variadic(field))
|
val = traverse_obj(obj, *variadic(field))
|
||||||
if (not val and val != 0) if ignore is NO_DEFAULT else val in ignore:
|
if (not val and val != 0) if ignore is NO_DEFAULT else val in variadic(ignore):
|
||||||
return default
|
return default
|
||||||
return template % (func(val) if func else val)
|
return template % func(val)
|
||||||
|
|
||||||
|
|
||||||
def clean_podcast_url(url):
|
def clean_podcast_url(url):
|
||||||
@ -5207,10 +5228,8 @@ def _traverse_obj(obj, path, _current_depth=0):
|
|||||||
|
|
||||||
if isinstance(expected_type, type):
|
if isinstance(expected_type, type):
|
||||||
type_test = lambda val: val if isinstance(val, expected_type) else None
|
type_test = lambda val: val if isinstance(val, expected_type) else None
|
||||||
elif expected_type is not None:
|
|
||||||
type_test = expected_type
|
|
||||||
else:
|
else:
|
||||||
type_test = lambda val: val
|
type_test = expected_type or IDENTITY
|
||||||
|
|
||||||
for path in path_list:
|
for path in path_list:
|
||||||
depth = 0
|
depth = 0
|
||||||
@ -5243,17 +5262,6 @@ def variadic(x, allowed_types=(str, bytes, dict)):
|
|||||||
return x if isinstance(x, collections.abc.Iterable) and not isinstance(x, allowed_types) else (x,)
|
return x if isinstance(x, collections.abc.Iterable) and not isinstance(x, allowed_types) else (x,)
|
||||||
|
|
||||||
|
|
||||||
def decode_base(value, digits):
|
|
||||||
# This will convert given base-x string to scalar (long or int)
|
|
||||||
table = {char: index for index, char in enumerate(digits)}
|
|
||||||
result = 0
|
|
||||||
base = len(digits)
|
|
||||||
for chr in value:
|
|
||||||
result *= base
|
|
||||||
result += table[chr]
|
|
||||||
return result
|
|
||||||
|
|
||||||
|
|
||||||
def time_seconds(**kwargs):
|
def time_seconds(**kwargs):
|
||||||
t = datetime.datetime.now(datetime.timezone(datetime.timedelta(**kwargs)))
|
t = datetime.datetime.now(datetime.timezone(datetime.timedelta(**kwargs)))
|
||||||
return t.timestamp()
|
return t.timestamp()
|
||||||
@ -5327,7 +5335,7 @@ def number_of_digits(number):
|
|||||||
|
|
||||||
def join_nonempty(*values, delim='-', from_dict=None):
|
def join_nonempty(*values, delim='-', from_dict=None):
|
||||||
if from_dict is not None:
|
if from_dict is not None:
|
||||||
values = map(from_dict.get, values)
|
values = (traverse_obj(from_dict, variadic(v)) for v in values)
|
||||||
return delim.join(map(str, filter(None, values)))
|
return delim.join(map(str, filter(None, values)))
|
||||||
|
|
||||||
|
|
||||||
|
Loading…
Reference in New Issue
Block a user