[docs,cleanup] Fix linter and misc cleanup

Closes #2419
This commit is contained in:
pukkandan
2022-01-24 01:21:39 +05:30
parent bb66c24797
commit 88f23a18e0
10 changed files with 16 additions and 12 deletions

View File

@@ -484,6 +484,7 @@ class YoutubeDL(object):
extractor_args: A dictionary of arguments to be passed to the extractors.
See "EXTRACTOR ARGUMENTS" for details.
Eg: {'youtube': {'skip': ['dash', 'hls']}}
mark_watched: Mark videos watched (even with --simulate). Only for YouTube
youtube_include_dash_manifest: Deprecated - Use extractor_args instead.
If True (default), DASH manifests and related
data will be downloaded and processed by extractor.
@@ -3504,7 +3505,7 @@ class YoutubeDL(object):
delim=self._format_screen('\u2500', self.Styles.DELIM, '-', test_encoding=True))
def render_thumbnails_table(self, info_dict):
thumbnails = list(info_dict.get('thumbnails'))
thumbnails = list(info_dict.get('thumbnails') or [])
if not thumbnails:
return None
return render_table(

View File

@@ -300,7 +300,7 @@ class ABCIViewShowSeriesIE(InfoExtractor):
unescapeHTML(webpage_data).encode('utf-8').decode('unicode_escape'), show_id)
video_data = video_data['route']['pageData']['_embedded']
highlight = try_get(video_data, lambda x: ['highlightVideo']['shareUrl'])
highlight = try_get(video_data, lambda x: x['highlightVideo']['shareUrl'])
if not self._yes_playlist(show_id, bool(highlight), video_label='highlight video'):
return self.url_result(highlight, ie=ABCIViewIE.ie_key())

View File

@@ -440,7 +440,7 @@ class YoutubeWebArchiveIE(InfoExtractor):
}, {
'url': 'ytarchive:BaW_jenozKc:20050214000000',
'only_matching': True
},{
}, {
'url': 'ytarchive:BaW_jenozKc',
'only_matching': True
},

View File

@@ -65,7 +65,7 @@ class LiTVIE(InfoExtractor):
return self.playlist_result(all_episodes, content_id, episode_title)
def _real_extract(self, url):
url, data = unsmuggle_url(url, {})
url, smuggled_data = unsmuggle_url(url, {})
video_id = self._match_id(url)

View File

@@ -22,7 +22,6 @@ from ..utils import (
unified_strdate,
unsmuggle_url,
url_or_none,
urljoin,
)

View File

@@ -345,7 +345,7 @@ def create_parser():
help=(
'Use the specified HTTP/HTTPS/SOCKS proxy. To enable '
'SOCKS proxy, specify a proper scheme. For example '
'socks5://127.0.0.1:1080/. Pass in an empty string (--proxy "") '
'socks5://user:pass@127.0.0.1:1080/. Pass in an empty string (--proxy "") '
'for direct connection'))
network.add_option(
'--socket-timeout',

View File

@@ -66,7 +66,7 @@ class MetadataParserPP(PostProcessor):
self.write_debug(f'Searching for {out_re.pattern!r} in {template!r}')
match = out_re.search(data_to_parse)
if match is None:
self.report_warning(f'Could not interpret {inp!r} as {out!r}')
self.to_screen(f'Could not interpret {inp!r} as {out!r}')
return
for attribute, value in match.groupdict().items():
info[attribute] = value
@@ -80,7 +80,7 @@ class MetadataParserPP(PostProcessor):
def f(info):
val = info.get(field)
if val is None:
self.report_warning(f'Video does not have a {field}')
self.to_screen(f'Video does not have a {field}')
return
elif not isinstance(val, str):
self.report_warning(f'Cannot replace in field {field} since it is a {type(val).__name__}')