mirror of
https://github.com/yt-dlp/yt-dlp.git
synced 2024-11-07 20:30:41 -05:00
[youtube] Improve video upload date handling (#3029)
* Don't prefer UTC upload date for past live streams/premieres * Improve regex (fixes a regression) Authored-by: coletdjnz
This commit is contained in:
parent
5ca764c506
commit
17322130a9
1 changed files with 90 additions and 91 deletions
|
@ -730,11 +730,11 @@ def _extract_time_text(self, renderer, *path_list):
|
||||||
timestamp = (
|
timestamp = (
|
||||||
unified_timestamp(text) or unified_timestamp(
|
unified_timestamp(text) or unified_timestamp(
|
||||||
self._search_regex(
|
self._search_regex(
|
||||||
(r'(?:.+|^)(?:live|premieres|ed|ing)(?:\s*on)?\s*(.+\d)', r'\w+[\s,\.-]*\w+[\s,\.-]+20\d{2}'),
|
(r'([a-z]+\s*\d{1,2},?\s*20\d{2})', r'(?:.+|^)(?:live|premieres|ed|ing)(?:\s*(?:on|for))?\s*(.+\d)'),
|
||||||
text.lower(), 'time text', default=None)))
|
text.lower(), 'time text', default=None)))
|
||||||
|
|
||||||
if text and timestamp is None:
|
if text and timestamp is None:
|
||||||
self.report_warning('Cannot parse localized time text' + bug_reports_message(), only_once=True)
|
self.report_warning(f"Cannot parse localized time text '{text}'" + bug_reports_message(), only_once=True)
|
||||||
return timestamp, text
|
return timestamp, text
|
||||||
|
|
||||||
def _extract_response(self, item_id, query, note='Downloading API JSON', headers=None,
|
def _extract_response(self, item_id, query, note='Downloading API JSON', headers=None,
|
||||||
|
@ -1204,7 +1204,7 @@ class YoutubeIE(YoutubeBaseInfoExtractor):
|
||||||
'id': 'Tq92D6wQ1mg',
|
'id': 'Tq92D6wQ1mg',
|
||||||
'title': '[MMD] Adios - EVERGLOW [+Motion DL]',
|
'title': '[MMD] Adios - EVERGLOW [+Motion DL]',
|
||||||
'ext': 'mp4',
|
'ext': 'mp4',
|
||||||
'upload_date': '20191227',
|
'upload_date': '20191228',
|
||||||
'uploader_id': 'UC1yoRdFoFJaCY-AGfD9W0wQ',
|
'uploader_id': 'UC1yoRdFoFJaCY-AGfD9W0wQ',
|
||||||
'uploader': 'Projekt Melody',
|
'uploader': 'Projekt Melody',
|
||||||
'description': 'md5:17eccca93a786d51bc67646756894066',
|
'description': 'md5:17eccca93a786d51bc67646756894066',
|
||||||
|
@ -1297,6 +1297,7 @@ class YoutubeIE(YoutubeBaseInfoExtractor):
|
||||||
},
|
},
|
||||||
'expected_warnings': [
|
'expected_warnings': [
|
||||||
'DASH manifest missing',
|
'DASH manifest missing',
|
||||||
|
'Some formats are possibly damaged'
|
||||||
]
|
]
|
||||||
},
|
},
|
||||||
# Olympics (https://github.com/ytdl-org/youtube-dl/issues/4431)
|
# Olympics (https://github.com/ytdl-org/youtube-dl/issues/4431)
|
||||||
|
@ -1569,7 +1570,7 @@ class YoutubeIE(YoutubeBaseInfoExtractor):
|
||||||
'title': 'md5:e41008789470fc2533a3252216f1c1d1',
|
'title': 'md5:e41008789470fc2533a3252216f1c1d1',
|
||||||
'description': 'md5:a677553cf0840649b731a3024aeff4cc',
|
'description': 'md5:a677553cf0840649b731a3024aeff4cc',
|
||||||
'duration': 721,
|
'duration': 721,
|
||||||
'upload_date': '20150127',
|
'upload_date': '20150128',
|
||||||
'uploader_id': 'BerkmanCenter',
|
'uploader_id': 'BerkmanCenter',
|
||||||
'uploader_url': r're:https?://(?:www\.)?youtube\.com/user/BerkmanCenter',
|
'uploader_url': r're:https?://(?:www\.)?youtube\.com/user/BerkmanCenter',
|
||||||
'uploader': 'The Berkman Klein Center for Internet & Society',
|
'uploader': 'The Berkman Klein Center for Internet & Society',
|
||||||
|
@ -1601,7 +1602,7 @@ class YoutubeIE(YoutubeBaseInfoExtractor):
|
||||||
'title': 'Democratic Socialism and Foreign Policy | Bernie Sanders',
|
'title': 'Democratic Socialism and Foreign Policy | Bernie Sanders',
|
||||||
'description': 'md5:13a2503d7b5904ef4b223aa101628f39',
|
'description': 'md5:13a2503d7b5904ef4b223aa101628f39',
|
||||||
'duration': 4060,
|
'duration': 4060,
|
||||||
'upload_date': '20151119',
|
'upload_date': '20151120',
|
||||||
'uploader': 'Bernie Sanders',
|
'uploader': 'Bernie Sanders',
|
||||||
'uploader_id': 'UCH1dpzjCEiGAt8CXkryhkZg',
|
'uploader_id': 'UCH1dpzjCEiGAt8CXkryhkZg',
|
||||||
'uploader_url': r're:https?://(?:www\.)?youtube\.com/channel/UCH1dpzjCEiGAt8CXkryhkZg',
|
'uploader_url': r're:https?://(?:www\.)?youtube\.com/channel/UCH1dpzjCEiGAt8CXkryhkZg',
|
||||||
|
@ -3565,86 +3566,84 @@ def process_language(container, base_url, lang_code, sub_name, query):
|
||||||
or self._extract_chapters_from_engagement_panel(initial_data, duration)
|
or self._extract_chapters_from_engagement_panel(initial_data, duration)
|
||||||
or None)
|
or None)
|
||||||
|
|
||||||
contents = try_get(
|
contents = traverse_obj(
|
||||||
initial_data,
|
initial_data, ('contents', 'twoColumnWatchNextResults', 'results', 'results', 'contents'),
|
||||||
lambda x: x['contents']['twoColumnWatchNextResults']['results']['results']['contents'],
|
expected_type=list, default=[])
|
||||||
list) or []
|
|
||||||
for content in contents:
|
|
||||||
vpir = content.get('videoPrimaryInfoRenderer')
|
|
||||||
if vpir:
|
|
||||||
info['upload_date'] = strftime_or_none(self._extract_time_text(vpir, 'dateText')[0], '%Y%m%d')
|
|
||||||
stl = vpir.get('superTitleLink')
|
|
||||||
if stl:
|
|
||||||
stl = self._get_text(stl)
|
|
||||||
if try_get(
|
|
||||||
vpir,
|
|
||||||
lambda x: x['superTitleIcon']['iconType']) == 'LOCATION_PIN':
|
|
||||||
info['location'] = stl
|
|
||||||
else:
|
|
||||||
mobj = re.search(r'(.+?)\s*S(\d+)\s*•\s*E(\d+)', stl)
|
|
||||||
if mobj:
|
|
||||||
info.update({
|
|
||||||
'series': mobj.group(1),
|
|
||||||
'season_number': int(mobj.group(2)),
|
|
||||||
'episode_number': int(mobj.group(3)),
|
|
||||||
})
|
|
||||||
for tlb in (try_get(
|
|
||||||
vpir,
|
|
||||||
lambda x: x['videoActions']['menuRenderer']['topLevelButtons'],
|
|
||||||
list) or []):
|
|
||||||
tbr = tlb.get('toggleButtonRenderer') or {}
|
|
||||||
for getter, regex in [(
|
|
||||||
lambda x: x['defaultText']['accessibility']['accessibilityData'],
|
|
||||||
r'(?P<count>[\d,]+)\s*(?P<type>(?:dis)?like)'), ([
|
|
||||||
lambda x: x['accessibility'],
|
|
||||||
lambda x: x['accessibilityData']['accessibilityData'],
|
|
||||||
], r'(?P<type>(?:dis)?like) this video along with (?P<count>[\d,]+) other people')]:
|
|
||||||
label = (try_get(tbr, getter, dict) or {}).get('label')
|
|
||||||
if label:
|
|
||||||
mobj = re.match(regex, label)
|
|
||||||
if mobj:
|
|
||||||
info[mobj.group('type') + '_count'] = str_to_int(mobj.group('count'))
|
|
||||||
break
|
|
||||||
sbr_tooltip = try_get(
|
|
||||||
vpir, lambda x: x['sentimentBar']['sentimentBarRenderer']['tooltip'])
|
|
||||||
if sbr_tooltip:
|
|
||||||
like_count, dislike_count = sbr_tooltip.split(' / ')
|
|
||||||
info.update({
|
|
||||||
'like_count': str_to_int(like_count),
|
|
||||||
'dislike_count': str_to_int(dislike_count),
|
|
||||||
})
|
|
||||||
vsir = content.get('videoSecondaryInfoRenderer')
|
|
||||||
if vsir:
|
|
||||||
vor = traverse_obj(vsir, ('owner', 'videoOwnerRenderer'))
|
|
||||||
info.update({
|
|
||||||
'channel': self._get_text(vor, 'title'),
|
|
||||||
'channel_follower_count': self._get_count(vor, 'subscriberCountText')})
|
|
||||||
|
|
||||||
rows = try_get(
|
vpir = get_first(contents, 'videoPrimaryInfoRenderer')
|
||||||
vsir,
|
if vpir:
|
||||||
lambda x: x['metadataRowContainer']['metadataRowContainerRenderer']['rows'],
|
stl = vpir.get('superTitleLink')
|
||||||
list) or []
|
if stl:
|
||||||
multiple_songs = False
|
stl = self._get_text(stl)
|
||||||
for row in rows:
|
if try_get(
|
||||||
if try_get(row, lambda x: x['metadataRowRenderer']['hasDividerLine']) is True:
|
vpir,
|
||||||
multiple_songs = True
|
lambda x: x['superTitleIcon']['iconType']) == 'LOCATION_PIN':
|
||||||
|
info['location'] = stl
|
||||||
|
else:
|
||||||
|
mobj = re.search(r'(.+?)\s*S(\d+)\s*•\s*E(\d+)', stl)
|
||||||
|
if mobj:
|
||||||
|
info.update({
|
||||||
|
'series': mobj.group(1),
|
||||||
|
'season_number': int(mobj.group(2)),
|
||||||
|
'episode_number': int(mobj.group(3)),
|
||||||
|
})
|
||||||
|
for tlb in (try_get(
|
||||||
|
vpir,
|
||||||
|
lambda x: x['videoActions']['menuRenderer']['topLevelButtons'],
|
||||||
|
list) or []):
|
||||||
|
tbr = tlb.get('toggleButtonRenderer') or {}
|
||||||
|
for getter, regex in [(
|
||||||
|
lambda x: x['defaultText']['accessibility']['accessibilityData'],
|
||||||
|
r'(?P<count>[\d,]+)\s*(?P<type>(?:dis)?like)'), ([
|
||||||
|
lambda x: x['accessibility'],
|
||||||
|
lambda x: x['accessibilityData']['accessibilityData'],
|
||||||
|
], r'(?P<type>(?:dis)?like) this video along with (?P<count>[\d,]+) other people')]:
|
||||||
|
label = (try_get(tbr, getter, dict) or {}).get('label')
|
||||||
|
if label:
|
||||||
|
mobj = re.match(regex, label)
|
||||||
|
if mobj:
|
||||||
|
info[mobj.group('type') + '_count'] = str_to_int(mobj.group('count'))
|
||||||
break
|
break
|
||||||
for row in rows:
|
sbr_tooltip = try_get(
|
||||||
mrr = row.get('metadataRowRenderer') or {}
|
vpir, lambda x: x['sentimentBar']['sentimentBarRenderer']['tooltip'])
|
||||||
mrr_title = mrr.get('title')
|
if sbr_tooltip:
|
||||||
if not mrr_title:
|
like_count, dislike_count = sbr_tooltip.split(' / ')
|
||||||
continue
|
info.update({
|
||||||
mrr_title = self._get_text(mrr, 'title')
|
'like_count': str_to_int(like_count),
|
||||||
mrr_contents_text = self._get_text(mrr, ('contents', 0))
|
'dislike_count': str_to_int(dislike_count),
|
||||||
if mrr_title == 'License':
|
})
|
||||||
info['license'] = mrr_contents_text
|
vsir = get_first(contents, 'videoSecondaryInfoRenderer')
|
||||||
elif not multiple_songs:
|
if vsir:
|
||||||
if mrr_title == 'Album':
|
vor = traverse_obj(vsir, ('owner', 'videoOwnerRenderer'))
|
||||||
info['album'] = mrr_contents_text
|
info.update({
|
||||||
elif mrr_title == 'Artist':
|
'channel': self._get_text(vor, 'title'),
|
||||||
info['artist'] = mrr_contents_text
|
'channel_follower_count': self._get_count(vor, 'subscriberCountText')})
|
||||||
elif mrr_title == 'Song':
|
|
||||||
info['track'] = mrr_contents_text
|
rows = try_get(
|
||||||
|
vsir,
|
||||||
|
lambda x: x['metadataRowContainer']['metadataRowContainerRenderer']['rows'],
|
||||||
|
list) or []
|
||||||
|
multiple_songs = False
|
||||||
|
for row in rows:
|
||||||
|
if try_get(row, lambda x: x['metadataRowRenderer']['hasDividerLine']) is True:
|
||||||
|
multiple_songs = True
|
||||||
|
break
|
||||||
|
for row in rows:
|
||||||
|
mrr = row.get('metadataRowRenderer') or {}
|
||||||
|
mrr_title = mrr.get('title')
|
||||||
|
if not mrr_title:
|
||||||
|
continue
|
||||||
|
mrr_title = self._get_text(mrr, 'title')
|
||||||
|
mrr_contents_text = self._get_text(mrr, ('contents', 0))
|
||||||
|
if mrr_title == 'License':
|
||||||
|
info['license'] = mrr_contents_text
|
||||||
|
elif not multiple_songs:
|
||||||
|
if mrr_title == 'Album':
|
||||||
|
info['album'] = mrr_contents_text
|
||||||
|
elif mrr_title == 'Artist':
|
||||||
|
info['artist'] = mrr_contents_text
|
||||||
|
elif mrr_title == 'Song':
|
||||||
|
info['track'] = mrr_contents_text
|
||||||
|
|
||||||
fallbacks = {
|
fallbacks = {
|
||||||
'channel': 'uploader',
|
'channel': 'uploader',
|
||||||
|
@ -3652,15 +3651,15 @@ def process_language(container, base_url, lang_code, sub_name, query):
|
||||||
'channel_url': 'uploader_url',
|
'channel_url': 'uploader_url',
|
||||||
}
|
}
|
||||||
|
|
||||||
# The upload date for scheduled and current live streams / premieres in microformats
|
# The upload date for scheduled, live and past live streams / premieres in microformats
|
||||||
# is generally the true upload date. Although not in UTC, we will prefer that in this case.
|
# may be different from the stream date. Although not in UTC, we will prefer it in this case.
|
||||||
# Note this changes to the published date when the stream/premiere has finished.
|
|
||||||
# See: https://github.com/yt-dlp/yt-dlp/pull/2223#issuecomment-1008485139
|
# See: https://github.com/yt-dlp/yt-dlp/pull/2223#issuecomment-1008485139
|
||||||
if not info.get('upload_date') or info.get('is_live') or info.get('live_status') == 'is_upcoming':
|
upload_date = (
|
||||||
info['upload_date'] = (
|
unified_strdate(get_first(microformats, 'uploadDate'))
|
||||||
unified_strdate(get_first(microformats, 'uploadDate'))
|
or unified_strdate(search_meta('uploadDate')))
|
||||||
or unified_strdate(search_meta('uploadDate'))
|
if not upload_date or (not info.get('is_live') and not info.get('was_live') and info.get('live_status') != 'is_upcoming'):
|
||||||
or info.get('upload_date'))
|
upload_date = strftime_or_none(self._extract_time_text(vpir, 'dateText')[0], '%Y%m%d')
|
||||||
|
info['upload_date'] = upload_date
|
||||||
|
|
||||||
for to, frm in fallbacks.items():
|
for to, frm in fallbacks.items():
|
||||||
if not info.get(to):
|
if not info.get(to):
|
||||||
|
|
Loading…
Reference in a new issue