2014-11-26 07:05:11 -05:00
|
|
|
from __future__ import unicode_literals
|
|
|
|
|
2022-01-01 17:01:49 -05:00
|
|
|
import collections
|
2015-02-28 08:43:24 -05:00
|
|
|
import io
|
2021-07-10 14:56:35 -04:00
|
|
|
import itertools
|
2014-01-06 23:59:22 -05:00
|
|
|
import os
|
|
|
|
import subprocess
|
|
|
|
import time
|
2016-05-05 16:41:30 -04:00
|
|
|
import re
|
2021-01-27 10:02:51 -05:00
|
|
|
import json
|
2014-01-06 23:59:22 -05:00
|
|
|
|
|
|
|
from .common import AudioConversionError, PostProcessor
|
|
|
|
|
2021-10-17 23:49:25 -04:00
|
|
|
from ..compat import compat_str
|
2014-11-02 05:23:40 -05:00
|
|
|
from ..utils import (
|
2022-01-11 22:22:09 -05:00
|
|
|
determine_ext,
|
2021-09-01 16:55:16 -04:00
|
|
|
dfxp2srt,
|
2014-05-16 09:47:54 -04:00
|
|
|
encodeArgument,
|
2014-01-06 23:59:22 -05:00
|
|
|
encodeFilename,
|
2021-09-03 16:07:41 -04:00
|
|
|
float_or_none,
|
2021-11-03 14:53:48 -04:00
|
|
|
_get_exe_version_output,
|
|
|
|
detect_exe_version,
|
2014-10-26 11:46:34 -04:00
|
|
|
is_outdated_version,
|
2021-09-01 16:55:16 -04:00
|
|
|
ISO639Utils,
|
|
|
|
orderedSet,
|
2021-10-20 12:19:40 -04:00
|
|
|
Popen,
|
2014-01-06 23:59:22 -05:00
|
|
|
PostProcessingError,
|
|
|
|
prepend_extension,
|
2021-01-27 10:02:51 -05:00
|
|
|
replace_extension,
|
2021-09-01 16:55:16 -04:00
|
|
|
shell_quote,
|
2021-06-08 04:53:56 -04:00
|
|
|
traverse_obj,
|
2021-07-10 17:59:44 -04:00
|
|
|
variadic,
|
2021-11-14 17:33:41 -05:00
|
|
|
write_json_file,
|
2014-01-06 23:59:22 -05:00
|
|
|
)
|
|
|
|
|
|
|
|
|
2016-03-13 07:15:29 -04:00
|
|
|
EXT_TO_OUT_FORMATS = {
|
2017-03-16 07:50:45 -04:00
|
|
|
'aac': 'adts',
|
|
|
|
'flac': 'flac',
|
|
|
|
'm4a': 'ipod',
|
|
|
|
'mka': 'matroska',
|
|
|
|
'mkv': 'matroska',
|
|
|
|
'mpg': 'mpeg',
|
|
|
|
'ogv': 'ogg',
|
|
|
|
'ts': 'mpegts',
|
|
|
|
'wma': 'asf',
|
|
|
|
'wmv': 'asf',
|
2021-10-26 10:42:30 -04:00
|
|
|
'vtt': 'webvtt',
|
2017-03-16 07:50:45 -04:00
|
|
|
}
|
|
|
|
ACODECS = {
|
|
|
|
'mp3': 'libmp3lame',
|
|
|
|
'aac': 'aac',
|
|
|
|
'flac': 'flac',
|
|
|
|
'm4a': 'aac',
|
2017-10-01 17:43:25 -04:00
|
|
|
'opus': 'libopus',
|
2017-03-16 07:50:45 -04:00
|
|
|
'vorbis': 'libvorbis',
|
|
|
|
'wav': None,
|
2021-11-18 18:50:13 -05:00
|
|
|
'alac': None,
|
2016-03-13 07:15:29 -04:00
|
|
|
}
|
|
|
|
|
|
|
|
|
2014-01-06 23:59:22 -05:00
|
|
|
class FFmpegPostProcessorError(PostProcessingError):
|
|
|
|
pass
|
|
|
|
|
2014-07-22 20:55:06 -04:00
|
|
|
|
2014-01-06 23:59:22 -05:00
|
|
|
class FFmpegPostProcessor(PostProcessor):
|
2015-04-18 05:52:36 -04:00
|
|
|
def __init__(self, downloader=None):
|
2014-01-06 23:59:22 -05:00
|
|
|
PostProcessor.__init__(self, downloader)
|
2015-02-13 05:14:01 -05:00
|
|
|
self._determine_executables()
|
2014-01-06 23:59:22 -05:00
|
|
|
|
2014-10-26 11:46:34 -04:00
|
|
|
def check_version(self):
|
2015-02-17 11:26:41 -05:00
|
|
|
if not self.available:
|
2021-03-19 23:20:08 -04:00
|
|
|
raise FFmpegPostProcessorError('ffmpeg not found. Please install or provide the path using --ffmpeg-location')
|
2014-10-26 11:46:34 -04:00
|
|
|
|
2015-02-17 11:27:29 -05:00
|
|
|
required_version = '10-0' if self.basename == 'avconv' else '1.0'
|
2014-10-26 11:46:34 -04:00
|
|
|
if is_outdated_version(
|
2015-02-13 05:14:01 -05:00
|
|
|
self._versions[self.basename], required_version):
|
2014-11-26 07:05:11 -05:00
|
|
|
warning = 'Your copy of %s is outdated, update %s to version %s or newer if you encounter any errors.' % (
|
2015-02-13 05:14:01 -05:00
|
|
|
self.basename, self.basename, required_version)
|
2021-01-10 08:44:54 -05:00
|
|
|
self.report_warning(warning)
|
2014-10-26 11:46:34 -04:00
|
|
|
|
2021-11-03 17:40:49 -04:00
|
|
|
@staticmethod
|
|
|
|
def get_versions_and_features(downloader=None):
|
|
|
|
pp = FFmpegPostProcessor(downloader)
|
|
|
|
return pp._versions, pp._features
|
|
|
|
|
2014-01-06 23:59:22 -05:00
|
|
|
@staticmethod
|
2015-02-13 05:14:01 -05:00
|
|
|
def get_versions(downloader=None):
|
2021-11-03 17:40:49 -04:00
|
|
|
return FFmpegPostProcessor.get_version_and_features(downloader)[0]
|
2015-01-09 23:45:51 -05:00
|
|
|
|
2015-02-13 05:14:01 -05:00
|
|
|
def _determine_executables(self):
|
|
|
|
programs = ['avprobe', 'avconv', 'ffmpeg', 'ffprobe']
|
|
|
|
|
2021-11-03 14:53:48 -04:00
|
|
|
def get_ffmpeg_version(path, prog):
|
|
|
|
out = _get_exe_version_output(path, ['-bsfs'])
|
|
|
|
ver = detect_exe_version(out) if out else False
|
2019-01-11 09:09:44 -05:00
|
|
|
if ver:
|
|
|
|
regexs = [
|
2019-01-11 12:30:06 -05:00
|
|
|
r'(?:\d+:)?([0-9.]+)-[0-9]+ubuntu[0-9.]+$', # Ubuntu, see [1]
|
2019-01-11 11:47:23 -05:00
|
|
|
r'n([0-9.]+)$', # Arch Linux
|
2019-01-11 12:30:06 -05:00
|
|
|
# 1. http://www.ducea.com/2006/06/17/ubuntu-package-version-naming-explanation/
|
2019-01-11 09:09:44 -05:00
|
|
|
]
|
|
|
|
for regex in regexs:
|
|
|
|
mobj = re.match(regex, ver)
|
|
|
|
if mobj:
|
|
|
|
ver = mobj.group(1)
|
2021-11-03 14:53:48 -04:00
|
|
|
self._versions[prog] = ver
|
|
|
|
if prog != 'ffmpeg' or not out:
|
|
|
|
return
|
|
|
|
|
2021-11-03 17:40:49 -04:00
|
|
|
mobj = re.search(r'(?m)^\s+libavformat\s+(?:[0-9. ]+)\s+/\s+(?P<runtime>[0-9. ]+)', out)
|
|
|
|
lavf_runtime_version = mobj.group('runtime').replace(' ', '') if mobj else None
|
2021-11-03 16:54:12 -04:00
|
|
|
self._features = {
|
|
|
|
'fdk': '--enable-libfdk-aac' in out,
|
|
|
|
'setts': 'setts' in out.splitlines(),
|
2021-11-03 17:40:49 -04:00
|
|
|
'needs_adtstoasc': is_outdated_version(lavf_runtime_version, '57.56.100', False),
|
2021-11-03 16:54:12 -04:00
|
|
|
}
|
2019-01-11 09:09:44 -05:00
|
|
|
|
2015-02-13 05:14:01 -05:00
|
|
|
self.basename = None
|
|
|
|
self.probe_basename = None
|
|
|
|
self._paths = None
|
|
|
|
self._versions = None
|
2021-11-03 14:53:48 -04:00
|
|
|
self._features = {}
|
|
|
|
|
|
|
|
prefer_ffmpeg = self.get_param('prefer_ffmpeg', True)
|
|
|
|
location = self.get_param('ffmpeg_location')
|
|
|
|
if location is None:
|
|
|
|
self._paths = {p: p for p in programs}
|
|
|
|
else:
|
|
|
|
if not os.path.exists(location):
|
|
|
|
self.report_warning(
|
|
|
|
'ffmpeg-location %s does not exist! '
|
|
|
|
'Continuing without ffmpeg.' % (location))
|
|
|
|
self._versions = {}
|
|
|
|
return
|
|
|
|
elif os.path.isdir(location):
|
|
|
|
dirname, basename = location, None
|
|
|
|
else:
|
|
|
|
basename = os.path.splitext(os.path.basename(location))[0]
|
|
|
|
basename = next((p for p in programs if basename.startswith(p)), 'ffmpeg')
|
|
|
|
dirname = os.path.dirname(os.path.abspath(location))
|
|
|
|
if basename in ('ffmpeg', 'ffprobe'):
|
|
|
|
prefer_ffmpeg = True
|
|
|
|
|
|
|
|
self._paths = dict(
|
|
|
|
(p, os.path.join(dirname, p)) for p in programs)
|
|
|
|
if basename:
|
|
|
|
self._paths[basename] = location
|
|
|
|
|
|
|
|
self._versions = {}
|
|
|
|
for p in programs:
|
|
|
|
get_ffmpeg_version(self._paths[p], p)
|
2015-02-13 05:14:01 -05:00
|
|
|
|
2018-06-28 14:09:14 -04:00
|
|
|
if prefer_ffmpeg is False:
|
2014-10-26 11:31:52 -04:00
|
|
|
prefs = ('avconv', 'ffmpeg')
|
2018-06-28 14:09:14 -04:00
|
|
|
else:
|
|
|
|
prefs = ('ffmpeg', 'avconv')
|
2014-10-26 11:31:52 -04:00
|
|
|
for p in prefs:
|
|
|
|
if self._versions[p]:
|
2015-02-13 05:14:01 -05:00
|
|
|
self.basename = p
|
|
|
|
break
|
2014-01-08 11:53:34 -05:00
|
|
|
|
2018-06-28 14:09:14 -04:00
|
|
|
if prefer_ffmpeg is False:
|
2014-10-26 16:03:16 -04:00
|
|
|
prefs = ('avprobe', 'ffprobe')
|
2018-06-28 14:09:14 -04:00
|
|
|
else:
|
|
|
|
prefs = ('ffprobe', 'avprobe')
|
2014-10-26 16:03:16 -04:00
|
|
|
for p in prefs:
|
|
|
|
if self._versions[p]:
|
2015-02-13 05:14:01 -05:00
|
|
|
self.probe_basename = p
|
|
|
|
break
|
|
|
|
|
2021-11-29 12:46:06 -05:00
|
|
|
if self.basename == 'avconv':
|
|
|
|
self.deprecation_warning(
|
|
|
|
'Support for avconv is deprecated and may be removed in a future version. Use ffmpeg instead')
|
|
|
|
if self.probe_basename == 'avprobe':
|
|
|
|
self.deprecation_warning(
|
|
|
|
'Support for avprobe is deprecated and may be removed in a future version. Use ffprobe instead')
|
|
|
|
|
2015-02-17 11:26:41 -05:00
|
|
|
@property
|
2015-02-13 05:14:01 -05:00
|
|
|
def available(self):
|
|
|
|
return self.basename is not None
|
2014-10-26 16:03:16 -04:00
|
|
|
|
2015-02-13 05:14:01 -05:00
|
|
|
@property
|
|
|
|
def executable(self):
|
|
|
|
return self._paths[self.basename]
|
|
|
|
|
2015-04-03 08:09:50 -04:00
|
|
|
@property
|
|
|
|
def probe_available(self):
|
|
|
|
return self.probe_basename is not None
|
|
|
|
|
2015-02-13 05:14:01 -05:00
|
|
|
@property
|
|
|
|
def probe_executable(self):
|
|
|
|
return self._paths[self.probe_basename]
|
2014-01-08 11:53:34 -05:00
|
|
|
|
2022-01-11 22:22:09 -05:00
|
|
|
@staticmethod
|
|
|
|
def stream_copy_opts(copy=True, *, ext=None):
|
|
|
|
yield from ('-map', '0')
|
|
|
|
# Don't copy Apple TV chapters track, bin_data
|
|
|
|
# See https://github.com/yt-dlp/yt-dlp/issues/2, #19042, #19024, https://trac.ffmpeg.org/ticket/6016
|
2022-01-13 05:39:19 -05:00
|
|
|
yield from ('-dn', '-ignore_unknown')
|
2022-01-11 22:22:09 -05:00
|
|
|
if copy:
|
|
|
|
yield from ('-c', 'copy')
|
|
|
|
# For some reason, '-c copy -map 0' is not enough to copy subtitles
|
|
|
|
if ext in ('mp4', 'mov'):
|
|
|
|
yield from ('-c:s', 'mov_text')
|
|
|
|
|
2016-09-16 17:06:55 -04:00
|
|
|
def get_audio_codec(self, path):
|
2019-01-24 14:23:04 -05:00
|
|
|
if not self.probe_available and not self.available:
|
2021-03-19 23:20:08 -04:00
|
|
|
raise PostProcessingError('ffprobe and ffmpeg not found. Please install or provide the path using --ffmpeg-location')
|
2016-09-16 17:06:55 -04:00
|
|
|
try:
|
2019-01-24 14:23:04 -05:00
|
|
|
if self.probe_available:
|
|
|
|
cmd = [
|
|
|
|
encodeFilename(self.probe_executable, True),
|
|
|
|
encodeArgument('-show_streams')]
|
|
|
|
else:
|
|
|
|
cmd = [
|
|
|
|
encodeFilename(self.executable, True),
|
|
|
|
encodeArgument('-i')]
|
|
|
|
cmd.append(encodeFilename(self._ffmpeg_filename_argument(path), True))
|
2021-01-10 08:44:54 -05:00
|
|
|
self.write_debug('%s command line: %s' % (self.basename, shell_quote(cmd)))
|
2021-10-20 12:19:40 -04:00
|
|
|
handle = Popen(cmd, stdin=subprocess.PIPE, stdout=subprocess.PIPE, stderr=subprocess.PIPE)
|
|
|
|
stdout_data, stderr_data = handle.communicate_or_kill()
|
2019-01-24 14:23:04 -05:00
|
|
|
expected_ret = 0 if self.probe_available else 1
|
|
|
|
if handle.wait() != expected_ret:
|
2016-09-16 17:06:55 -04:00
|
|
|
return None
|
|
|
|
except (IOError, OSError):
|
|
|
|
return None
|
2019-01-24 14:23:04 -05:00
|
|
|
output = (stdout_data if self.probe_available else stderr_data).decode('ascii', 'ignore')
|
|
|
|
if self.probe_available:
|
|
|
|
audio_codec = None
|
|
|
|
for line in output.split('\n'):
|
|
|
|
if line.startswith('codec_name='):
|
|
|
|
audio_codec = line.split('=')[1].strip()
|
|
|
|
elif line.strip() == 'codec_type=audio' and audio_codec is not None:
|
|
|
|
return audio_codec
|
|
|
|
else:
|
|
|
|
# Stream #FILE_INDEX:STREAM_INDEX[STREAM_ID](LANGUAGE): CODEC_TYPE: CODEC_NAME
|
|
|
|
mobj = re.search(
|
|
|
|
r'Stream\s*#\d+:\d+(?:\[0x[0-9a-f]+\])?(?:\([a-z]{3}\))?:\s*Audio:\s*([0-9a-z]+)',
|
|
|
|
output)
|
|
|
|
if mobj:
|
|
|
|
return mobj.group(1)
|
2016-09-16 17:06:55 -04:00
|
|
|
return None
|
|
|
|
|
2021-01-27 10:02:51 -05:00
|
|
|
def get_metadata_object(self, path, opts=[]):
|
|
|
|
if self.probe_basename != 'ffprobe':
|
|
|
|
if self.probe_available:
|
|
|
|
self.report_warning('Only ffprobe is supported for metadata extraction')
|
2021-03-19 23:20:08 -04:00
|
|
|
raise PostProcessingError('ffprobe not found. Please install or provide the path using --ffmpeg-location')
|
2021-01-27 10:02:51 -05:00
|
|
|
self.check_version()
|
|
|
|
|
|
|
|
cmd = [
|
|
|
|
encodeFilename(self.probe_executable, True),
|
|
|
|
encodeArgument('-hide_banner'),
|
|
|
|
encodeArgument('-show_format'),
|
|
|
|
encodeArgument('-show_streams'),
|
|
|
|
encodeArgument('-print_format'),
|
|
|
|
encodeArgument('json'),
|
|
|
|
]
|
|
|
|
|
|
|
|
cmd += opts
|
|
|
|
cmd.append(encodeFilename(self._ffmpeg_filename_argument(path), True))
|
2021-02-11 23:34:04 -05:00
|
|
|
self.write_debug('ffprobe command line: %s' % shell_quote(cmd))
|
2021-10-20 12:19:40 -04:00
|
|
|
p = Popen(cmd, stdout=subprocess.PIPE, stderr=subprocess.PIPE, stdin=subprocess.PIPE)
|
2021-01-27 10:02:51 -05:00
|
|
|
stdout, stderr = p.communicate()
|
|
|
|
return json.loads(stdout.decode('utf-8', 'replace'))
|
|
|
|
|
|
|
|
def get_stream_number(self, path, keys, value):
|
|
|
|
streams = self.get_metadata_object(path)['streams']
|
|
|
|
num = next(
|
2021-06-08 04:53:56 -04:00
|
|
|
(i for i, stream in enumerate(streams) if traverse_obj(stream, keys, casesense=False) == value),
|
2021-01-27 10:02:51 -05:00
|
|
|
None)
|
|
|
|
return num, len(streams)
|
|
|
|
|
2021-11-27 08:49:02 -05:00
|
|
|
def _get_real_video_duration(self, filepath, fatal=True):
|
2021-09-03 16:07:41 -04:00
|
|
|
try:
|
2021-11-27 08:49:02 -05:00
|
|
|
duration = float_or_none(
|
|
|
|
traverse_obj(self.get_metadata_object(filepath), ('format', 'duration')))
|
|
|
|
if not duration:
|
2021-09-03 16:07:41 -04:00
|
|
|
raise PostProcessingError('ffprobe returned empty duration')
|
2021-11-27 08:49:02 -05:00
|
|
|
return duration
|
2021-09-03 16:07:41 -04:00
|
|
|
except PostProcessingError as e:
|
|
|
|
if fatal:
|
2021-11-27 08:49:02 -05:00
|
|
|
raise PostProcessingError(f'Unable to determine video duration: {e.msg}')
|
2021-09-03 16:07:41 -04:00
|
|
|
|
|
|
|
def _duration_mismatch(self, d1, d2):
|
|
|
|
if not d1 or not d2:
|
|
|
|
return None
|
2021-11-27 08:49:02 -05:00
|
|
|
# The duration is often only known to nearest second. So there can be <1sec disparity natually.
|
|
|
|
# Further excuse an additional <1sec difference.
|
|
|
|
return abs(d1 - d2) > 2
|
2021-09-03 16:07:41 -04:00
|
|
|
|
2021-07-10 18:37:25 -04:00
|
|
|
def run_ffmpeg_multiple_files(self, input_paths, out_path, opts, **kwargs):
|
2021-03-08 21:17:21 -05:00
|
|
|
return self.real_run_ffmpeg(
|
|
|
|
[(path, []) for path in input_paths],
|
2021-07-10 18:37:25 -04:00
|
|
|
[(out_path, opts)], **kwargs)
|
2021-03-08 21:17:21 -05:00
|
|
|
|
2021-07-10 18:37:25 -04:00
|
|
|
def real_run_ffmpeg(self, input_path_opts, output_path_opts, *, expected_retcodes=(0,)):
|
2014-10-26 11:46:34 -04:00
|
|
|
self.check_version()
|
2014-01-06 23:59:22 -05:00
|
|
|
|
2015-01-10 00:13:18 -05:00
|
|
|
oldest_mtime = min(
|
2021-07-10 14:56:35 -04:00
|
|
|
os.stat(encodeFilename(path)).st_mtime for path, _ in input_path_opts if path)
|
2015-01-10 00:10:18 -05:00
|
|
|
|
2021-10-09 17:26:30 -04:00
|
|
|
cmd = [encodeFilename(self.executable, True), encodeArgument('-y')]
|
2019-01-29 18:15:23 -05:00
|
|
|
# avconv does not have repeat option
|
|
|
|
if self.basename == 'ffmpeg':
|
|
|
|
cmd += [encodeArgument('-loglevel'), encodeArgument('repeat+info')]
|
2021-02-24 11:05:18 -05:00
|
|
|
|
2021-03-08 21:17:21 -05:00
|
|
|
def make_args(file, args, name, number):
|
|
|
|
keys = ['_%s%d' % (name, number), '_%s' % name]
|
2021-12-19 21:56:03 -05:00
|
|
|
if name == 'o':
|
|
|
|
args += ['-movflags', '+faststart']
|
2021-12-24 22:12:08 -05:00
|
|
|
if number == 1:
|
|
|
|
keys.append('')
|
2021-03-08 21:17:21 -05:00
|
|
|
args += self._configuration_args(self.basename, keys)
|
|
|
|
if name == 'i':
|
|
|
|
args.append('-i')
|
2021-02-24 11:05:18 -05:00
|
|
|
return (
|
2021-03-08 21:17:21 -05:00
|
|
|
[encodeArgument(arg) for arg in args]
|
2021-02-24 11:05:18 -05:00
|
|
|
+ [encodeFilename(self._ffmpeg_filename_argument(file), True)])
|
|
|
|
|
2021-03-08 21:17:21 -05:00
|
|
|
for arg_type, path_opts in (('i', input_path_opts), ('o', output_path_opts)):
|
2021-07-10 14:56:35 -04:00
|
|
|
cmd += itertools.chain.from_iterable(
|
|
|
|
make_args(path, list(opts), arg_type, i + 1)
|
|
|
|
for i, (path, opts) in enumerate(path_opts) if path)
|
2014-01-06 23:59:22 -05:00
|
|
|
|
2021-01-10 08:44:54 -05:00
|
|
|
self.write_debug('ffmpeg command line: %s' % shell_quote(cmd))
|
2021-10-20 12:19:40 -04:00
|
|
|
p = Popen(cmd, stdout=subprocess.PIPE, stderr=subprocess.PIPE, stdin=subprocess.PIPE)
|
|
|
|
stdout, stderr = p.communicate_or_kill()
|
2021-07-10 18:37:25 -04:00
|
|
|
if p.returncode not in variadic(expected_retcodes):
|
2021-01-27 10:02:51 -05:00
|
|
|
stderr = stderr.decode('utf-8', 'replace').strip()
|
2021-09-23 20:21:54 -04:00
|
|
|
self.write_debug(stderr)
|
2021-01-27 10:02:51 -05:00
|
|
|
raise FFmpegPostProcessorError(stderr.split('\n')[-1])
|
2021-03-08 21:17:21 -05:00
|
|
|
for out_path, _ in output_path_opts:
|
2021-07-10 14:56:35 -04:00
|
|
|
if out_path:
|
|
|
|
self.try_utime(out_path, oldest_mtime, oldest_mtime)
|
2021-01-27 10:02:51 -05:00
|
|
|
return stderr.decode('utf-8', 'replace')
|
2015-04-07 17:33:18 -04:00
|
|
|
|
2021-07-10 18:37:25 -04:00
|
|
|
def run_ffmpeg(self, path, out_path, opts, **kwargs):
|
|
|
|
return self.run_ffmpeg_multiple_files([path], out_path, opts, **kwargs)
|
2014-01-06 23:59:22 -05:00
|
|
|
|
2021-09-01 16:55:16 -04:00
|
|
|
@staticmethod
|
|
|
|
def _ffmpeg_filename_argument(fn):
|
2015-09-17 09:22:19 -04:00
|
|
|
# Always use 'file:' because the filename may contain ':' (ffmpeg
|
|
|
|
# interprets that as a protocol) or can start with '-' (-- is broken in
|
|
|
|
# ffmpeg, see https://ffmpeg.org/trac/ffmpeg/ticket/2127 for details)
|
2016-04-16 16:49:13 -04:00
|
|
|
# Also leave '-' intact in order not to break streaming to stdout.
|
2021-01-27 10:02:51 -05:00
|
|
|
if fn.startswith(('http://', 'https://')):
|
|
|
|
return fn
|
2016-04-16 14:45:56 -04:00
|
|
|
return 'file:' + fn if fn != '-' else fn
|
2014-01-06 23:59:22 -05:00
|
|
|
|
2021-09-01 16:55:16 -04:00
|
|
|
@staticmethod
|
|
|
|
def _quote_for_ffmpeg(string):
|
|
|
|
# See https://ffmpeg.org/ffmpeg-utils.html#toc-Quoting-and-escaping
|
|
|
|
# A sequence of '' produces '\'''\'';
|
|
|
|
# final replace removes the empty '' between \' \'.
|
|
|
|
string = string.replace("'", r"'\''").replace("'''", "'")
|
|
|
|
# Handle potential ' at string boundaries.
|
|
|
|
string = string[1:] if string[0] == "'" else "'" + string
|
|
|
|
return string[:-1] if string[-1] == "'" else string + "'"
|
|
|
|
|
|
|
|
def force_keyframes(self, filename, timestamps):
|
|
|
|
timestamps = orderedSet(timestamps)
|
|
|
|
if timestamps[0] == 0:
|
|
|
|
timestamps = timestamps[1:]
|
|
|
|
keyframe_file = prepend_extension(filename, 'keyframes.temp')
|
|
|
|
self.to_screen(f'Re-encoding "{filename}" with appropriate keyframes')
|
2022-01-11 22:22:09 -05:00
|
|
|
self.run_ffmpeg(filename, keyframe_file, [
|
|
|
|
*self.stream_copy_opts(False, ext=determine_ext(filename)),
|
|
|
|
'-force_key_frames', ','.join(f'{t:.6f}' for t in timestamps)])
|
2021-09-01 16:55:16 -04:00
|
|
|
return keyframe_file
|
|
|
|
|
|
|
|
def concat_files(self, in_files, out_file, concat_opts=None):
|
|
|
|
"""
|
|
|
|
Use concat demuxer to concatenate multiple files having identical streams.
|
|
|
|
|
|
|
|
Only inpoint, outpoint, and duration concat options are supported.
|
|
|
|
See https://ffmpeg.org/ffmpeg-formats.html#concat-1 for details
|
|
|
|
"""
|
|
|
|
concat_file = f'{out_file}.concat'
|
|
|
|
self.write_debug(f'Writing concat spec to {concat_file}')
|
|
|
|
with open(concat_file, 'wt', encoding='utf-8') as f:
|
|
|
|
f.writelines(self._concat_spec(in_files, concat_opts))
|
|
|
|
|
2022-01-11 22:22:09 -05:00
|
|
|
out_flags = list(self.stream_copy_opts(ext=determine_ext(out_file)))
|
2021-09-01 16:55:16 -04:00
|
|
|
|
2022-02-17 12:10:34 -05:00
|
|
|
self.real_run_ffmpeg(
|
|
|
|
[(concat_file, ['-hide_banner', '-nostdin', '-f', 'concat', '-safe', '0'])],
|
|
|
|
[(out_file, out_flags)])
|
|
|
|
os.remove(concat_file)
|
2021-09-01 16:55:16 -04:00
|
|
|
|
|
|
|
@classmethod
|
|
|
|
def _concat_spec(cls, in_files, concat_opts=None):
|
|
|
|
if concat_opts is None:
|
|
|
|
concat_opts = [{}] * len(in_files)
|
|
|
|
yield 'ffconcat version 1.0\n'
|
|
|
|
for file, opts in zip(in_files, concat_opts):
|
|
|
|
yield f'file {cls._quote_for_ffmpeg(cls._ffmpeg_filename_argument(file))}\n'
|
|
|
|
# Iterate explicitly to yield the following directives in order, ignoring the rest.
|
|
|
|
for directive in 'inpoint', 'outpoint', 'duration':
|
|
|
|
if directive in opts:
|
|
|
|
yield f'{directive} {opts[directive]}\n'
|
|
|
|
|
2014-01-06 23:59:22 -05:00
|
|
|
|
|
|
|
class FFmpegExtractAudioPP(FFmpegPostProcessor):
|
2021-05-22 04:24:12 -04:00
|
|
|
COMMON_AUDIO_EXTS = ('wav', 'flac', 'm4a', 'aiff', 'mp3', 'ogg', 'mka', 'opus', 'wma')
|
2021-11-18 18:50:13 -05:00
|
|
|
SUPPORTED_EXTS = ('best', 'aac', 'flac', 'mp3', 'm4a', 'opus', 'vorbis', 'wav', 'alac')
|
2021-02-15 12:46:11 -05:00
|
|
|
|
2014-01-06 23:59:22 -05:00
|
|
|
def __init__(self, downloader=None, preferredcodec=None, preferredquality=None, nopostoverwrites=False):
|
|
|
|
FFmpegPostProcessor.__init__(self, downloader)
|
2021-05-22 04:24:12 -04:00
|
|
|
self._preferredcodec = preferredcodec or 'best'
|
2021-11-03 14:35:53 -04:00
|
|
|
self._preferredquality = float_or_none(preferredquality)
|
2014-01-06 23:59:22 -05:00
|
|
|
self._nopostoverwrites = nopostoverwrites
|
|
|
|
|
2021-11-03 14:35:53 -04:00
|
|
|
def _quality_args(self, codec):
|
|
|
|
if self._preferredquality is None:
|
|
|
|
return []
|
|
|
|
elif self._preferredquality > 10:
|
|
|
|
return ['-b:a', f'{self._preferredquality}k']
|
|
|
|
|
|
|
|
limits = {
|
|
|
|
'libmp3lame': (10, 0),
|
2021-11-18 18:50:13 -05:00
|
|
|
'libvorbis': (0, 10),
|
2021-11-03 14:53:48 -04:00
|
|
|
# FFmpeg's AAC encoder does not have an upper limit for the value of -q:a.
|
|
|
|
# Experimentally, with values over 4, bitrate changes were minimal or non-existent
|
|
|
|
'aac': (0.1, 4),
|
2021-11-03 16:53:40 -04:00
|
|
|
'libfdk_aac': (1, 5),
|
2021-11-13 04:41:33 -05:00
|
|
|
}.get(codec)
|
2021-11-03 14:35:53 -04:00
|
|
|
if not limits:
|
|
|
|
return []
|
|
|
|
|
|
|
|
q = limits[1] + (limits[0] - limits[1]) * (self._preferredquality / 10)
|
2021-11-03 16:53:40 -04:00
|
|
|
if codec == 'libfdk_aac':
|
|
|
|
return ['-vbr', f'{int(q)}']
|
2021-11-03 14:35:53 -04:00
|
|
|
return ['-q:a', f'{q}']
|
|
|
|
|
2014-01-06 23:59:22 -05:00
|
|
|
def run_ffmpeg(self, path, out_path, codec, more_opts):
|
|
|
|
if codec is None:
|
|
|
|
acodec_opts = []
|
|
|
|
else:
|
|
|
|
acodec_opts = ['-acodec', codec]
|
|
|
|
opts = ['-vn'] + acodec_opts + more_opts
|
|
|
|
try:
|
|
|
|
FFmpegPostProcessor.run_ffmpeg(self, path, out_path, opts)
|
|
|
|
except FFmpegPostProcessorError as err:
|
|
|
|
raise AudioConversionError(err.msg)
|
|
|
|
|
2021-06-12 16:02:19 -04:00
|
|
|
@PostProcessor._restrict_to(images=False)
|
2014-01-06 23:59:22 -05:00
|
|
|
def run(self, information):
|
2021-11-18 18:50:13 -05:00
|
|
|
orig_path = path = information['filepath']
|
2021-02-15 12:46:11 -05:00
|
|
|
orig_ext = information['ext']
|
|
|
|
|
2021-05-22 04:24:12 -04:00
|
|
|
if self._preferredcodec == 'best' and orig_ext in self.COMMON_AUDIO_EXTS:
|
2021-02-15 12:46:11 -05:00
|
|
|
self.to_screen('Skipping audio extraction since the file is already in a common audio format')
|
2021-02-16 04:28:12 -05:00
|
|
|
return [], information
|
2014-01-06 23:59:22 -05:00
|
|
|
|
|
|
|
filecodec = self.get_audio_codec(path)
|
|
|
|
if filecodec is None:
|
2014-11-26 07:05:11 -05:00
|
|
|
raise PostProcessingError('WARNING: unable to obtain file audio codec with ffprobe')
|
2014-01-06 23:59:22 -05:00
|
|
|
|
|
|
|
more_opts = []
|
|
|
|
if self._preferredcodec == 'best' or self._preferredcodec == filecodec or (self._preferredcodec == 'm4a' and filecodec == 'aac'):
|
|
|
|
if filecodec == 'aac' and self._preferredcodec in ['m4a', 'best']:
|
|
|
|
# Lossless, but in another container
|
|
|
|
acodec = 'copy'
|
|
|
|
extension = 'm4a'
|
2015-02-06 16:05:11 -05:00
|
|
|
more_opts = ['-bsf:a', 'aac_adtstoasc']
|
2017-03-16 07:50:45 -04:00
|
|
|
elif filecodec in ['aac', 'flac', 'mp3', 'vorbis', 'opus']:
|
2014-01-06 23:59:22 -05:00
|
|
|
# Lossless if possible
|
|
|
|
acodec = 'copy'
|
|
|
|
extension = filecodec
|
|
|
|
if filecodec == 'aac':
|
|
|
|
more_opts = ['-f', 'adts']
|
|
|
|
if filecodec == 'vorbis':
|
|
|
|
extension = 'ogg'
|
2021-11-18 18:50:13 -05:00
|
|
|
elif filecodec == 'alac':
|
|
|
|
acodec = None
|
|
|
|
extension = 'm4a'
|
|
|
|
more_opts += ['-acodec', 'alac']
|
2014-01-06 23:59:22 -05:00
|
|
|
else:
|
|
|
|
# MP3 otherwise.
|
|
|
|
acodec = 'libmp3lame'
|
|
|
|
extension = 'mp3'
|
2021-11-03 14:35:53 -04:00
|
|
|
more_opts = self._quality_args(acodec)
|
2014-01-06 23:59:22 -05:00
|
|
|
else:
|
2017-03-16 07:50:45 -04:00
|
|
|
# We convert the audio (lossy if codec is lossy)
|
|
|
|
acodec = ACODECS[self._preferredcodec]
|
2021-11-03 16:53:40 -04:00
|
|
|
if acodec == 'aac' and self._features.get('fdk'):
|
|
|
|
acodec = 'libfdk_aac'
|
2014-01-06 23:59:22 -05:00
|
|
|
extension = self._preferredcodec
|
2021-11-03 14:35:53 -04:00
|
|
|
more_opts = self._quality_args(acodec)
|
2014-01-06 23:59:22 -05:00
|
|
|
if self._preferredcodec == 'aac':
|
|
|
|
more_opts += ['-f', 'adts']
|
2021-11-18 18:50:13 -05:00
|
|
|
elif self._preferredcodec == 'm4a':
|
2015-02-06 16:05:11 -05:00
|
|
|
more_opts += ['-bsf:a', 'aac_adtstoasc']
|
2021-11-18 18:50:13 -05:00
|
|
|
elif self._preferredcodec == 'vorbis':
|
2014-01-06 23:59:22 -05:00
|
|
|
extension = 'ogg'
|
2021-11-18 18:50:13 -05:00
|
|
|
elif self._preferredcodec == 'wav':
|
2014-01-06 23:59:22 -05:00
|
|
|
extension = 'wav'
|
|
|
|
more_opts += ['-f', 'wav']
|
2021-11-18 18:50:13 -05:00
|
|
|
elif self._preferredcodec == 'alac':
|
|
|
|
extension = 'm4a'
|
|
|
|
more_opts += ['-acodec', 'alac']
|
2014-01-06 23:59:22 -05:00
|
|
|
|
2014-11-26 07:05:11 -05:00
|
|
|
prefix, sep, ext = path.rpartition('.') # not os.path.splitext, since the latter does not work on unicode in all setups
|
2021-11-18 18:50:13 -05:00
|
|
|
temp_path = new_path = prefix + sep + extension
|
2014-01-06 23:59:22 -05:00
|
|
|
|
2021-11-18 18:50:13 -05:00
|
|
|
if new_path == path:
|
|
|
|
orig_path = prepend_extension(path, 'orig')
|
|
|
|
temp_path = prepend_extension(path, 'temp')
|
|
|
|
if (self._nopostoverwrites and os.path.exists(encodeFilename(new_path))
|
|
|
|
and os.path.exists(encodeFilename(orig_path))):
|
2021-01-07 14:28:41 -05:00
|
|
|
self.to_screen('Post-process file %s exists, skipping' % new_path)
|
2015-04-18 05:36:42 -04:00
|
|
|
return [], information
|
2014-01-06 23:59:22 -05:00
|
|
|
|
|
|
|
try:
|
2021-11-18 18:50:13 -05:00
|
|
|
self.to_screen(f'Destination: {new_path}')
|
|
|
|
self.run_ffmpeg(path, temp_path, acodec, more_opts)
|
2015-03-27 08:02:20 -04:00
|
|
|
except AudioConversionError as e:
|
|
|
|
raise PostProcessingError(
|
|
|
|
'audio conversion failed: ' + e.msg)
|
|
|
|
except Exception:
|
|
|
|
raise PostProcessingError('error running ' + self.basename)
|
2014-01-06 23:59:22 -05:00
|
|
|
|
2021-11-18 18:50:13 -05:00
|
|
|
os.replace(path, orig_path)
|
|
|
|
os.replace(temp_path, new_path)
|
|
|
|
information['filepath'] = new_path
|
|
|
|
information['ext'] = extension
|
|
|
|
|
2014-01-06 23:59:22 -05:00
|
|
|
# Try to update the date time for extracted audio file.
|
|
|
|
if information.get('filetime') is not None:
|
2015-04-08 11:40:31 -04:00
|
|
|
self.try_utime(
|
|
|
|
new_path, time.time(), information['filetime'],
|
|
|
|
errnote='Cannot update utime of audio file')
|
2014-01-06 23:59:22 -05:00
|
|
|
|
2021-11-18 18:50:13 -05:00
|
|
|
return [orig_path], information
|
2014-01-06 23:59:22 -05:00
|
|
|
|
|
|
|
|
2021-05-22 03:38:12 -04:00
|
|
|
class FFmpegVideoConvertorPP(FFmpegPostProcessor):
|
2021-05-22 04:24:12 -04:00
|
|
|
SUPPORTED_EXTS = ('mp4', 'mkv', 'flv', 'webm', 'mov', 'avi', 'mp3', 'mka', 'm4a', 'ogg', 'opus')
|
|
|
|
FORMAT_RE = re.compile(r'{0}(?:/{0})*$'.format(r'(?:\w+>)?(?:%s)' % '|'.join(SUPPORTED_EXTS)))
|
2021-09-17 14:23:55 -04:00
|
|
|
_ACTION = 'converting'
|
2021-05-22 03:38:12 -04:00
|
|
|
|
2020-05-16 12:09:12 -04:00
|
|
|
def __init__(self, downloader=None, preferedformat=None):
|
2021-05-22 03:38:12 -04:00
|
|
|
super(FFmpegVideoConvertorPP, self).__init__(downloader)
|
2021-01-27 10:02:51 -05:00
|
|
|
self._preferedformats = preferedformat.lower().split('/')
|
2020-05-16 12:09:12 -04:00
|
|
|
|
2021-05-22 03:38:12 -04:00
|
|
|
def _target_ext(self, source_ext):
|
2021-01-27 10:02:51 -05:00
|
|
|
for pair in self._preferedformats:
|
|
|
|
kv = pair.split('>')
|
2021-05-22 03:38:12 -04:00
|
|
|
if len(kv) == 1 or kv[0].strip() == source_ext:
|
|
|
|
return kv[-1].strip()
|
2021-01-27 10:02:51 -05:00
|
|
|
|
2021-05-22 03:38:12 -04:00
|
|
|
@staticmethod
|
|
|
|
def _options(target_ext):
|
|
|
|
if target_ext == 'avi':
|
|
|
|
return ['-c:v', 'libxvid', '-vtag', 'XVID']
|
|
|
|
return []
|
|
|
|
|
2021-06-12 16:02:19 -04:00
|
|
|
@PostProcessor._restrict_to(images=False)
|
2021-09-17 14:23:55 -04:00
|
|
|
def run(self, info):
|
|
|
|
filename, source_ext = info['filepath'], info['ext'].lower()
|
2021-05-22 04:24:12 -04:00
|
|
|
target_ext = self._target_ext(source_ext)
|
2021-01-27 10:02:51 -05:00
|
|
|
_skip_msg = (
|
2021-09-17 14:23:55 -04:00
|
|
|
f'could not find a mapping for {source_ext}' if not target_ext
|
|
|
|
else f'already is in target format {source_ext}' if source_ext == target_ext
|
2021-01-27 10:02:51 -05:00
|
|
|
else None)
|
|
|
|
if _skip_msg:
|
2022-01-19 17:57:36 -05:00
|
|
|
self.to_screen(f'Not {self._ACTION} media file "{filename}"; {_skip_msg}')
|
2021-09-17 14:23:55 -04:00
|
|
|
return [], info
|
2021-01-27 10:02:51 -05:00
|
|
|
|
2021-09-17 14:23:55 -04:00
|
|
|
outpath = replace_extension(filename, target_ext, source_ext)
|
|
|
|
self.to_screen(f'{self._ACTION.title()} video from {source_ext} to {target_ext}; Destination: {outpath}')
|
|
|
|
self.run_ffmpeg(filename, outpath, self._options(target_ext))
|
2021-05-22 03:38:12 -04:00
|
|
|
|
2021-09-17 14:23:55 -04:00
|
|
|
info['filepath'] = outpath
|
|
|
|
info['format'] = info['ext'] = target_ext
|
|
|
|
return [filename], info
|
2020-05-16 12:09:12 -04:00
|
|
|
|
|
|
|
|
2021-05-22 03:38:12 -04:00
|
|
|
class FFmpegVideoRemuxerPP(FFmpegVideoConvertorPP):
|
2021-09-17 14:23:55 -04:00
|
|
|
_ACTION = 'remuxing'
|
2014-01-06 23:59:22 -05:00
|
|
|
|
2021-05-22 03:38:12 -04:00
|
|
|
@staticmethod
|
|
|
|
def _options(target_ext):
|
2022-01-11 22:30:21 -05:00
|
|
|
return FFmpegPostProcessor.stream_copy_opts()
|
2014-01-06 23:59:22 -05:00
|
|
|
|
|
|
|
|
|
|
|
class FFmpegEmbedSubtitlePP(FFmpegPostProcessor):
|
2021-02-09 13:37:10 -05:00
|
|
|
def __init__(self, downloader=None, already_have_subtitle=False):
|
|
|
|
super(FFmpegEmbedSubtitlePP, self).__init__(downloader)
|
|
|
|
self._already_have_subtitle = already_have_subtitle
|
|
|
|
|
2021-06-12 16:02:19 -04:00
|
|
|
@PostProcessor._restrict_to(images=False)
|
2021-11-27 08:49:02 -05:00
|
|
|
def run(self, info):
|
|
|
|
if info['ext'] not in ('mp4', 'webm', 'mkv'):
|
2021-01-07 14:28:41 -05:00
|
|
|
self.to_screen('Subtitles can only be embedded in mp4, webm or mkv files')
|
2021-11-27 08:49:02 -05:00
|
|
|
return [], info
|
|
|
|
subtitles = info.get('requested_subtitles')
|
2015-02-16 15:12:31 -05:00
|
|
|
if not subtitles:
|
2021-01-07 14:28:41 -05:00
|
|
|
self.to_screen('There aren\'t any subtitles to embed')
|
2021-11-27 08:49:02 -05:00
|
|
|
return [], info
|
2014-01-06 23:59:22 -05:00
|
|
|
|
2021-11-27 08:49:02 -05:00
|
|
|
filename = info['filepath']
|
2021-12-02 09:13:41 -05:00
|
|
|
|
|
|
|
# Disabled temporarily. There needs to be a way to overide this
|
|
|
|
# in case of duration actually mismatching in extractor
|
|
|
|
# See: https://github.com/yt-dlp/yt-dlp/issues/1870, https://github.com/yt-dlp/yt-dlp/issues/1385
|
|
|
|
'''
|
2021-11-27 08:49:02 -05:00
|
|
|
if info.get('duration') and not info.get('__real_download') and self._duration_mismatch(
|
|
|
|
self._get_real_video_duration(filename, False), info['duration']):
|
2021-09-03 16:07:41 -04:00
|
|
|
self.to_screen(f'Skipping {self.pp_key()} since the real and expected durations mismatch')
|
2021-11-27 08:49:02 -05:00
|
|
|
return [], info
|
2021-12-02 09:13:41 -05:00
|
|
|
'''
|
2016-03-19 18:12:34 -04:00
|
|
|
|
2021-11-27 08:49:02 -05:00
|
|
|
ext = info['ext']
|
2021-05-12 15:37:58 -04:00
|
|
|
sub_langs, sub_names, sub_filenames = [], [], []
|
2016-03-19 18:12:34 -04:00
|
|
|
webm_vtt_warn = False
|
2021-01-27 10:02:51 -05:00
|
|
|
mp4_ass_warn = False
|
2016-03-19 18:12:34 -04:00
|
|
|
|
|
|
|
for lang, sub_info in subtitles.items():
|
2021-10-02 13:06:31 -04:00
|
|
|
if not os.path.exists(sub_info.get('filepath', '')):
|
2021-08-31 23:15:56 -04:00
|
|
|
self.report_warning(f'Skipping embedding {lang} subtitle because the file is missing')
|
|
|
|
continue
|
2016-03-19 18:12:34 -04:00
|
|
|
sub_ext = sub_info['ext']
|
2020-11-04 15:17:52 -05:00
|
|
|
if sub_ext == 'json':
|
2021-01-27 10:02:51 -05:00
|
|
|
self.report_warning('JSON subtitles cannot be embedded')
|
2020-11-04 15:17:52 -05:00
|
|
|
elif ext != 'webm' or ext == 'webm' and sub_ext == 'vtt':
|
2016-03-19 18:12:34 -04:00
|
|
|
sub_langs.append(lang)
|
2021-05-12 15:37:58 -04:00
|
|
|
sub_names.append(sub_info.get('name'))
|
2021-03-18 11:24:53 -04:00
|
|
|
sub_filenames.append(sub_info['filepath'])
|
2016-03-19 18:12:34 -04:00
|
|
|
else:
|
|
|
|
if not webm_vtt_warn and ext == 'webm' and sub_ext != 'vtt':
|
|
|
|
webm_vtt_warn = True
|
2021-01-27 10:02:51 -05:00
|
|
|
self.report_warning('Only WebVTT subtitles can be embedded in webm files')
|
|
|
|
if not mp4_ass_warn and ext == 'mp4' and sub_ext == 'ass':
|
|
|
|
mp4_ass_warn = True
|
|
|
|
self.report_warning('ASS subtitles cannot be properly embedded in mp4 files; expect issues')
|
2016-03-19 18:12:34 -04:00
|
|
|
|
|
|
|
if not sub_langs:
|
2021-11-27 08:49:02 -05:00
|
|
|
return [], info
|
2016-03-19 18:12:34 -04:00
|
|
|
|
2015-04-18 05:44:42 -04:00
|
|
|
input_files = [filename] + sub_filenames
|
2014-01-06 23:59:22 -05:00
|
|
|
|
2015-01-16 07:29:01 -05:00
|
|
|
opts = [
|
2022-01-11 22:22:09 -05:00
|
|
|
*self.stream_copy_opts(ext=info['ext']),
|
2015-01-16 07:29:01 -05:00
|
|
|
# Don't copy the existing subtitles, we may be running the
|
|
|
|
# postprocessor a second time
|
|
|
|
'-map', '-0:s',
|
|
|
|
]
|
2021-05-12 15:37:58 -04:00
|
|
|
for i, (lang, name) in enumerate(zip(sub_langs, sub_names)):
|
2015-01-16 07:37:37 -05:00
|
|
|
opts.extend(['-map', '%d:0' % (i + 1)])
|
2019-01-06 12:57:24 -05:00
|
|
|
lang_code = ISO639Utils.short2long(lang) or lang
|
|
|
|
opts.extend(['-metadata:s:s:%d' % i, 'language=%s' % lang_code])
|
2021-05-12 15:37:58 -04:00
|
|
|
if name:
|
|
|
|
opts.extend(['-metadata:s:s:%d' % i, 'handler_name=%s' % name,
|
|
|
|
'-metadata:s:s:%d' % i, 'title=%s' % name])
|
2014-01-06 23:59:22 -05:00
|
|
|
|
2015-01-16 07:37:37 -05:00
|
|
|
temp_filename = prepend_extension(filename, 'temp')
|
2021-01-27 10:02:51 -05:00
|
|
|
self.to_screen('Embedding subtitles in "%s"' % filename)
|
2014-01-06 23:59:22 -05:00
|
|
|
self.run_ffmpeg_multiple_files(input_files, temp_filename, opts)
|
2021-08-26 22:27:20 -04:00
|
|
|
os.replace(temp_filename, filename)
|
2014-01-06 23:59:22 -05:00
|
|
|
|
2021-02-09 13:37:10 -05:00
|
|
|
files_to_delete = [] if self._already_have_subtitle else sub_filenames
|
2021-11-27 08:49:02 -05:00
|
|
|
return files_to_delete, info
|
2014-01-06 23:59:22 -05:00
|
|
|
|
|
|
|
|
|
|
|
class FFmpegMetadataPP(FFmpegPostProcessor):
|
2021-07-10 14:56:35 -04:00
|
|
|
|
2021-11-14 17:33:41 -05:00
|
|
|
def __init__(self, downloader, add_metadata=True, add_chapters=True, add_infojson='if_exists'):
|
2021-09-01 16:55:16 -04:00
|
|
|
FFmpegPostProcessor.__init__(self, downloader)
|
|
|
|
self._add_metadata = add_metadata
|
|
|
|
self._add_chapters = add_chapters
|
2021-11-14 17:33:41 -05:00
|
|
|
self._add_infojson = add_infojson
|
2021-09-01 16:55:16 -04:00
|
|
|
|
2021-07-10 14:56:35 -04:00
|
|
|
@staticmethod
|
|
|
|
def _options(target_ext):
|
2022-01-11 22:22:09 -05:00
|
|
|
audio_only = target_ext == 'm4a'
|
2022-01-11 22:30:21 -05:00
|
|
|
yield from FFmpegPostProcessor.stream_copy_opts(not audio_only)
|
2022-01-11 22:22:09 -05:00
|
|
|
if audio_only:
|
2021-07-10 14:56:35 -04:00
|
|
|
yield from ('-vn', '-acodec', 'copy')
|
|
|
|
|
2021-06-12 16:02:19 -04:00
|
|
|
@PostProcessor._restrict_to(images=False)
|
2014-01-06 23:59:22 -05:00
|
|
|
def run(self, info):
|
2021-09-01 16:55:16 -04:00
|
|
|
filename, metadata_filename = info['filepath'], None
|
2021-11-14 17:33:41 -05:00
|
|
|
files_to_delete, options = [], []
|
2021-09-01 16:55:16 -04:00
|
|
|
if self._add_chapters and info.get('chapters'):
|
|
|
|
metadata_filename = replace_extension(filename, 'meta')
|
|
|
|
options.extend(self._get_chapter_opts(info['chapters'], metadata_filename))
|
2021-11-14 17:33:41 -05:00
|
|
|
files_to_delete.append(metadata_filename)
|
2021-09-01 16:55:16 -04:00
|
|
|
if self._add_metadata:
|
|
|
|
options.extend(self._get_metadata_opts(info))
|
|
|
|
|
2021-11-14 17:33:41 -05:00
|
|
|
if self._add_infojson:
|
|
|
|
if info['ext'] in ('mkv', 'mka'):
|
|
|
|
infojson_filename = info.get('infojson_filename')
|
|
|
|
options.extend(self._get_infojson_opts(info, infojson_filename))
|
|
|
|
if not infojson_filename:
|
|
|
|
files_to_delete.append(info.get('infojson_filename'))
|
|
|
|
elif self._add_infojson is True:
|
|
|
|
self.to_screen('The info-json can only be attached to mkv/mka files')
|
|
|
|
|
2021-09-01 16:55:16 -04:00
|
|
|
if not options:
|
|
|
|
self.to_screen('There isn\'t any metadata to add')
|
|
|
|
return [], info
|
|
|
|
|
|
|
|
temp_filename = prepend_extension(filename, 'temp')
|
|
|
|
self.to_screen('Adding metadata to "%s"' % filename)
|
|
|
|
self.run_ffmpeg_multiple_files(
|
|
|
|
(filename, metadata_filename), temp_filename,
|
|
|
|
itertools.chain(self._options(info['ext']), *options))
|
2021-11-14 17:33:41 -05:00
|
|
|
for file in filter(None, files_to_delete):
|
|
|
|
os.remove(file) # Don't obey --keep-files
|
2021-09-01 16:55:16 -04:00
|
|
|
os.replace(temp_filename, filename)
|
|
|
|
return [], info
|
|
|
|
|
|
|
|
@staticmethod
|
|
|
|
def _get_chapter_opts(chapters, metadata_filename):
|
|
|
|
with io.open(metadata_filename, 'wt', encoding='utf-8') as f:
|
|
|
|
def ffmpeg_escape(text):
|
|
|
|
return re.sub(r'([\\=;#\n])', r'\\\1', text)
|
|
|
|
|
|
|
|
metadata_file_content = ';FFMETADATA1\n'
|
|
|
|
for chapter in chapters:
|
|
|
|
metadata_file_content += '[CHAPTER]\nTIMEBASE=1/1000\n'
|
|
|
|
metadata_file_content += 'START=%d\n' % (chapter['start_time'] * 1000)
|
|
|
|
metadata_file_content += 'END=%d\n' % (chapter['end_time'] * 1000)
|
|
|
|
chapter_title = chapter.get('title')
|
|
|
|
if chapter_title:
|
|
|
|
metadata_file_content += 'title=%s\n' % ffmpeg_escape(chapter_title)
|
|
|
|
f.write(metadata_file_content)
|
|
|
|
yield ('-map_metadata', '1')
|
|
|
|
|
|
|
|
def _get_metadata_opts(self, info):
|
2022-01-01 17:01:49 -05:00
|
|
|
meta_prefix = 'meta'
|
|
|
|
metadata = collections.defaultdict(dict)
|
2016-05-01 00:56:54 -04:00
|
|
|
|
|
|
|
def add(meta_list, info_list=None):
|
2021-10-17 23:49:25 -04:00
|
|
|
value = next((
|
2022-01-01 17:01:49 -05:00
|
|
|
str(info[key]) for key in [f'{meta_prefix}_'] + list(variadic(info_list or meta_list))
|
2021-10-17 23:49:25 -04:00
|
|
|
if info.get(key) is not None), None)
|
|
|
|
if value not in ('', None):
|
2022-01-01 17:01:49 -05:00
|
|
|
metadata['common'].update({meta_f: value for meta_f in variadic(meta_list)})
|
2016-05-01 00:56:54 -04:00
|
|
|
|
2020-05-23 01:26:21 -04:00
|
|
|
# See [1-4] for some info on media metadata/metadata supported
|
|
|
|
# by ffmpeg.
|
|
|
|
# 1. https://kdenlive.org/en/project/adding-meta-data-to-mp4-video/
|
|
|
|
# 2. https://wiki.multimedia.cx/index.php/FFmpeg_Metadata
|
|
|
|
# 3. https://kodi.wiki/view/Video_file_tagging
|
|
|
|
|
2016-05-01 00:56:54 -04:00
|
|
|
add('title', ('track', 'title'))
|
|
|
|
add('date', 'upload_date')
|
2021-03-19 08:12:29 -04:00
|
|
|
add(('description', 'synopsis'), 'description')
|
|
|
|
add(('purl', 'comment'), 'webpage_url')
|
2016-05-01 00:56:54 -04:00
|
|
|
add('track', 'track_number')
|
|
|
|
add('artist', ('artist', 'creator', 'uploader', 'uploader_id'))
|
|
|
|
add('genre')
|
|
|
|
add('album')
|
|
|
|
add('album_artist')
|
|
|
|
add('disc', 'disc_number')
|
2020-05-23 01:26:21 -04:00
|
|
|
add('show', 'series')
|
|
|
|
add('season_number')
|
|
|
|
add('episode_id', ('episode', 'episode_id'))
|
|
|
|
add('episode_sort', 'episode_number')
|
2021-11-14 14:55:47 -05:00
|
|
|
if 'embed-metadata' in self.get_param('compat_opts', []):
|
|
|
|
add('comment', 'description')
|
2022-01-01 17:01:49 -05:00
|
|
|
metadata['common'].pop('synopsis', None)
|
2014-01-06 23:59:22 -05:00
|
|
|
|
2022-01-01 17:01:49 -05:00
|
|
|
meta_regex = rf'{re.escape(meta_prefix)}(?P<i>\d+)?_(?P<key>.+)'
|
2021-10-17 23:49:25 -04:00
|
|
|
for key, value in info.items():
|
2022-01-01 17:01:49 -05:00
|
|
|
mobj = re.fullmatch(meta_regex, key)
|
|
|
|
if value is not None and mobj:
|
|
|
|
metadata[mobj.group('i') or 'common'][mobj.group('key')] = value
|
2021-04-03 04:29:55 -04:00
|
|
|
|
2022-01-01 17:01:49 -05:00
|
|
|
for name, value in metadata['common'].items():
|
2021-09-01 16:55:16 -04:00
|
|
|
yield ('-metadata', f'{name}={value}')
|
2014-02-22 12:23:30 -05:00
|
|
|
|
2021-07-10 14:56:35 -04:00
|
|
|
stream_idx = 0
|
|
|
|
for fmt in info.get('requested_formats') or []:
|
|
|
|
stream_count = 2 if 'none' not in (fmt.get('vcodec'), fmt.get('acodec')) else 1
|
2022-01-03 10:09:46 -05:00
|
|
|
lang = ISO639Utils.short2long(fmt.get('language') or '') or fmt.get('language')
|
2022-01-01 17:01:49 -05:00
|
|
|
for i in range(stream_idx, stream_idx + stream_count):
|
|
|
|
if lang:
|
|
|
|
metadata[str(i)].setdefault('language', lang)
|
|
|
|
for name, value in metadata[str(i)].items():
|
|
|
|
yield (f'-metadata:s:{i}', f'{name}={value}')
|
2021-07-10 14:56:35 -04:00
|
|
|
stream_idx += stream_count
|
2014-01-06 23:59:22 -05:00
|
|
|
|
2021-11-14 17:33:41 -05:00
|
|
|
def _get_infojson_opts(self, info, infofn):
|
|
|
|
if not infofn or not os.path.exists(infofn):
|
|
|
|
if self._add_infojson is not True:
|
|
|
|
return
|
|
|
|
infofn = infofn or '%s.temp' % (
|
|
|
|
self._downloader.prepare_filename(info, 'infojson')
|
|
|
|
or replace_extension(self._downloader.prepare_filename(info), 'info.json', info['ext']))
|
|
|
|
if not self._downloader._ensure_dir_exists(infofn):
|
|
|
|
return
|
|
|
|
self.write_debug(f'Writing info-json to: {infofn}')
|
|
|
|
write_json_file(self._downloader.sanitize_info(info, self.get_param('clean_infojson', True)), infofn)
|
|
|
|
info['infojson_filename'] = infofn
|
|
|
|
|
|
|
|
old_stream, new_stream = self.get_stream_number(info['filepath'], ('tags', 'mimetype'), 'application/json')
|
|
|
|
if old_stream is not None:
|
|
|
|
yield ('-map', '-0:%d' % old_stream)
|
|
|
|
new_stream -= 1
|
2021-01-27 10:02:51 -05:00
|
|
|
|
2021-11-14 17:33:41 -05:00
|
|
|
yield ('-attach', infofn,
|
|
|
|
'-metadata:s:%d' % new_stream, 'mimetype=application/json')
|
2014-01-06 23:59:22 -05:00
|
|
|
|
|
|
|
|
|
|
|
class FFmpegMergerPP(FFmpegPostProcessor):
|
2021-06-12 16:02:19 -04:00
|
|
|
@PostProcessor._restrict_to(images=False)
|
2014-01-06 23:59:22 -05:00
|
|
|
def run(self, info):
|
|
|
|
filename = info['filepath']
|
2015-04-19 10:56:22 -04:00
|
|
|
temp_filename = prepend_extension(filename, 'temp')
|
2015-08-04 03:07:44 -04:00
|
|
|
args = ['-c', 'copy']
|
2021-09-23 02:18:49 -04:00
|
|
|
audio_streams = 0
|
2015-08-04 03:07:44 -04:00
|
|
|
for (i, fmt) in enumerate(info['requested_formats']):
|
|
|
|
if fmt.get('acodec') != 'none':
|
2021-09-22 10:21:40 -04:00
|
|
|
args.extend(['-map', f'{i}:a:0'])
|
2021-10-11 05:57:00 -04:00
|
|
|
aac_fixup = fmt['protocol'].startswith('m3u8') and self.get_audio_codec(fmt['filepath']) == 'aac'
|
|
|
|
if aac_fixup:
|
2021-09-23 02:18:49 -04:00
|
|
|
args.extend([f'-bsf:a:{audio_streams}', 'aac_adtstoasc'])
|
|
|
|
audio_streams += 1
|
2015-08-04 03:07:44 -04:00
|
|
|
if fmt.get('vcodec') != 'none':
|
|
|
|
args.extend(['-map', '%u:v:0' % (i)])
|
2021-01-07 14:28:41 -05:00
|
|
|
self.to_screen('Merging formats into "%s"' % filename)
|
2015-04-19 10:56:22 -04:00
|
|
|
self.run_ffmpeg_multiple_files(info['__files_to_merge'], temp_filename, args)
|
|
|
|
os.rename(encodeFilename(temp_filename), encodeFilename(filename))
|
2015-04-18 05:52:36 -04:00
|
|
|
return info['__files_to_merge'], info
|
2014-01-06 23:59:22 -05:00
|
|
|
|
2015-05-10 16:00:31 -04:00
|
|
|
def can_merge(self):
|
|
|
|
# TODO: figure out merge-capable ffmpeg version
|
|
|
|
if self.basename != 'avconv':
|
|
|
|
return True
|
|
|
|
|
|
|
|
required_version = '10-0'
|
|
|
|
if is_outdated_version(
|
|
|
|
self._versions[self.basename], required_version):
|
|
|
|
warning = ('Your copy of %s is outdated and unable to properly mux separate video and audio files, '
|
2021-02-24 13:45:56 -05:00
|
|
|
'yt-dlp will download single file media. '
|
2015-05-10 16:00:31 -04:00
|
|
|
'Update %s to version %s or newer to fix this.') % (
|
|
|
|
self.basename, self.basename, required_version)
|
2021-01-10 08:44:54 -05:00
|
|
|
self.report_warning(warning)
|
2015-05-10 16:00:31 -04:00
|
|
|
return False
|
|
|
|
return True
|
|
|
|
|
2014-02-22 07:55:51 -05:00
|
|
|
|
2021-06-19 18:19:23 -04:00
|
|
|
class FFmpegFixupPostProcessor(FFmpegPostProcessor):
|
|
|
|
def _fixup(self, msg, filename, options):
|
2015-01-09 23:45:51 -05:00
|
|
|
temp_filename = prepend_extension(filename, 'temp')
|
|
|
|
|
2021-06-19 18:45:19 -04:00
|
|
|
self.to_screen(f'{msg} of "{filename}"')
|
2015-01-09 23:45:51 -05:00
|
|
|
self.run_ffmpeg(filename, temp_filename, options)
|
|
|
|
|
2021-08-26 22:27:20 -04:00
|
|
|
os.replace(temp_filename, filename)
|
2015-01-09 23:45:51 -05:00
|
|
|
|
2021-06-19 18:19:23 -04:00
|
|
|
|
|
|
|
class FFmpegFixupStretchedPP(FFmpegFixupPostProcessor):
|
|
|
|
@PostProcessor._restrict_to(images=False, audio=False)
|
|
|
|
def run(self, info):
|
|
|
|
stretched_ratio = info.get('stretched_ratio')
|
|
|
|
if stretched_ratio not in (None, 1):
|
|
|
|
self._fixup('Fixing aspect ratio', info['filepath'], [
|
2022-01-11 22:22:09 -05:00
|
|
|
*self.stream_copy_opts(), '-aspect', '%f' % stretched_ratio])
|
2015-04-18 05:36:42 -04:00
|
|
|
return [], info
|
2015-01-23 12:39:12 -05:00
|
|
|
|
|
|
|
|
2021-06-19 18:19:23 -04:00
|
|
|
class FFmpegFixupM4aPP(FFmpegFixupPostProcessor):
|
2021-06-12 16:02:19 -04:00
|
|
|
@PostProcessor._restrict_to(images=False, video=False)
|
2015-01-23 12:39:12 -05:00
|
|
|
def run(self, info):
|
2021-06-19 18:19:23 -04:00
|
|
|
if info.get('container') == 'm4a_dash':
|
2022-01-11 22:22:09 -05:00
|
|
|
self._fixup('Correcting container', info['filepath'], [*self.stream_copy_opts(), '-f', 'mp4'])
|
2015-04-18 05:36:42 -04:00
|
|
|
return [], info
|
2015-02-28 08:43:24 -05:00
|
|
|
|
|
|
|
|
2021-06-19 18:19:23 -04:00
|
|
|
class FFmpegFixupM3u8PP(FFmpegFixupPostProcessor):
|
2021-11-27 08:50:39 -05:00
|
|
|
def _needs_fixup(self, info):
|
|
|
|
yield info['ext'] in ('mp4', 'm4a')
|
|
|
|
yield info['protocol'].startswith('m3u8')
|
|
|
|
try:
|
|
|
|
metadata = self.get_metadata_object(info['filepath'])
|
|
|
|
except PostProcessingError as e:
|
|
|
|
self.report_warning(f'Unable to extract metadata: {e.msg}')
|
|
|
|
yield True
|
|
|
|
else:
|
|
|
|
yield traverse_obj(metadata, ('format', 'format_name'), casesense=False) == 'mpegts'
|
|
|
|
|
2021-06-12 16:02:19 -04:00
|
|
|
@PostProcessor._restrict_to(images=False)
|
2016-03-01 15:08:50 -05:00
|
|
|
def run(self, info):
|
2021-11-27 08:50:39 -05:00
|
|
|
if all(self._needs_fixup(info)):
|
|
|
|
self._fixup('Fixing MPEG-TS in MP4 container', info['filepath'], [
|
2022-01-11 22:22:09 -05:00
|
|
|
*self.stream_copy_opts(), '-f', 'mp4', '-bsf:a', 'aac_adtstoasc'])
|
2016-03-01 15:08:50 -05:00
|
|
|
return [], info
|
|
|
|
|
|
|
|
|
2021-06-21 13:23:17 -04:00
|
|
|
class FFmpegFixupTimestampPP(FFmpegFixupPostProcessor):
|
|
|
|
|
|
|
|
def __init__(self, downloader=None, trim=0.001):
|
|
|
|
# "trim" should be used when the video contains unintended packets
|
|
|
|
super(FFmpegFixupTimestampPP, self).__init__(downloader)
|
|
|
|
assert isinstance(trim, (int, float))
|
|
|
|
self.trim = str(trim)
|
|
|
|
|
|
|
|
@PostProcessor._restrict_to(images=False)
|
|
|
|
def run(self, info):
|
2021-11-03 16:54:12 -04:00
|
|
|
if not self._features.get('setts'):
|
2021-06-21 13:23:17 -04:00
|
|
|
self.report_warning(
|
|
|
|
'A re-encode is needed to fix timestamps in older versions of ffmpeg. '
|
2021-11-03 16:54:12 -04:00
|
|
|
'Please install ffmpeg 4.4 or later to fixup without re-encoding')
|
2021-06-21 13:23:17 -04:00
|
|
|
opts = ['-vf', 'setpts=PTS-STARTPTS']
|
|
|
|
else:
|
|
|
|
opts = ['-c', 'copy', '-bsf', 'setts=ts=TS-STARTPTS']
|
2022-01-11 22:22:09 -05:00
|
|
|
self._fixup('Fixing frame timestamp', info['filepath'], opts + [*self.stream_copy_opts(False), '-ss', self.trim])
|
2021-06-21 13:23:17 -04:00
|
|
|
return [], info
|
|
|
|
|
|
|
|
|
2022-01-19 17:57:36 -05:00
|
|
|
class FFmpegCopyStreamPP(FFmpegFixupPostProcessor):
|
2021-12-20 01:06:46 -05:00
|
|
|
MESSAGE = 'Copying stream'
|
|
|
|
|
2021-06-21 13:23:17 -04:00
|
|
|
@PostProcessor._restrict_to(images=False)
|
|
|
|
def run(self, info):
|
2022-01-11 22:22:09 -05:00
|
|
|
self._fixup(self.MESSAGE, info['filepath'], self.stream_copy_opts())
|
2021-06-21 13:23:17 -04:00
|
|
|
return [], info
|
|
|
|
|
|
|
|
|
2022-01-19 17:57:36 -05:00
|
|
|
class FFmpegFixupDurationPP(FFmpegCopyStreamPP):
|
2021-12-20 01:06:46 -05:00
|
|
|
MESSAGE = 'Fixing video duration'
|
|
|
|
|
|
|
|
|
2022-01-19 17:57:36 -05:00
|
|
|
class FFmpegFixupDuplicateMoovPP(FFmpegCopyStreamPP):
|
2021-12-20 01:06:46 -05:00
|
|
|
MESSAGE = 'Fixing duplicate MOOV atoms'
|
|
|
|
|
|
|
|
|
2015-02-28 08:43:24 -05:00
|
|
|
class FFmpegSubtitlesConvertorPP(FFmpegPostProcessor):
|
2021-05-22 04:24:12 -04:00
|
|
|
SUPPORTED_EXTS = ('srt', 'vtt', 'ass', 'lrc')
|
|
|
|
|
2015-02-28 08:43:24 -05:00
|
|
|
def __init__(self, downloader=None, format=None):
|
|
|
|
super(FFmpegSubtitlesConvertorPP, self).__init__(downloader)
|
|
|
|
self.format = format
|
|
|
|
|
|
|
|
def run(self, info):
|
|
|
|
subs = info.get('requested_subtitles')
|
|
|
|
new_ext = self.format
|
|
|
|
new_format = new_ext
|
|
|
|
if new_format == 'vtt':
|
|
|
|
new_format = 'webvtt'
|
|
|
|
if subs is None:
|
2021-01-07 14:28:41 -05:00
|
|
|
self.to_screen('There aren\'t any subtitles to convert')
|
2015-04-18 05:36:42 -04:00
|
|
|
return [], info
|
2021-01-07 14:28:41 -05:00
|
|
|
self.to_screen('Converting subtitles')
|
2016-01-31 08:22:36 -05:00
|
|
|
sub_filenames = []
|
2015-02-28 08:43:24 -05:00
|
|
|
for lang, sub in subs.items():
|
2021-10-02 13:06:31 -04:00
|
|
|
if not os.path.exists(sub.get('filepath', '')):
|
|
|
|
self.report_warning(f'Skipping embedding {lang} subtitle because the file is missing')
|
|
|
|
continue
|
2015-02-28 08:43:24 -05:00
|
|
|
ext = sub['ext']
|
|
|
|
if ext == new_ext:
|
2021-01-07 14:28:41 -05:00
|
|
|
self.to_screen('Subtitle file for %s is already in the requested format' % new_ext)
|
2015-02-28 08:43:24 -05:00
|
|
|
continue
|
2020-11-04 15:17:52 -05:00
|
|
|
elif ext == 'json':
|
2021-01-07 14:28:41 -05:00
|
|
|
self.to_screen(
|
|
|
|
'You have requested to convert json subtitles into another format, '
|
2020-11-04 15:17:52 -05:00
|
|
|
'which is currently not possible')
|
|
|
|
continue
|
2021-03-18 11:24:53 -04:00
|
|
|
old_file = sub['filepath']
|
2016-01-31 08:22:36 -05:00
|
|
|
sub_filenames.append(old_file)
|
2021-03-18 11:24:53 -04:00
|
|
|
new_file = replace_extension(old_file, new_ext)
|
2015-04-25 11:15:05 -04:00
|
|
|
|
2017-04-12 15:38:43 -04:00
|
|
|
if ext in ('dfxp', 'ttml', 'tt'):
|
2021-01-10 08:44:54 -05:00
|
|
|
self.report_warning(
|
2021-01-07 14:28:41 -05:00
|
|
|
'You have requested to convert dfxp (TTML) subtitles into another format, '
|
2015-04-25 11:15:05 -04:00
|
|
|
'which results in style information loss')
|
|
|
|
|
2016-01-31 08:22:36 -05:00
|
|
|
dfxp_file = old_file
|
2021-03-18 11:24:53 -04:00
|
|
|
srt_file = replace_extension(old_file, 'srt')
|
2015-04-25 11:15:05 -04:00
|
|
|
|
2017-09-16 00:18:38 -04:00
|
|
|
with open(dfxp_file, 'rb') as f:
|
2015-04-25 11:15:05 -04:00
|
|
|
srt_data = dfxp2srt(f.read())
|
|
|
|
|
|
|
|
with io.open(srt_file, 'wt', encoding='utf-8') as f:
|
|
|
|
f.write(srt_data)
|
2016-02-06 12:51:05 -05:00
|
|
|
old_file = srt_file
|
2015-04-25 11:15:05 -04:00
|
|
|
|
|
|
|
subs[lang] = {
|
|
|
|
'ext': 'srt',
|
2021-03-18 11:24:53 -04:00
|
|
|
'data': srt_data,
|
|
|
|
'filepath': srt_file,
|
2015-04-25 11:15:05 -04:00
|
|
|
}
|
|
|
|
|
|
|
|
if new_ext == 'srt':
|
|
|
|
continue
|
2016-02-06 13:04:18 -05:00
|
|
|
else:
|
|
|
|
sub_filenames.append(srt_file)
|
2015-04-25 11:15:05 -04:00
|
|
|
|
2016-01-31 08:22:36 -05:00
|
|
|
self.run_ffmpeg(old_file, new_file, ['-f', new_format])
|
2015-02-28 08:43:24 -05:00
|
|
|
|
|
|
|
with io.open(new_file, 'rt', encoding='utf-8') as f:
|
|
|
|
subs[lang] = {
|
2016-02-06 12:58:18 -05:00
|
|
|
'ext': new_ext,
|
2015-02-28 08:43:24 -05:00
|
|
|
'data': f.read(),
|
2021-03-18 11:24:53 -04:00
|
|
|
'filepath': new_file,
|
2015-02-28 08:43:24 -05:00
|
|
|
}
|
|
|
|
|
2021-03-18 11:24:53 -04:00
|
|
|
info['__files_to_move'][new_file] = replace_extension(
|
2021-08-13 11:10:13 -04:00
|
|
|
info['__files_to_move'][sub['filepath']], new_ext)
|
2021-03-18 11:24:53 -04:00
|
|
|
|
2016-01-31 08:22:36 -05:00
|
|
|
return sub_filenames, info
|
2021-03-14 19:02:13 -04:00
|
|
|
|
|
|
|
|
|
|
|
class FFmpegSplitChaptersPP(FFmpegPostProcessor):
|
2021-09-01 16:55:16 -04:00
|
|
|
def __init__(self, downloader, force_keyframes=False):
|
|
|
|
FFmpegPostProcessor.__init__(self, downloader)
|
|
|
|
self._force_keyframes = force_keyframes
|
2021-03-14 19:02:13 -04:00
|
|
|
|
|
|
|
def _prepare_filename(self, number, chapter, info):
|
|
|
|
info = info.copy()
|
|
|
|
info.update({
|
|
|
|
'section_number': number,
|
|
|
|
'section_title': chapter.get('title'),
|
|
|
|
'section_start': chapter.get('start_time'),
|
|
|
|
'section_end': chapter.get('end_time'),
|
|
|
|
})
|
|
|
|
return self._downloader.prepare_filename(info, 'chapter')
|
|
|
|
|
|
|
|
def _ffmpeg_args_for_chapter(self, number, chapter, info):
|
|
|
|
destination = self._prepare_filename(number, chapter, info)
|
|
|
|
if not self._downloader._ensure_dir_exists(encodeFilename(destination)):
|
|
|
|
return
|
|
|
|
|
2021-03-18 11:24:53 -04:00
|
|
|
chapter['filepath'] = destination
|
2021-03-14 19:02:13 -04:00
|
|
|
self.to_screen('Chapter %03d; Destination: %s' % (number, destination))
|
|
|
|
return (
|
|
|
|
destination,
|
|
|
|
['-ss', compat_str(chapter['start_time']),
|
2021-03-19 23:45:45 -04:00
|
|
|
'-t', compat_str(chapter['end_time'] - chapter['start_time'])])
|
2021-03-14 19:02:13 -04:00
|
|
|
|
2021-06-12 16:02:19 -04:00
|
|
|
@PostProcessor._restrict_to(images=False)
|
2021-03-14 19:02:13 -04:00
|
|
|
def run(self, info):
|
|
|
|
chapters = info.get('chapters') or []
|
|
|
|
if not chapters:
|
2021-09-01 16:55:16 -04:00
|
|
|
self.to_screen('Chapter information is unavailable')
|
2021-03-14 19:02:13 -04:00
|
|
|
return [], info
|
|
|
|
|
2021-09-01 16:55:16 -04:00
|
|
|
in_file = info['filepath']
|
|
|
|
if self._force_keyframes and len(chapters) > 1:
|
|
|
|
in_file = self.force_keyframes(in_file, (c['start_time'] for c in chapters))
|
2021-03-14 19:02:13 -04:00
|
|
|
self.to_screen('Splitting video by chapters; %d chapters found' % len(chapters))
|
|
|
|
for idx, chapter in enumerate(chapters):
|
|
|
|
destination, opts = self._ffmpeg_args_for_chapter(idx + 1, chapter, info)
|
2022-01-11 22:22:09 -05:00
|
|
|
self.real_run_ffmpeg([(in_file, opts)], [(destination, self.stream_copy_opts())])
|
2021-09-01 16:55:16 -04:00
|
|
|
if in_file != info['filepath']:
|
|
|
|
os.remove(in_file)
|
2021-03-14 19:02:13 -04:00
|
|
|
return [], info
|
2021-04-10 18:18:52 -04:00
|
|
|
|
|
|
|
|
|
|
|
class FFmpegThumbnailsConvertorPP(FFmpegPostProcessor):
|
2022-02-21 10:01:49 -05:00
|
|
|
SUPPORTED_EXTS = ('jpg', 'png', 'webp')
|
2021-05-22 04:24:12 -04:00
|
|
|
|
2021-04-10 18:18:52 -04:00
|
|
|
def __init__(self, downloader=None, format=None):
|
|
|
|
super(FFmpegThumbnailsConvertorPP, self).__init__(downloader)
|
|
|
|
self.format = format
|
|
|
|
|
|
|
|
@staticmethod
|
|
|
|
def is_webp(path):
|
|
|
|
with open(encodeFilename(path), 'rb') as f:
|
|
|
|
b = f.read(12)
|
|
|
|
return b[0:4] == b'RIFF' and b[8:] == b'WEBP'
|
|
|
|
|
|
|
|
def fixup_webp(self, info, idx=-1):
|
|
|
|
thumbnail_filename = info['thumbnails'][idx]['filepath']
|
|
|
|
_, thumbnail_ext = os.path.splitext(thumbnail_filename)
|
|
|
|
if thumbnail_ext:
|
|
|
|
thumbnail_ext = thumbnail_ext[1:].lower()
|
|
|
|
if thumbnail_ext != 'webp' and self.is_webp(thumbnail_filename):
|
|
|
|
self.to_screen('Correcting thumbnail "%s" extension to webp' % thumbnail_filename)
|
|
|
|
webp_filename = replace_extension(thumbnail_filename, 'webp')
|
2021-08-26 22:27:20 -04:00
|
|
|
os.replace(thumbnail_filename, webp_filename)
|
2021-04-10 18:18:52 -04:00
|
|
|
info['thumbnails'][idx]['filepath'] = webp_filename
|
|
|
|
info['__files_to_move'][webp_filename] = replace_extension(
|
|
|
|
info['__files_to_move'].pop(thumbnail_filename), 'webp')
|
|
|
|
|
2021-05-22 04:24:12 -04:00
|
|
|
@staticmethod
|
|
|
|
def _options(target_ext):
|
|
|
|
if target_ext == 'jpg':
|
|
|
|
return ['-bsf:v', 'mjpeg2jpeg']
|
|
|
|
return []
|
|
|
|
|
|
|
|
def convert_thumbnail(self, thumbnail_filename, target_ext):
|
|
|
|
thumbnail_conv_filename = replace_extension(thumbnail_filename, target_ext)
|
2021-05-28 16:09:07 -04:00
|
|
|
|
|
|
|
self.to_screen('Converting thumbnail "%s" to %s' % (thumbnail_filename, target_ext))
|
|
|
|
self.real_run_ffmpeg(
|
|
|
|
[(thumbnail_filename, ['-f', 'image2', '-pattern_type', 'none'])],
|
|
|
|
[(thumbnail_conv_filename.replace('%', '%%'), self._options(target_ext))])
|
2021-05-21 14:09:48 -04:00
|
|
|
return thumbnail_conv_filename
|
2021-04-10 18:18:52 -04:00
|
|
|
|
|
|
|
def run(self, info):
|
|
|
|
files_to_delete = []
|
|
|
|
has_thumbnail = False
|
|
|
|
|
2021-12-26 16:48:11 -05:00
|
|
|
for idx, thumbnail_dict in enumerate(info.get('thumbnails') or []):
|
|
|
|
original_thumbnail = thumbnail_dict.get('filepath')
|
|
|
|
if not original_thumbnail:
|
2021-04-10 18:18:52 -04:00
|
|
|
continue
|
|
|
|
has_thumbnail = True
|
|
|
|
self.fixup_webp(info, idx)
|
|
|
|
_, thumbnail_ext = os.path.splitext(original_thumbnail)
|
|
|
|
if thumbnail_ext:
|
|
|
|
thumbnail_ext = thumbnail_ext[1:].lower()
|
2021-06-24 20:06:35 -04:00
|
|
|
if thumbnail_ext == 'jpeg':
|
|
|
|
thumbnail_ext = 'jpg'
|
2021-04-10 18:18:52 -04:00
|
|
|
if thumbnail_ext == self.format:
|
|
|
|
self.to_screen('Thumbnail "%s" is already in the requested format' % original_thumbnail)
|
|
|
|
continue
|
|
|
|
thumbnail_dict['filepath'] = self.convert_thumbnail(original_thumbnail, self.format)
|
|
|
|
files_to_delete.append(original_thumbnail)
|
|
|
|
info['__files_to_move'][thumbnail_dict['filepath']] = replace_extension(
|
|
|
|
info['__files_to_move'][original_thumbnail], self.format)
|
|
|
|
|
|
|
|
if not has_thumbnail:
|
|
|
|
self.to_screen('There aren\'t any thumbnails to convert')
|
|
|
|
return files_to_delete, info
|
2022-01-13 06:01:08 -05:00
|
|
|
|
|
|
|
|
|
|
|
class FFmpegConcatPP(FFmpegPostProcessor):
|
|
|
|
def __init__(self, downloader, only_multi_video=False):
|
|
|
|
self._only_multi_video = only_multi_video
|
|
|
|
super().__init__(downloader)
|
|
|
|
|
|
|
|
def concat_files(self, in_files, out_file):
|
|
|
|
if len(in_files) == 1:
|
2022-01-19 17:57:36 -05:00
|
|
|
if os.path.realpath(in_files[0]) != os.path.realpath(out_file):
|
|
|
|
self.to_screen(f'Moving "{in_files[0]}" to "{out_file}"')
|
2022-01-13 06:01:08 -05:00
|
|
|
os.replace(in_files[0], out_file)
|
2022-01-19 17:57:36 -05:00
|
|
|
return []
|
2022-01-13 06:01:08 -05:00
|
|
|
|
|
|
|
codecs = [traverse_obj(self.get_metadata_object(file), ('streams', ..., 'codec_name')) for file in in_files]
|
|
|
|
if len(set(map(tuple, codecs))) > 1:
|
|
|
|
raise PostProcessingError(
|
|
|
|
'The files have different streams/codecs and cannot be concatenated. '
|
|
|
|
'Either select different formats or --recode-video them to a common format')
|
2022-01-19 17:57:36 -05:00
|
|
|
|
|
|
|
self.to_screen(f'Concatenating {len(in_files)} files; Destination: {out_file}')
|
2022-01-13 06:01:08 -05:00
|
|
|
super().concat_files(in_files, out_file)
|
2022-01-19 17:57:36 -05:00
|
|
|
return in_files
|
2022-01-13 06:01:08 -05:00
|
|
|
|
2022-02-18 12:46:16 -05:00
|
|
|
@PostProcessor._restrict_to(images=False, simulated=False)
|
2022-01-13 06:01:08 -05:00
|
|
|
def run(self, info):
|
2022-02-03 09:56:27 -05:00
|
|
|
entries = info.get('entries') or []
|
2022-02-18 12:46:16 -05:00
|
|
|
if not any(entries) or (self._only_multi_video and info['_type'] != 'multi_video'):
|
2022-01-13 06:01:08 -05:00
|
|
|
return [], info
|
2022-02-03 09:56:27 -05:00
|
|
|
elif any(len(entry) > 1 for entry in traverse_obj(entries, (..., 'requested_downloads')) or []):
|
2022-01-13 06:01:08 -05:00
|
|
|
raise PostProcessingError('Concatenation is not supported when downloading multiple separate formats')
|
|
|
|
|
2022-02-18 12:46:16 -05:00
|
|
|
in_files = traverse_obj(entries, (..., 'requested_downloads', 0, 'filepath')) or []
|
2022-02-03 09:56:27 -05:00
|
|
|
if len(in_files) < len(entries):
|
|
|
|
raise PostProcessingError('Aborting concatenation because some downloads failed')
|
2022-01-13 06:01:08 -05:00
|
|
|
|
|
|
|
ie_copy = self._downloader._playlist_infodict(info)
|
2022-02-03 09:56:27 -05:00
|
|
|
exts = traverse_obj(entries, (..., 'requested_downloads', 0, 'ext'), (..., 'ext'))
|
2022-01-13 06:01:08 -05:00
|
|
|
ie_copy['ext'] = exts[0] if len(set(exts)) == 1 else 'mkv'
|
|
|
|
out_file = self._downloader.prepare_filename(ie_copy, 'pl_video')
|
|
|
|
|
2022-01-19 17:57:36 -05:00
|
|
|
files_to_delete = self.concat_files(in_files, out_file)
|
2022-01-13 06:01:08 -05:00
|
|
|
|
|
|
|
info['requested_downloads'] = [{
|
|
|
|
'filepath': out_file,
|
|
|
|
'ext': ie_copy['ext'],
|
|
|
|
}]
|
2022-01-19 17:57:36 -05:00
|
|
|
return files_to_delete, info
|