using MediaBrowser.Common.Extensions; using MediaBrowser.Common.IO; using MediaBrowser.Controller.Configuration; using MediaBrowser.Controller.Dto; using MediaBrowser.Controller.Entities; using MediaBrowser.Controller.Library; using MediaBrowser.Controller.LiveTv; using MediaBrowser.Controller.MediaEncoding; using MediaBrowser.Controller.Persistence; using MediaBrowser.Model.Configuration; using MediaBrowser.Model.Drawing; using MediaBrowser.Model.Dto; using MediaBrowser.Model.Entities; using MediaBrowser.Model.IO; using MediaBrowser.Model.Library; using MediaBrowser.Model.LiveTv; using System; using System.Collections.Generic; using System.Diagnostics; using System.Globalization; using System.IO; using System.Linq; using System.Threading; using System.Threading.Tasks; namespace MediaBrowser.Api.Playback { /// /// Class BaseStreamingService /// public abstract class BaseStreamingService : BaseApiService { /// /// Gets or sets the application paths. /// /// The application paths. protected IServerConfigurationManager ServerConfigurationManager { get; private set; } /// /// Gets or sets the user manager. /// /// The user manager. protected IUserManager UserManager { get; private set; } /// /// Gets or sets the library manager. /// /// The library manager. protected ILibraryManager LibraryManager { get; private set; } /// /// Gets or sets the iso manager. /// /// The iso manager. protected IIsoManager IsoManager { get; private set; } /// /// Gets or sets the media encoder. /// /// The media encoder. protected IMediaEncoder MediaEncoder { get; private set; } protected IEncodingManager EncodingManager { get; private set; } protected IDtoService DtoService { get; private set; } protected IFileSystem FileSystem { get; private set; } protected IItemRepository ItemRepository { get; private set; } protected ILiveTvManager LiveTvManager { get; private set; } /// /// Initializes a new instance of the class. /// /// The server configuration. /// The user manager. /// The library manager. /// The iso manager. /// The media encoder. /// The dto service. /// The file system. /// The item repository. protected BaseStreamingService(IServerConfigurationManager serverConfig, IUserManager userManager, ILibraryManager libraryManager, IIsoManager isoManager, IMediaEncoder mediaEncoder, IDtoService dtoService, IFileSystem fileSystem, IItemRepository itemRepository, ILiveTvManager liveTvManager, IEncodingManager encodingManager) { EncodingManager = encodingManager; LiveTvManager = liveTvManager; ItemRepository = itemRepository; FileSystem = fileSystem; DtoService = dtoService; ServerConfigurationManager = serverConfig; UserManager = userManager; LibraryManager = libraryManager; IsoManager = isoManager; MediaEncoder = mediaEncoder; } /// /// Gets the command line arguments. /// /// The output path. /// The state. /// if set to true [perform subtitle conversions]. /// System.String. protected abstract string GetCommandLineArguments(string outputPath, StreamState state, bool performSubtitleConversions); /// /// Gets the type of the transcoding job. /// /// The type of the transcoding job. protected abstract TranscodingJobType TranscodingJobType { get; } /// /// Gets the output file extension. /// /// The state. /// System.String. protected virtual string GetOutputFileExtension(StreamState state) { return Path.GetExtension(state.RequestedUrl); } /// /// Gets the output file path. /// /// The state. /// System.String. protected virtual string GetOutputFilePath(StreamState state) { var folder = ServerConfigurationManager.ApplicationPaths.TranscodingTempPath; var outputFileExtension = GetOutputFileExtension(state); return Path.Combine(folder, GetCommandLineArguments("dummy\\dummy", state, false).GetMD5() + (outputFileExtension ?? string.Empty).ToLower()); } protected readonly CultureInfo UsCulture = new CultureInfo("en-US"); /// /// The fast seek offset seconds /// private const int FastSeekOffsetSeconds = 1; /// /// Gets the fast seek command line parameter. /// /// The request. /// System.String. /// The fast seek command line parameter. protected string GetFastSeekCommandLineParameter(StreamRequest request) { var time = request.StartTimeTicks; if (time.HasValue) { var seconds = TimeSpan.FromTicks(time.Value).TotalSeconds - FastSeekOffsetSeconds; if (seconds > 0) { return string.Format("-ss {0}", seconds.ToString(UsCulture)); } } return string.Empty; } /// /// Gets the slow seek command line parameter. /// /// The request. /// System.String. /// The slow seek command line parameter. protected string GetSlowSeekCommandLineParameter(StreamRequest request) { var time = request.StartTimeTicks; if (time.HasValue) { if (TimeSpan.FromTicks(time.Value).TotalSeconds - FastSeekOffsetSeconds > 0) { return string.Format(" -ss {0}", FastSeekOffsetSeconds.ToString(UsCulture)); } } return string.Empty; } /// /// Gets the map args. /// /// The state. /// System.String. protected virtual string GetMapArgs(StreamState state) { var args = string.Empty; if (state.IsRemote || !state.HasMediaStreams) { return string.Empty; } if (state.VideoStream != null) { args += string.Format("-map 0:{0}", state.VideoStream.Index); } else { args += "-map -0:v"; } if (state.AudioStream != null) { args += string.Format(" -map 0:{0}", state.AudioStream.Index); } else { args += " -map -0:a"; } if (state.SubtitleStream == null) { args += " -map -0:s"; } return args; } /// /// Determines which stream will be used for playback /// /// All stream. /// Index of the desired. /// The type. /// if set to true [return first if no index]. /// MediaStream. private MediaStream GetMediaStream(IEnumerable allStream, int? desiredIndex, MediaStreamType type, bool returnFirstIfNoIndex = true) { var streams = allStream.Where(s => s.Type == type).OrderBy(i => i.Index).ToList(); if (desiredIndex.HasValue) { var stream = streams.FirstOrDefault(s => s.Index == desiredIndex.Value); if (stream != null) { return stream; } } if (returnFirstIfNoIndex && type == MediaStreamType.Audio) { return streams.FirstOrDefault(i => i.Channels.HasValue && i.Channels.Value > 0) ?? streams.FirstOrDefault(); } // Just return the first one return returnFirstIfNoIndex ? streams.FirstOrDefault() : null; } protected EncodingQuality GetQualitySetting() { var quality = ServerConfigurationManager.Configuration.MediaEncodingQuality; if (quality == EncodingQuality.Auto) { var cpuCount = Environment.ProcessorCount; if (cpuCount >= 4) { //return EncodingQuality.HighQuality; } return EncodingQuality.HighSpeed; } return quality; } /// /// Gets the number of threads. /// /// System.Int32. /// Unrecognized MediaEncodingQuality value. protected int GetNumberOfThreads(StreamState state, bool isWebm) { // Use more when this is true. -re will keep cpu usage under control if (state.ReadInputAtNativeFramerate) { if (isWebm) { return Math.Max(Environment.ProcessorCount - 1, 1); } return 0; } // Webm: http://www.webmproject.org/docs/encoder-parameters/ // The decoder will usually automatically use an appropriate number of threads according to how many cores are available but it can only use multiple threads // for the coefficient data if the encoder selected --token-parts > 0 at encode time. switch (GetQualitySetting()) { case EncodingQuality.HighSpeed: return 2; case EncodingQuality.HighQuality: return 2; case EncodingQuality.MaxQuality: return isWebm ? 2 : 0; default: throw new Exception("Unrecognized MediaEncodingQuality value."); } } /// /// Gets the video bitrate to specify on the command line /// /// The state. /// The video codec. /// System.String. protected string GetVideoQualityParam(StreamState state, string videoCodec, bool isHls) { var param = string.Empty; var hasFixedResolution = state.VideoRequest.HasFixedResolution; var qualitySetting = GetQualitySetting(); if (string.Equals(videoCodec, "libx264", StringComparison.OrdinalIgnoreCase)) { switch (qualitySetting) { case EncodingQuality.HighSpeed: param = "-preset ultrafast"; break; case EncodingQuality.HighQuality: param = "-preset superfast"; break; case EncodingQuality.MaxQuality: param = "-preset superfast"; break; } if (!isHls) { switch (qualitySetting) { case EncodingQuality.HighSpeed: param += " -crf 23"; break; case EncodingQuality.HighQuality: param += " -crf 20"; break; case EncodingQuality.MaxQuality: param += " -crf 18"; break; } } } // webm else if (string.Equals(videoCodec, "libvpx", StringComparison.OrdinalIgnoreCase)) { // http://www.webmproject.org/docs/encoder-parameters/ param = "-speed 16 -quality good -profile:v 0 -slices 8"; if (!hasFixedResolution) { switch (qualitySetting) { case EncodingQuality.HighSpeed: param += " -crf 18"; break; case EncodingQuality.HighQuality: param += " -crf 10"; break; case EncodingQuality.MaxQuality: param += " -crf 4"; break; } } } else if (string.Equals(videoCodec, "mpeg4", StringComparison.OrdinalIgnoreCase)) { param = "-mbd rd -flags +mv4+aic -trellis 2 -cmp 2 -subcmp 2 -bf 2"; } // asf/wmv else if (string.Equals(videoCodec, "wmv2", StringComparison.OrdinalIgnoreCase)) { param = "-qmin 2"; } else if (string.Equals(videoCodec, "msmpeg4", StringComparison.OrdinalIgnoreCase)) { param = "-mbd 2"; } param += GetVideoBitrateParam(state, videoCodec, isHls); var framerate = GetFramerateParam(state); if (framerate.HasValue) { param += string.Format(" -r {0}", framerate.Value.ToString(UsCulture)); } if (!string.IsNullOrEmpty(state.VideoSync)) { param += " -vsync " + state.VideoSync; } if (!string.IsNullOrEmpty(state.VideoRequest.Profile)) { param += " -profile:v " + state.VideoRequest.Profile; } if (!string.IsNullOrEmpty(state.VideoRequest.Level)) { param += " -level " + state.VideoRequest.Level; } return param; } protected string GetAudioFilterParam(StreamState state, bool isHls) { var volParam = string.Empty; var audioSampleRate = string.Empty; var channels = GetNumAudioChannelsParam(state.Request, state.AudioStream); // Boost volume to 200% when downsampling from 6ch to 2ch if (channels.HasValue && channels.Value <= 2 && state.AudioStream.Channels.HasValue && state.AudioStream.Channels.Value > 5) { volParam = ",volume=2.000000"; } if (state.Request.AudioSampleRate.HasValue) { audioSampleRate = state.Request.AudioSampleRate.Value + ":"; } var adelay = isHls ? "adelay=1," : string.Empty; var pts = string.Empty; if (state.SubtitleStream != null) { if (state.SubtitleStream.Codec.IndexOf("srt", StringComparison.OrdinalIgnoreCase) != -1 || state.SubtitleStream.Codec.IndexOf("subrip", StringComparison.OrdinalIgnoreCase) != -1 || string.Equals(state.SubtitleStream.Codec, "ass", StringComparison.OrdinalIgnoreCase) || string.Equals(state.SubtitleStream.Codec, "ssa", StringComparison.OrdinalIgnoreCase)) { var seconds = TimeSpan.FromTicks(state.Request.StartTimeTicks ?? 0).TotalSeconds; pts = string.Format(",asetpts=PTS-{0}/TB", Math.Round(seconds).ToString(UsCulture)); } } return string.Format("-af \"{0}aresample={1}async={4}{2}{3}\"", adelay, audioSampleRate, volParam, pts, state.AudioSync); } /// /// If we're going to put a fixed size on the command line, this will calculate it /// /// The state. /// The output video codec. /// if set to true [perform text subtitle conversion]. /// System.String. protected string GetOutputSizeParam(StreamState state, string outputVideoCodec, bool performTextSubtitleConversion) { // http://sonnati.wordpress.com/2012/10/19/ffmpeg-the-swiss-army-knife-of-internet-streaming-part-vi/ var assSubtitleParam = string.Empty; var copyTsParam = string.Empty; var yadifParam = state.DeInterlace ? "yadif=0:-1:0," : string.Empty; var request = state.VideoRequest; if (state.SubtitleStream != null) { if (state.SubtitleStream.Codec.IndexOf("srt", StringComparison.OrdinalIgnoreCase) != -1 || state.SubtitleStream.Codec.IndexOf("subrip", StringComparison.OrdinalIgnoreCase) != -1 || string.Equals(state.SubtitleStream.Codec, "ass", StringComparison.OrdinalIgnoreCase) || string.Equals(state.SubtitleStream.Codec, "ssa", StringComparison.OrdinalIgnoreCase)) { assSubtitleParam = GetTextSubtitleParam(state, performTextSubtitleConversion); copyTsParam = " -copyts"; } } // If fixed dimensions were supplied if (request.Width.HasValue && request.Height.HasValue) { var widthParam = request.Width.Value.ToString(UsCulture); var heightParam = request.Height.Value.ToString(UsCulture); return string.Format("{4} -vf \"{0}scale=trunc({1}/2)*2:trunc({2}/2)*2{3}\"", yadifParam, widthParam, heightParam, assSubtitleParam, copyTsParam); } // If Max dimensions were supplied //this makes my brain hurt. For width selects lowest even number between input width and width req size and selects lowest even number from in width*display aspect and requested size if (request.MaxWidth.HasValue && request.MaxHeight.HasValue) { var MaxwidthParam = request.MaxWidth.Value.ToString(UsCulture); var MaxheightParam = request.MaxHeight.Value.ToString(UsCulture); return string.Format("{4} -vf \"{0}scale=trunc(min(iw\\,{1})/2)*2:trunc(min((iw/dar)\\,{2})/2)*2{3}\"", yadifParam, MaxwidthParam, MaxheightParam, assSubtitleParam, copyTsParam); } var isH264Output = outputVideoCodec.Equals("libx264", StringComparison.OrdinalIgnoreCase); // If a fixed width was requested if (request.Width.HasValue) { var widthParam = request.Width.Value.ToString(UsCulture); return isH264Output ? string.Format("{3} -vf \"{0}scale={1}:trunc(ow/a/2)*2{2}\"", yadifParam, widthParam, assSubtitleParam, copyTsParam) : string.Format("{3} -vf \"{0}scale={1}:-1{2}\"", yadifParam, widthParam, assSubtitleParam, copyTsParam); } // If a fixed height was requested if (request.Height.HasValue) { var heightParam = request.Height.Value.ToString(UsCulture); return isH264Output ? string.Format("{3} -vf \"{0}scale=trunc(oh*a*2)/2:{1}{2}\"", yadifParam, heightParam, assSubtitleParam, copyTsParam) : string.Format("{3} -vf \"{0}scale=-1:{1}{2}\"", yadifParam, heightParam, assSubtitleParam, copyTsParam); } // If a max width was requested if (request.MaxWidth.HasValue && (!request.MaxHeight.HasValue || state.VideoStream == null)) { var maxWidthParam = request.MaxWidth.Value.ToString(UsCulture); return isH264Output ? string.Format("{3} -vf \"{0}scale=min(iw\\,{1}):trunc(ow/a/2)*2{2}\"", yadifParam, maxWidthParam, assSubtitleParam, copyTsParam) : string.Format("{3} -vf \"{0}scale=min(iw\\,{1}):-1{2}\"", yadifParam, maxWidthParam, assSubtitleParam, copyTsParam); } // If a max height was requested if (request.MaxHeight.HasValue && (!request.MaxWidth.HasValue || state.VideoStream == null)) { var maxHeightParam = request.MaxHeight.Value.ToString(UsCulture); return isH264Output ? string.Format("{3} -vf \"{0}scale=trunc(oh*a*2)/2:min(ih\\,{1}){2}\"", yadifParam, maxHeightParam, assSubtitleParam, copyTsParam) : string.Format("{3} -vf \"{0}scale=-1:min(ih\\,{1}){2}\"", yadifParam, maxHeightParam, assSubtitleParam, copyTsParam); } if (state.VideoStream == null) { // No way to figure this out return string.Empty; } // Need to perform calculations manually // Try to account for bad media info var currentHeight = state.VideoStream.Height ?? request.MaxHeight ?? request.Height ?? 0; var currentWidth = state.VideoStream.Width ?? request.MaxWidth ?? request.Width ?? 0; var outputSize = DrawingUtils.Resize(currentWidth, currentHeight, request.Width, request.Height, request.MaxWidth, request.MaxHeight); // If we're encoding with libx264, it can't handle odd numbered widths or heights, so we'll have to fix that if (isH264Output) { var widthParam = outputSize.Width.ToString(UsCulture); var heightParam = outputSize.Height.ToString(UsCulture); return string.Format("{4} -vf \"{0}scale=trunc({1}/2)*2:trunc({2}/2)*2{3}\"", yadifParam, widthParam, heightParam, assSubtitleParam, copyTsParam); } // Otherwise use -vf scale since ffmpeg will ensure internally that the aspect ratio is preserved return string.Format("{3} -vf \"{0}scale={1}:-1{2}\"", yadifParam, Convert.ToInt32(outputSize.Width), assSubtitleParam, copyTsParam); } /// /// Gets the text subtitle param. /// /// The state. /// if set to true [perform conversion]. /// System.String. protected string GetTextSubtitleParam(StreamState state, bool performConversion) { var path = state.SubtitleStream.IsExternal ? GetConvertedAssPath(state.MediaPath, state.SubtitleStream, performConversion) : GetExtractedAssPath(state, performConversion); if (string.IsNullOrEmpty(path)) { return string.Empty; } var seconds = TimeSpan.FromTicks(state.Request.StartTimeTicks ?? 0).TotalSeconds; return string.Format(",ass='{0}',setpts=PTS -{1}/TB", path.Replace('\\', '/').Replace(":/", "\\:/"), Math.Round(seconds).ToString(UsCulture)); } /// /// Gets the extracted ass path. /// /// The state. /// if set to true [perform conversion]. /// System.String. private string GetExtractedAssPath(StreamState state, bool performConversion) { var path = EncodingManager.GetSubtitleCachePath(state.MediaPath, state.SubtitleStream.Index, ".ass"); if (performConversion) { InputType type; var inputPath = MediaEncoderHelpers.GetInputArgument(state.MediaPath, state.IsRemote, state.VideoType, state.IsoType, null, state.PlayableStreamFileNames, out type); try { var parentPath = Path.GetDirectoryName(path); Directory.CreateDirectory(parentPath); // Don't re-encode ass/ssa to ass because ffmpeg ass encoder fails if there's more than one ass rectangle. Affect Anime mostly. // See https://lists.ffmpeg.org/pipermail/ffmpeg-cvslog/2013-April/063616.html var isAssSubtitle = string.Equals(state.SubtitleStream.Codec, "ass", StringComparison.OrdinalIgnoreCase) || string.Equals(state.SubtitleStream.Codec, "ssa", StringComparison.OrdinalIgnoreCase); var task = MediaEncoder.ExtractTextSubtitle(inputPath, type, state.SubtitleStream.Index, isAssSubtitle, path, CancellationToken.None); Task.WaitAll(task); } catch { return null; } } return path; } /// /// Gets the converted ass path. /// /// The media path. /// The subtitle stream. /// if set to true [perform conversion]. /// System.String. private string GetConvertedAssPath(string mediaPath, MediaStream subtitleStream, bool performConversion) { var path = EncodingManager.GetSubtitleCachePath(subtitleStream.Path, ".ass"); if (performConversion) { try { var parentPath = Path.GetDirectoryName(path); Directory.CreateDirectory(parentPath); var task = MediaEncoder.ConvertTextSubtitleToAss(subtitleStream.Path, path, subtitleStream.Language, CancellationToken.None); Task.WaitAll(task); } catch { return null; } } return path; } /// /// Gets the internal graphical subtitle param. /// /// The state. /// The output video codec. /// System.String. protected string GetInternalGraphicalSubtitleParam(StreamState state, string outputVideoCodec) { var outputSizeParam = string.Empty; var request = state.VideoRequest; // Add resolution params, if specified if (request.Width.HasValue || request.Height.HasValue || request.MaxHeight.HasValue || request.MaxWidth.HasValue) { outputSizeParam = GetOutputSizeParam(state, outputVideoCodec, false).TrimEnd('"'); outputSizeParam = "," + outputSizeParam.Substring(outputSizeParam.IndexOf("scale", StringComparison.OrdinalIgnoreCase)); } var videoSizeParam = string.Empty; if (state.VideoStream != null && state.VideoStream.Width.HasValue && state.VideoStream.Height.HasValue) { videoSizeParam = string.Format(",scale={0}:{1}", state.VideoStream.Width.Value.ToString(UsCulture), state.VideoStream.Height.Value.ToString(UsCulture)); } return string.Format(" -filter_complex \"[0:{0}]format=yuva444p{3},lut=u=128:v=128:y=gammaval(.3)[sub] ; [0:{1}] [sub] overlay{2}\"", state.SubtitleStream.Index, state.VideoStream.Index, outputSizeParam, videoSizeParam); } /// /// Gets the probe size argument. /// /// The media path. /// if set to true [is video]. /// Type of the video. /// Type of the iso. /// System.String. private string GetProbeSizeArgument(string mediaPath, bool isVideo, VideoType? videoType, IsoType? isoType) { var type = !isVideo ? MediaEncoderHelpers.GetInputType(null, null) : MediaEncoderHelpers.GetInputType(videoType, isoType); return MediaEncoder.GetProbeSizeArgument(type); } /// /// Gets the number of audio channels to specify on the command line /// /// The request. /// The audio stream. /// System.Nullable{System.Int32}. protected int? GetNumAudioChannelsParam(StreamRequest request, MediaStream audioStream) { if (audioStream != null) { if (audioStream.Channels > 2 && request.AudioCodec.HasValue) { if (request.AudioCodec.Value == AudioCodecs.Wma) { // wmav2 currently only supports two channel output return 2; } } } return request.AudioChannels; } /// /// Determines whether the specified stream is H264. /// /// The stream. /// true if the specified stream is H264; otherwise, false. protected bool IsH264(MediaStream stream) { return stream.Codec.IndexOf("264", StringComparison.OrdinalIgnoreCase) != -1 || stream.Codec.IndexOf("avc", StringComparison.OrdinalIgnoreCase) != -1; } /// /// Gets the name of the output audio codec /// /// The request. /// System.String. protected string GetAudioCodec(StreamRequest request) { var codec = request.AudioCodec; if (codec.HasValue) { if (codec == AudioCodecs.Aac) { return "aac -strict experimental"; } if (codec == AudioCodecs.Mp3) { return "libmp3lame"; } if (codec == AudioCodecs.Vorbis) { return "libvorbis"; } if (codec == AudioCodecs.Wma) { return "wmav2"; } return codec.ToString().ToLower(); } return "copy"; } /// /// Gets the name of the output video codec /// /// The request. /// System.String. protected string GetVideoCodec(VideoStreamRequest request) { var codec = request.VideoCodec; if (codec.HasValue) { if (codec == VideoCodecs.H264) { return "libx264"; } if (codec == VideoCodecs.Vpx) { return "libvpx"; } if (codec == VideoCodecs.Wmv) { return "msmpeg4"; } if (codec == VideoCodecs.Theora) { return "libtheora"; } return codec.ToString().ToLower(); } return "copy"; } /// /// Gets the input argument. /// /// The state. /// System.String. protected string GetInputArgument(StreamState state) { if (state.SendInputOverStandardInput) { return "-"; } var type = InputType.File; var inputPath = new[] { state.MediaPath }; if (state.IsInputVideo) { if (!(state.VideoType == VideoType.Iso && state.IsoMount == null)) { inputPath = MediaEncoderHelpers.GetInputArgument(state.MediaPath, state.IsRemote, state.VideoType, state.IsoType, state.IsoMount, state.PlayableStreamFileNames, out type); } } return MediaEncoder.GetInputArgument(inputPath, type); } /// /// Starts the FFMPEG. /// /// The state. /// The output path. /// Task. protected async Task StartFfMpeg(StreamState state, string outputPath) { if (!File.Exists(MediaEncoder.EncoderPath)) { throw new InvalidOperationException("ffmpeg was not found at " + MediaEncoder.EncoderPath); } Directory.CreateDirectory(Path.GetDirectoryName(outputPath)); if (state.IsInputVideo && state.VideoType == VideoType.Iso && state.IsoType.HasValue && IsoManager.CanMount(state.MediaPath)) { state.IsoMount = await IsoManager.Mount(state.MediaPath, CancellationToken.None).ConfigureAwait(false); } var commandLineArgs = GetCommandLineArguments(outputPath, state, true); if (ServerConfigurationManager.Configuration.EnableDebugEncodingLogging) { commandLineArgs = "-loglevel debug " + commandLineArgs; } var process = new Process { StartInfo = new ProcessStartInfo { CreateNoWindow = true, UseShellExecute = false, // Must consume both stdout and stderr or deadlocks may occur RedirectStandardOutput = true, RedirectStandardError = true, FileName = MediaEncoder.EncoderPath, WorkingDirectory = Path.GetDirectoryName(MediaEncoder.EncoderPath), Arguments = commandLineArgs, WindowStyle = ProcessWindowStyle.Hidden, ErrorDialog = false, RedirectStandardInput = state.SendInputOverStandardInput }, EnableRaisingEvents = true }; ApiEntryPoint.Instance.OnTranscodeBeginning(outputPath, TranscodingJobType, process, state.IsInputVideo, state.Request.StartTimeTicks, state.MediaPath, state.Request.DeviceId); Logger.Info(process.StartInfo.FileName + " " + process.StartInfo.Arguments); var logFilePath = Path.Combine(ServerConfigurationManager.ApplicationPaths.LogDirectoryPath, "ffmpeg-" + Guid.NewGuid() + ".txt"); Directory.CreateDirectory(Path.GetDirectoryName(logFilePath)); // FFMpeg writes debug/error info to stderr. This is useful when debugging so let's put it in the log directory. state.LogFileStream = FileSystem.GetFileStream(logFilePath, FileMode.Create, FileAccess.Write, FileShare.Read, true); process.Exited += (sender, args) => OnFfMpegProcessExited(process, state); try { process.Start(); } catch (Exception ex) { Logger.ErrorException("Error starting ffmpeg", ex); ApiEntryPoint.Instance.OnTranscodeFailedToStart(outputPath, TranscodingJobType); state.LogFileStream.Dispose(); throw; } if (state.SendInputOverStandardInput) { StreamToStandardInput(process, state); } // MUST read both stdout and stderr asynchronously or a deadlock may occurr process.BeginOutputReadLine(); #pragma warning disable 4014 // Important - don't await the log task or we won't be able to kill ffmpeg when the user stops playback process.StandardError.BaseStream.CopyToAsync(state.LogFileStream); #pragma warning restore 4014 // Wait for the file to exist before proceeeding while (!File.Exists(outputPath)) { await Task.Delay(100).ConfigureAwait(false); } // Allow a small amount of time to buffer a little if (state.IsInputVideo) { await Task.Delay(500).ConfigureAwait(false); } // This is arbitrary, but add a little buffer time when internet streaming if (state.IsRemote) { await Task.Delay(3000).ConfigureAwait(false); } } private async void StreamToStandardInput(Process process, StreamState state) { state.StandardInputCancellationTokenSource = new CancellationTokenSource(); try { await StreamToStandardInputInternal(process, state).ConfigureAwait(false); } catch (OperationCanceledException) { Logger.Debug("Stream to standard input closed normally."); } catch (Exception ex) { Logger.ErrorException("Error writing to standard input", ex); } } private async Task StreamToStandardInputInternal(Process process, StreamState state) { state.StandardInputCancellationTokenSource = new CancellationTokenSource(); using (var fileStream = FileSystem.GetFileStream(state.MediaPath, FileMode.Open, FileAccess.Read, FileShare.ReadWrite, true)) { await new EndlessStreamCopy().CopyStream(fileStream, process.StandardInput.BaseStream, state.StandardInputCancellationTokenSource.Token).ConfigureAwait(false); } } protected int? GetVideoBitrateParamValue(StreamState state) { var bitrate = state.VideoRequest.VideoBitRate; if (state.VideoStream != null) { var isUpscaling = state.VideoRequest.Height.HasValue && state.VideoStream.Height.HasValue && state.VideoRequest.Height.Value > state.VideoStream.Height.Value; if (state.VideoRequest.Width.HasValue && state.VideoStream.Width.HasValue && state.VideoRequest.Width.Value > state.VideoStream.Width.Value) { isUpscaling = true; } // Don't allow bitrate increases unless upscaling if (!isUpscaling) { if (bitrate.HasValue && state.VideoStream.BitRate.HasValue) { bitrate = Math.Min(bitrate.Value, state.VideoStream.BitRate.Value); } } } return bitrate; } protected string GetVideoBitrateParam(StreamState state, string videoCodec, bool isHls) { var bitrate = GetVideoBitrateParamValue(state); if (bitrate.HasValue) { var hasFixedResolution = state.VideoRequest.HasFixedResolution; if (isHls) { return string.Format(" -b:v {0} -maxrate ({0}*.80) -bufsize {0}", bitrate.Value.ToString(UsCulture)); } if (string.Equals(videoCodec, "libvpx", StringComparison.OrdinalIgnoreCase)) { if (hasFixedResolution) { return string.Format(" -minrate:v ({0}*.90) -maxrate:v ({0}*1.10) -bufsize:v {0} -b:v {0}", bitrate.Value.ToString(UsCulture)); } // With vpx when crf is used, b:v becomes a max rate // https://trac.ffmpeg.org/wiki/vpxEncodingGuide return string.Format(" -b:v {0}", bitrate.Value.ToString(UsCulture)); //return string.Format(" -minrate:v ({0}*.95) -maxrate:v ({0}*1.05) -bufsize:v {0} -b:v {0}", bitrate.Value.ToString(UsCulture)); } if (string.Equals(videoCodec, "msmpeg4", StringComparison.OrdinalIgnoreCase)) { return string.Format(" -b:v {0}", bitrate.Value.ToString(UsCulture)); } // H264 if (hasFixedResolution) { return string.Format(" -b:v {0}", bitrate.Value.ToString(UsCulture)); } return string.Format(" -maxrate {0} -bufsize {1}", bitrate.Value.ToString(UsCulture), (bitrate.Value * 2).ToString(UsCulture)); } return string.Empty; } protected int? GetAudioBitrateParam(StreamState state) { if (state.Request.AudioBitRate.HasValue) { // Make sure we don't request a bitrate higher than the source var currentBitrate = state.AudioStream == null ? state.Request.AudioBitRate.Value : state.AudioStream.BitRate ?? state.Request.AudioBitRate.Value; return Math.Min(currentBitrate, state.Request.AudioBitRate.Value); } return null; } /// /// Gets the user agent param. /// /// The path. /// System.String. private string GetUserAgentParam(string path) { var useragent = GetUserAgent(path); if (!string.IsNullOrEmpty(useragent)) { return "-user-agent \"" + useragent + "\""; } return string.Empty; } /// /// Gets the user agent. /// /// The path. /// System.String. protected string GetUserAgent(string path) { if (string.IsNullOrEmpty(path)) { throw new ArgumentNullException("path"); } if (path.IndexOf("apple.com", StringComparison.OrdinalIgnoreCase) != -1) { return "QuickTime/7.7.4"; } return string.Empty; } /// /// Processes the exited. /// /// The process. /// The state. protected async void OnFfMpegProcessExited(Process process, StreamState state) { if (state.IsoMount != null) { state.IsoMount.Dispose(); state.IsoMount = null; } if (state.StandardInputCancellationTokenSource != null) { state.StandardInputCancellationTokenSource.Cancel(); } var outputFilePath = GetOutputFilePath(state); state.LogFileStream.Dispose(); try { Logger.Info("FFMpeg exited with code {0} for {1}", process.ExitCode, outputFilePath); } catch { Logger.Info("FFMpeg exited with an error for {0}", outputFilePath); } if (!string.IsNullOrEmpty(state.LiveTvStreamId)) { try { await LiveTvManager.CloseLiveStream(state.LiveTvStreamId, CancellationToken.None).ConfigureAwait(false); } catch (Exception ex) { Logger.ErrorException("Error closing live tv stream", ex); } } } protected double? GetFramerateParam(StreamState state) { if (state.VideoRequest != null && state.VideoRequest.Framerate.HasValue) { return state.VideoRequest.Framerate.Value; } if (state.VideoStream != null) { var contentRate = state.VideoStream.AverageFrameRate ?? state.VideoStream.RealFrameRate; if (contentRate.HasValue && contentRate.Value > 23.976) { return 23.976; } } return null; } /// /// Parses the parameters. /// /// The request. private void ParseParams(StreamRequest request) { var vals = request.Params.Split(';'); var videoRequest = request as VideoStreamRequest; for (var i = 0; i < vals.Length; i++) { var val = vals[i]; if (string.IsNullOrWhiteSpace(val)) { continue; } if (i == 0) { request.DeviceId = val; } else if (i == 1) { request.Static = string.Equals("true", val, StringComparison.OrdinalIgnoreCase); } else if (i == 2) { if (videoRequest != null) { videoRequest.VideoCodec = (VideoCodecs)Enum.Parse(typeof(VideoCodecs), val, true); } } else if (i == 3) { request.AudioCodec = (AudioCodecs)Enum.Parse(typeof(AudioCodecs), val, true); } else if (i == 4) { if (videoRequest != null) { videoRequest.AudioStreamIndex = int.Parse(val, UsCulture); } } else if (i == 5) { if (videoRequest != null) { videoRequest.SubtitleStreamIndex = int.Parse(val, UsCulture); } } else if (i == 6) { if (videoRequest != null) { videoRequest.VideoBitRate = int.Parse(val, UsCulture); } } else if (i == 7) { request.AudioBitRate = int.Parse(val, UsCulture); } else if (i == 8) { request.AudioChannels = int.Parse(val, UsCulture); } else if (i == 9) { if (videoRequest != null) { request.StartTimeTicks = long.Parse(val, UsCulture); } } else if (i == 10) { if (videoRequest != null) { videoRequest.Profile = val; } } else if (i == 11) { if (videoRequest != null) { videoRequest.Level = val; } } else if (i == 12) { request.ForcedMimeType = val; } } } /// /// Gets the state. /// /// The request. /// The cancellation token. /// StreamState. protected async Task GetState(StreamRequest request, CancellationToken cancellationToken) { if (!string.IsNullOrWhiteSpace(request.Params)) { ParseParams(request); } if (request.ThrowDebugError) { throw new InvalidOperationException("You asked for a debug error, you got one."); } var user = AuthorizationRequestFilterAttribute.GetCurrentUser(Request, UserManager); var url = Request.PathInfo; if (!request.AudioCodec.HasValue) { request.AudioCodec = InferAudioCodec(url); } var state = new StreamState { Request = request, RequestedUrl = url }; var item = DtoService.GetItemByDtoId(request.Id); if (user != null && item.GetPlayAccess(user) != PlayAccess.Full) { throw new ArgumentException(string.Format("{0} is not allowed to play media.", user.Name)); } if (item is ILiveTvRecording) { var recording = await LiveTvManager.GetInternalRecording(request.Id, cancellationToken).ConfigureAwait(false); state.VideoType = VideoType.VideoFile; state.IsInputVideo = string.Equals(recording.MediaType, MediaType.Video, StringComparison.OrdinalIgnoreCase); state.PlayableStreamFileNames = new List(); var path = recording.RecordingInfo.Path; var mediaUrl = recording.RecordingInfo.Url; if (string.IsNullOrWhiteSpace(path) && string.IsNullOrWhiteSpace(mediaUrl)) { var streamInfo = await LiveTvManager.GetRecordingStream(request.Id, cancellationToken).ConfigureAwait(false); state.LiveTvStreamId = streamInfo.Id; path = streamInfo.Path; mediaUrl = streamInfo.Url; } if (!string.IsNullOrEmpty(path) && File.Exists(path)) { state.MediaPath = path; state.IsRemote = false; state.SendInputOverStandardInput = recording.RecordingInfo.Status == RecordingStatus.InProgress; } else if (!string.IsNullOrEmpty(mediaUrl)) { state.MediaPath = mediaUrl; state.IsRemote = true; } //state.RunTimeTicks = recording.RunTimeTicks; if (recording.RecordingInfo.Status == RecordingStatus.InProgress && !state.IsRemote) { await Task.Delay(1000, cancellationToken).ConfigureAwait(false); } state.ReadInputAtNativeFramerate = recording.RecordingInfo.Status == RecordingStatus.InProgress; state.AudioSync = "1000"; state.DeInterlace = true; } else if (item is LiveTvChannel) { var channel = LiveTvManager.GetInternalChannel(request.Id); state.VideoType = VideoType.VideoFile; state.IsInputVideo = string.Equals(channel.MediaType, MediaType.Video, StringComparison.OrdinalIgnoreCase); state.PlayableStreamFileNames = new List(); var streamInfo = await LiveTvManager.GetChannelStream(request.Id, cancellationToken).ConfigureAwait(false); state.LiveTvStreamId = streamInfo.Id; if (!string.IsNullOrEmpty(streamInfo.Path) && File.Exists(streamInfo.Path)) { state.MediaPath = streamInfo.Path; state.IsRemote = false; await Task.Delay(1000, cancellationToken).ConfigureAwait(false); } else if (!string.IsNullOrEmpty(streamInfo.Url)) { state.MediaPath = streamInfo.Url; state.IsRemote = true; } state.ReadInputAtNativeFramerate = true; state.AudioSync = "1000"; state.DeInterlace = true; } else { state.MediaPath = item.Path; state.IsRemote = item.LocationType == LocationType.Remote; var video = item as Video; if (video != null) { state.IsInputVideo = true; state.VideoType = video.VideoType; state.IsoType = video.IsoType; state.PlayableStreamFileNames = video.PlayableStreamFileNames == null ? new List() : video.PlayableStreamFileNames.ToList(); } state.RunTimeTicks = item.RunTimeTicks; } var videoRequest = request as VideoStreamRequest; var mediaStreams = ItemRepository.GetMediaStreams(new MediaStreamQuery { ItemId = item.Id }).ToList(); if (videoRequest != null) { if (!videoRequest.VideoCodec.HasValue) { videoRequest.VideoCodec = InferVideoCodec(url); } state.VideoStream = GetMediaStream(mediaStreams, videoRequest.VideoStreamIndex, MediaStreamType.Video); state.SubtitleStream = GetMediaStream(mediaStreams, videoRequest.SubtitleStreamIndex, MediaStreamType.Subtitle, false); state.AudioStream = GetMediaStream(mediaStreams, videoRequest.AudioStreamIndex, MediaStreamType.Audio); if (state.VideoStream != null && state.VideoStream.IsInterlaced) { state.DeInterlace = true; } EnforceResolutionLimit(state, videoRequest); } else { state.AudioStream = GetMediaStream(mediaStreams, null, MediaStreamType.Audio, true); } state.HasMediaStreams = mediaStreams.Count > 0; state.SegmentLength = state.ReadInputAtNativeFramerate ? 5 : 10; state.HlsListSize = state.ReadInputAtNativeFramerate ? 100 : 1440; return state; } /// /// Enforces the resolution limit. /// /// The state. /// The video request. private void EnforceResolutionLimit(StreamState state, VideoStreamRequest videoRequest) { // If enabled, allow whatever the client asks for if (ServerConfigurationManager.Configuration.AllowVideoUpscaling) { return; } // Switch the incoming params to be ceilings rather than fixed values videoRequest.MaxWidth = videoRequest.MaxWidth ?? videoRequest.Width; videoRequest.MaxHeight = videoRequest.MaxHeight ?? videoRequest.Height; videoRequest.Width = null; videoRequest.Height = null; } protected string GetInputModifier(StreamState state) { var inputModifier = string.Empty; var probeSize = GetProbeSizeArgument(state.MediaPath, state.IsInputVideo, state.VideoType, state.IsoType); inputModifier += " " + probeSize; inputModifier = inputModifier.Trim(); inputModifier += " " + GetUserAgentParam(state.MediaPath); inputModifier = inputModifier.Trim(); inputModifier += " " + GetFastSeekCommandLineParameter(state.Request); inputModifier = inputModifier.Trim(); if (state.VideoRequest != null) { inputModifier += " -fflags genpts"; } if (!string.IsNullOrEmpty(state.InputFormat)) { inputModifier += " -f " + state.InputFormat; } if (!string.IsNullOrEmpty(state.InputVideoCodec)) { inputModifier += " -vcodec " + state.InputVideoCodec; } if (!string.IsNullOrEmpty(state.InputAudioCodec)) { inputModifier += " -acodec " + state.InputAudioCodec; } if (!string.IsNullOrEmpty(state.InputAudioSync)) { inputModifier += " -async " + state.InputAudioSync; } if (!string.IsNullOrEmpty(state.InputVideoSync)) { inputModifier += " -vsync " + state.InputVideoSync; } if (state.ReadInputAtNativeFramerate) { inputModifier += " -re"; } return inputModifier; } /// /// Infers the audio codec based on the url /// /// The URL. /// System.Nullable{AudioCodecs}. private AudioCodecs? InferAudioCodec(string url) { var ext = Path.GetExtension(url); if (string.Equals(ext, ".mp3", StringComparison.OrdinalIgnoreCase)) { return AudioCodecs.Mp3; } if (string.Equals(ext, ".aac", StringComparison.OrdinalIgnoreCase)) { return AudioCodecs.Aac; } if (string.Equals(ext, ".wma", StringComparison.OrdinalIgnoreCase)) { return AudioCodecs.Wma; } if (string.Equals(ext, ".ogg", StringComparison.OrdinalIgnoreCase)) { return AudioCodecs.Vorbis; } if (string.Equals(ext, ".oga", StringComparison.OrdinalIgnoreCase)) { return AudioCodecs.Vorbis; } if (string.Equals(ext, ".ogv", StringComparison.OrdinalIgnoreCase)) { return AudioCodecs.Vorbis; } if (string.Equals(ext, ".webm", StringComparison.OrdinalIgnoreCase)) { return AudioCodecs.Vorbis; } if (string.Equals(ext, ".webma", StringComparison.OrdinalIgnoreCase)) { return AudioCodecs.Vorbis; } return null; } /// /// Infers the video codec. /// /// The URL. /// System.Nullable{VideoCodecs}. private VideoCodecs? InferVideoCodec(string url) { var ext = Path.GetExtension(url); if (string.Equals(ext, ".asf", StringComparison.OrdinalIgnoreCase)) { return VideoCodecs.Wmv; } if (string.Equals(ext, ".webm", StringComparison.OrdinalIgnoreCase)) { return VideoCodecs.Vpx; } if (string.Equals(ext, ".ogg", StringComparison.OrdinalIgnoreCase) || string.Equals(ext, ".ogv", StringComparison.OrdinalIgnoreCase)) { return VideoCodecs.Theora; } if (string.Equals(ext, ".m3u8", StringComparison.OrdinalIgnoreCase) || string.Equals(ext, ".ts", StringComparison.OrdinalIgnoreCase)) { return VideoCodecs.H264; } return VideoCodecs.Copy; } } }