!function(e){if("object"==typeof exports&&"undefined"!=typeof module)module.exports=e();else if("function"==typeof define&&define.amd)define([],e);else{var f;"undefined"!=typeof window?f=window:"undefined"!=typeof global?f=global:"undefined"!=typeof self&&(f=self),f.Hls=e()}}(function(){var define,module,exports;return (function e(t,n,r){function s(o,u){if(!n[o]){if(!t[o]){var a=typeof require=="function"&&require;if(!u&&a)return a(o,!0);if(i)return i(o,!0);var f=new Error("Cannot find module '"+o+"'");throw f.code="MODULE_NOT_FOUND",f}var l=n[o]={exports:{}};t[o][0].call(l.exports,function(e){var n=t[o][1][e];return s(n?n:e)},l,l.exports,e,t,n,r)}return n[o].exports}var i=typeof require=="function"&&require;for(var o=0;o 0 && this._events[type].length > m) { this._events[type].warned = true; console.error('(node) warning: possible EventEmitter memory ' + 'leak detected. %d listeners added. ' + 'Use emitter.setMaxListeners() to increase limit.', this._events[type].length); if (typeof console.trace === 'function') { // not supported in IE 10 console.trace(); } } } return this; }; EventEmitter.prototype.on = EventEmitter.prototype.addListener; EventEmitter.prototype.once = function(type, listener) { if (!isFunction(listener)) throw TypeError('listener must be a function'); var fired = false; function g() { this.removeListener(type, g); if (!fired) { fired = true; listener.apply(this, arguments); } } g.listener = listener; this.on(type, g); return this; }; // emits a 'removeListener' event iff the listener was removed EventEmitter.prototype.removeListener = function(type, listener) { var list, position, length, i; if (!isFunction(listener)) throw TypeError('listener must be a function'); if (!this._events || !this._events[type]) return this; list = this._events[type]; length = list.length; position = -1; if (list === listener || (isFunction(list.listener) && list.listener === listener)) { delete this._events[type]; if (this._events.removeListener) this.emit('removeListener', type, listener); } else if (isObject(list)) { for (i = length; i-- > 0;) { if (list[i] === listener || (list[i].listener && list[i].listener === listener)) { position = i; break; } } if (position < 0) return this; if (list.length === 1) { list.length = 0; delete this._events[type]; } else { list.splice(position, 1); } if (this._events.removeListener) this.emit('removeListener', type, listener); } return this; }; EventEmitter.prototype.removeAllListeners = function(type) { var key, listeners; if (!this._events) return this; // not listening for removeListener, no need to emit if (!this._events.removeListener) { if (arguments.length === 0) this._events = {}; else if (this._events[type]) delete this._events[type]; return this; } // emit removeListener for all listeners on all events if (arguments.length === 0) { for (key in this._events) { if (key === 'removeListener') continue; this.removeAllListeners(key); } this.removeAllListeners('removeListener'); this._events = {}; return this; } listeners = this._events[type]; if (isFunction(listeners)) { this.removeListener(type, listeners); } else { // LIFO order while (listeners.length) this.removeListener(type, listeners[listeners.length - 1]); } delete this._events[type]; return this; }; EventEmitter.prototype.listeners = function(type) { var ret; if (!this._events || !this._events[type]) ret = []; else if (isFunction(this._events[type])) ret = [this._events[type]]; else ret = this._events[type].slice(); return ret; }; EventEmitter.listenerCount = function(emitter, type) { var ret; if (!emitter._events || !emitter._events[type]) ret = 0; else if (isFunction(emitter._events[type])) ret = 1; else ret = emitter._events[type].length; return ret; }; function isFunction(arg) { return typeof arg === 'function'; } function isNumber(arg) { return typeof arg === 'number'; } function isObject(arg) { return typeof arg === 'object' && arg !== null; } function isUndefined(arg) { return arg === void 0; } },{}],2:[function(require,module,exports){ var bundleFn = arguments[3]; var sources = arguments[4]; var cache = arguments[5]; var stringify = JSON.stringify; module.exports = function (fn) { var keys = []; var wkey; var cacheKeys = Object.keys(cache); for (var i = 0, l = cacheKeys.length; i < l; i++) { var key = cacheKeys[i]; if (cache[key].exports === fn) { wkey = key; break; } } if (!wkey) { wkey = Math.floor(Math.pow(16, 8) * Math.random()).toString(16); var wcache = {}; for (var i = 0, l = cacheKeys.length; i < l; i++) { var key = cacheKeys[i]; wcache[key] = key; } sources[wkey] = [ Function(['require','module','exports'], '(' + fn + ')(self)'), wcache ]; } var skey = Math.floor(Math.pow(16, 8) * Math.random()).toString(16); var scache = {}; scache[wkey] = wkey; sources[skey] = [ Function(['require'],'require(' + stringify(wkey) + ')(self)'), scache ]; var src = '(' + bundleFn + ')({' + Object.keys(sources).map(function (key) { return stringify(key) + ':[' + sources[key][0] + ',' + stringify(sources[key][1]) + ']' ; }).join(',') + '},{},[' + stringify(skey) + '])' ; var URL = window.URL || window.webkitURL || window.mozURL || window.msURL; return new Worker(URL.createObjectURL( new Blob([src], { type: 'text/javascript' }) )); }; },{}],3:[function(require,module,exports){ /* * simple ABR Controller */ 'use strict'; Object.defineProperty(exports, '__esModule', { value: true }); var _createClass = (function () { function defineProperties(target, props) { for (var i = 0; i < props.length; i++) { var descriptor = props[i]; descriptor.enumerable = descriptor.enumerable || false; descriptor.configurable = true; if ('value' in descriptor) descriptor.writable = true; Object.defineProperty(target, descriptor.key, descriptor); } } return function (Constructor, protoProps, staticProps) { if (protoProps) defineProperties(Constructor.prototype, protoProps); if (staticProps) defineProperties(Constructor, staticProps); return Constructor; }; })(); function _interopRequireDefault(obj) { return obj && obj.__esModule ? obj : { 'default': obj }; } function _classCallCheck(instance, Constructor) { if (!(instance instanceof Constructor)) { throw new TypeError('Cannot call a class as a function'); } } var _events = require('../events'); var _events2 = _interopRequireDefault(_events); var AbrController = (function () { function AbrController(hls) { _classCallCheck(this, AbrController); this.hls = hls; this.lastfetchlevel = 0; this._autoLevelCapping = -1; this._nextAutoLevel = -1; this.onflp = this.onFragmentLoadProgress.bind(this); hls.on(_events2['default'].FRAG_LOAD_PROGRESS, this.onflp); } _createClass(AbrController, [{ key: 'destroy', value: function destroy() { this.hls.off(_events2['default'].FRAG_LOAD_PROGRESS, this.onflp); } }, { key: 'onFragmentLoadProgress', value: function onFragmentLoadProgress(event, data) { var stats = data.stats; if (stats.aborted === undefined) { this.lastfetchduration = (performance.now() - stats.trequest) / 1000; this.lastfetchlevel = data.frag.level; this.lastbw = stats.loaded * 8 / this.lastfetchduration; //console.log(`fetchDuration:${this.lastfetchduration},bw:${(this.lastbw/1000).toFixed(0)}/${stats.aborted}`); } } /** Return the capping/max level value that could be used by automatic level selection algorithm **/ }, { key: 'autoLevelCapping', get: function get() { return this._autoLevelCapping; }, /** set the capping/max level value that could be used by automatic level selection algorithm **/ set: function set(newLevel) { this._autoLevelCapping = newLevel; } }, { key: 'nextAutoLevel', get: function get() { var lastbw = this.lastbw, hls = this.hls, adjustedbw, i, maxAutoLevel; if (this._autoLevelCapping === -1) { maxAutoLevel = hls.levels.length - 1; } else { maxAutoLevel = this._autoLevelCapping; } if (this._nextAutoLevel !== -1) { var nextLevel = Math.min(this._nextAutoLevel, maxAutoLevel); if (nextLevel === this.lastfetchlevel) { this._nextAutoLevel = -1; } else { return nextLevel; } } // follow algorithm captured from stagefright : // https://android.googlesource.com/platform/frameworks/av/+/master/media/libstagefright/httplive/LiveSession.cpp // Pick the highest bandwidth stream below or equal to estimated bandwidth. for (i = 0; i <= maxAutoLevel; i++) { // consider only 80% of the available bandwidth, but if we are switching up, // be even more conservative (70%) to avoid overestimating and immediately // switching back. if (i <= this.lastfetchlevel) { adjustedbw = 0.8 * lastbw; } else { adjustedbw = 0.7 * lastbw; } if (adjustedbw < hls.levels[i].bitrate) { return Math.max(0, i - 1); } } return i - 1; }, set: function set(nextLevel) { this._nextAutoLevel = nextLevel; } }]); return AbrController; })(); exports['default'] = AbrController; module.exports = exports['default']; },{"../events":12}],4:[function(require,module,exports){ /* * Level Controller */ 'use strict'; Object.defineProperty(exports, '__esModule', { value: true }); var _createClass = (function () { function defineProperties(target, props) { for (var i = 0; i < props.length; i++) { var descriptor = props[i]; descriptor.enumerable = descriptor.enumerable || false; descriptor.configurable = true; if ('value' in descriptor) descriptor.writable = true; Object.defineProperty(target, descriptor.key, descriptor); } } return function (Constructor, protoProps, staticProps) { if (protoProps) defineProperties(Constructor.prototype, protoProps); if (staticProps) defineProperties(Constructor, staticProps); return Constructor; }; })(); function _interopRequireDefault(obj) { return obj && obj.__esModule ? obj : { 'default': obj }; } function _classCallCheck(instance, Constructor) { if (!(instance instanceof Constructor)) { throw new TypeError('Cannot call a class as a function'); } } var _events = require('../events'); var _events2 = _interopRequireDefault(_events); var _utilsLogger = require('../utils/logger'); var _errors = require('../errors'); var LevelController = (function () { function LevelController(hls) { _classCallCheck(this, LevelController); this.hls = hls; this.onml = this.onManifestLoaded.bind(this); this.onll = this.onLevelLoaded.bind(this); this.onerr = this.onError.bind(this); this.ontick = this.tick.bind(this); hls.on(_events2['default'].MANIFEST_LOADED, this.onml); hls.on(_events2['default'].LEVEL_LOADED, this.onll); hls.on(_events2['default'].ERROR, this.onerr); this._manualLevel = this._autoLevelCapping = -1; } _createClass(LevelController, [{ key: 'destroy', value: function destroy() { var hls = this.hls; hls.off(_events2['default'].MANIFEST_LOADED, this.onml); hls.off(_events2['default'].LEVEL_LOADED, this.onll); hls.off(_events2['default'].ERROR, this.onerr); if (this.timer) { clearInterval(this.timer); } this._manualLevel = -1; } }, { key: 'onManifestLoaded', value: function onManifestLoaded(event, data) { var levels0 = [], levels = [], bitrateStart, i, bitrateSet = {}, videoCodecFound = false, audioCodecFound = false; // regroup redundant level together data.levels.forEach(function (level) { if (level.videoCodec) { videoCodecFound = true; } if (level.audioCodec) { audioCodecFound = true; } var redundantLevelId = bitrateSet[level.bitrate]; if (redundantLevelId === undefined) { bitrateSet[level.bitrate] = levels.length; level.url = [level.url]; level.urlId = 0; levels0.push(level); } else { levels0[redundantLevelId].url.push(level.url); } }); // remove audio-only level if we also have levels with audio+video codecs signalled if (videoCodecFound && audioCodecFound) { levels0.forEach(function (level) { if (level.videoCodec) { levels.push(level); } }); } else { levels = levels0; } // start bitrate is the first bitrate of the manifest bitrateStart = levels[0].bitrate; // sort level on bitrate levels.sort(function (a, b) { return a.bitrate - b.bitrate; }); this._levels = levels; // find index of first level in sorted levels for (i = 0; i < levels.length; i++) { if (levels[i].bitrate === bitrateStart) { this._firstLevel = i; _utilsLogger.logger.log('manifest loaded,' + levels.length + ' level(s) found, first bitrate:' + bitrateStart); break; } } this.hls.trigger(_events2['default'].MANIFEST_PARSED, { levels: this._levels, firstLevel: this._firstLevel, stats: data.stats }); return; } }, { key: 'setLevelInternal', value: function setLevelInternal(newLevel) { // check if level idx is valid if (newLevel >= 0 && newLevel < this._levels.length) { // stopping live reloading timer if any if (this.timer) { clearInterval(this.timer); this.timer = null; } this._level = newLevel; _utilsLogger.logger.log('switching to level ' + newLevel); this.hls.trigger(_events2['default'].LEVEL_SWITCH, { level: newLevel }); var level = this._levels[newLevel]; // check if we need to load playlist for this level if (level.details === undefined || level.details.live === true) { // level not retrieved yet, or live playlist we need to (re)load it _utilsLogger.logger.log('(re)loading playlist for level ' + newLevel); var urlId = level.urlId; this.hls.trigger(_events2['default'].LEVEL_LOADING, { url: level.url[urlId], level: newLevel, id: urlId }); } } else { // invalid level id given, trigger error this.hls.trigger(_events2['default'].ERROR, { type: _errors.ErrorTypes.OTHER_ERROR, details: _errors.ErrorDetails.LEVEL_SWITCH_ERROR, level: newLevel, fatal: false, reason: 'invalid level idx' }); } } }, { key: 'onError', value: function onError(event, data) { if (data.fatal) { return; } var details = data.details, hls = this.hls, levelId, level; // try to recover not fatal errors switch (details) { case _errors.ErrorDetails.FRAG_LOAD_ERROR: case _errors.ErrorDetails.FRAG_LOAD_TIMEOUT: case _errors.ErrorDetails.FRAG_LOOP_LOADING_ERROR: case _errors.ErrorDetails.KEY_LOAD_ERROR: case _errors.ErrorDetails.KEY_LOAD_TIMEOUT: levelId = data.frag.level; break; case _errors.ErrorDetails.LEVEL_LOAD_ERROR: case _errors.ErrorDetails.LEVEL_LOAD_TIMEOUT: levelId = data.level; break; default: break; } /* try to switch to a redundant stream if any available. * if no redundant stream available, emergency switch down (if in auto mode and current level not 0) * otherwise, we cannot recover this network error .... */ if (levelId !== undefined) { level = this._levels[levelId]; if (level.urlId < level.url.length - 1) { level.urlId++; level.details = undefined; _utilsLogger.logger.warn('level controller,' + details + ' for level ' + levelId + ': switching to redundant stream id ' + level.urlId); } else { // we could try to recover if in auto mode and current level not lowest level (0) var recoverable = this._manualLevel === -1 && levelId; if (recoverable) { _utilsLogger.logger.warn('level controller,' + details + ': emergency switch-down for next fragment'); hls.abrController.nextAutoLevel = 0; } else if (level && level.details && level.details.live) { _utilsLogger.logger.warn('level controller,' + details + ' on live stream, discard'); } else { _utilsLogger.logger.error('cannot recover ' + details + ' error'); this._level = undefined; // stopping live reloading timer if any if (this.timer) { clearInterval(this.timer); this.timer = null; } // redispatch same error but with fatal set to true data.fatal = true; hls.trigger(event, data); } } } } }, { key: 'onLevelLoaded', value: function onLevelLoaded(event, data) { // check if current playlist is a live playlist if (data.details.live && !this.timer) { // if live playlist we will have to reload it periodically // set reload period to playlist target duration this.timer = setInterval(this.ontick, 1000 * data.details.targetduration); } if (!data.details.live && this.timer) { // playlist is not live and timer is armed : stopping it clearInterval(this.timer); this.timer = null; } } }, { key: 'tick', value: function tick() { var levelId = this._level; if (levelId !== undefined) { var level = this._levels[levelId], urlId = level.urlId; this.hls.trigger(_events2['default'].LEVEL_LOADING, { url: level.url[urlId], level: levelId, id: urlId }); } } }, { key: 'nextLoadLevel', value: function nextLoadLevel() { if (this._manualLevel !== -1) { return this._manualLevel; } else { return this.hls.abrController.nextAutoLevel; } } }, { key: 'levels', get: function get() { return this._levels; } }, { key: 'level', get: function get() { return this._level; }, set: function set(newLevel) { if (this._level !== newLevel || this._levels[newLevel].details === undefined) { this.setLevelInternal(newLevel); } } }, { key: 'manualLevel', get: function get() { return this._manualLevel; }, set: function set(newLevel) { this._manualLevel = newLevel; if (newLevel !== -1) { this.level = newLevel; } } }, { key: 'firstLevel', get: function get() { return this._firstLevel; }, set: function set(newLevel) { this._firstLevel = newLevel; } }, { key: 'startLevel', get: function get() { if (this._startLevel === undefined) { return this._firstLevel; } else { return this._startLevel; } }, set: function set(newLevel) { this._startLevel = newLevel; } }]); return LevelController; })(); exports['default'] = LevelController; module.exports = exports['default']; },{"../errors":11,"../events":12,"../utils/logger":21}],5:[function(require,module,exports){ /* * MSE Media Controller */ 'use strict'; Object.defineProperty(exports, '__esModule', { value: true }); var _createClass = (function () { function defineProperties(target, props) { for (var i = 0; i < props.length; i++) { var descriptor = props[i]; descriptor.enumerable = descriptor.enumerable || false; descriptor.configurable = true; if ('value' in descriptor) descriptor.writable = true; Object.defineProperty(target, descriptor.key, descriptor); } } return function (Constructor, protoProps, staticProps) { if (protoProps) defineProperties(Constructor.prototype, protoProps); if (staticProps) defineProperties(Constructor, staticProps); return Constructor; }; })(); function _interopRequireDefault(obj) { return obj && obj.__esModule ? obj : { 'default': obj }; } function _classCallCheck(instance, Constructor) { if (!(instance instanceof Constructor)) { throw new TypeError('Cannot call a class as a function'); } } var _demuxDemuxer = require('../demux/demuxer'); var _demuxDemuxer2 = _interopRequireDefault(_demuxDemuxer); var _events = require('../events'); var _events2 = _interopRequireDefault(_events); var _utilsLogger = require('../utils/logger'); var _utilsBinarySearch = require('../utils/binary-search'); var _utilsBinarySearch2 = _interopRequireDefault(_utilsBinarySearch); var _helperLevelHelper = require('../helper/level-helper'); var _helperLevelHelper2 = _interopRequireDefault(_helperLevelHelper); var _errors = require('../errors'); var State = { ERROR: -2, STARTING: -1, IDLE: 0, KEY_LOADING: 1, FRAG_LOADING: 2, WAITING_LEVEL: 3, PARSING: 4, PARSED: 5, APPENDING: 6, BUFFER_FLUSHING: 7 }; var MSEMediaController = (function () { function MSEMediaController(hls) { _classCallCheck(this, MSEMediaController); this.config = hls.config; this.hls = hls; // Source Buffer listeners this.onsbue = this.onSBUpdateEnd.bind(this); this.onsbe = this.onSBUpdateError.bind(this); // internal listeners this.onmediaatt0 = this.onMediaAttaching.bind(this); this.onmediadet0 = this.onMediaDetaching.bind(this); this.onmp = this.onManifestParsed.bind(this); this.onll = this.onLevelLoaded.bind(this); this.onfl = this.onFragLoaded.bind(this); this.onkl = this.onKeyLoaded.bind(this); this.onis = this.onInitSegment.bind(this); this.onfpg = this.onFragParsing.bind(this); this.onfp = this.onFragParsed.bind(this); this.onerr = this.onError.bind(this); this.ontick = this.tick.bind(this); hls.on(_events2['default'].MEDIA_ATTACHING, this.onmediaatt0); hls.on(_events2['default'].MEDIA_DETACHING, this.onmediadet0); hls.on(_events2['default'].MANIFEST_PARSED, this.onmp); } _createClass(MSEMediaController, [{ key: 'destroy', value: function destroy() { this.stop(); var hls = this.hls; hls.off(_events2['default'].MEDIA_ATTACHING, this.onmediaatt0); hls.off(_events2['default'].MEDIA_DETACHING, this.onmediadet0); hls.off(_events2['default'].MANIFEST_PARSED, this.onmp); this.state = State.IDLE; } }, { key: 'startLoad', value: function startLoad() { if (this.levels && this.media) { this.startInternal(); if (this.lastCurrentTime) { _utilsLogger.logger.log('seeking @ ' + this.lastCurrentTime); if (!this.lastPaused) { _utilsLogger.logger.log('resuming video'); this.media.play(); } this.state = State.IDLE; } else { this.lastCurrentTime = 0; this.state = State.STARTING; } this.nextLoadPosition = this.startPosition = this.lastCurrentTime; this.tick(); } else { _utilsLogger.logger.warn('cannot start loading as either manifest not parsed or video not attached'); } } }, { key: 'startInternal', value: function startInternal() { var hls = this.hls; this.stop(); this.demuxer = new _demuxDemuxer2['default'](hls); this.timer = setInterval(this.ontick, 100); this.level = -1; hls.on(_events2['default'].FRAG_LOADED, this.onfl); hls.on(_events2['default'].FRAG_PARSING_INIT_SEGMENT, this.onis); hls.on(_events2['default'].FRAG_PARSING_DATA, this.onfpg); hls.on(_events2['default'].FRAG_PARSED, this.onfp); hls.on(_events2['default'].ERROR, this.onerr); hls.on(_events2['default'].LEVEL_LOADED, this.onll); hls.on(_events2['default'].KEY_LOADED, this.onkl); } }, { key: 'stop', value: function stop() { this.mp4segments = []; this.flushRange = []; this.bufferRange = []; var frag = this.fragCurrent; if (frag) { if (frag.loader) { frag.loader.abort(); } this.fragCurrent = null; } this.fragPrevious = null; if (this.sourceBuffer) { for (var type in this.sourceBuffer) { var sb = this.sourceBuffer[type]; try { this.mediaSource.removeSourceBuffer(sb); sb.removeEventListener('updateend', this.onsbue); sb.removeEventListener('error', this.onsbe); } catch (err) {} } this.sourceBuffer = null; } if (this.timer) { clearInterval(this.timer); this.timer = null; } if (this.demuxer) { this.demuxer.destroy(); this.demuxer = null; } var hls = this.hls; hls.off(_events2['default'].FRAG_LOADED, this.onfl); hls.off(_events2['default'].FRAG_PARSED, this.onfp); hls.off(_events2['default'].FRAG_PARSING_DATA, this.onfpg); hls.off(_events2['default'].LEVEL_LOADED, this.onll); hls.off(_events2['default'].KEY_LOADED, this.onkl); hls.off(_events2['default'].FRAG_PARSING_INIT_SEGMENT, this.onis); hls.off(_events2['default'].ERROR, this.onerr); } }, { key: 'tick', value: function tick() { var pos, level, levelDetails, hls = this.hls; switch (this.state) { case State.ERROR: //don't do anything in error state to avoid breaking further ... break; case State.STARTING: // determine load level this.startLevel = hls.startLevel; if (this.startLevel === -1) { // -1 : guess start Level by doing a bitrate test by loading first fragment of lowest quality level this.startLevel = 0; this.fragBitrateTest = true; } // set new level to playlist loader : this will trigger start level load this.level = hls.nextLoadLevel = this.startLevel; this.state = State.WAITING_LEVEL; this.loadedmetadata = false; break; case State.IDLE: // if video detached or unbound exit loop if (!this.media) { break; } // determine next candidate fragment to be loaded, based on current position and // end of buffer position // ensure 60s of buffer upfront // if we have not yet loaded any fragment, start loading from start position if (this.loadedmetadata) { pos = this.media.currentTime; } else { pos = this.nextLoadPosition; } // determine next load level if (this.startFragmentRequested === false) { level = this.startLevel; } else { // we are not at playback start, get next load level from level Controller level = hls.nextLoadLevel; } var bufferInfo = this.bufferInfo(pos, 0.3), bufferLen = bufferInfo.len, bufferEnd = bufferInfo.end, fragPrevious = this.fragPrevious, maxBufLen; // compute max Buffer Length that we could get from this load level, based on level bitrate. don't buffer more than 60 MB and more than 30s if (this.levels[level].hasOwnProperty('bitrate')) { maxBufLen = Math.max(8 * this.config.maxBufferSize / this.levels[level].bitrate, this.config.maxBufferLength); maxBufLen = Math.min(maxBufLen, this.config.maxMaxBufferLength); } else { maxBufLen = this.config.maxBufferLength; } // if buffer length is less than maxBufLen try to load a new fragment if (bufferLen < maxBufLen) { // set next load level : this will trigger a playlist load if needed hls.nextLoadLevel = level; this.level = level; levelDetails = this.levels[level].details; // if level info not retrieved yet, switch state and wait for level retrieval if (typeof levelDetails === 'undefined') { this.state = State.WAITING_LEVEL; break; } // find fragment index, contiguous with end of buffer position var fragments = levelDetails.fragments, fragLen = fragments.length, start = fragments[0].start, end = fragments[fragLen - 1].start + fragments[fragLen - 1].duration, _frag = undefined; // in case of live playlist we need to ensure that requested position is not located before playlist start if (levelDetails.live) { // check if requested position is within seekable boundaries : //logger.log(`start/pos/bufEnd/seeking:${start.toFixed(3)}/${pos.toFixed(3)}/${bufferEnd.toFixed(3)}/${this.media.seeking}`); if (bufferEnd < Math.max(start, end - this.config.liveMaxLatencyDurationCount * levelDetails.targetduration)) { this.seekAfterBuffered = start + Math.max(0, levelDetails.totalduration - this.config.liveSyncDurationCount * levelDetails.targetduration); _utilsLogger.logger.log('buffer end: ' + bufferEnd + ' is located too far from the end of live sliding playlist, media position will be reseted to: ' + this.seekAfterBuffered.toFixed(3)); bufferEnd = this.seekAfterBuffered; } if (this.startFragmentRequested && !levelDetails.PTSKnown) { /* we are switching level on live playlist, but we don't have any PTS info for that quality level ... try to load frag matching with next SN. even if SN are not synchronized between playlists, loading this frag will help us compute playlist sliding and find the right one after in case it was not the right consecutive one */ if (fragPrevious) { var targetSN = fragPrevious.sn + 1; if (targetSN >= levelDetails.startSN && targetSN <= levelDetails.endSN) { _frag = fragments[targetSN - levelDetails.startSN]; _utilsLogger.logger.log('live playlist, switching playlist, load frag with next SN: ' + _frag.sn); } } if (!_frag) { /* we have no idea about which fragment should be loaded. so let's load mid fragment. it will help computing playlist sliding and find the right one */ _frag = fragments[Math.round(fragLen / 2)]; _utilsLogger.logger.log('live playlist, switching playlist, unknown, load middle frag : ' + _frag.sn); } } } else { // VoD playlist: if bufferEnd before start of playlist, load first fragment if (bufferEnd < start) { _frag = fragments[0]; } } if (!_frag) { if (bufferEnd > end) { // reach end of playlist break; } var foundFrag = _utilsBinarySearch2['default'].search(fragments, function (candidate) { //logger.log('level/sn/sliding/start/end/bufEnd:${level}/${candidate.sn}/${sliding.toFixed(3)}/${candidate.start.toFixed(3)}/${(candidate.start+candidate.duration).toFixed(3)}/${bufferEnd.toFixed(3)}'); // offset should be within fragment boundary if (candidate.start + candidate.duration <= bufferEnd) { return 1; } else if (candidate.start > bufferEnd) { return -1; } return 0; }); if (foundFrag) { _frag = foundFrag; start = foundFrag.start; //logger.log('find SN matching with pos:' + bufferEnd + ':' + frag.sn); if (fragPrevious && _frag.level === fragPrevious.level && _frag.sn === fragPrevious.sn) { if (_frag.sn < levelDetails.endSN) { _frag = fragments[_frag.sn + 1 - levelDetails.startSN]; _utilsLogger.logger.log('SN just loaded, load next one: ' + _frag.sn); } else { // end of VOD playlist reached _frag = null; } } } } if (_frag) { //logger.log(' loading frag ' + i +',pos/bufEnd:' + pos.toFixed(3) + '/' + bufferEnd.toFixed(3)); if (_frag.decryptdata.uri != null && _frag.decryptdata.key == null) { _utilsLogger.logger.log('Loading key for ' + _frag.sn + ' of [' + levelDetails.startSN + ' ,' + levelDetails.endSN + '],level ' + level); this.state = State.KEY_LOADING; hls.trigger(_events2['default'].KEY_LOADING, { frag: _frag }); } else { _utilsLogger.logger.log('Loading ' + _frag.sn + ' of [' + levelDetails.startSN + ' ,' + levelDetails.endSN + '],level ' + level + ', currentTime:' + pos + ',bufferEnd:' + bufferEnd.toFixed(3)); _frag.autoLevel = hls.autoLevelEnabled; if (this.levels.length > 1) { _frag.expectedLen = Math.round(_frag.duration * this.levels[level].bitrate / 8); _frag.trequest = performance.now(); } // ensure that we are not reloading the same fragments in loop ... if (this.fragLoadIdx !== undefined) { this.fragLoadIdx++; } else { this.fragLoadIdx = 0; } if (_frag.loadCounter) { _frag.loadCounter++; var maxThreshold = this.config.fragLoadingLoopThreshold; // if this frag has already been loaded 3 times, and if it has been reloaded recently if (_frag.loadCounter > maxThreshold && Math.abs(this.fragLoadIdx - _frag.loadIdx) < maxThreshold) { hls.trigger(_events2['default'].ERROR, { type: _errors.ErrorTypes.MEDIA_ERROR, details: _errors.ErrorDetails.FRAG_LOOP_LOADING_ERROR, fatal: false, frag: _frag }); return; } } else { _frag.loadCounter = 1; } _frag.loadIdx = this.fragLoadIdx; this.fragCurrent = _frag; this.startFragmentRequested = true; hls.trigger(_events2['default'].FRAG_LOADING, { frag: _frag }); this.state = State.FRAG_LOADING; } } } break; case State.WAITING_LEVEL: level = this.levels[this.level]; // check if playlist is already loaded if (level && level.details) { this.state = State.IDLE; } break; case State.FRAG_LOADING: /* monitor fragment retrieval time... we compute expected time of arrival of the complete fragment. we compare it to expected time of buffer starvation */ var v = this.media, frag = this.fragCurrent; /* only monitor frag retrieval time if (video not paused OR first fragment being loaded) AND autoswitching enabled AND not lowest level AND multiple levels */ if (v && (!v.paused || this.loadedmetadata === false) && frag.autoLevel && this.level && this.levels.length > 1) { var requestDelay = performance.now() - frag.trequest; // monitor fragment load progress after half of expected fragment duration,to stabilize bitrate if (requestDelay > 500 * frag.duration) { var loadRate = frag.loaded * 1000 / requestDelay; // byte/s if (frag.expectedLen < frag.loaded) { frag.expectedLen = frag.loaded; } pos = v.currentTime; var fragLoadedDelay = (frag.expectedLen - frag.loaded) / loadRate; var bufferStarvationDelay = this.bufferInfo(pos, 0.3).end - pos; var fragLevelNextLoadedDelay = frag.duration * this.levels[hls.nextLoadLevel].bitrate / (8 * loadRate); //bps/Bps /* if we have less than 2 frag duration in buffer and if frag loaded delay is greater than buffer starvation delay ... and also bigger than duration needed to load fragment at next level ...*/ if (bufferStarvationDelay < 2 * frag.duration && fragLoadedDelay > bufferStarvationDelay && fragLoadedDelay > fragLevelNextLoadedDelay) { // abort fragment loading ... _utilsLogger.logger.warn('loading too slow, abort fragment loading'); _utilsLogger.logger.log('fragLoadedDelay/bufferStarvationDelay/fragLevelNextLoadedDelay :' + fragLoadedDelay.toFixed(1) + '/' + bufferStarvationDelay.toFixed(1) + '/' + fragLevelNextLoadedDelay.toFixed(1)); //abort fragment loading frag.loader.abort(); hls.trigger(_events2['default'].FRAG_LOAD_EMERGENCY_ABORTED, { frag: frag }); // switch back to IDLE state to request new fragment at lowest level this.state = State.IDLE; } } } break; case State.PARSING: // nothing to do, wait for fragment being parsed break; case State.PARSED: case State.APPENDING: if (this.sourceBuffer) { if (this.media.error) { _utilsLogger.logger.error('trying to append although a media error occured, switch to ERROR state'); this.state = State.ERROR; return; } // if MP4 segment appending in progress nothing to do else if (this.sourceBuffer.audio && this.sourceBuffer.audio.updating || this.sourceBuffer.video && this.sourceBuffer.video.updating) { //logger.log('sb append in progress'); // check if any MP4 segments left to append } else if (this.mp4segments.length) { var segment = this.mp4segments.shift(); try { //logger.log(`appending ${segment.type} SB, size:${segment.data.length}); this.sourceBuffer[segment.type].appendBuffer(segment.data); this.appendError = 0; } catch (err) { // in case any error occured while appending, put back segment in mp4segments table //logger.error(`error while trying to append buffer:${err.message},try appending later`); this.mp4segments.unshift(segment); if (this.appendError) { this.appendError++; } else { this.appendError = 1; } var event = { type: _errors.ErrorTypes.MEDIA_ERROR, details: _errors.ErrorDetails.BUFFER_APPEND_ERROR, frag: this.fragCurrent }; /* with UHD content, we could get loop of quota exceeded error until browser is able to evict some data from sourcebuffer. retrying help recovering this */ if (this.appendError > this.config.appendErrorMaxRetry) { _utilsLogger.logger.log('fail ' + this.config.appendErrorMaxRetry + ' times to append segment in sourceBuffer'); event.fatal = true; hls.trigger(_events2['default'].ERROR, event); this.state = State.ERROR; return; } else { event.fatal = false; hls.trigger(_events2['default'].ERROR, event); } } this.state = State.APPENDING; } } else { // sourceBuffer undefined, switch back to IDLE state this.state = State.IDLE; } break; case State.BUFFER_FLUSHING: // loop through all buffer ranges to flush while (this.flushRange.length) { var range = this.flushRange[0]; // flushBuffer will abort any buffer append in progress and flush Audio/Video Buffer if (this.flushBuffer(range.start, range.end)) { // range flushed, remove from flush array this.flushRange.shift(); } else { // flush in progress, come back later break; } } if (this.flushRange.length === 0) { // handle end of immediate switching if needed if (this.immediateSwitch) { this.immediateLevelSwitchEnd(); } // move to IDLE once flush complete. this should trigger new fragment loading this.state = State.IDLE; // reset reference to frag this.fragPrevious = null; } /* if not everything flushed, stay in BUFFER_FLUSHING state. we will come back here each time sourceBuffer updateend() callback will be triggered */ break; default: break; } // check/update current fragment this._checkFragmentChanged(); // check buffer this._checkBuffer(); } }, { key: 'bufferInfo', value: function bufferInfo(pos, maxHoleDuration) { var media = this.media, vbuffered = media.buffered, buffered = [], i; for (i = 0; i < vbuffered.length; i++) { buffered.push({ start: vbuffered.start(i), end: vbuffered.end(i) }); } return this.bufferedInfo(buffered, pos, maxHoleDuration); } }, { key: 'bufferedInfo', value: function bufferedInfo(buffered, pos, maxHoleDuration) { var buffered2 = [], // bufferStart and bufferEnd are buffer boundaries around current video position bufferLen, bufferStart, bufferEnd, bufferStartNext, i; // sort on buffer.start/smaller end (IE does not always return sorted buffered range) buffered.sort(function (a, b) { var diff = a.start - b.start; if (diff) { return diff; } else { return b.end - a.end; } }); // there might be some small holes between buffer time range // consider that holes smaller than maxHoleDuration are irrelevant and build another // buffer time range representations that discards those holes for (i = 0; i < buffered.length; i++) { var buf2len = buffered2.length; if (buf2len) { var buf2end = buffered2[buf2len - 1].end; // if small hole (value between 0 or maxHoleDuration ) or overlapping (negative) if (buffered[i].start - buf2end < maxHoleDuration) { // merge overlapping time ranges // update lastRange.end only if smaller than item.end // e.g. [ 1, 15] with [ 2,8] => [ 1,15] (no need to modify lastRange.end) // whereas [ 1, 8] with [ 2,15] => [ 1,15] ( lastRange should switch from [1,8] to [1,15]) if (buffered[i].end > buf2end) { buffered2[buf2len - 1].end = buffered[i].end; } } else { // big hole buffered2.push(buffered[i]); } } else { // first value buffered2.push(buffered[i]); } } for (i = 0, bufferLen = 0, bufferStart = bufferEnd = pos; i < buffered2.length; i++) { var start = buffered2[i].start, end = buffered2[i].end; //logger.log('buf start/end:' + buffered.start(i) + '/' + buffered.end(i)); if (pos + maxHoleDuration >= start && pos < end) { // play position is inside this buffer TimeRange, retrieve end of buffer position and buffer length bufferStart = start; bufferEnd = end; bufferLen = bufferEnd - pos; } else if (pos + maxHoleDuration < start) { bufferStartNext = start; } } return { len: bufferLen, start: bufferStart, end: bufferEnd, nextStart: bufferStartNext }; } }, { key: 'getBufferRange', value: function getBufferRange(position) { var i, range; for (i = this.bufferRange.length - 1; i >= 0; i--) { range = this.bufferRange[i]; if (position >= range.start && position <= range.end) { return range; } } return null; } }, { key: 'followingBufferRange', value: function followingBufferRange(range) { if (range) { // try to get range of next fragment (500ms after this range) return this.getBufferRange(range.end + 0.5); } return null; } }, { key: 'isBuffered', value: function isBuffered(position) { var v = this.media, buffered = v.buffered; for (var i = 0; i < buffered.length; i++) { if (position >= buffered.start(i) && position <= buffered.end(i)) { return true; } } return false; } }, { key: '_checkFragmentChanged', value: function _checkFragmentChanged() { var rangeCurrent, currentTime, video = this.media; if (video && video.seeking === false) { currentTime = video.currentTime; /* if video element is in seeked state, currentTime can only increase. (assuming that playback rate is positive ...) As sometimes currentTime jumps back to zero after a media decode error, check this, to avoid seeking back to wrong position after a media decode error */ if (currentTime > video.playbackRate * this.lastCurrentTime) { this.lastCurrentTime = currentTime; } if (this.isBuffered(currentTime)) { rangeCurrent = this.getBufferRange(currentTime); } else if (this.isBuffered(currentTime + 0.1)) { /* ensure that FRAG_CHANGED event is triggered at startup, when first video frame is displayed and playback is paused. add a tolerance of 100ms, in case current position is not buffered, check if current pos+100ms is buffered and use that buffer range for FRAG_CHANGED event reporting */ rangeCurrent = this.getBufferRange(currentTime + 0.1); } if (rangeCurrent) { var fragPlaying = rangeCurrent.frag; if (fragPlaying !== this.fragPlaying) { this.fragPlaying = fragPlaying; this.hls.trigger(_events2['default'].FRAG_CHANGED, { frag: fragPlaying }); } } } } /* abort any buffer append in progress, and flush all buffered data return true once everything has been flushed. sourceBuffer.abort() and sourceBuffer.remove() are asynchronous operations the idea is to call this function from tick() timer and call it again until all resources have been cleaned the timer is rearmed upon sourceBuffer updateend() event, so this should be optimal */ }, { key: 'flushBuffer', value: function flushBuffer(startOffset, endOffset) { var sb, i, bufStart, bufEnd, flushStart, flushEnd; //logger.log('flushBuffer,pos/start/end: ' + this.media.currentTime + '/' + startOffset + '/' + endOffset); // safeguard to avoid infinite looping if (this.flushBufferCounter++ < 2 * this.bufferRange.length && this.sourceBuffer) { for (var type in this.sourceBuffer) { sb = this.sourceBuffer[type]; if (!sb.updating) { for (i = 0; i < sb.buffered.length; i++) { bufStart = sb.buffered.start(i); bufEnd = sb.buffered.end(i); // workaround firefox not able to properly flush multiple buffered range. if (navigator.userAgent.toLowerCase().indexOf('firefox') !== -1 && endOffset === Number.POSITIVE_INFINITY) { flushStart = startOffset; flushEnd = endOffset; } else { flushStart = Math.max(bufStart, startOffset); flushEnd = Math.min(bufEnd, endOffset); } /* sometimes sourcebuffer.remove() does not flush the exact expected time range. to avoid rounding issues/infinite loop, only flush buffer range of length greater than 500ms. */ if (flushEnd - flushStart > 0.5) { _utilsLogger.logger.log('flush ' + type + ' [' + flushStart + ',' + flushEnd + '], of [' + bufStart + ',' + bufEnd + '], pos:' + this.media.currentTime); sb.remove(flushStart, flushEnd); return false; } } } else { //logger.log('abort ' + type + ' append in progress'); // this will abort any appending in progress //sb.abort(); return false; } } } /* after successful buffer flushing, rebuild buffer Range array loop through existing buffer range and check if corresponding range is still buffered. only push to new array already buffered range */ var newRange = [], range; for (i = 0; i < this.bufferRange.length; i++) { range = this.bufferRange[i]; if (this.isBuffered((range.start + range.end) / 2)) { newRange.push(range); } } this.bufferRange = newRange; _utilsLogger.logger.log('buffer flushed'); // everything flushed ! return true; } /* on immediate level switch : - pause playback if playing - cancel any pending load request - and trigger a buffer flush */ }, { key: 'immediateLevelSwitch', value: function immediateLevelSwitch() { _utilsLogger.logger.log('immediateLevelSwitch'); if (!this.immediateSwitch) { this.immediateSwitch = true; this.previouslyPaused = this.media.paused; this.media.pause(); } var fragCurrent = this.fragCurrent; if (fragCurrent && fragCurrent.loader) { fragCurrent.loader.abort(); } this.fragCurrent = null; // flush everything this.flushBufferCounter = 0; this.flushRange.push({ start: 0, end: Number.POSITIVE_INFINITY }); // trigger a sourceBuffer flush this.state = State.BUFFER_FLUSHING; // increase fragment load Index to avoid frag loop loading error after buffer flush this.fragLoadIdx += 2 * this.config.fragLoadingLoopThreshold; // speed up switching, trigger timer function this.tick(); } /* on immediate level switch end, after new fragment has been buffered : - nudge video decoder by slightly adjusting video currentTime - resume the playback if needed */ }, { key: 'immediateLevelSwitchEnd', value: function immediateLevelSwitchEnd() { this.immediateSwitch = false; this.media.currentTime -= 0.0001; if (!this.previouslyPaused) { this.media.play(); } } }, { key: 'nextLevelSwitch', value: function nextLevelSwitch() { /* try to switch ASAP without breaking video playback : in order to ensure smooth but quick level switching, we need to find the next flushable buffer range we should take into account new segment fetch time */ var fetchdelay, currentRange, nextRange; currentRange = this.getBufferRange(this.media.currentTime); if (currentRange) { // flush buffer preceding current fragment (flush until current fragment start offset) // minus 1s to avoid video freezing, that could happen if we flush keyframe of current video ... this.flushRange.push({ start: 0, end: currentRange.start - 1 }); } if (!this.media.paused) { // add a safety delay of 1s var nextLevelId = this.hls.nextLoadLevel, nextLevel = this.levels[nextLevelId], fragLastKbps = this.fragLastKbps; if (fragLastKbps && this.fragCurrent) { fetchdelay = this.fragCurrent.duration * nextLevel.bitrate / (1000 * fragLastKbps) + 1; } else { fetchdelay = 0; } } else { fetchdelay = 0; } //logger.log('fetchdelay:'+fetchdelay); // find buffer range that will be reached once new fragment will be fetched nextRange = this.getBufferRange(this.media.currentTime + fetchdelay); if (nextRange) { // we can flush buffer range following this one without stalling playback nextRange = this.followingBufferRange(nextRange); if (nextRange) { // flush position is the start position of this new buffer this.flushRange.push({ start: nextRange.start, end: Number.POSITIVE_INFINITY }); // if we are here, we can also cancel any loading/demuxing in progress, as they are useless var fragCurrent = this.fragCurrent; if (fragCurrent && fragCurrent.loader) { fragCurrent.loader.abort(); } this.fragCurrent = null; } } if (this.flushRange.length) { this.flushBufferCounter = 0; // trigger a sourceBuffer flush this.state = State.BUFFER_FLUSHING; // increase fragment load Index to avoid frag loop loading error after buffer flush this.fragLoadIdx += 2 * this.config.fragLoadingLoopThreshold; // speed up switching, trigger timer function this.tick(); } } }, { key: 'onMediaAttaching', value: function onMediaAttaching(event, data) { var media = this.media = data.media; // setup the media source var ms = this.mediaSource = new MediaSource(); //Media Source listeners this.onmso = this.onMediaSourceOpen.bind(this); this.onmse = this.onMediaSourceEnded.bind(this); this.onmsc = this.onMediaSourceClose.bind(this); ms.addEventListener('sourceopen', this.onmso); ms.addEventListener('sourceended', this.onmse); ms.addEventListener('sourceclose', this.onmsc); // link video and media Source media.src = URL.createObjectURL(ms); // FIXME: this was in code before but onverror was never set! can be removed or fixed? //media.addEventListener('error', this.onverror); } }, { key: 'onMediaDetaching', value: function onMediaDetaching() { var media = this.media; if (media && media.ended) { _utilsLogger.logger.log('MSE detaching and video ended, reset startPosition'); this.startPosition = this.lastCurrentTime = 0; } // reset fragment loading counter on MSE detaching to avoid reporting FRAG_LOOP_LOADING_ERROR after error recovery var levels = this.levels; if (levels) { // reset fragment load counter levels.forEach(function (level) { if (level.details) { level.details.fragments.forEach(function (fragment) { fragment.loadCounter = undefined; }); } }); } var ms = this.mediaSource; if (ms) { if (ms.readyState === 'open') { ms.endOfStream(); } ms.removeEventListener('sourceopen', this.onmso); ms.removeEventListener('sourceended', this.onmse); ms.removeEventListener('sourceclose', this.onmsc); // unlink MediaSource from video tag this.media.src = ''; this.mediaSource = null; // remove video listeners if (media) { media.removeEventListener('seeking', this.onvseeking); media.removeEventListener('seeked', this.onvseeked); media.removeEventListener('loadedmetadata', this.onvmetadata); media.removeEventListener('ended', this.onvended); this.onvseeking = this.onvseeked = this.onvmetadata = null; } this.media = null; this.loadedmetadata = false; this.stop(); } this.onmso = this.onmse = this.onmsc = null; this.hls.trigger(_events2['default'].MEDIA_DETACHED); } }, { key: 'onMediaSeeking', value: function onMediaSeeking() { if (this.state === State.FRAG_LOADING) { // check if currently loaded fragment is inside buffer. //if outside, cancel fragment loading, otherwise do nothing if (this.bufferInfo(this.media.currentTime, 0.3).len === 0) { _utilsLogger.logger.log('seeking outside of buffer while fragment load in progress, cancel fragment load'); var fragCurrent = this.fragCurrent; if (fragCurrent) { if (fragCurrent.loader) { fragCurrent.loader.abort(); } this.fragCurrent = null; } this.fragPrevious = null; // switch to IDLE state to load new fragment this.state = State.IDLE; } } if (this.media) { this.lastCurrentTime = this.media.currentTime; } // avoid reporting fragment loop loading error in case user is seeking several times on same position if (this.fragLoadIdx !== undefined) { this.fragLoadIdx += 2 * this.config.fragLoadingLoopThreshold; } // tick to speed up processing this.tick(); } }, { key: 'onMediaSeeked', value: function onMediaSeeked() { // tick to speed up FRAGMENT_PLAYING triggering this.tick(); } }, { key: 'onMediaMetadata', value: function onMediaMetadata() { if (this.media.currentTime !== this.startPosition) { this.media.currentTime = this.startPosition; } this.loadedmetadata = true; this.tick(); } }, { key: 'onMediaEnded', value: function onMediaEnded() { _utilsLogger.logger.log('media ended'); // reset startPosition and lastCurrentTime to restart playback @ stream beginning this.startPosition = this.lastCurrentTime = 0; } }, { key: 'onManifestParsed', value: function onManifestParsed(event, data) { var aac = false, heaac = false, codecs; data.levels.forEach(function (level) { // detect if we have different kind of audio codecs used amongst playlists codecs = level.codecs; if (codecs) { if (codecs.indexOf('mp4a.40.2') !== -1) { aac = true; } if (codecs.indexOf('mp4a.40.5') !== -1) { heaac = true; } } }); this.audiocodecswitch = aac && heaac; if (this.audiocodecswitch) { _utilsLogger.logger.log('both AAC/HE-AAC audio found in levels; declaring audio codec as HE-AAC'); } this.levels = data.levels; this.startLevelLoaded = false; this.startFragmentRequested = false; if (this.media && this.config.autoStartLoad) { this.startLoad(); } } }, { key: 'onLevelLoaded', value: function onLevelLoaded(event, data) { var newDetails = data.details, newLevelId = data.level, curLevel = this.levels[newLevelId], duration = newDetails.totalduration; _utilsLogger.logger.log('level ' + newLevelId + ' loaded [' + newDetails.startSN + ',' + newDetails.endSN + '],duration:' + duration); if (newDetails.live) { var curDetails = curLevel.details; if (curDetails) { // we already have details for that level, merge them _helperLevelHelper2['default'].mergeDetails(curDetails, newDetails); if (newDetails.PTSKnown) { _utilsLogger.logger.log('live playlist sliding:' + newDetails.fragments[0].start.toFixed(3)); } else { _utilsLogger.logger.log('live playlist - outdated PTS, unknown sliding'); } } else { newDetails.PTSKnown = false; _utilsLogger.logger.log('live playlist - first load, unknown sliding'); } } else { newDetails.PTSKnown = false; } // override level info curLevel.details = newDetails; this.hls.trigger(_events2['default'].LEVEL_UPDATED, { details: newDetails, level: newLevelId }); // compute start position if (this.startLevelLoaded === false) { // if live playlist, set start position to be fragment N-this.config.liveSyncDurationCount (usually 3) if (newDetails.live) { this.startPosition = Math.max(0, duration - this.config.liveSyncDurationCount * newDetails.targetduration); } this.nextLoadPosition = this.startPosition; this.startLevelLoaded = true; } // only switch batck to IDLE state if we were waiting for level to start downloading a new fragment if (this.state === State.WAITING_LEVEL) { this.state = State.IDLE; } //trigger handler right now this.tick(); } }, { key: 'onKeyLoaded', value: function onKeyLoaded() { if (this.state === State.KEY_LOADING) { this.state = State.IDLE; this.tick(); } } }, { key: 'onFragLoaded', value: function onFragLoaded(event, data) { var fragCurrent = this.fragCurrent; if (this.state === State.FRAG_LOADING && fragCurrent && data.frag.level === fragCurrent.level && data.frag.sn === fragCurrent.sn) { if (this.fragBitrateTest === true) { // switch back to IDLE state ... we just loaded a fragment to determine adequate start bitrate and initialize autoswitch algo this.state = State.IDLE; this.fragBitrateTest = false; data.stats.tparsed = data.stats.tbuffered = performance.now(); this.hls.trigger(_events2['default'].FRAG_BUFFERED, { stats: data.stats, frag: fragCurrent }); } else { this.state = State.PARSING; // transmux the MPEG-TS data to ISO-BMFF segments this.stats = data.stats; var currentLevel = this.levels[this.level], details = currentLevel.details, duration = details.totalduration, start = fragCurrent.start, level = fragCurrent.level, sn = fragCurrent.sn; _utilsLogger.logger.log('Demuxing ' + sn + ' of [' + details.startSN + ' ,' + details.endSN + '],level ' + level); this.demuxer.push(data.payload, currentLevel.audioCodec, currentLevel.videoCodec, start, fragCurrent.cc, level, sn, duration, fragCurrent.decryptdata); } } } }, { key: 'onInitSegment', value: function onInitSegment(event, data) { if (this.state === State.PARSING) { // check if codecs have been explicitely defined in the master playlist for this level; // if yes use these ones instead of the ones parsed from the demux var audioCodec = this.levels[this.level].audioCodec, videoCodec = this.levels[this.level].videoCodec, sb; //logger.log('playlist level A/V codecs:' + audioCodec + ',' + videoCodec); //logger.log('playlist codecs:' + codec); // if playlist does not specify codecs, use codecs found while parsing fragment if (audioCodec === undefined || data.audiocodec === undefined) { audioCodec = data.audioCodec; } if (videoCodec === undefined || data.videocodec === undefined) { videoCodec = data.videoCodec; } // in case several audio codecs might be used, force HE-AAC for audio (some browsers don't support audio codec switch) //don't do it for mono streams ... var ua = navigator.userAgent.toLowerCase(); if (this.audiocodecswitch && data.audioChannelCount !== 1 && ua.indexOf('android') === -1 && ua.indexOf('firefox') === -1) { audioCodec = 'mp4a.40.5'; } if (!this.sourceBuffer) { this.sourceBuffer = {}; _utilsLogger.logger.log('selected A/V codecs for sourceBuffers:' + audioCodec + ',' + videoCodec); // create source Buffer and link them to MediaSource if (audioCodec) { sb = this.sourceBuffer.audio = this.mediaSource.addSourceBuffer('video/mp4;codecs=' + audioCodec); sb.addEventListener('updateend', this.onsbue); sb.addEventListener('error', this.onsbe); } if (videoCodec) { sb = this.sourceBuffer.video = this.mediaSource.addSourceBuffer('video/mp4;codecs=' + videoCodec); sb.addEventListener('updateend', this.onsbue); sb.addEventListener('error', this.onsbe); } } if (audioCodec) { this.mp4segments.push({ type: 'audio', data: data.audioMoov }); } if (videoCodec) { this.mp4segments.push({ type: 'video', data: data.videoMoov }); } //trigger handler right now this.tick(); } } }, { key: 'onFragParsing', value: function onFragParsing(event, data) { if (this.state === State.PARSING) { this.tparse2 = Date.now(); var level = this.levels[this.level], frag = this.fragCurrent; _utilsLogger.logger.log('parsed data, type/startPTS/endPTS/startDTS/endDTS/nb:' + data.type + '/' + data.startPTS.toFixed(3) + '/' + data.endPTS.toFixed(3) + '/' + data.startDTS.toFixed(3) + '/' + data.endDTS.toFixed(3) + '/' + data.nb); var drift = _helperLevelHelper2['default'].updateFragPTS(level.details, frag.sn, data.startPTS, data.endPTS); this.hls.trigger(_events2['default'].LEVEL_PTS_UPDATED, { details: level.details, level: this.level, drift: drift }); this.mp4segments.push({ type: data.type, data: data.moof }); this.mp4segments.push({ type: data.type, data: data.mdat }); this.nextLoadPosition = data.endPTS; this.bufferRange.push({ type: data.type, start: data.startPTS, end: data.endPTS, frag: frag }); //trigger handler right now this.tick(); } else { _utilsLogger.logger.warn('not in PARSING state, discarding ' + event); } } }, { key: 'onFragParsed', value: function onFragParsed() { if (this.state === State.PARSING) { this.state = State.PARSED; this.stats.tparsed = performance.now(); //trigger handler right now this.tick(); } } }, { key: 'onError', value: function onError(event, data) { switch (data.details) { // abort fragment loading on errors case _errors.ErrorDetails.FRAG_LOAD_ERROR: case _errors.ErrorDetails.FRAG_LOAD_TIMEOUT: case _errors.ErrorDetails.FRAG_LOOP_LOADING_ERROR: case _errors.ErrorDetails.LEVEL_LOAD_ERROR: case _errors.ErrorDetails.LEVEL_LOAD_TIMEOUT: case _errors.ErrorDetails.KEY_LOAD_ERROR: case _errors.ErrorDetails.KEY_LOAD_TIMEOUT: // if fatal error, stop processing, otherwise move to IDLE to retry loading _utilsLogger.logger.warn('buffer controller: ' + data.details + ' while loading frag,switch to ' + (data.fatal ? 'ERROR' : 'IDLE') + ' state ...'); this.state = data.fatal ? State.ERROR : State.IDLE; break; default: break; } } }, { key: 'onSBUpdateEnd', value: function onSBUpdateEnd() { //trigger handler right now if (this.state === State.APPENDING && this.mp4segments.length === 0) { var frag = this.fragCurrent, stats = this.stats; if (frag) { this.fragPrevious = frag; stats.tbuffered = performance.now(); this.fragLastKbps = Math.round(8 * stats.length / (stats.tbuffered - stats.tfirst)); this.hls.trigger(_events2['default'].FRAG_BUFFERED, { stats: stats, frag: frag }); _utilsLogger.logger.log('media buffered : ' + this.timeRangesToString(this.media.buffered)); // if stream is VOD (not live) and we reach End of Stream var levelDetails = this.levels[this.level].details; if (levelDetails && !levelDetails.live) { // have we buffered last fragment ? if (frag.sn === levelDetails.endSN) { var mediaSource = this.mediaSource; if (mediaSource && mediaSource.readyState === 'open') { _utilsLogger.logger.log('all media data available, signal endOfStream() to MediaSource'); //Notify the media element that it now has all of the media data mediaSource.endOfStream(); } } } this.state = State.IDLE; } } this.tick(); } }, { key: '_checkBuffer', value: function _checkBuffer() { var media = this.media; if (media) { // compare readyState var readyState = media.readyState; //logger.log(`readyState:${readyState}`); // if ready state different from HAVE_NOTHING (numeric value 0), we are allowed to seek if (readyState) { // if seek after buffered defined, let's seek if within acceptable range var seekAfterBuffered = this.seekAfterBuffered; if (seekAfterBuffered) { if (media.duration >= seekAfterBuffered) { media.currentTime = seekAfterBuffered; this.seekAfterBuffered = undefined; } } else if (readyState < 3) { // readyState = 1 or 2 // HAVE_METADATA (numeric value 1) Enough of the resource has been obtained that the duration of the resource is available. // The API will no longer throw an exception when seeking. // HAVE_CURRENT_DATA (numeric value 2) Data for the immediate current playback position is available, // but either not enough data is available that the user agent could // successfully advance the current playback position var currentTime = media.currentTime; var bufferInfo = this.bufferInfo(currentTime, 0); // check if current time is buffered or not if (bufferInfo.len === 0) { // no buffer available @ currentTime, check if next buffer is close (in a 300 ms range) var nextBufferStart = bufferInfo.nextStart; if (nextBufferStart && nextBufferStart - currentTime < 0.3) { // next buffer is close ! adjust currentTime to nextBufferStart // this will ensure effective video decoding _utilsLogger.logger.log('adjust currentTime from ' + currentTime + ' to ' + nextBufferStart); media.currentTime = nextBufferStart; } } } } } } }, { key: 'onSBUpdateError', value: function onSBUpdateError(event) { _utilsLogger.logger.error('sourceBuffer error:' + event); this.state = State.ERROR; this.hls.trigger(_events2['default'].ERROR, { type: _errors.ErrorTypes.MEDIA_ERROR, details: _errors.ErrorDetails.BUFFER_APPENDING_ERROR, fatal: true, frag: this.fragCurrent }); } }, { key: 'timeRangesToString', value: function timeRangesToString(r) { var log = '', len = r.length; for (var i = 0; i < len; i++) { log += '[' + r.start(i) + ',' + r.end(i) + ']'; } return log; } }, { key: 'onMediaSourceOpen', value: function onMediaSourceOpen() { _utilsLogger.logger.log('media source opened'); this.hls.trigger(_events2['default'].MEDIA_ATTACHED); this.onvseeking = this.onMediaSeeking.bind(this); this.onvseeked = this.onMediaSeeked.bind(this); this.onvmetadata = this.onMediaMetadata.bind(this); this.onvended = this.onMediaEnded.bind(this); var media = this.media; media.addEventListener('seeking', this.onvseeking); media.addEventListener('seeked', this.onvseeked); media.addEventListener('loadedmetadata', this.onvmetadata); media.addEventListener('ended', this.onvended); if (this.levels && this.config.autoStartLoad) { this.startLoad(); } // once received, don't listen anymore to sourceopen event this.mediaSource.removeEventListener('sourceopen', this.onmso); } }, { key: 'onMediaSourceClose', value: function onMediaSourceClose() { _utilsLogger.logger.log('media source closed'); } }, { key: 'onMediaSourceEnded', value: function onMediaSourceEnded() { _utilsLogger.logger.log('media source ended'); } }, { key: 'currentLevel', get: function get() { if (this.media) { var range = this.getBufferRange(this.media.currentTime); if (range) { return range.frag.level; } } return -1; } }, { key: 'nextBufferRange', get: function get() { if (this.media) { // first get end range of current fragment return this.followingBufferRange(this.getBufferRange(this.media.currentTime)); } else { return null; } } }, { key: 'nextLevel', get: function get() { var range = this.nextBufferRange; if (range) { return range.frag.level; } else { return -1; } } }]); return MSEMediaController; })(); exports['default'] = MSEMediaController; module.exports = exports['default']; },{"../demux/demuxer":8,"../errors":11,"../events":12,"../helper/level-helper":13,"../utils/binary-search":20,"../utils/logger":21}],6:[function(require,module,exports){ /* inline demuxer. * probe fragments and instantiate appropriate demuxer depending on content type (TSDemuxer, AACDemuxer, ...) */ 'use strict'; Object.defineProperty(exports, '__esModule', { value: true }); var _createClass = (function () { function defineProperties(target, props) { for (var i = 0; i < props.length; i++) { var descriptor = props[i]; descriptor.enumerable = descriptor.enumerable || false; descriptor.configurable = true; if ('value' in descriptor) descriptor.writable = true; Object.defineProperty(target, descriptor.key, descriptor); } } return function (Constructor, protoProps, staticProps) { if (protoProps) defineProperties(Constructor.prototype, protoProps); if (staticProps) defineProperties(Constructor, staticProps); return Constructor; }; })(); function _interopRequireDefault(obj) { return obj && obj.__esModule ? obj : { 'default': obj }; } function _classCallCheck(instance, Constructor) { if (!(instance instanceof Constructor)) { throw new TypeError('Cannot call a class as a function'); } } var _events = require('../events'); var _events2 = _interopRequireDefault(_events); var _errors = require('../errors'); var _demuxTsdemuxer = require('../demux/tsdemuxer'); var _demuxTsdemuxer2 = _interopRequireDefault(_demuxTsdemuxer); var DemuxerInline = (function () { function DemuxerInline(hls, remuxer) { _classCallCheck(this, DemuxerInline); this.hls = hls; this.remuxer = remuxer; } _createClass(DemuxerInline, [{ key: 'destroy', value: function destroy() { var demuxer = this.demuxer; if (demuxer) { demuxer.destroy(); } } }, { key: 'push', value: function push(data, audioCodec, videoCodec, timeOffset, cc, level, sn, duration) { var demuxer = this.demuxer; if (!demuxer) { // probe for content type if (_demuxTsdemuxer2['default'].probe(data)) { demuxer = this.demuxer = new _demuxTsdemuxer2['default'](this.hls, this.remuxer); } else { this.hls.trigger(_events2['default'].ERROR, { type: _errors.ErrorTypes.MEDIA_ERROR, details: _errors.ErrorDetails.FRAG_PARSING_ERROR, fatal: true, reason: 'no demux matching with content found' }); return; } } demuxer.push(data, audioCodec, videoCodec, timeOffset, cc, level, sn, duration); } }, { key: 'remux', value: function remux() { var demuxer = this.demuxer; if (demuxer) { demuxer.remux(); } } }]); return DemuxerInline; })(); exports['default'] = DemuxerInline; module.exports = exports['default']; },{"../demux/tsdemuxer":10,"../errors":11,"../events":12}],7:[function(require,module,exports){ /* demuxer web worker. * - listen to worker message, and trigger DemuxerInline upon reception of Fragments. * - provides MP4 Boxes back to main thread using [transferable objects](https://developers.google.com/web/updates/2011/12/Transferable-Objects-Lightning-Fast) in order to minimize message passing overhead. */ 'use strict'; Object.defineProperty(exports, '__esModule', { value: true }); function _interopRequireDefault(obj) { return obj && obj.__esModule ? obj : { 'default': obj }; } var _demuxDemuxerInline = require('../demux/demuxer-inline'); var _demuxDemuxerInline2 = _interopRequireDefault(_demuxDemuxerInline); var _events = require('../events'); var _events2 = _interopRequireDefault(_events); var _events3 = require('events'); var _events4 = _interopRequireDefault(_events3); var _remuxMp4Remuxer = require('../remux/mp4-remuxer'); var _remuxMp4Remuxer2 = _interopRequireDefault(_remuxMp4Remuxer); var DemuxerWorker = function DemuxerWorker(self) { // observer setup var observer = new _events4['default'](); observer.trigger = function trigger(event) { for (var _len = arguments.length, data = Array(_len > 1 ? _len - 1 : 0), _key = 1; _key < _len; _key++) { data[_key - 1] = arguments[_key]; } observer.emit.apply(observer, [event, event].concat(data)); }; observer.off = function off(event) { for (var _len2 = arguments.length, data = Array(_len2 > 1 ? _len2 - 1 : 0), _key2 = 1; _key2 < _len2; _key2++) { data[_key2 - 1] = arguments[_key2]; } observer.removeListener.apply(observer, [event].concat(data)); }; self.addEventListener('message', function (ev) { //console.log('demuxer cmd:' + ev.data.cmd); switch (ev.data.cmd) { case 'init': self.demuxer = new _demuxDemuxerInline2['default'](observer, _remuxMp4Remuxer2['default']); break; case 'demux': var data = ev.data; self.demuxer.push(new Uint8Array(data.data), data.audioCodec, data.videoCodec, data.timeOffset, data.cc, data.level, data.sn, data.duration); break; default: break; } }); // listen to events triggered by TS Demuxer observer.on(_events2['default'].FRAG_PARSING_INIT_SEGMENT, function (ev, data) { var objData = { event: ev }; var objTransferable = []; if (data.audioCodec) { objData.audioCodec = data.audioCodec; objData.audioMoov = data.audioMoov.buffer; objData.audioChannelCount = data.audioChannelCount; objTransferable.push(objData.audioMoov); } if (data.videoCodec) { objData.videoCodec = data.videoCodec; objData.videoMoov = data.videoMoov.buffer; objData.videoWidth = data.videoWidth; objData.videoHeight = data.videoHeight; objTransferable.push(objData.videoMoov); } // pass moov as transferable object (no copy) self.postMessage(objData, objTransferable); }); observer.on(_events2['default'].FRAG_PARSING_DATA, function (ev, data) { var objData = { event: ev, type: data.type, startPTS: data.startPTS, endPTS: data.endPTS, startDTS: data.startDTS, endDTS: data.endDTS, moof: data.moof.buffer, mdat: data.mdat.buffer, nb: data.nb }; // pass moof/mdat data as transferable object (no copy) self.postMessage(objData, [objData.moof, objData.mdat]); }); observer.on(_events2['default'].FRAG_PARSED, function (event) { self.postMessage({ event: event }); }); observer.on(_events2['default'].ERROR, function (event, data) { self.postMessage({ event: event, data: data }); }); observer.on(_events2['default'].FRAG_PARSING_METADATA, function (event, data) { var objData = { event: event, samples: data.samples }; self.postMessage(objData); }); }; exports['default'] = DemuxerWorker; module.exports = exports['default']; },{"../demux/demuxer-inline":6,"../events":12,"../remux/mp4-remuxer":19,"events":1}],8:[function(require,module,exports){ 'use strict'; Object.defineProperty(exports, '__esModule', { value: true }); var _createClass = (function () { function defineProperties(target, props) { for (var i = 0; i < props.length; i++) { var descriptor = props[i]; descriptor.enumerable = descriptor.enumerable || false; descriptor.configurable = true; if ('value' in descriptor) descriptor.writable = true; Object.defineProperty(target, descriptor.key, descriptor); } } return function (Constructor, protoProps, staticProps) { if (protoProps) defineProperties(Constructor.prototype, protoProps); if (staticProps) defineProperties(Constructor, staticProps); return Constructor; }; })(); function _interopRequireDefault(obj) { return obj && obj.__esModule ? obj : { 'default': obj }; } function _classCallCheck(instance, Constructor) { if (!(instance instanceof Constructor)) { throw new TypeError('Cannot call a class as a function'); } } var _events = require('../events'); var _events2 = _interopRequireDefault(_events); var _errors = require('../errors'); var _demuxDemuxerInline = require('../demux/demuxer-inline'); var _demuxDemuxerInline2 = _interopRequireDefault(_demuxDemuxerInline); var _demuxDemuxerWorker = require('../demux/demuxer-worker'); var _demuxDemuxerWorker2 = _interopRequireDefault(_demuxDemuxerWorker); var _utilsLogger = require('../utils/logger'); var _remuxMp4Remuxer = require('../remux/mp4-remuxer'); var _remuxMp4Remuxer2 = _interopRequireDefault(_remuxMp4Remuxer); var Demuxer = (function () { function Demuxer(hls) { _classCallCheck(this, Demuxer); this.hls = hls; if (hls.config.enableWorker && typeof Worker !== 'undefined') { _utilsLogger.logger.log('demuxing in webworker'); try { var work = require('webworkify'); this.w = work(_demuxDemuxerWorker2['default']); this.onwmsg = this.onWorkerMessage.bind(this); this.w.addEventListener('message', this.onwmsg); this.w.postMessage({ cmd: 'init' }); } catch (err) { _utilsLogger.logger.error('error while initializing DemuxerWorker, fallback on DemuxerInline'); this.demuxer = new _demuxDemuxerInline2['default'](hls, _remuxMp4Remuxer2['default']); } } else { this.demuxer = new _demuxDemuxerInline2['default'](hls, _remuxMp4Remuxer2['default']); } this.demuxInitialized = true; } _createClass(Demuxer, [{ key: 'destroy', value: function destroy() { if (this.w) { this.w.removeEventListener('message', this.onwmsg); this.w.terminate(); this.w = null; } else { this.demuxer.destroy(); } } }, { key: 'pushDecrypted', value: function pushDecrypted(data, audioCodec, videoCodec, timeOffset, cc, level, sn, duration) { if (this.w) { // post fragment payload as transferable objects (no copy) this.w.postMessage({ cmd: 'demux', data: data, audioCodec: audioCodec, videoCodec: videoCodec, timeOffset: timeOffset, cc: cc, level: level, sn: sn, duration: duration }, [data]); } else { this.demuxer.push(new Uint8Array(data), audioCodec, videoCodec, timeOffset, cc, level, sn, duration); } } }, { key: 'push', value: function push(data, audioCodec, videoCodec, timeOffset, cc, level, sn, duration, decryptdata) { if (data.byteLength > 0 && decryptdata != null && decryptdata.key != null && decryptdata.method === 'AES-128') { var localthis = this; window.crypto.subtle.importKey('raw', decryptdata.key, { name: 'AES-CBC', length: 128 }, false, ['decrypt']).then(function (importedKey) { window.crypto.subtle.decrypt({ name: 'AES-CBC', iv: decryptdata.iv.buffer }, importedKey, data).then(function (result) { localthis.pushDecrypted(result, audioCodec, videoCodec, timeOffset, cc, level, sn, duration); })['catch'](function (err) { _utilsLogger.logger.error('decrypting error : ' + err.message); localthis.hls.trigger(_events2['default'].ERROR, { type: _errors.ErrorTypes.MEDIA_ERROR, details: _errors.ErrorDetails.FRAG_DECRYPT_ERROR, fatal: true, reason: err.message }); return; }); })['catch'](function (err) { _utilsLogger.logger.error('decrypting error : ' + err.message); localthis.hls.trigger(_events2['default'].ERROR, { type: _errors.ErrorTypes.MEDIA_ERROR, details: _errors.ErrorDetails.FRAG_DECRYPT_ERROR, fatal: true, reason: err.message }); return; }); } else { this.pushDecrypted(data, audioCodec, videoCodec, timeOffset, cc, level, sn, duration); } } }, { key: 'onWorkerMessage', value: function onWorkerMessage(ev) { //console.log('onWorkerMessage:' + ev.data.event); switch (ev.data.event) { case _events2['default'].FRAG_PARSING_INIT_SEGMENT: var obj = {}; if (ev.data.audioMoov) { obj.audioMoov = new Uint8Array(ev.data.audioMoov); obj.audioCodec = ev.data.audioCodec; obj.audioChannelCount = ev.data.audioChannelCount; } if (ev.data.videoMoov) { obj.videoMoov = new Uint8Array(ev.data.videoMoov); obj.videoCodec = ev.data.videoCodec; obj.videoWidth = ev.data.videoWidth; obj.videoHeight = ev.data.videoHeight; } this.hls.trigger(_events2['default'].FRAG_PARSING_INIT_SEGMENT, obj); break; case _events2['default'].FRAG_PARSING_DATA: this.hls.trigger(_events2['default'].FRAG_PARSING_DATA, { moof: new Uint8Array(ev.data.moof), mdat: new Uint8Array(ev.data.mdat), startPTS: ev.data.startPTS, endPTS: ev.data.endPTS, startDTS: ev.data.startDTS, endDTS: ev.data.endDTS, type: ev.data.type, nb: ev.data.nb }); break; case _events2['default'].FRAG_PARSING_METADATA: this.hls.trigger(_events2['default'].FRAG_PARSING_METADATA, { samples: ev.data.samples }); break; default: this.hls.trigger(ev.data.event, ev.data.data); break; } } }]); return Demuxer; })(); exports['default'] = Demuxer; module.exports = exports['default']; },{"../demux/demuxer-inline":6,"../demux/demuxer-worker":7,"../errors":11,"../events":12,"../remux/mp4-remuxer":19,"../utils/logger":21,"webworkify":2}],9:[function(require,module,exports){ /** * Parser for exponential Golomb codes, a variable-bitwidth number encoding scheme used by h264. */ 'use strict'; Object.defineProperty(exports, '__esModule', { value: true }); var _createClass = (function () { function defineProperties(target, props) { for (var i = 0; i < props.length; i++) { var descriptor = props[i]; descriptor.enumerable = descriptor.enumerable || false; descriptor.configurable = true; if ('value' in descriptor) descriptor.writable = true; Object.defineProperty(target, descriptor.key, descriptor); } } return function (Constructor, protoProps, staticProps) { if (protoProps) defineProperties(Constructor.prototype, protoProps); if (staticProps) defineProperties(Constructor, staticProps); return Constructor; }; })(); function _classCallCheck(instance, Constructor) { if (!(instance instanceof Constructor)) { throw new TypeError('Cannot call a class as a function'); } } var _utilsLogger = require('../utils/logger'); var ExpGolomb = (function () { function ExpGolomb(data) { _classCallCheck(this, ExpGolomb); this.data = data; // the number of bytes left to examine in this.data this.bytesAvailable = this.data.byteLength; // the current word being examined this.word = 0; // :uint // the number of bits left to examine in the current word this.bitsAvailable = 0; // :uint } // ():void _createClass(ExpGolomb, [{ key: 'loadWord', value: function loadWord() { var position = this.data.byteLength - this.bytesAvailable, workingBytes = new Uint8Array(4), availableBytes = Math.min(4, this.bytesAvailable); if (availableBytes === 0) { throw new Error('no bytes available'); } workingBytes.set(this.data.subarray(position, position + availableBytes)); this.word = new DataView(workingBytes.buffer).getUint32(0); // track the amount of this.data that has been processed this.bitsAvailable = availableBytes * 8; this.bytesAvailable -= availableBytes; } // (count:int):void }, { key: 'skipBits', value: function skipBits(count) { var skipBytes; // :int if (this.bitsAvailable > count) { this.word <<= count; this.bitsAvailable -= count; } else { count -= this.bitsAvailable; skipBytes = count >> 3; count -= skipBytes >> 3; this.bytesAvailable -= skipBytes; this.loadWord(); this.word <<= count; this.bitsAvailable -= count; } } // (size:int):uint }, { key: 'readBits', value: function readBits(size) { var bits = Math.min(this.bitsAvailable, size), // :uint valu = this.word >>> 32 - bits; // :uint if (size > 32) { _utilsLogger.logger.error('Cannot read more than 32 bits at a time'); } this.bitsAvailable -= bits; if (this.bitsAvailable > 0) { this.word <<= bits; } else if (this.bytesAvailable > 0) { this.loadWord(); } bits = size - bits; if (bits > 0) { return valu << bits | this.readBits(bits); } else { return valu; } } // ():uint }, { key: 'skipLZ', value: function skipLZ() { var leadingZeroCount; // :uint for (leadingZeroCount = 0; leadingZeroCount < this.bitsAvailable; ++leadingZeroCount) { if (0 !== (this.word & 0x80000000 >>> leadingZeroCount)) { // the first bit of working word is 1 this.word <<= leadingZeroCount; this.bitsAvailable -= leadingZeroCount; return leadingZeroCount; } } // we exhausted word and still have not found a 1 this.loadWord(); return leadingZeroCount + this.skipLZ(); } // ():void }, { key: 'skipUEG', value: function skipUEG() { this.skipBits(1 + this.skipLZ()); } // ():void }, { key: 'skipEG', value: function skipEG() { this.skipBits(1 + this.skipLZ()); } // ():uint }, { key: 'readUEG', value: function readUEG() { var clz = this.skipLZ(); // :uint return this.readBits(clz + 1) - 1; } // ():int }, { key: 'readEG', value: function readEG() { var valu = this.readUEG(); // :int if (0x01 & valu) { // the number is odd if the low order bit is set return 1 + valu >>> 1; // add 1 to make it even, and divide by 2 } else { return -1 * (valu >>> 1); // divide by two then make it negative } } // Some convenience functions // :Boolean }, { key: 'readBoolean', value: function readBoolean() { return 1 === this.readBits(1); } // ():int }, { key: 'readUByte', value: function readUByte() { return this.readBits(8); } /** * Advance the ExpGolomb decoder past a scaling list. The scaling * list is optionally transmitted as part of a sequence parameter * set and is not relevant to transmuxing. * @param count {number} the number of entries in this scaling list * @see Recommendation ITU-T H.264, Section 7.3.2.1.1.1 */ }, { key: 'skipScalingList', value: function skipScalingList(count) { var lastScale = 8, nextScale = 8, j, deltaScale; for (j = 0; j < count; j++) { if (nextScale !== 0) { deltaScale = this.readEG(); nextScale = (lastScale + deltaScale + 256) % 256; } lastScale = nextScale === 0 ? lastScale : nextScale; } } /** * Read a sequence parameter set and return some interesting video * properties. A sequence parameter set is the H264 metadata that * describes the properties of upcoming video frames. * @param data {Uint8Array} the bytes of a sequence parameter set * @return {object} an object with configuration parsed from the * sequence parameter set, including the dimensions of the * associated video frames. */ }, { key: 'readSPS', value: function readSPS() { var frameCropLeftOffset = 0, frameCropRightOffset = 0, frameCropTopOffset = 0, frameCropBottomOffset = 0, profileIdc, profileCompat, levelIdc, numRefFramesInPicOrderCntCycle, picWidthInMbsMinus1, picHeightInMapUnitsMinus1, frameMbsOnlyFlag, scalingListCount, i; this.readUByte(); profileIdc = this.readUByte(); // profile_idc profileCompat = this.readBits(5); // constraint_set[0-4]_flag, u(5) this.skipBits(3); // reserved_zero_3bits u(3), levelIdc = this.readUByte(); //level_idc u(8) this.skipUEG(); // seq_parameter_set_id // some profiles have more optional data we don't need if (profileIdc === 100 || profileIdc === 110 || profileIdc === 122 || profileIdc === 144) { var chromaFormatIdc = this.readUEG(); if (chromaFormatIdc === 3) { this.skipBits(1); // separate_colour_plane_flag } this.skipUEG(); // bit_depth_luma_minus8 this.skipUEG(); // bit_depth_chroma_minus8 this.skipBits(1); // qpprime_y_zero_transform_bypass_flag if (this.readBoolean()) { // seq_scaling_matrix_present_flag scalingListCount = chromaFormatIdc !== 3 ? 8 : 12; for (i = 0; i < scalingListCount; i++) { if (this.readBoolean()) { // seq_scaling_list_present_flag[ i ] if (i < 6) { this.skipScalingList(16); } else { this.skipScalingList(64); } } } } } this.skipUEG(); // log2_max_frame_num_minus4 var picOrderCntType = this.readUEG(); if (picOrderCntType === 0) { this.readUEG(); //log2_max_pic_order_cnt_lsb_minus4 } else if (picOrderCntType === 1) { this.skipBits(1); // delta_pic_order_always_zero_flag this.skipEG(); // offset_for_non_ref_pic this.skipEG(); // offset_for_top_to_bottom_field numRefFramesInPicOrderCntCycle = this.readUEG(); for (i = 0; i < numRefFramesInPicOrderCntCycle; i++) { this.skipEG(); // offset_for_ref_frame[ i ] } } this.skipUEG(); // max_num_ref_frames this.skipBits(1); // gaps_in_frame_num_value_allowed_flag picWidthInMbsMinus1 = this.readUEG(); picHeightInMapUnitsMinus1 = this.readUEG(); frameMbsOnlyFlag = this.readBits(1); if (frameMbsOnlyFlag === 0) { this.skipBits(1); // mb_adaptive_frame_field_flag } this.skipBits(1); // direct_8x8_inference_flag if (this.readBoolean()) { // frame_cropping_flag frameCropLeftOffset = this.readUEG(); frameCropRightOffset = this.readUEG(); frameCropTopOffset = this.readUEG(); frameCropBottomOffset = this.readUEG(); } return { width: (picWidthInMbsMinus1 + 1) * 16 - frameCropLeftOffset * 2 - frameCropRightOffset * 2, height: (2 - frameMbsOnlyFlag) * (picHeightInMapUnitsMinus1 + 1) * 16 - frameCropTopOffset * 2 - frameCropBottomOffset * 2 }; } }, { key: 'readSliceType', value: function readSliceType() { // skip NALu type this.readUByte(); // discard first_mb_in_slice this.readUEG(); // return slice_type return this.readUEG(); } }]); return ExpGolomb; })(); exports['default'] = ExpGolomb; module.exports = exports['default']; },{"../utils/logger":21}],10:[function(require,module,exports){ /** * highly optimized TS demuxer: * parse PAT, PMT * extract PES packet from audio and video PIDs * extract AVC/H264 NAL units and AAC/ADTS samples from PES packet * trigger the remuxer upon parsing completion * it also tries to workaround as best as it can audio codec switch (HE-AAC to AAC and vice versa), without having to restart the MediaSource. * it also controls the remuxing process : * upon discontinuity or level switch detection, it will also notifies the remuxer so that it can reset its state. */ 'use strict'; Object.defineProperty(exports, '__esModule', { value: true }); var _createClass = (function () { function defineProperties(target, props) { for (var i = 0; i < props.length; i++) { var descriptor = props[i]; descriptor.enumerable = descriptor.enumerable || false; descriptor.configurable = true; if ('value' in descriptor) descriptor.writable = true; Object.defineProperty(target, descriptor.key, descriptor); } } return function (Constructor, protoProps, staticProps) { if (protoProps) defineProperties(Constructor.prototype, protoProps); if (staticProps) defineProperties(Constructor, staticProps); return Constructor; }; })(); function _interopRequireDefault(obj) { return obj && obj.__esModule ? obj : { 'default': obj }; } function _classCallCheck(instance, Constructor) { if (!(instance instanceof Constructor)) { throw new TypeError('Cannot call a class as a function'); } } var _events = require('../events'); var _events2 = _interopRequireDefault(_events); var _expGolomb = require('./exp-golomb'); var _expGolomb2 = _interopRequireDefault(_expGolomb); // import Hex from '../utils/hex'; var _utilsLogger = require('../utils/logger'); var _errors = require('../errors'); var TSDemuxer = (function () { function TSDemuxer(observer, remuxerClass) { _classCallCheck(this, TSDemuxer); this.observer = observer; this.remuxerClass = remuxerClass; this.lastCC = 0; this.PES_TIMESCALE = 90000; this.remuxer = new this.remuxerClass(observer); } _createClass(TSDemuxer, [{ key: 'switchLevel', value: function switchLevel() { this.pmtParsed = false; this._pmtId = -1; this._avcTrack = { type: 'video', id: -1, sequenceNumber: 0, samples: [], len: 0, nbNalu: 0 }; this._aacTrack = { type: 'audio', id: -1, sequenceNumber: 0, samples: [], len: 0 }; this._id3Track = { type: 'id3', id: -1, sequenceNumber: 0, samples: [], len: 0 }; this.remuxer.switchLevel(); } }, { key: 'insertDiscontinuity', value: function insertDiscontinuity() { this.switchLevel(); this.remuxer.insertDiscontinuity(); } // feed incoming data to the front of the parsing pipeline }, { key: 'push', value: function push(data, audioCodec, videoCodec, timeOffset, cc, level, sn, duration) { var avcData, aacData, id3Data, start, len = data.length, stt, pid, atf, offset; this.audioCodec = audioCodec; this.videoCodec = videoCodec; this.timeOffset = timeOffset; this._duration = duration; this.contiguous = false; if (cc !== this.lastCC) { _utilsLogger.logger.log('discontinuity detected'); this.insertDiscontinuity(); this.lastCC = cc; } else if (level !== this.lastLevel) { _utilsLogger.logger.log('level switch detected'); this.switchLevel(); this.lastLevel = level; } else if (sn === this.lastSN + 1) { this.contiguous = true; } this.lastSN = sn; if (!this.contiguous) { // flush any partial content this.aacOverFlow = null; } var pmtParsed = this.pmtParsed, avcId = this._avcTrack.id, aacId = this._aacTrack.id, id3Id = this._id3Track.id; // loop through TS packets for (start = 0; start < len; start += 188) { if (data[start] === 0x47) { stt = !!(data[start + 1] & 0x40); // pid is a 13-bit field starting at the last bit of TS[1] pid = ((data[start + 1] & 0x1f) << 8) + data[start + 2]; atf = (data[start + 3] & 0x30) >> 4; // if an adaption field is present, its length is specified by the fifth byte of the TS packet header. if (atf > 1) { offset = start + 5 + data[start + 4]; // continue if there is only adaptation field if (offset === start + 188) { continue; } } else { offset = start + 4; } if (pmtParsed) { if (pid === avcId) { if (stt) { if (avcData) { this._parseAVCPES(this._parsePES(avcData)); } avcData = { data: [], size: 0 }; } if (avcData) { avcData.data.push(data.subarray(offset, start + 188)); avcData.size += start + 188 - offset; } } else if (pid === aacId) { if (stt) { if (aacData) { this._parseAACPES(this._parsePES(aacData)); } aacData = { data: [], size: 0 }; } if (aacData) { aacData.data.push(data.subarray(offset, start + 188)); aacData.size += start + 188 - offset; } } else if (pid === id3Id) { if (stt) { if (id3Data) { this._parseID3PES(this._parsePES(id3Data)); } id3Data = { data: [], size: 0 }; } if (id3Data) { id3Data.data.push(data.subarray(offset, start + 188)); id3Data.size += start + 188 - offset; } } } else { if (stt) { offset += data[offset] + 1; } if (pid === 0) { this._parsePAT(data, offset); } else if (pid === this._pmtId) { this._parsePMT(data, offset); pmtParsed = this.pmtParsed = true; avcId = this._avcTrack.id; aacId = this._aacTrack.id; id3Id = this._id3Track.id; } } } else { this.observer.trigger(_events2['default'].ERROR, { type: _errors.ErrorTypes.MEDIA_ERROR, details: _errors.ErrorDetails.FRAG_PARSING_ERROR, fatal: false, reason: 'TS packet did not start with 0x47' }); } } // parse last PES packet if (avcData) { this._parseAVCPES(this._parsePES(avcData)); } if (aacData) { this._parseAACPES(this._parsePES(aacData)); } if (id3Data) { this._parseID3PES(this._parsePES(id3Data)); } this.remux(); } }, { key: 'remux', value: function remux() { this.remuxer.remux(this._aacTrack, this._avcTrack, this._id3Track, this.timeOffset, this.contiguous); } }, { key: 'destroy', value: function destroy() { this.switchLevel(); this._initPTS = this._initDTS = undefined; this._duration = 0; } }, { key: '_parsePAT', value: function _parsePAT(data, offset) { // skip the PSI header and parse the first PMT entry this._pmtId = (data[offset + 10] & 0x1F) << 8 | data[offset + 11]; //logger.log('PMT PID:' + this._pmtId); } }, { key: '_parsePMT', value: function _parsePMT(data, offset) { var sectionLength, tableEnd, programInfoLength, pid; sectionLength = (data[offset + 1] & 0x0f) << 8 | data[offset + 2]; tableEnd = offset + 3 + sectionLength - 4; // to determine where the table is, we have to figure out how // long the program info descriptors are programInfoLength = (data[offset + 10] & 0x0f) << 8 | data[offset + 11]; // advance the offset to the first entry in the mapping table offset += 12 + programInfoLength; while (offset < tableEnd) { pid = (data[offset + 1] & 0x1F) << 8 | data[offset + 2]; switch (data[offset]) { // ISO/IEC 13818-7 ADTS AAC (MPEG-2 lower bit-rate audio) case 0x0f: //logger.log('AAC PID:' + pid); this._aacTrack.id = pid; break; // Packetized metadata (ID3) case 0x15: //logger.log('ID3 PID:' + pid); this._id3Track.id = pid; break; // ITU-T Rec. H.264 and ISO/IEC 14496-10 (lower bit-rate video) case 0x1b: //logger.log('AVC PID:' + pid); this._avcTrack.id = pid; break; default: _utilsLogger.logger.log('unkown stream type:' + data[offset]); break; } // move to the next table entry // skip past the elementary stream descriptors, if present offset += ((data[offset + 3] & 0x0F) << 8 | data[offset + 4]) + 5; } } }, { key: '_parsePES', value: function _parsePES(stream) { var i = 0, frag, pesFlags, pesPrefix, pesLen, pesHdrLen, pesData, pesPts, pesDts, payloadStartOffset; //retrieve PTS/DTS from first fragment frag = stream.data[0]; pesPrefix = (frag[0] << 16) + (frag[1] << 8) + frag[2]; if (pesPrefix === 1) { pesLen = (frag[4] << 8) + frag[5]; pesFlags = frag[7]; if (pesFlags & 0xC0) { /* PES header described here : http://dvd.sourceforge.net/dvdinfo/pes-hdr.html as PTS / DTS is 33 bit we cannot use bitwise operator in JS, as Bitwise operators treat their operands as a sequence of 32 bits */ pesPts = (frag[9] & 0x0E) * 536870912 + // 1 << 29 (frag[10] & 0xFF) * 4194304 + // 1 << 22 (frag[11] & 0xFE) * 16384 + // 1 << 14 (frag[12] & 0xFF) * 128 + // 1 << 7 (frag[13] & 0xFE) / 2; // check if greater than 2^32 -1 if (pesPts > 4294967295) { // decrement 2^33 pesPts -= 8589934592; } if (pesFlags & 0x40) { pesDts = (frag[14] & 0x0E) * 536870912 + // 1 << 29 (frag[15] & 0xFF) * 4194304 + // 1 << 22 (frag[16] & 0xFE) * 16384 + // 1 << 14 (frag[17] & 0xFF) * 128 + // 1 << 7 (frag[18] & 0xFE) / 2; // check if greater than 2^32 -1 if (pesDts > 4294967295) { // decrement 2^33 pesDts -= 8589934592; } } else { pesDts = pesPts; } } pesHdrLen = frag[8]; payloadStartOffset = pesHdrLen + 9; // trim PES header stream.data[0] = stream.data[0].subarray(payloadStartOffset); stream.size -= payloadStartOffset; //reassemble PES packet pesData = new Uint8Array(stream.size); // reassemble the packet while (stream.data.length) { frag = stream.data.shift(); pesData.set(frag, i); i += frag.byteLength; } return { data: pesData, pts: pesPts, dts: pesDts, len: pesLen }; } else { return null; } } }, { key: '_parseAVCPES', value: function _parseAVCPES(pes) { var _this = this; var track = this._avcTrack, samples = track.samples, units = this._parseAVCNALu(pes.data), units2 = [], debug = false, key = false, length = 0, avcSample, push; // no NALu found if (units.length === 0 && samples.length > 0) { // append pes.data to previous NAL unit var lastavcSample = samples[samples.length - 1]; var lastUnit = lastavcSample.units.units[lastavcSample.units.units.length - 1]; var tmp = new Uint8Array(lastUnit.data.byteLength + pes.data.byteLength); tmp.set(lastUnit.data, 0); tmp.set(pes.data, lastUnit.data.byteLength); lastUnit.data = tmp; lastavcSample.units.length += pes.data.byteLength; track.len += pes.data.byteLength; } //free pes.data to save up some memory pes.data = null; var debugString = ''; units.forEach(function (unit) { switch (unit.type) { //NDR case 1: push = true; if (debug) { debugString += 'NDR '; } break; //IDR case 5: push = true; if (debug) { debugString += 'IDR '; } key = true; break; case 6: push = true; if (debug) { debugString += 'SEI '; } break; //SPS case 7: push = true; if (debug) { debugString += 'SPS '; } if (!track.sps) { var expGolombDecoder = new _expGolomb2['default'](unit.data); var config = expGolombDecoder.readSPS(); track.width = config.width; track.height = config.height; track.sps = [unit.data]; track.timescale = _this.remuxer.timescale; track.duration = _this.remuxer.timescale * _this._duration; var codecarray = unit.data.subarray(1, 4); var codecstring = 'avc1.'; for (var i = 0; i < 3; i++) { var h = codecarray[i].toString(16); if (h.length < 2) { h = '0' + h; } codecstring += h; } track.codec = codecstring; } break; //PPS case 8: push = true; if (debug) { debugString += 'PPS '; } if (!track.pps) { track.pps = [unit.data]; } break; case 9: push = true; if (debug) { debugString += 'AUD '; } break; default: push = false; debugString += 'unknown NAL ' + unit.type + ' '; break; } if (push) { units2.push(unit); length += unit.data.byteLength; } }); if (debug || debugString.length) { _utilsLogger.logger.log(debugString); } //build sample from PES // Annex B to MP4 conversion to be done if (units2.length) { // only push AVC sample if keyframe already found. browsers expect a keyframe at first to start decoding if (key === true || track.sps) { avcSample = { units: { units: units2, length: length }, pts: pes.pts, dts: pes.dts, key: key }; samples.push(avcSample); track.len += length; track.nbNalu += units2.length; } } } }, { key: '_parseAVCNALu', value: function _parseAVCNALu(array) { var i = 0, len = array.byteLength, value, overflow, state = 0; var units = [], unit, unitType, lastUnitStart, lastUnitType; //logger.log('PES:' + Hex.hexDump(array)); while (i < len) { value = array[i++]; // finding 3 or 4-byte start codes (00 00 01 OR 00 00 00 01) switch (state) { case 0: if (value === 0) { state = 1; } break; case 1: if (value === 0) { state = 2; } else { state = 0; } break; case 2: case 3: if (value === 0) { state = 3; } else if (value === 1) { unitType = array[i] & 0x1f; //logger.log('find NALU @ offset:' + i + ',type:' + unitType); if (lastUnitStart) { unit = { data: array.subarray(lastUnitStart, i - state - 1), type: lastUnitType }; //logger.log('pushing NALU, type/size:' + unit.type + '/' + unit.data.byteLength); units.push(unit); } else { // If NAL units are not starting right at the beginning of the PES packet, push preceding data into previous NAL unit. overflow = i - state - 1; if (overflow) { //logger.log('first NALU found with overflow:' + overflow); if (this._avcTrack.samples.length) { var lastavcSample = this._avcTrack.samples[this._avcTrack.samples.length - 1]; var lastUnit = lastavcSample.units.units[lastavcSample.units.units.length - 1]; var tmp = new Uint8Array(lastUnit.data.byteLength + overflow); tmp.set(lastUnit.data, 0); tmp.set(array.subarray(0, overflow), lastUnit.data.byteLength); lastUnit.data = tmp; lastavcSample.units.length += overflow; this._avcTrack.len += overflow; } } } lastUnitStart = i; lastUnitType = unitType; if (unitType === 1 || unitType === 5) { // OPTI !!! if IDR/NDR unit, consider it is last NALu i = len; } state = 0; } else { state = 0; } break; default: break; } } if (lastUnitStart) { unit = { data: array.subarray(lastUnitStart, len), type: lastUnitType }; units.push(unit); //logger.log('pushing NALU, type/size:' + unit.type + '/' + unit.data.byteLength); } return units; } }, { key: '_parseAACPES', value: function _parseAACPES(pes) { var track = this._aacTrack, aacSample, data = pes.data, config, adtsFrameSize, adtsStartOffset, adtsHeaderLen, stamp, nbSamples, len; if (this.aacOverFlow) { var tmp = new Uint8Array(this.aacOverFlow.byteLength + data.byteLength); tmp.set(this.aacOverFlow, 0); tmp.set(data, this.aacOverFlow.byteLength); data = tmp; } // look for ADTS header (0xFFFx) for (adtsStartOffset = 0, len = data.length; adtsStartOffset < len - 1; adtsStartOffset++) { if (data[adtsStartOffset] === 0xff && (data[adtsStartOffset + 1] & 0xf0) === 0xf0) { break; } } // if ADTS header does not start straight from the beginning of the PES payload, raise an error if (adtsStartOffset) { var reason, fatal; if (adtsStartOffset < len - 1) { reason = 'AAC PES did not start with ADTS header,offset:' + adtsStartOffset; fatal = false; } else { reason = 'no ADTS header found in AAC PES'; fatal = true; } this.observer.trigger(_events2['default'].ERROR, { type: _errors.ErrorTypes.MEDIA_ERROR, details: _errors.ErrorDetails.FRAG_PARSING_ERROR, fatal: fatal, reason: reason }); if (fatal) { return; } } if (!track.audiosamplerate) { config = this._ADTStoAudioConfig(data, adtsStartOffset, this.audioCodec); track.config = config.config; track.audiosamplerate = config.samplerate; track.channelCount = config.channelCount; track.codec = config.codec; track.timescale = this.remuxer.timescale; track.duration = this.remuxer.timescale * this._duration; _utilsLogger.logger.log('parsed codec:' + track.codec + ',rate:' + config.samplerate + ',nb channel:' + config.channelCount); } nbSamples = 0; while (adtsStartOffset + 5 < len) { // retrieve frame size adtsFrameSize = (data[adtsStartOffset + 3] & 0x03) << 11; // byte 4 adtsFrameSize |= data[adtsStartOffset + 4] << 3; // byte 5 adtsFrameSize |= (data[adtsStartOffset + 5] & 0xE0) >>> 5; adtsHeaderLen = !!(data[adtsStartOffset + 1] & 0x01) ? 7 : 9; adtsFrameSize -= adtsHeaderLen; stamp = Math.round(pes.pts + nbSamples * 1024 * this.PES_TIMESCALE / track.audiosamplerate); //stamp = pes.pts; //console.log('AAC frame, offset/length/pts:' + (adtsStartOffset+7) + '/' + adtsFrameSize + '/' + stamp.toFixed(0)); if (adtsFrameSize > 0 && adtsStartOffset + adtsHeaderLen + adtsFrameSize <= len) { aacSample = { unit: data.subarray(adtsStartOffset + adtsHeaderLen, adtsStartOffset + adtsHeaderLen + adtsFrameSize), pts: stamp, dts: stamp }; this._aacTrack.samples.push(aacSample); this._aacTrack.len += adtsFrameSize; adtsStartOffset += adtsFrameSize + adtsHeaderLen; nbSamples++; // look for ADTS header (0xFFFx) for (; adtsStartOffset < len - 1; adtsStartOffset++) { if (data[adtsStartOffset] === 0xff && (data[adtsStartOffset + 1] & 0xf0) === 0xf0) { break; } } } else { break; } } if (adtsStartOffset < len) { this.aacOverFlow = data.subarray(adtsStartOffset, len); } else { this.aacOverFlow = null; } } }, { key: '_ADTStoAudioConfig', value: function _ADTStoAudioConfig(data, offset, audioCodec) { var adtsObjectType, // :int adtsSampleingIndex, // :int adtsExtensionSampleingIndex, // :int adtsChanelConfig, // :int config, userAgent = navigator.userAgent.toLowerCase(), adtsSampleingRates = [96000, 88200, 64000, 48000, 44100, 32000, 24000, 22050, 16000, 12000, 11025, 8000, 7350]; // byte 2 adtsObjectType = ((data[offset + 2] & 0xC0) >>> 6) + 1; adtsSampleingIndex = (data[offset + 2] & 0x3C) >>> 2; if (adtsSampleingIndex > adtsSampleingRates.length - 1) { this.observer.trigger(_events2['default'].ERROR, { type: _errors.ErrorTypes.MEDIA_ERROR, details: _errors.ErrorDetails.FRAG_PARSING_ERROR, fatal: true, reason: 'invalid ADTS sampling index:' + adtsSampleingIndex }); return; } adtsChanelConfig = (data[offset + 2] & 0x01) << 2; // byte 3 adtsChanelConfig |= (data[offset + 3] & 0xC0) >>> 6; _utilsLogger.logger.log('manifest codec:' + audioCodec + ',ADTS data:type:' + adtsObjectType + ',sampleingIndex:' + adtsSampleingIndex + '[' + adtsSampleingRates[adtsSampleingIndex] + 'kHz],channelConfig:' + adtsChanelConfig); // firefox: freq less than 24kHz = AAC SBR (HE-AAC) if (userAgent.indexOf('firefox') !== -1) { if (adtsSampleingIndex >= 6) { adtsObjectType = 5; config = new Array(4); // HE-AAC uses SBR (Spectral Band Replication) , high frequencies are constructed from low frequencies // there is a factor 2 between frame sample rate and output sample rate // multiply frequency by 2 (see table below, equivalent to substract 3) adtsExtensionSampleingIndex = adtsSampleingIndex - 3; } else { adtsObjectType = 2; config = new Array(2); adtsExtensionSampleingIndex = adtsSampleingIndex; } // Android : always use AAC } else if (userAgent.indexOf('android') !== -1) { adtsObjectType = 2; config = new Array(2); adtsExtensionSampleingIndex = adtsSampleingIndex; } else { /* for other browsers (chrome ...) always force audio type to be HE-AAC SBR, as some browsers do not support audio codec switch properly (like Chrome ...) */ adtsObjectType = 5; config = new Array(4); // if (manifest codec is HE-AAC) OR (manifest codec not specified AND frequency less than 24kHz) if (audioCodec && audioCodec.indexOf('mp4a.40.5') !== -1 || !audioCodec && adtsSampleingIndex >= 6) { // HE-AAC uses SBR (Spectral Band Replication) , high frequencies are constructed from low frequencies // there is a factor 2 between frame sample rate and output sample rate // multiply frequency by 2 (see table below, equivalent to substract 3) adtsExtensionSampleingIndex = adtsSampleingIndex - 3; } else { // if (manifest codec is AAC) AND (frequency less than 24kHz OR nb channel is 1) if (audioCodec && audioCodec.indexOf('mp4a.40.2') !== -1 && (adtsSampleingIndex >= 6 || adtsChanelConfig === 1)) { adtsObjectType = 2; config = new Array(2); } adtsExtensionSampleingIndex = adtsSampleingIndex; } } /* refer to http://wiki.multimedia.cx/index.php?title=MPEG-4_Audio#Audio_Specific_Config ISO 14496-3 (AAC).pdf - Table 1.13 — Syntax of AudioSpecificConfig() Audio Profile / Audio Object Type 0: Null 1: AAC Main 2: AAC LC (Low Complexity) 3: AAC SSR (Scalable Sample Rate) 4: AAC LTP (Long Term Prediction) 5: SBR (Spectral Band Replication) 6: AAC Scalable sampling freq 0: 96000 Hz 1: 88200 Hz 2: 64000 Hz 3: 48000 Hz 4: 44100 Hz 5: 32000 Hz 6: 24000 Hz 7: 22050 Hz 8: 16000 Hz 9: 12000 Hz 10: 11025 Hz 11: 8000 Hz 12: 7350 Hz 13: Reserved 14: Reserved 15: frequency is written explictly Channel Configurations These are the channel configurations: 0: Defined in AOT Specifc Config 1: 1 channel: front-center 2: 2 channels: front-left, front-right */ // audioObjectType = profile => profile, the MPEG-4 Audio Object Type minus 1 config[0] = adtsObjectType << 3; // samplingFrequencyIndex config[0] |= (adtsSampleingIndex & 0x0E) >> 1; config[1] |= (adtsSampleingIndex & 0x01) << 7; // channelConfiguration config[1] |= adtsChanelConfig << 3; if (adtsObjectType === 5) { // adtsExtensionSampleingIndex config[1] |= (adtsExtensionSampleingIndex & 0x0E) >> 1; config[2] = (adtsExtensionSampleingIndex & 0x01) << 7; // adtsObjectType (force to 2, chrome is checking that object type is less than 5 ??? // https://chromium.googlesource.com/chromium/src.git/+/master/media/formats/mp4/aac.cc config[2] |= 2 << 2; config[3] = 0; } return { config: config, samplerate: adtsSampleingRates[adtsSampleingIndex], channelCount: adtsChanelConfig, codec: 'mp4a.40.' + adtsObjectType }; } }, { key: '_parseID3PES', value: function _parseID3PES(pes) { this._id3Track.samples.push(pes); } }], [{ key: 'probe', value: function probe(data) { // a TS fragment should contain at least 3 TS packets, a PAT, a PMT, and one PID, each starting with 0x47 if (data.length >= 3 * 188 && data[0] === 0x47 && data[188] === 0x47 && data[2 * 188] === 0x47) { return true; } else { return false; } } }]); return TSDemuxer; })(); exports['default'] = TSDemuxer; module.exports = exports['default']; },{"../errors":11,"../events":12,"../utils/logger":21,"./exp-golomb":9}],11:[function(require,module,exports){ 'use strict'; Object.defineProperty(exports, '__esModule', { value: true }); var ErrorTypes = { // Identifier for a network error (loading error / timeout ...) NETWORK_ERROR: 'hlsNetworkError', // Identifier for a media Error (video/parsing/mediasource error) MEDIA_ERROR: 'hlsMediaError', // Identifier for all other errors OTHER_ERROR: 'hlsOtherError' }; exports.ErrorTypes = ErrorTypes; var ErrorDetails = { // Identifier for a manifest load error - data: { url : faulty URL, response : XHR response} MANIFEST_LOAD_ERROR: 'manifestLoadError', // Identifier for a manifest load timeout - data: { url : faulty URL, response : XHR response} MANIFEST_LOAD_TIMEOUT: 'manifestLoadTimeOut', // Identifier for a manifest parsing error - data: { url : faulty URL, reason : error reason} MANIFEST_PARSING_ERROR: 'manifestParsingError', // Identifier for playlist load error - data: { url : faulty URL, response : XHR response} LEVEL_LOAD_ERROR: 'levelLoadError', // Identifier for playlist load timeout - data: { url : faulty URL, response : XHR response} LEVEL_LOAD_TIMEOUT: 'levelLoadTimeOut', // Identifier for a level switch error - data: { level : faulty level Id, event : error description} LEVEL_SWITCH_ERROR: 'levelSwitchError', // Identifier for fragment load error - data: { frag : fragment object, response : XHR response} FRAG_LOAD_ERROR: 'fragLoadError', // Identifier for fragment loop loading error - data: { frag : fragment object} FRAG_LOOP_LOADING_ERROR: 'fragLoopLoadingError', // Identifier for fragment load timeout error - data: { frag : fragment object} FRAG_LOAD_TIMEOUT: 'fragLoadTimeOut', // Identifier for a fragment decryption error event - data: parsing error description FRAG_DECRYPT_ERROR: 'fragDecryptError', // Identifier for a fragment parsing error event - data: parsing error description FRAG_PARSING_ERROR: 'fragParsingError', // Identifier for decrypt key load error - data: { frag : fragment object, response : XHR response} KEY_LOAD_ERROR: 'keyLoadError', // Identifier for decrypt key load timeout error - data: { frag : fragment object} KEY_LOAD_TIMEOUT: 'keyLoadTimeOut', // Identifier for a buffer append error - data: append error description BUFFER_APPEND_ERROR: 'bufferAppendError', // Identifier for a buffer appending error event - data: appending error description BUFFER_APPENDING_ERROR: 'bufferAppendingError' }; exports.ErrorDetails = ErrorDetails; },{}],12:[function(require,module,exports){ 'use strict'; Object.defineProperty(exports, '__esModule', { value: true }); exports['default'] = { // fired before MediaSource is attaching to media element - data: { media } MEDIA_ATTACHING: 'hlsMediaAttaching', // fired when MediaSource has been succesfully attached to media element - data: { } MEDIA_ATTACHED: 'hlsMediaAttached', // fired before detaching MediaSource from media element - data: { } MEDIA_DETACHING: 'hlsMediaDetaching', // fired when MediaSource has been detached from media element - data: { } MEDIA_DETACHED: 'hlsMediaDetached', // fired to signal that a manifest loading starts - data: { url : manifestURL} MANIFEST_LOADING: 'hlsManifestLoading', // fired after manifest has been loaded - data: { levels : [available quality levels] , url : manifestURL, stats : { trequest, tfirst, tload, mtime}} MANIFEST_LOADED: 'hlsManifestLoaded', // fired after manifest has been parsed - data: { levels : [available quality levels] , firstLevel : index of first quality level appearing in Manifest} MANIFEST_PARSED: 'hlsManifestParsed', // fired when a level playlist loading starts - data: { url : level URL level : id of level being loaded} LEVEL_LOADING: 'hlsLevelLoading', // fired when a level playlist loading finishes - data: { details : levelDetails object, level : id of loaded level, stats : { trequest, tfirst, tload, mtime} } LEVEL_LOADED: 'hlsLevelLoaded', // fired when a level's details have been updated based on previous details, after it has been loaded. - data: { details : levelDetails object, level : id of updated level } LEVEL_UPDATED: 'hlsLevelUpdated', // fired when a level's PTS information has been updated after parsing a fragment - data: { details : levelDetails object, level : id of updated level, drift: PTS drift observed when parsing last fragment } LEVEL_PTS_UPDATED: 'hlsPTSUpdated', // fired when a level switch is requested - data: { level : id of new level } LEVEL_SWITCH: 'hlsLevelSwitch', // fired when a fragment loading starts - data: { frag : fragment object} FRAG_LOADING: 'hlsFragLoading', // fired when a fragment loading is progressing - data: { frag : fragment object, { trequest, tfirst, loaded}} FRAG_LOAD_PROGRESS: 'hlsFragLoadProgress', // Identifier for fragment load aborting for emergency switch down - data: {frag : fragment object} FRAG_LOAD_EMERGENCY_ABORTED: 'hlsFragLoadEmergencyAborted', // fired when a fragment loading is completed - data: { frag : fragment object, payload : fragment payload, stats : { trequest, tfirst, tload, length}} FRAG_LOADED: 'hlsFragLoaded', // fired when Init Segment has been extracted from fragment - data: { moov : moov MP4 box, codecs : codecs found while parsing fragment} FRAG_PARSING_INIT_SEGMENT: 'hlsFragParsingInitSegment', // fired when parsing id3 is completed - data: { samples : [ id3 samples pes ] } FRAG_PARSING_METADATA: 'hlsFraParsingMetadata', // fired when moof/mdat have been extracted from fragment - data: { moof : moof MP4 box, mdat : mdat MP4 box} FRAG_PARSING_DATA: 'hlsFragParsingData', // fired when fragment parsing is completed - data: undefined FRAG_PARSED: 'hlsFragParsed', // fired when fragment remuxed MP4 boxes have all been appended into SourceBuffer - data: { frag : fragment object, stats : { trequest, tfirst, tload, tparsed, tbuffered, length} } FRAG_BUFFERED: 'hlsFragBuffered', // fired when fragment matching with current media position is changing - data : { frag : fragment object } FRAG_CHANGED: 'hlsFragChanged', // Identifier for a FPS drop event - data: {curentDropped, currentDecoded, totalDroppedFrames} FPS_DROP: 'hlsFPSDrop', // Identifier for an error event - data: { type : error type, details : error details, fatal : if true, hls.js cannot/will not try to recover, if false, hls.js will try to recover,other error specific data} ERROR: 'hlsError', // fired when hls.js instance starts destroying. Different from MSE_DETACHED as one could want to detach and reattach a media to the instance of hls.js to handle mid-rolls for example DESTROYING: 'hlsDestroying', // fired when a decrypt key loading starts - data: { frag : fragment object} KEY_LOADING: 'hlsKeyLoading', // fired when a decrypt key loading is completed - data: { frag : fragment object, payload : key payload, stats : { trequest, tfirst, tload, length}} KEY_LOADED: 'hlsKeyLoaded' }; module.exports = exports['default']; },{}],13:[function(require,module,exports){ /** * Level Helper class, providing methods dealing with playlist sliding and drift */ 'use strict'; Object.defineProperty(exports, '__esModule', { value: true }); var _createClass = (function () { function defineProperties(target, props) { for (var i = 0; i < props.length; i++) { var descriptor = props[i]; descriptor.enumerable = descriptor.enumerable || false; descriptor.configurable = true; if ('value' in descriptor) descriptor.writable = true; Object.defineProperty(target, descriptor.key, descriptor); } } return function (Constructor, protoProps, staticProps) { if (protoProps) defineProperties(Constructor.prototype, protoProps); if (staticProps) defineProperties(Constructor, staticProps); return Constructor; }; })(); function _classCallCheck(instance, Constructor) { if (!(instance instanceof Constructor)) { throw new TypeError('Cannot call a class as a function'); } } var _utilsLogger = require('../utils/logger'); var LevelHelper = (function () { function LevelHelper() { _classCallCheck(this, LevelHelper); } _createClass(LevelHelper, null, [{ key: 'mergeDetails', value: function mergeDetails(oldDetails, newDetails) { var start = Math.max(oldDetails.startSN, newDetails.startSN) - newDetails.startSN, end = Math.min(oldDetails.endSN, newDetails.endSN) - newDetails.startSN, delta = newDetails.startSN - oldDetails.startSN, oldfragments = oldDetails.fragments, newfragments = newDetails.fragments, ccOffset = 0, PTSFrag; // check if old/new playlists have fragments in common if (end < start) { newDetails.PTSKnown = false; return; } // loop through overlapping SN and update startPTS , cc, and duration if any found for (var i = start; i <= end; i++) { var oldFrag = oldfragments[delta + i], newFrag = newfragments[i]; ccOffset = oldFrag.cc - newFrag.cc; if (!isNaN(oldFrag.startPTS)) { newFrag.start = newFrag.startPTS = oldFrag.startPTS; newFrag.endPTS = oldFrag.endPTS; newFrag.duration = oldFrag.duration; PTSFrag = newFrag; } } if (ccOffset) { _utilsLogger.logger.log('discontinuity sliding from playlist, take drift into account'); for (i = 0; i < newfragments.length; i++) { newfragments[i].cc += ccOffset; } } // if at least one fragment contains PTS info, recompute PTS information for all fragments if (PTSFrag) { LevelHelper.updateFragPTS(newDetails, PTSFrag.sn, PTSFrag.startPTS, PTSFrag.endPTS); } else { // adjust start by sliding offset var sliding = oldfragments[delta].start; for (i = 0; i < newfragments.length; i++) { newfragments[i].start += sliding; } } // if we are here, it means we have fragments overlapping between // old and new level. reliable PTS info is thus relying on old level newDetails.PTSKnown = oldDetails.PTSKnown; return; } }, { key: 'updateFragPTS', value: function updateFragPTS(details, sn, startPTS, endPTS) { var fragIdx, fragments, frag, i; // exit if sn out of range if (sn < details.startSN || sn > details.endSN) { return 0; } fragIdx = sn - details.startSN; fragments = details.fragments; frag = fragments[fragIdx]; if (!isNaN(frag.startPTS)) { startPTS = Math.max(startPTS, frag.startPTS); endPTS = Math.min(endPTS, frag.endPTS); } var drift = startPTS - frag.start; frag.start = frag.startPTS = startPTS; frag.endPTS = endPTS; frag.duration = endPTS - startPTS; // adjust fragment PTS/duration from seqnum-1 to frag 0 for (i = fragIdx; i > 0; i--) { LevelHelper.updatePTS(fragments, i, i - 1); } // adjust fragment PTS/duration from seqnum to last frag for (i = fragIdx; i < fragments.length - 1; i++) { LevelHelper.updatePTS(fragments, i, i + 1); } details.PTSKnown = true; //logger.log(` frag start/end:${startPTS.toFixed(3)}/${endPTS.toFixed(3)}`); return drift; } }, { key: 'updatePTS', value: function updatePTS(fragments, fromIdx, toIdx) { var fragFrom = fragments[fromIdx], fragTo = fragments[toIdx], fragToPTS = fragTo.startPTS; // if we know startPTS[toIdx] if (!isNaN(fragToPTS)) { // update fragment duration. // it helps to fix drifts between playlist reported duration and fragment real duration if (toIdx > fromIdx) { fragFrom.duration = fragToPTS - fragFrom.start; if (fragFrom.duration < 0) { _utilsLogger.logger.error('negative duration computed for ' + fragFrom + ', there should be some duration drift between playlist and fragment!'); } } else { fragTo.duration = fragFrom.start - fragToPTS; if (fragTo.duration < 0) { _utilsLogger.logger.error('negative duration computed for ' + fragTo + ', there should be some duration drift between playlist and fragment!'); } } } else { // we dont know startPTS[toIdx] if (toIdx > fromIdx) { fragTo.start = fragFrom.start + fragFrom.duration; } else { fragTo.start = fragFrom.start - fragTo.duration; } } } }]); return LevelHelper; })(); exports['default'] = LevelHelper; module.exports = exports['default']; },{"../utils/logger":21}],14:[function(require,module,exports){ /** * HLS interface */ 'use strict'; Object.defineProperty(exports, '__esModule', { value: true }); var _createClass = (function () { function defineProperties(target, props) { for (var i = 0; i < props.length; i++) { var descriptor = props[i]; descriptor.enumerable = descriptor.enumerable || false; descriptor.configurable = true; if ('value' in descriptor) descriptor.writable = true; Object.defineProperty(target, descriptor.key, descriptor); } } return function (Constructor, protoProps, staticProps) { if (protoProps) defineProperties(Constructor.prototype, protoProps); if (staticProps) defineProperties(Constructor, staticProps); return Constructor; }; })(); function _interopRequireDefault(obj) { return obj && obj.__esModule ? obj : { 'default': obj }; } function _classCallCheck(instance, Constructor) { if (!(instance instanceof Constructor)) { throw new TypeError('Cannot call a class as a function'); } } var _events = require('./events'); var _events2 = _interopRequireDefault(_events); var _errors = require('./errors'); var _loaderPlaylistLoader = require('./loader/playlist-loader'); var _loaderPlaylistLoader2 = _interopRequireDefault(_loaderPlaylistLoader); var _loaderFragmentLoader = require('./loader/fragment-loader'); var _loaderFragmentLoader2 = _interopRequireDefault(_loaderFragmentLoader); var _controllerAbrController = require('./controller/abr-controller'); var _controllerAbrController2 = _interopRequireDefault(_controllerAbrController); var _controllerMseMediaController = require('./controller/mse-media-controller'); var _controllerMseMediaController2 = _interopRequireDefault(_controllerMseMediaController); var _controllerLevelController = require('./controller/level-controller'); var _controllerLevelController2 = _interopRequireDefault(_controllerLevelController); //import FPSController from './controller/fps-controller'; var _utilsLogger = require('./utils/logger'); var _utilsXhrLoader = require('./utils/xhr-loader'); var _utilsXhrLoader2 = _interopRequireDefault(_utilsXhrLoader); var _events3 = require('events'); var _events4 = _interopRequireDefault(_events3); var _loaderKeyLoader = require('./loader/key-loader'); var _loaderKeyLoader2 = _interopRequireDefault(_loaderKeyLoader); var Hls = (function () { _createClass(Hls, null, [{ key: 'isSupported', value: function isSupported() { return window.MediaSource && window.MediaSource.isTypeSupported('video/mp4; codecs="avc1.42E01E,mp4a.40.2"'); } }, { key: 'Events', get: function get() { return _events2['default']; } }, { key: 'ErrorTypes', get: function get() { return _errors.ErrorTypes; } }, { key: 'ErrorDetails', get: function get() { return _errors.ErrorDetails; } }]); function Hls() { var config = arguments.length <= 0 || arguments[0] === undefined ? {} : arguments[0]; _classCallCheck(this, Hls); var configDefault = { autoStartLoad: true, debug: false, maxBufferLength: 30, maxBufferSize: 60 * 1000 * 1000, liveSyncDurationCount: 3, liveMaxLatencyDurationCount: Infinity, maxMaxBufferLength: 600, enableWorker: true, fragLoadingTimeOut: 20000, fragLoadingMaxRetry: 1, fragLoadingRetryDelay: 1000, fragLoadingLoopThreshold: 3, manifestLoadingTimeOut: 10000, manifestLoadingMaxRetry: 1, manifestLoadingRetryDelay: 1000, fpsDroppedMonitoringPeriod: 5000, fpsDroppedMonitoringThreshold: 0.2, appendErrorMaxRetry: 200, loader: _utilsXhrLoader2['default'], fLoader: undefined, pLoader: undefined, abrController: _controllerAbrController2['default'], mediaController: _controllerMseMediaController2['default'] }; for (var prop in configDefault) { if (prop in config) { continue; } config[prop] = configDefault[prop]; } if (config.liveMaxLatencyDurationCount !== undefined && config.liveMaxLatencyDurationCount <= config.liveSyncDurationCount) { throw new Error('Illegal hls.js configuration: "liveMaxLatencyDurationCount" must be strictly superior to "liveSyncDurationCount" in player configuration'); } (0, _utilsLogger.enableLogs)(config.debug); this.config = config; // observer setup var observer = this.observer = new _events4['default'](); observer.trigger = function trigger(event) { for (var _len = arguments.length, data = Array(_len > 1 ? _len - 1 : 0), _key = 1; _key < _len; _key++) { data[_key - 1] = arguments[_key]; } observer.emit.apply(observer, [event, event].concat(data)); }; observer.off = function off(event) { for (var _len2 = arguments.length, data = Array(_len2 > 1 ? _len2 - 1 : 0), _key2 = 1; _key2 < _len2; _key2++) { data[_key2 - 1] = arguments[_key2]; } observer.removeListener.apply(observer, [event].concat(data)); }; this.on = observer.on.bind(observer); this.off = observer.off.bind(observer); this.trigger = observer.trigger.bind(observer); this.playlistLoader = new _loaderPlaylistLoader2['default'](this); this.fragmentLoader = new _loaderFragmentLoader2['default'](this); this.levelController = new _controllerLevelController2['default'](this); this.abrController = new config.abrController(this); this.mediaController = new config.mediaController(this); this.keyLoader = new _loaderKeyLoader2['default'](this); //this.fpsController = new FPSController(this); } _createClass(Hls, [{ key: 'destroy', value: function destroy() { _utilsLogger.logger.log('destroy'); this.detachMedia(); this.trigger(_events2['default'].DESTROYING); this.playlistLoader.destroy(); this.fragmentLoader.destroy(); this.levelController.destroy(); this.mediaController.destroy(); this.keyLoader.destroy(); //this.fpsController.destroy(); this.url = null; this.observer.removeAllListeners(); } }, { key: 'attachMedia', value: function attachMedia(media) { _utilsLogger.logger.log('attachMedia'); this.media = media; this.trigger(_events2['default'].MEDIA_ATTACHING, { media: media }); } }, { key: 'detachMedia', value: function detachMedia() { _utilsLogger.logger.log('detachMedia'); this.trigger(_events2['default'].MEDIA_DETACHING); this.media = null; } }, { key: 'loadSource', value: function loadSource(url) { _utilsLogger.logger.log('loadSource:' + url); this.url = url; // when attaching to a source URL, trigger a playlist load this.trigger(_events2['default'].MANIFEST_LOADING, { url: url }); } }, { key: 'startLoad', value: function startLoad() { _utilsLogger.logger.log('startLoad'); this.mediaController.startLoad(); } }, { key: 'recoverMediaError', value: function recoverMediaError() { _utilsLogger.logger.log('recoverMediaError'); var media = this.media; this.detachMedia(); this.attachMedia(media); } /** Return all quality levels **/ }, { key: 'levels', get: function get() { return this.levelController.levels; } /** Return current playback quality level **/ }, { key: 'currentLevel', get: function get() { return this.mediaController.currentLevel; }, /* set quality level immediately (-1 for automatic level selection) */ set: function set(newLevel) { _utilsLogger.logger.log('set currentLevel:' + newLevel); this.loadLevel = newLevel; this.mediaController.immediateLevelSwitch(); } /** Return next playback quality level (quality level of next fragment) **/ }, { key: 'nextLevel', get: function get() { return this.mediaController.nextLevel; }, /* set quality level for next fragment (-1 for automatic level selection) */ set: function set(newLevel) { _utilsLogger.logger.log('set nextLevel:' + newLevel); this.levelController.manualLevel = newLevel; this.mediaController.nextLevelSwitch(); } /** Return the quality level of current/last loaded fragment **/ }, { key: 'loadLevel', get: function get() { return this.levelController.level; }, /* set quality level for current/next loaded fragment (-1 for automatic level selection) */ set: function set(newLevel) { _utilsLogger.logger.log('set loadLevel:' + newLevel); this.levelController.manualLevel = newLevel; } /** Return the quality level of next loaded fragment **/ }, { key: 'nextLoadLevel', get: function get() { return this.levelController.nextLoadLevel(); }, /** set quality level of next loaded fragment **/ set: function set(level) { this.levelController.level = level; } /** Return first level (index of first level referenced in manifest) **/ }, { key: 'firstLevel', get: function get() { return this.levelController.firstLevel; }, /** set first level (index of first level referenced in manifest) **/ set: function set(newLevel) { _utilsLogger.logger.log('set firstLevel:' + newLevel); this.levelController.firstLevel = newLevel; } /** Return start level (level of first fragment that will be played back) if not overrided by user, first level appearing in manifest will be used as start level if -1 : automatic start level selection, playback will start from level matching download bandwidth (determined from download of first segment) **/ }, { key: 'startLevel', get: function get() { return this.levelController.startLevel; }, /** set start level (level of first fragment that will be played back) if not overrided by user, first level appearing in manifest will be used as start level if -1 : automatic start level selection, playback will start from level matching download bandwidth (determined from download of first segment) **/ set: function set(newLevel) { _utilsLogger.logger.log('set startLevel:' + newLevel); this.levelController.startLevel = newLevel; } /** Return the capping/max level value that could be used by automatic level selection algorithm **/ }, { key: 'autoLevelCapping', get: function get() { return this.abrController.autoLevelCapping; }, /** set the capping/max level value that could be used by automatic level selection algorithm **/ set: function set(newLevel) { _utilsLogger.logger.log('set autoLevelCapping:' + newLevel); this.abrController.autoLevelCapping = newLevel; } /* check if we are in automatic level selection mode */ }, { key: 'autoLevelEnabled', get: function get() { return this.levelController.manualLevel === -1; } /* return manual level */ }, { key: 'manualLevel', get: function get() { return this.levelController.manualLevel; } }]); return Hls; })(); exports['default'] = Hls; module.exports = exports['default']; },{"./controller/abr-controller":3,"./controller/level-controller":4,"./controller/mse-media-controller":5,"./errors":11,"./events":12,"./loader/fragment-loader":15,"./loader/key-loader":16,"./loader/playlist-loader":17,"./utils/logger":21,"./utils/xhr-loader":23,"events":1}],15:[function(require,module,exports){ /* * Fragment Loader */ 'use strict'; Object.defineProperty(exports, '__esModule', { value: true }); var _createClass = (function () { function defineProperties(target, props) { for (var i = 0; i < props.length; i++) { var descriptor = props[i]; descriptor.enumerable = descriptor.enumerable || false; descriptor.configurable = true; if ('value' in descriptor) descriptor.writable = true; Object.defineProperty(target, descriptor.key, descriptor); } } return function (Constructor, protoProps, staticProps) { if (protoProps) defineProperties(Constructor.prototype, protoProps); if (staticProps) defineProperties(Constructor, staticProps); return Constructor; }; })(); function _interopRequireDefault(obj) { return obj && obj.__esModule ? obj : { 'default': obj }; } function _classCallCheck(instance, Constructor) { if (!(instance instanceof Constructor)) { throw new TypeError('Cannot call a class as a function'); } } var _events = require('../events'); var _events2 = _interopRequireDefault(_events); var _errors = require('../errors'); var FragmentLoader = (function () { function FragmentLoader(hls) { _classCallCheck(this, FragmentLoader); this.hls = hls; this.onfl = this.onFragLoading.bind(this); hls.on(_events2['default'].FRAG_LOADING, this.onfl); } _createClass(FragmentLoader, [{ key: 'destroy', value: function destroy() { if (this.loader) { this.loader.destroy(); this.loader = null; } this.hls.off(_events2['default'].FRAG_LOADING, this.onfl); } }, { key: 'onFragLoading', value: function onFragLoading(event, data) { var frag = data.frag; this.frag = frag; this.frag.loaded = 0; var config = this.hls.config; frag.loader = this.loader = typeof config.fLoader !== 'undefined' ? new config.fLoader(config) : new config.loader(config); this.loader.load(frag.url, 'arraybuffer', this.loadsuccess.bind(this), this.loaderror.bind(this), this.loadtimeout.bind(this), config.fragLoadingTimeOut, config.fragLoadingMaxRetry, config.fragLoadingRetryDelay, this.loadprogress.bind(this), frag); } }, { key: 'loadsuccess', value: function loadsuccess(event, stats) { var payload = event.currentTarget.response; stats.length = payload.byteLength; // detach fragment loader on load success this.frag.loader = undefined; this.hls.trigger(_events2['default'].FRAG_LOADED, { payload: payload, frag: this.frag, stats: stats }); } }, { key: 'loaderror', value: function loaderror(event) { this.loader.abort(); this.hls.trigger(_events2['default'].ERROR, { type: _errors.ErrorTypes.NETWORK_ERROR, details: _errors.ErrorDetails.FRAG_LOAD_ERROR, fatal: false, frag: this.frag, response: event }); } }, { key: 'loadtimeout', value: function loadtimeout() { this.loader.abort(); this.hls.trigger(_events2['default'].ERROR, { type: _errors.ErrorTypes.NETWORK_ERROR, details: _errors.ErrorDetails.FRAG_LOAD_TIMEOUT, fatal: false, frag: this.frag }); } }, { key: 'loadprogress', value: function loadprogress(event, stats) { this.frag.loaded = stats.loaded; this.hls.trigger(_events2['default'].FRAG_LOAD_PROGRESS, { frag: this.frag, stats: stats }); } }]); return FragmentLoader; })(); exports['default'] = FragmentLoader; module.exports = exports['default']; },{"../errors":11,"../events":12}],16:[function(require,module,exports){ /* * Decrypt key Loader */ 'use strict'; Object.defineProperty(exports, '__esModule', { value: true }); var _createClass = (function () { function defineProperties(target, props) { for (var i = 0; i < props.length; i++) { var descriptor = props[i]; descriptor.enumerable = descriptor.enumerable || false; descriptor.configurable = true; if ('value' in descriptor) descriptor.writable = true; Object.defineProperty(target, descriptor.key, descriptor); } } return function (Constructor, protoProps, staticProps) { if (protoProps) defineProperties(Constructor.prototype, protoProps); if (staticProps) defineProperties(Constructor, staticProps); return Constructor; }; })(); function _interopRequireDefault(obj) { return obj && obj.__esModule ? obj : { 'default': obj }; } function _classCallCheck(instance, Constructor) { if (!(instance instanceof Constructor)) { throw new TypeError('Cannot call a class as a function'); } } var _events = require('../events'); var _events2 = _interopRequireDefault(_events); var _errors = require('../errors'); var KeyLoader = (function () { function KeyLoader(hls) { _classCallCheck(this, KeyLoader); this.hls = hls; this.decryptkey = null; this.decrypturl = null; this.ondkl = this.onDecryptKeyLoading.bind(this); hls.on(_events2['default'].KEY_LOADING, this.ondkl); } _createClass(KeyLoader, [{ key: 'destroy', value: function destroy() { if (this.loader) { this.loader.destroy(); this.loader = null; } this.hls.off(_events2['default'].KEY_LOADING, this.ondkl); } }, { key: 'onDecryptKeyLoading', value: function onDecryptKeyLoading(event, data) { var frag = this.frag = data.frag, decryptdata = frag.decryptdata, uri = decryptdata.uri; // if uri is different from previous one or if decrypt key not retrieved yet if (uri !== this.decrypturl || this.decryptkey === null) { var config = this.hls.config; frag.loader = this.loader = new config.loader(config); this.decrypturl = uri; this.decryptkey = null; frag.loader.load(uri, 'arraybuffer', this.loadsuccess.bind(this), this.loaderror.bind(this), this.loadtimeout.bind(this), config.fragLoadingTimeOut, config.fragLoadingMaxRetry, config.fragLoadingRetryDelay, this.loadprogress.bind(this), frag); } else if (this.decryptkey) { // we already loaded this key, return it decryptdata.key = this.decryptkey; this.hls.trigger(_events2['default'].KEY_LOADED, { frag: frag }); } } }, { key: 'loadsuccess', value: function loadsuccess(event) { var frag = this.frag; this.decryptkey = frag.decryptdata.key = new Uint8Array(event.currentTarget.response); // detach fragment loader on load success frag.loader = undefined; this.hls.trigger(_events2['default'].KEY_LOADED, { frag: frag }); } }, { key: 'loaderror', value: function loaderror(event) { this.loader.abort(); this.hls.trigger(_events2['default'].ERROR, { type: _errors.ErrorTypes.NETWORK_ERROR, details: _errors.ErrorDetails.KEY_LOAD_ERROR, fatal: false, frag: this.frag, response: event }); } }, { key: 'loadtimeout', value: function loadtimeout() { this.loader.abort(); this.hls.trigger(_events2['default'].ERROR, { type: _errors.ErrorTypes.NETWORK_ERROR, details: _errors.ErrorDetails.KEY_LOAD_TIMEOUT, fatal: false, frag: this.frag }); } }, { key: 'loadprogress', value: function loadprogress() {} }]); return KeyLoader; })(); exports['default'] = KeyLoader; module.exports = exports['default']; },{"../errors":11,"../events":12}],17:[function(require,module,exports){ /** * Playlist Loader */ 'use strict'; Object.defineProperty(exports, '__esModule', { value: true }); var _createClass = (function () { function defineProperties(target, props) { for (var i = 0; i < props.length; i++) { var descriptor = props[i]; descriptor.enumerable = descriptor.enumerable || false; descriptor.configurable = true; if ('value' in descriptor) descriptor.writable = true; Object.defineProperty(target, descriptor.key, descriptor); } } return function (Constructor, protoProps, staticProps) { if (protoProps) defineProperties(Constructor.prototype, protoProps); if (staticProps) defineProperties(Constructor, staticProps); return Constructor; }; })(); function _interopRequireDefault(obj) { return obj && obj.__esModule ? obj : { 'default': obj }; } function _classCallCheck(instance, Constructor) { if (!(instance instanceof Constructor)) { throw new TypeError('Cannot call a class as a function'); } } var _events = require('../events'); var _events2 = _interopRequireDefault(_events); var _errors = require('../errors'); var _utilsUrl = require('../utils/url'); var _utilsUrl2 = _interopRequireDefault(_utilsUrl); //import {logger} from '../utils/logger'; var PlaylistLoader = (function () { function PlaylistLoader(hls) { _classCallCheck(this, PlaylistLoader); this.hls = hls; this.onml = this.onManifestLoading.bind(this); this.onll = this.onLevelLoading.bind(this); hls.on(_events2['default'].MANIFEST_LOADING, this.onml); hls.on(_events2['default'].LEVEL_LOADING, this.onll); } _createClass(PlaylistLoader, [{ key: 'destroy', value: function destroy() { if (this.loader) { this.loader.destroy(); this.loader = null; } this.url = this.id = null; this.hls.off(_events2['default'].MANIFEST_LOADING, this.onml); this.hls.off(_events2['default'].LEVEL_LOADING, this.onll); } }, { key: 'onManifestLoading', value: function onManifestLoading(event, data) { this.load(data.url, null); } }, { key: 'onLevelLoading', value: function onLevelLoading(event, data) { this.load(data.url, data.level, data.id); } }, { key: 'load', value: function load(url, id1, id2) { var config = this.hls.config; this.url = url; this.id = id1; this.id2 = id2; this.loader = typeof config.pLoader !== 'undefined' ? new config.pLoader(config) : new config.loader(config); this.loader.load(url, '', this.loadsuccess.bind(this), this.loaderror.bind(this), this.loadtimeout.bind(this), config.manifestLoadingTimeOut, config.manifestLoadingMaxRetry, config.manifestLoadingRetryDelay); } }, { key: 'resolve', value: function resolve(url, baseUrl) { return _utilsUrl2['default'].buildAbsoluteURL(baseUrl, url); } }, { key: 'parseMasterPlaylist', value: function parseMasterPlaylist(string, baseurl) { var levels = [], level = {}, result, codecs, codec; // https://regex101.com is your friend var re = /#EXT-X-STREAM-INF:([^\n\r]*(BAND)WIDTH=(\d+))?([^\n\r]*(CODECS)=\"([^\"\n\r]*)\",?)?([^\n\r]*(RES)OLUTION=(\d+)x(\d+))?([^\n\r]*(NAME)=\"(.*)\")?[^\n\r]*[\r\n]+([^\r\n]+)/g; while ((result = re.exec(string)) != null) { result.shift(); result = result.filter(function (n) { return n !== undefined; }); level.url = this.resolve(result.pop(), baseurl); while (result.length > 0) { switch (result.shift()) { case 'RES': level.width = parseInt(result.shift()); level.height = parseInt(result.shift()); break; case 'BAND': level.bitrate = parseInt(result.shift()); break; case 'NAME': level.name = result.shift(); break; case 'CODECS': codecs = result.shift().split(','); while (codecs.length > 0) { codec = codecs.shift(); if (codec.indexOf('avc1') !== -1) { level.videoCodec = this.avc1toavcoti(codec); } else { level.audioCodec = codec; } } break; default: break; } } levels.push(level); level = {}; } return levels; } }, { key: 'avc1toavcoti', value: function avc1toavcoti(codec) { var result, avcdata = codec.split('.'); if (avcdata.length > 2) { result = avcdata.shift() + '.'; result += parseInt(avcdata.shift()).toString(16); result += ('00' + parseInt(avcdata.shift()).toString(16)).substr(-4); } else { result = codec; } return result; } }, { key: 'parseKeyParamsByRegex', value: function parseKeyParamsByRegex(string, regexp) { var result = regexp.exec(string); if (result) { result.shift(); result = result.filter(function (n) { return n !== undefined; }); if (result.length === 2) { return result[1]; } } return null; } }, { key: 'cloneObj', value: function cloneObj(obj) { return JSON.parse(JSON.stringify(obj)); } }, { key: 'parseLevelPlaylist', value: function parseLevelPlaylist(string, baseurl, id) { var currentSN = 0, totalduration = 0, level = { url: baseurl, fragments: [], live: true, startSN: 0 }, result, regexp, cc = 0, frag, byteRangeEndOffset, byteRangeStartOffset; var levelkey = { method: null, key: null, iv: null, uri: null }; regexp = /(?:#EXT-X-(MEDIA-SEQUENCE):(\d+))|(?:#EXT-X-(TARGETDURATION):(\d+))|(?:#EXT-X-(KEY):(.*))|(?:#EXT(INF):([\d\.]+)[^\r\n]*([\r\n]+[^#|\r\n]+)?)|(?:#EXT-X-(BYTERANGE):([\d]+[@[\d]*)]*[\r\n]+([^#|\r\n]+)?|(?:#EXT-X-(ENDLIST))|(?:#EXT-X-(DIS)CONTINUITY))/g; while ((result = regexp.exec(string)) !== null) { result.shift(); result = result.filter(function (n) { return n !== undefined; }); switch (result[0]) { case 'MEDIA-SEQUENCE': currentSN = level.startSN = parseInt(result[1]); break; case 'TARGETDURATION': level.targetduration = parseFloat(result[1]); break; case 'ENDLIST': level.live = false; break; case 'DIS': cc++; break; case 'BYTERANGE': var params = result[1].split('@'); if (params.length === 1) { byteRangeStartOffset = byteRangeEndOffset; } else { byteRangeStartOffset = parseInt(params[1]); } byteRangeEndOffset = parseInt(params[0]) + byteRangeStartOffset; frag = level.fragments.length ? level.fragments[level.fragments.length - 1] : null; if (frag && !frag.url) { frag.byteRangeStartOffset = byteRangeStartOffset; frag.byteRangeEndOffset = byteRangeEndOffset; frag.url = this.resolve(result[2], baseurl); } break; case 'INF': var duration = parseFloat(result[1]); if (!isNaN(duration)) { var fragdecryptdata, sn = currentSN++; if (levelkey.method && levelkey.uri && !levelkey.iv) { fragdecryptdata = this.cloneObj(levelkey); var uint8View = new Uint8Array(16); for (var i = 12; i < 16; i++) { uint8View[i] = sn >> 8 * (15 - i) & 0xff; } fragdecryptdata.iv = uint8View; } else { fragdecryptdata = levelkey; } level.fragments.push({ url: result[2] ? this.resolve(result[2], baseurl) : null, duration: duration, start: totalduration, sn: sn, level: id, cc: cc, byteRangeStartOffset: byteRangeStartOffset, byteRangeEndOffset: byteRangeEndOffset, decryptdata: fragdecryptdata }); totalduration += duration; byteRangeStartOffset = null; } break; case 'KEY': // https://tools.ietf.org/html/draft-pantos-http-live-streaming-08#section-3.4.4 var decryptparams = result[1]; var decryptmethod = this.parseKeyParamsByRegex(decryptparams, /(METHOD)=([^,]*)/), decrypturi = this.parseKeyParamsByRegex(decryptparams, /(URI)=["]([^,]*)["]/), decryptiv = this.parseKeyParamsByRegex(decryptparams, /(IV)=([^,]*)/); if (decryptmethod) { levelkey = { method: null, key: null, iv: null, uri: null }; if (decrypturi && decryptmethod === 'AES-128') { levelkey.method = decryptmethod; // URI to get the key levelkey.uri = this.resolve(decrypturi, baseurl); levelkey.key = null; // Initialization Vector (IV) if (decryptiv) { levelkey.iv = decryptiv; if (levelkey.iv.substring(0, 2) === '0x') { levelkey.iv = levelkey.iv.substring(2); } levelkey.iv = levelkey.iv.match(/.{8}/g); levelkey.iv[0] = parseInt(levelkey.iv[0], 16); levelkey.iv[1] = parseInt(levelkey.iv[1], 16); levelkey.iv[2] = parseInt(levelkey.iv[2], 16); levelkey.iv[3] = parseInt(levelkey.iv[3], 16); levelkey.iv = new Uint32Array(levelkey.iv); } } } break; default: break; } } //logger.log('found ' + level.fragments.length + ' fragments'); level.totalduration = totalduration; level.endSN = currentSN - 1; return level; } }, { key: 'loadsuccess', value: function loadsuccess(event, stats) { var string = event.currentTarget.responseText, url = event.currentTarget.responseURL, id = this.id, id2 = this.id2, hls = this.hls, levels; // responseURL not supported on some browsers (it is used to detect URL redirection) if (url === undefined) { // fallback to initial URL url = this.url; } stats.tload = performance.now(); stats.mtime = new Date(event.currentTarget.getResponseHeader('Last-Modified')); if (string.indexOf('#EXTM3U') === 0) { if (string.indexOf('#EXTINF:') > 0) { // 1 level playlist // if first request, fire manifest loaded event, level will be reloaded afterwards // (this is to have a uniform logic for 1 level/multilevel playlists) if (this.id === null) { hls.trigger(_events2['default'].MANIFEST_LOADED, { levels: [{ url: url }], url: url, stats: stats }); } else { var levelDetails = this.parseLevelPlaylist(string, url, id); stats.tparsed = performance.now(); hls.trigger(_events2['default'].LEVEL_LOADED, { details: levelDetails, level: id, id: id2, stats: stats }); } } else { levels = this.parseMasterPlaylist(string, url); // multi level playlist, parse level info if (levels.length) { hls.trigger(_events2['default'].MANIFEST_LOADED, { levels: levels, url: url, stats: stats }); } else { hls.trigger(_events2['default'].ERROR, { type: _errors.ErrorTypes.NETWORK_ERROR, details: _errors.ErrorDetails.MANIFEST_PARSING_ERROR, fatal: true, url: url, reason: 'no level found in manifest' }); } } } else { hls.trigger(_events2['default'].ERROR, { type: _errors.ErrorTypes.NETWORK_ERROR, details: _errors.ErrorDetails.MANIFEST_PARSING_ERROR, fatal: true, url: url, reason: 'no EXTM3U delimiter' }); } } }, { key: 'loaderror', value: function loaderror(event) { var details, fatal; if (this.id === null) { details = _errors.ErrorDetails.MANIFEST_LOAD_ERROR; fatal = true; } else { details = _errors.ErrorDetails.LEVEL_LOAD_ERROR; fatal = false; } this.loader.abort(); this.hls.trigger(_events2['default'].ERROR, { type: _errors.ErrorTypes.NETWORK_ERROR, details: details, fatal: fatal, url: this.url, loader: this.loader, response: event.currentTarget, level: this.id, id: this.id2 }); } }, { key: 'loadtimeout', value: function loadtimeout() { var details, fatal; if (this.id === null) { details = _errors.ErrorDetails.MANIFEST_LOAD_TIMEOUT; fatal = true; } else { details = _errors.ErrorDetails.LEVEL_LOAD_TIMEOUT; fatal = false; } this.loader.abort(); this.hls.trigger(_events2['default'].ERROR, { type: _errors.ErrorTypes.NETWORK_ERROR, details: details, fatal: fatal, url: this.url, loader: this.loader, level: this.id, id: this.id2 }); } }]); return PlaylistLoader; })(); exports['default'] = PlaylistLoader; module.exports = exports['default']; },{"../errors":11,"../events":12,"../utils/url":22}],18:[function(require,module,exports){ /** * Generate MP4 Box */ //import Hex from '../utils/hex'; 'use strict'; Object.defineProperty(exports, '__esModule', { value: true }); var _createClass = (function () { function defineProperties(target, props) { for (var i = 0; i < props.length; i++) { var descriptor = props[i]; descriptor.enumerable = descriptor.enumerable || false; descriptor.configurable = true; if ('value' in descriptor) descriptor.writable = true; Object.defineProperty(target, descriptor.key, descriptor); } } return function (Constructor, protoProps, staticProps) { if (protoProps) defineProperties(Constructor.prototype, protoProps); if (staticProps) defineProperties(Constructor, staticProps); return Constructor; }; })(); function _classCallCheck(instance, Constructor) { if (!(instance instanceof Constructor)) { throw new TypeError('Cannot call a class as a function'); } } var MP4 = (function () { function MP4() { _classCallCheck(this, MP4); } _createClass(MP4, null, [{ key: 'init', value: function init() { MP4.types = { avc1: [], // codingname avcC: [], btrt: [], dinf: [], dref: [], esds: [], ftyp: [], hdlr: [], mdat: [], mdhd: [], mdia: [], mfhd: [], minf: [], moof: [], moov: [], mp4a: [], mvex: [], mvhd: [], sdtp: [], stbl: [], stco: [], stsc: [], stsd: [], stsz: [], stts: [], tfdt: [], tfhd: [], traf: [], trak: [], trun: [], trex: [], tkhd: [], vmhd: [], smhd: [] }; var i; for (i in MP4.types) { if (MP4.types.hasOwnProperty(i)) { MP4.types[i] = [i.charCodeAt(0), i.charCodeAt(1), i.charCodeAt(2), i.charCodeAt(3)]; } } MP4.MAJOR_BRAND = new Uint8Array(['i'.charCodeAt(0), 's'.charCodeAt(0), 'o'.charCodeAt(0), 'm'.charCodeAt(0)]); MP4.AVC1_BRAND = new Uint8Array(['a'.charCodeAt(0), 'v'.charCodeAt(0), 'c'.charCodeAt(0), '1'.charCodeAt(0)]); MP4.MINOR_VERSION = new Uint8Array([0, 0, 0, 1]); MP4.VIDEO_HDLR = new Uint8Array([0x00, // version 0 0x00, 0x00, 0x00, // flags 0x00, 0x00, 0x00, 0x00, // pre_defined 0x76, 0x69, 0x64, 0x65, // handler_type: 'vide' 0x00, 0x00, 0x00, 0x00, // reserved 0x00, 0x00, 0x00, 0x00, // reserved 0x00, 0x00, 0x00, 0x00, // reserved 0x56, 0x69, 0x64, 0x65, 0x6f, 0x48, 0x61, 0x6e, 0x64, 0x6c, 0x65, 0x72, 0x00 // name: 'VideoHandler' ]); MP4.AUDIO_HDLR = new Uint8Array([0x00, // version 0 0x00, 0x00, 0x00, // flags 0x00, 0x00, 0x00, 0x00, // pre_defined 0x73, 0x6f, 0x75, 0x6e, // handler_type: 'soun' 0x00, 0x00, 0x00, 0x00, // reserved 0x00, 0x00, 0x00, 0x00, // reserved 0x00, 0x00, 0x00, 0x00, // reserved 0x53, 0x6f, 0x75, 0x6e, 0x64, 0x48, 0x61, 0x6e, 0x64, 0x6c, 0x65, 0x72, 0x00 // name: 'SoundHandler' ]); MP4.HDLR_TYPES = { 'video': MP4.VIDEO_HDLR, 'audio': MP4.AUDIO_HDLR }; MP4.DREF = new Uint8Array([0x00, // version 0 0x00, 0x00, 0x00, // flags 0x00, 0x00, 0x00, 0x01, // entry_count 0x00, 0x00, 0x00, 0x0c, // entry_size 0x75, 0x72, 0x6c, 0x20, // 'url' type 0x00, // version 0 0x00, 0x00, 0x01 // entry_flags ]); MP4.STCO = new Uint8Array([0x00, // version 0x00, 0x00, 0x00, // flags 0x00, 0x00, 0x00, 0x00 // entry_count ]); MP4.STSC = MP4.STCO; MP4.STTS = MP4.STCO; MP4.STSZ = new Uint8Array([0x00, // version 0x00, 0x00, 0x00, // flags 0x00, 0x00, 0x00, 0x00, // sample_size 0x00, 0x00, 0x00, 0x00]); // sample_count MP4.VMHD = new Uint8Array([0x00, // version 0x00, 0x00, 0x01, // flags 0x00, 0x00, // graphicsmode 0x00, 0x00, 0x00, 0x00, 0x00, 0x00 // opcolor ]); MP4.SMHD = new Uint8Array([0x00, // version 0x00, 0x00, 0x00, // flags 0x00, 0x00, // balance 0x00, 0x00 // reserved ]); MP4.STSD = new Uint8Array([0x00, // version 0 0x00, 0x00, 0x00, // flags 0x00, 0x00, 0x00, 0x01]); // entry_count MP4.FTYP = MP4.box(MP4.types.ftyp, MP4.MAJOR_BRAND, MP4.MINOR_VERSION, MP4.MAJOR_BRAND, MP4.AVC1_BRAND); MP4.DINF = MP4.box(MP4.types.dinf, MP4.box(MP4.types.dref, MP4.DREF)); } }, { key: 'box', value: function box(type) { var payload = Array.prototype.slice.call(arguments, 1), size = 0, i = payload.length, result, view; // calculate the total size we need to allocate while (i--) { size += payload[i].byteLength; } result = new Uint8Array(size + 8); view = new DataView(result.buffer); view.setUint32(0, result.byteLength); result.set(type, 4); // copy the payload into the result for (i = 0, size = 8; i < payload.length; i++) { result.set(payload[i], size); size += payload[i].byteLength; } return result; } }, { key: 'hdlr', value: function hdlr(type) { return MP4.box(MP4.types.hdlr, MP4.HDLR_TYPES[type]); } }, { key: 'mdat', value: function mdat(data) { return MP4.box(MP4.types.mdat, data); } }, { key: 'mdhd', value: function mdhd(timescale, duration) { return MP4.box(MP4.types.mdhd, new Uint8Array([0x00, // version 0 0x00, 0x00, 0x00, // flags 0x00, 0x00, 0x00, 0x02, // creation_time 0x00, 0x00, 0x00, 0x03, // modification_time timescale >> 24 & 0xFF, timescale >> 16 & 0xFF, timescale >> 8 & 0xFF, timescale & 0xFF, // timescale duration >> 24, duration >> 16 & 0xFF, duration >> 8 & 0xFF, duration & 0xFF, // duration 0x55, 0xc4, // 'und' language (undetermined) 0x00, 0x00])); } }, { key: 'mdia', value: function mdia(track) { return MP4.box(MP4.types.mdia, MP4.mdhd(track.timescale, track.duration), MP4.hdlr(track.type), MP4.minf(track)); } }, { key: 'mfhd', value: function mfhd(sequenceNumber) { return MP4.box(MP4.types.mfhd, new Uint8Array([0x00, 0x00, 0x00, 0x00, // flags sequenceNumber >> 24, sequenceNumber >> 16 & 0xFF, sequenceNumber >> 8 & 0xFF, sequenceNumber & 0xFF])); } }, { key: 'minf', // sequence_number value: function minf(track) { if (track.type === 'audio') { return MP4.box(MP4.types.minf, MP4.box(MP4.types.smhd, MP4.SMHD), MP4.DINF, MP4.stbl(track)); } else { return MP4.box(MP4.types.minf, MP4.box(MP4.types.vmhd, MP4.VMHD), MP4.DINF, MP4.stbl(track)); } } }, { key: 'moof', value: function moof(sn, baseMediaDecodeTime, track) { return MP4.box(MP4.types.moof, MP4.mfhd(sn), MP4.traf(track, baseMediaDecodeTime)); } /** * @param tracks... (optional) {array} the tracks associated with this movie */ }, { key: 'moov', value: function moov(tracks) { var i = tracks.length, boxes = []; while (i--) { boxes[i] = MP4.trak(tracks[i]); } return MP4.box.apply(null, [MP4.types.moov, MP4.mvhd(tracks[0].timescale, tracks[0].duration)].concat(boxes).concat(MP4.mvex(tracks))); } }, { key: 'mvex', value: function mvex(tracks) { var i = tracks.length, boxes = []; while (i--) { boxes[i] = MP4.trex(tracks[i]); } return MP4.box.apply(null, [MP4.types.mvex].concat(boxes)); } }, { key: 'mvhd', value: function mvhd(timescale, duration) { var bytes = new Uint8Array([0x00, // version 0 0x00, 0x00, 0x00, // flags 0x00, 0x00, 0x00, 0x01, // creation_time 0x00, 0x00, 0x00, 0x02, // modification_time timescale >> 24 & 0xFF, timescale >> 16 & 0xFF, timescale >> 8 & 0xFF, timescale & 0xFF, // timescale duration >> 24 & 0xFF, duration >> 16 & 0xFF, duration >> 8 & 0xFF, duration & 0xFF, // duration 0x00, 0x01, 0x00, 0x00, // 1.0 rate 0x01, 0x00, // 1.0 volume 0x00, 0x00, // reserved 0x00, 0x00, 0x00, 0x00, // reserved 0x00, 0x00, 0x00, 0x00, // reserved 0x00, 0x01, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x01, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x40, 0x00, 0x00, 0x00, // transformation: unity matrix 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, // pre_defined 0xff, 0xff, 0xff, 0xff // next_track_ID ]); return MP4.box(MP4.types.mvhd, bytes); } }, { key: 'sdtp', value: function sdtp(track) { var samples = track.samples || [], bytes = new Uint8Array(4 + samples.length), flags, i; // leave the full box header (4 bytes) all zero // write the sample table for (i = 0; i < samples.length; i++) { flags = samples[i].flags; bytes[i + 4] = flags.dependsOn << 4 | flags.isDependedOn << 2 | flags.hasRedundancy; } return MP4.box(MP4.types.sdtp, bytes); } }, { key: 'stbl', value: function stbl(track) { return MP4.box(MP4.types.stbl, MP4.stsd(track), MP4.box(MP4.types.stts, MP4.STTS), MP4.box(MP4.types.stsc, MP4.STSC), MP4.box(MP4.types.stsz, MP4.STSZ), MP4.box(MP4.types.stco, MP4.STCO)); } }, { key: 'avc1', value: function avc1(track) { var sps = [], pps = [], i, data, len; // assemble the SPSs for (i = 0; i < track.sps.length; i++) { data = track.sps[i]; len = data.byteLength; sps.push(len >>> 8 & 0xFF); sps.push(len & 0xFF); sps = sps.concat(Array.prototype.slice.call(data)); // SPS } // assemble the PPSs for (i = 0; i < track.pps.length; i++) { data = track.pps[i]; len = data.byteLength; pps.push(len >>> 8 & 0xFF); pps.push(len & 0xFF); pps = pps.concat(Array.prototype.slice.call(data)); } var avcc = MP4.box(MP4.types.avcC, new Uint8Array([0x01, // version sps[3], // profile sps[4], // profile compat sps[5], // level 0xfc | 3, // lengthSizeMinusOne, hard-coded to 4 bytes 0xE0 | track.sps.length // 3bit reserved (111) + numOfSequenceParameterSets ].concat(sps).concat([track.pps.length // numOfPictureParameterSets ]).concat(pps))); // "PPS" //console.log('avcc:' + Hex.hexDump(avcc)); return MP4.box(MP4.types.avc1, new Uint8Array([0x00, 0x00, 0x00, // reserved 0x00, 0x00, 0x00, // reserved 0x00, 0x01, // data_reference_index 0x00, 0x00, // pre_defined 0x00, 0x00, // reserved 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, // pre_defined track.width >> 8 & 0xFF, track.width & 0xff, // width track.height >> 8 & 0xFF, track.height & 0xff, // height 0x00, 0x48, 0x00, 0x00, // horizresolution 0x00, 0x48, 0x00, 0x00, // vertresolution 0x00, 0x00, 0x00, 0x00, // reserved 0x00, 0x01, // frame_count 0x13, 0x76, 0x69, 0x64, 0x65, 0x6f, 0x6a, 0x73, 0x2d, 0x63, 0x6f, 0x6e, 0x74, 0x72, 0x69, 0x62, 0x2d, 0x68, 0x6c, 0x73, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, // compressorname 0x00, 0x18, // depth = 24 0x11, 0x11]), // pre_defined = -1 avcc, MP4.box(MP4.types.btrt, new Uint8Array([0x00, 0x1c, 0x9c, 0x80, // bufferSizeDB 0x00, 0x2d, 0xc6, 0xc0, // maxBitrate 0x00, 0x2d, 0xc6, 0xc0])) // avgBitrate ); } }, { key: 'esds', value: function esds(track) { return new Uint8Array([0x00, // version 0 0x00, 0x00, 0x00, // flags 0x03, // descriptor_type 0x17 + track.config.length, // length 0x00, 0x01, //es_id 0x00, // stream_priority 0x04, // descriptor_type 0x0f + track.config.length, // length 0x40, //codec : mpeg4_audio 0x15, // stream_type 0x00, 0x00, 0x00, // buffer_size 0x00, 0x00, 0x00, 0x00, // maxBitrate 0x00, 0x00, 0x00, 0x00, // avgBitrate 0x05 // descriptor_type ].concat([track.config.length]).concat(track.config).concat([0x06, 0x01, 0x02])); // GASpecificConfig)); // length + audio config descriptor } }, { key: 'mp4a', value: function mp4a(track) { return MP4.box(MP4.types.mp4a, new Uint8Array([0x00, 0x00, 0x00, // reserved 0x00, 0x00, 0x00, // reserved 0x00, 0x01, // data_reference_index 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, // reserved 0x00, track.channelCount, // channelcount 0x00, 0x10, // sampleSize:16bits 0x00, 0x00, 0x00, 0x00, // reserved2 track.audiosamplerate >> 8 & 0xFF, track.audiosamplerate & 0xff, // 0x00, 0x00]), MP4.box(MP4.types.esds, MP4.esds(track))); } }, { key: 'stsd', value: function stsd(track) { if (track.type === 'audio') { return MP4.box(MP4.types.stsd, MP4.STSD, MP4.mp4a(track)); } else { return MP4.box(MP4.types.stsd, MP4.STSD, MP4.avc1(track)); } } }, { key: 'tkhd', value: function tkhd(track) { return MP4.box(MP4.types.tkhd, new Uint8Array([0x00, // version 0 0x00, 0x00, 0x07, // flags 0x00, 0x00, 0x00, 0x00, // creation_time 0x00, 0x00, 0x00, 0x00, // modification_time track.id >> 24 & 0xFF, track.id >> 16 & 0xFF, track.id >> 8 & 0xFF, track.id & 0xFF, // track_ID 0x00, 0x00, 0x00, 0x00, // reserved track.duration >> 24, track.duration >> 16 & 0xFF, track.duration >> 8 & 0xFF, track.duration & 0xFF, // duration 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, // reserved 0x00, 0x00, // layer 0x00, 0x00, // alternate_group 0x00, 0x00, // non-audio track volume 0x00, 0x00, // reserved 0x00, 0x01, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x01, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x40, 0x00, 0x00, 0x00, // transformation: unity matrix track.width >> 8 & 0xFF, track.width & 0xFF, 0x00, 0x00, // width track.height >> 8 & 0xFF, track.height & 0xFF, 0x00, 0x00 // height ])); } }, { key: 'traf', value: function traf(track, baseMediaDecodeTime) { var sampleDependencyTable = MP4.sdtp(track); return MP4.box(MP4.types.traf, MP4.box(MP4.types.tfhd, new Uint8Array([0x00, // version 0 0x00, 0x00, 0x00, // flags track.id >> 24, track.id >> 16 & 0XFF, track.id >> 8 & 0XFF, track.id & 0xFF])), // track_ID MP4.box(MP4.types.tfdt, new Uint8Array([0x00, // version 0 0x00, 0x00, 0x00, // flags baseMediaDecodeTime >> 24, baseMediaDecodeTime >> 16 & 0XFF, baseMediaDecodeTime >> 8 & 0XFF, baseMediaDecodeTime & 0xFF])), // baseMediaDecodeTime MP4.trun(track, sampleDependencyTable.length + 16 + // tfhd 16 + // tfdt 8 + // traf header 16 + // mfhd 8 + // moof header 8), // mdat header sampleDependencyTable); } /** * Generate a track box. * @param track {object} a track definition * @return {Uint8Array} the track box */ }, { key: 'trak', value: function trak(track) { track.duration = track.duration || 0xffffffff; return MP4.box(MP4.types.trak, MP4.tkhd(track), MP4.mdia(track)); } }, { key: 'trex', value: function trex(track) { return MP4.box(MP4.types.trex, new Uint8Array([0x00, // version 0 0x00, 0x00, 0x00, // flags track.id >> 24, track.id >> 16 & 0XFF, track.id >> 8 & 0XFF, track.id & 0xFF, // track_ID 0x00, 0x00, 0x00, 0x01, // default_sample_description_index 0x00, 0x00, 0x00, 0x00, // default_sample_duration 0x00, 0x00, 0x00, 0x00, // default_sample_size 0x00, 0x01, 0x00, 0x01 // default_sample_flags ])); } }, { key: 'trun', value: function trun(track, offset) { var samples, sample, i, array; samples = track.samples || []; array = new Uint8Array(12 + 16 * samples.length); offset += 8 + array.byteLength; array.set([0x00, // version 0 0x00, 0x0f, 0x01, // flags samples.length >>> 24 & 0xFF, samples.length >>> 16 & 0xFF, samples.length >>> 8 & 0xFF, samples.length & 0xFF, // sample_count offset >>> 24 & 0xFF, offset >>> 16 & 0xFF, offset >>> 8 & 0xFF, offset & 0xFF // data_offset ], 0); for (i = 0; i < samples.length; i++) { sample = samples[i]; array.set([sample.duration >>> 24 & 0xFF, sample.duration >>> 16 & 0xFF, sample.duration >>> 8 & 0xFF, sample.duration & 0xFF, // sample_duration sample.size >>> 24 & 0xFF, sample.size >>> 16 & 0xFF, sample.size >>> 8 & 0xFF, sample.size & 0xFF, // sample_size sample.flags.isLeading << 2 | sample.flags.dependsOn, sample.flags.isDependedOn << 6 | sample.flags.hasRedundancy << 4 | sample.flags.paddingValue << 1 | sample.flags.isNonSync, sample.flags.degradPrio & 0xF0 << 8, sample.flags.degradPrio & 0x0F, // sample_flags sample.cts >>> 24 & 0xFF, sample.cts >>> 16 & 0xFF, sample.cts >>> 8 & 0xFF, sample.cts & 0xFF // sample_composition_time_offset ], 12 + 16 * i); } return MP4.box(MP4.types.trun, array); } }, { key: 'initSegment', value: function initSegment(tracks) { if (!MP4.types) { MP4.init(); } var movie = MP4.moov(tracks), result; result = new Uint8Array(MP4.FTYP.byteLength + movie.byteLength); result.set(MP4.FTYP); result.set(movie, MP4.FTYP.byteLength); return result; } }]); return MP4; })(); exports['default'] = MP4; module.exports = exports['default']; },{}],19:[function(require,module,exports){ /** * fMP4 remuxer */ 'use strict'; Object.defineProperty(exports, '__esModule', { value: true }); var _createClass = (function () { function defineProperties(target, props) { for (var i = 0; i < props.length; i++) { var descriptor = props[i]; descriptor.enumerable = descriptor.enumerable || false; descriptor.configurable = true; if ('value' in descriptor) descriptor.writable = true; Object.defineProperty(target, descriptor.key, descriptor); } } return function (Constructor, protoProps, staticProps) { if (protoProps) defineProperties(Constructor.prototype, protoProps); if (staticProps) defineProperties(Constructor, staticProps); return Constructor; }; })(); function _interopRequireDefault(obj) { return obj && obj.__esModule ? obj : { 'default': obj }; } function _classCallCheck(instance, Constructor) { if (!(instance instanceof Constructor)) { throw new TypeError('Cannot call a class as a function'); } } var _events = require('../events'); var _events2 = _interopRequireDefault(_events); var _utilsLogger = require('../utils/logger'); var _remuxMp4Generator = require('../remux/mp4-generator'); var _remuxMp4Generator2 = _interopRequireDefault(_remuxMp4Generator); var _errors = require('../errors'); var MP4Remuxer = (function () { function MP4Remuxer(observer) { _classCallCheck(this, MP4Remuxer); this.observer = observer; this.ISGenerated = false; this.PES2MP4SCALEFACTOR = 4; this.PES_TIMESCALE = 90000; this.MP4_TIMESCALE = this.PES_TIMESCALE / this.PES2MP4SCALEFACTOR; } _createClass(MP4Remuxer, [{ key: 'destroy', value: function destroy() {} }, { key: 'insertDiscontinuity', value: function insertDiscontinuity() { this._initPTS = this._initDTS = this.nextAacPts = this.nextAvcDts = undefined; } }, { key: 'switchLevel', value: function switchLevel() { this.ISGenerated = false; } }, { key: 'remux', value: function remux(audioTrack, videoTrack, id3Track, timeOffset, contiguous) { // generate Init Segment if needed if (!this.ISGenerated) { this.generateIS(audioTrack, videoTrack, timeOffset); } //logger.log('nb AVC samples:' + videoTrack.samples.length); if (videoTrack.samples.length) { this.remuxVideo(videoTrack, timeOffset, contiguous); } //logger.log('nb AAC samples:' + audioTrack.samples.length); if (audioTrack.samples.length) { this.remuxAudio(audioTrack, timeOffset, contiguous); } //logger.log('nb ID3 samples:' + audioTrack.samples.length); if (id3Track.samples.length) { this.remuxID3(id3Track, timeOffset); } //notify end of parsing this.observer.trigger(_events2['default'].FRAG_PARSED); } }, { key: 'generateIS', value: function generateIS(audioTrack, videoTrack, timeOffset) { var observer = this.observer, audioSamples = audioTrack.samples, videoSamples = videoTrack.samples, nbAudio = audioSamples.length, nbVideo = videoSamples.length, pesTimeScale = this.PES_TIMESCALE; if (nbAudio === 0 && nbVideo === 0) { observer.trigger(_events2['default'].ERROR, { type: _errors.ErrorTypes.MEDIA_ERROR, details: _errors.ErrorDetails.FRAG_PARSING_ERROR, fatal: false, reason: 'no audio/video samples found' }); } else if (nbVideo === 0) { //audio only if (audioTrack.config) { observer.trigger(_events2['default'].FRAG_PARSING_INIT_SEGMENT, { audioMoov: _remuxMp4Generator2['default'].initSegment([audioTrack]), audioCodec: audioTrack.codec, audioChannelCount: audioTrack.channelCount }); this.ISGenerated = true; } if (this._initPTS === undefined) { // remember first PTS of this demuxing context this._initPTS = audioSamples[0].pts - pesTimeScale * timeOffset; this._initDTS = audioSamples[0].dts - pesTimeScale * timeOffset; } } else if (nbAudio === 0) { //video only if (videoTrack.sps && videoTrack.pps) { observer.trigger(_events2['default'].FRAG_PARSING_INIT_SEGMENT, { videoMoov: _remuxMp4Generator2['default'].initSegment([videoTrack]), videoCodec: videoTrack.codec, videoWidth: videoTrack.width, videoHeight: videoTrack.height }); this.ISGenerated = true; if (this._initPTS === undefined) { // remember first PTS of this demuxing context this._initPTS = videoSamples[0].pts - pesTimeScale * timeOffset; this._initDTS = videoSamples[0].dts - pesTimeScale * timeOffset; } } } else { //audio and video if (audioTrack.config && videoTrack.sps && videoTrack.pps) { observer.trigger(_events2['default'].FRAG_PARSING_INIT_SEGMENT, { audioMoov: _remuxMp4Generator2['default'].initSegment([audioTrack]), audioCodec: audioTrack.codec, audioChannelCount: audioTrack.channelCount, videoMoov: _remuxMp4Generator2['default'].initSegment([videoTrack]), videoCodec: videoTrack.codec, videoWidth: videoTrack.width, videoHeight: videoTrack.height }); this.ISGenerated = true; if (this._initPTS === undefined) { // remember first PTS of this demuxing context this._initPTS = Math.min(videoSamples[0].pts, audioSamples[0].pts) - pesTimeScale * timeOffset; this._initDTS = Math.min(videoSamples[0].dts, audioSamples[0].dts) - pesTimeScale * timeOffset; } } } } }, { key: 'remuxVideo', value: function remuxVideo(track, timeOffset, contiguous) { var view, i = 8, pesTimeScale = this.PES_TIMESCALE, pes2mp4ScaleFactor = this.PES2MP4SCALEFACTOR, avcSample, mp4Sample, mp4SampleLength, unit, mdat, moof, firstPTS, firstDTS, lastDTS, pts, dts, ptsnorm, dtsnorm, samples = []; /* concatenate the video data and construct the mdat in place (need 8 more bytes to fill length and mpdat type) */ mdat = new Uint8Array(track.len + 4 * track.nbNalu + 8); view = new DataView(mdat.buffer); view.setUint32(0, mdat.byteLength); mdat.set(_remuxMp4Generator2['default'].types.mdat, 4); while (track.samples.length) { avcSample = track.samples.shift(); mp4SampleLength = 0; // convert NALU bitstream to MP4 format (prepend NALU with size field) while (avcSample.units.units.length) { unit = avcSample.units.units.shift(); view.setUint32(i, unit.data.byteLength); i += 4; mdat.set(unit.data, i); i += unit.data.byteLength; mp4SampleLength += 4 + unit.data.byteLength; } pts = avcSample.pts - this._initDTS; dts = avcSample.dts - this._initDTS; //logger.log('Video/PTS/DTS:' + pts + '/' + dts); // if not first AVC sample of video track, normalize PTS/DTS with previous sample value // and ensure that sample duration is positive if (lastDTS !== undefined) { ptsnorm = this._PTSNormalize(pts, lastDTS); dtsnorm = this._PTSNormalize(dts, lastDTS); mp4Sample.duration = (dtsnorm - lastDTS) / pes2mp4ScaleFactor; if (mp4Sample.duration < 0) { //logger.log('invalid sample duration at PTS/DTS::' + avcSample.pts + '/' + avcSample.dts + ':' + mp4Sample.duration); mp4Sample.duration = 0; } } else { var nextAvcDts = this.nextAvcDts, delta; // first AVC sample of video track, normalize PTS/DTS ptsnorm = this._PTSNormalize(pts, nextAvcDts); dtsnorm = this._PTSNormalize(dts, nextAvcDts); delta = Math.round((dtsnorm - nextAvcDts) / 90); // if fragment are contiguous, or delta less than 600ms, ensure there is no overlap/hole between fragments if (contiguous || Math.abs(delta) < 600) { if (delta) { if (delta > 1) { _utilsLogger.logger.log('AVC:' + delta + ' ms hole between fragments detected,filling it'); } else if (delta < -1) { _utilsLogger.logger.log('AVC:' + -delta + ' ms overlapping between fragments detected'); } // set DTS to next DTS dtsnorm = nextAvcDts; // offset PTS as well, ensure that PTS is smaller or equal than new DTS ptsnorm = Math.max(ptsnorm - delta, dtsnorm); _utilsLogger.logger.log('Video/PTS/DTS adjusted:' + ptsnorm + '/' + dtsnorm); } } // remember first PTS of our avcSamples, ensure value is positive firstPTS = Math.max(0, ptsnorm); firstDTS = Math.max(0, dtsnorm); } //console.log('PTS/DTS/initDTS/normPTS/normDTS/relative PTS : ${avcSample.pts}/${avcSample.dts}/${this._initDTS}/${ptsnorm}/${dtsnorm}/${(avcSample.pts/4294967296).toFixed(3)}'); mp4Sample = { size: mp4SampleLength, duration: 0, cts: (ptsnorm - dtsnorm) / pes2mp4ScaleFactor, flags: { isLeading: 0, isDependedOn: 0, hasRedundancy: 0, degradPrio: 0 } }; if (avcSample.key === true) { // the current sample is a key frame mp4Sample.flags.dependsOn = 2; mp4Sample.flags.isNonSync = 0; } else { mp4Sample.flags.dependsOn = 1; mp4Sample.flags.isNonSync = 1; } samples.push(mp4Sample); lastDTS = dtsnorm; } if (samples.length >= 2) { mp4Sample.duration = samples[samples.length - 2].duration; } // next AVC sample DTS should be equal to last sample DTS + last sample duration this.nextAvcDts = dtsnorm + mp4Sample.duration * pes2mp4ScaleFactor; track.len = 0; track.nbNalu = 0; if (navigator.userAgent.toLowerCase().indexOf('chrome') > -1) { // chrome workaround, mark first sample as being a Random Access Point to avoid sourcebuffer append issue // https://code.google.com/p/chromium/issues/detail?id=229412 samples[0].flags.dependsOn = 2; samples[0].flags.isNonSync = 0; } track.samples = samples; moof = _remuxMp4Generator2['default'].moof(track.sequenceNumber++, firstDTS / pes2mp4ScaleFactor, track); track.samples = []; this.observer.trigger(_events2['default'].FRAG_PARSING_DATA, { moof: moof, mdat: mdat, startPTS: firstPTS / pesTimeScale, endPTS: (ptsnorm + pes2mp4ScaleFactor * mp4Sample.duration) / pesTimeScale, startDTS: firstDTS / pesTimeScale, endDTS: (dtsnorm + pes2mp4ScaleFactor * mp4Sample.duration) / pesTimeScale, type: 'video', nb: samples.length }); } }, { key: 'remuxAudio', value: function remuxAudio(track, timeOffset, contiguous) { var view, i = 8, pesTimeScale = this.PES_TIMESCALE, pes2mp4ScaleFactor = this.PES2MP4SCALEFACTOR, aacSample, mp4Sample, unit, mdat, moof, firstPTS, firstDTS, lastDTS, pts, dts, ptsnorm, dtsnorm, samples = []; /* concatenate the audio data and construct the mdat in place (need 8 more bytes to fill length and mdat type) */ mdat = new Uint8Array(track.len + 8); view = new DataView(mdat.buffer); view.setUint32(0, mdat.byteLength); mdat.set(_remuxMp4Generator2['default'].types.mdat, 4); while (track.samples.length) { aacSample = track.samples.shift(); unit = aacSample.unit; mdat.set(unit, i); i += unit.byteLength; pts = aacSample.pts - this._initDTS; dts = aacSample.dts - this._initDTS; //logger.log('Audio/PTS:' + aacSample.pts.toFixed(0)); if (lastDTS !== undefined) { ptsnorm = this._PTSNormalize(pts, lastDTS); dtsnorm = this._PTSNormalize(dts, lastDTS); // we use DTS to compute sample duration, but we use PTS to compute initPTS which is used to sync audio and video mp4Sample.duration = (dtsnorm - lastDTS) / pes2mp4ScaleFactor; if (mp4Sample.duration < 0) { _utilsLogger.logger.log('invalid AAC sample duration at PTS:' + aacSample.pts + ':' + mp4Sample.duration); mp4Sample.duration = 0; } } else { var nextAacPts = this.nextAacPts, delta; ptsnorm = this._PTSNormalize(pts, nextAacPts); dtsnorm = this._PTSNormalize(dts, nextAacPts); delta = Math.round(1000 * (ptsnorm - nextAacPts) / pesTimeScale); // if fragment are contiguous, or delta less than 600ms, ensure there is no overlap/hole between fragments if (contiguous || Math.abs(delta) < 600) { // log delta if (delta) { if (delta > 1) { _utilsLogger.logger.log(delta + ' ms hole between AAC samples detected,filling it'); // set PTS to next PTS, and ensure PTS is greater or equal than last DTS } else if (delta < -1) { _utilsLogger.logger.log(-delta + ' ms overlapping between AAC samples detected'); } // set DTS to next DTS ptsnorm = dtsnorm = nextAacPts; } } // remember first PTS of our aacSamples, ensure value is positive firstPTS = Math.max(0, ptsnorm); firstDTS = Math.max(0, dtsnorm); } //console.log('PTS/DTS/initDTS/normPTS/normDTS/relative PTS : ${aacSample.pts}/${aacSample.dts}/${this._initDTS}/${ptsnorm}/${dtsnorm}/${(aacSample.pts/4294967296).toFixed(3)}'); mp4Sample = { size: unit.byteLength, cts: 0, duration: 0, flags: { isLeading: 0, isDependedOn: 0, hasRedundancy: 0, degradPrio: 0, dependsOn: 1 } }; samples.push(mp4Sample); lastDTS = dtsnorm; } //set last sample duration as being identical to previous sample if (samples.length >= 2) { mp4Sample.duration = samples[samples.length - 2].duration; } // next aac sample PTS should be equal to last sample PTS + duration this.nextAacPts = ptsnorm + pes2mp4ScaleFactor * mp4Sample.duration; //logger.log('Audio/PTS/PTSend:' + aacSample.pts.toFixed(0) + '/' + this.nextAacDts.toFixed(0)); track.len = 0; track.samples = samples; moof = _remuxMp4Generator2['default'].moof(track.sequenceNumber++, firstDTS / pes2mp4ScaleFactor, track); track.samples = []; this.observer.trigger(_events2['default'].FRAG_PARSING_DATA, { moof: moof, mdat: mdat, startPTS: firstPTS / pesTimeScale, endPTS: this.nextAacPts / pesTimeScale, startDTS: firstDTS / pesTimeScale, endDTS: (dtsnorm + pes2mp4ScaleFactor * mp4Sample.duration) / pesTimeScale, type: 'audio', nb: samples.length }); } }, { key: 'remuxID3', value: function remuxID3(track, timeOffset) { var length = track.samples.length, sample; // consume samples if (length) { for (var index = 0; index < length; index++) { sample = track.samples[index]; // setting id3 pts, dts to relative time // using this._initPTS and this._initDTS to calculate relative time sample.pts = (sample.pts - this._initPTS) / this.PES_TIMESCALE; sample.dts = (sample.dts - this._initDTS) / this.PES_TIMESCALE; } this.observer.trigger(_events2['default'].FRAG_PARSING_METADATA, { samples: track.samples }); } track.samples = []; timeOffset = timeOffset; } }, { key: '_PTSNormalize', value: function _PTSNormalize(value, reference) { var offset; if (reference === undefined) { return value; } if (reference < value) { // - 2^33 offset = -8589934592; } else { // + 2^33 offset = 8589934592; } /* PTS is 33bit (from 0 to 2^33 -1) if diff between value and reference is bigger than half of the amplitude (2^32) then it means that PTS looping occured. fill the gap */ while (Math.abs(value - reference) > 4294967296) { value += offset; } return value; } }, { key: 'timescale', get: function get() { return this.MP4_TIMESCALE; } }]); return MP4Remuxer; })(); exports['default'] = MP4Remuxer; module.exports = exports['default']; },{"../errors":11,"../events":12,"../remux/mp4-generator":18,"../utils/logger":21}],20:[function(require,module,exports){ "use strict"; var BinarySearch = { /** * Searches for an item in an array which matches a certain condition. * This requires the condition to only match one item in the array, * and for the array to be ordered. * * @param {Array} list The array to search. * @param {Function} comparisonFunction * Called and provided a candidate item as the first argument. * Should return: * > -1 if the item should be located at a lower index than the provided item. * > 1 if the item should be located at a higher index than the provided item. * > 0 if the item is the item you're looking for. * * @return {*} The object if it is found or null otherwise. */ search: function search(list, comparisonFunction) { var minIndex = 0; var maxIndex = list.length - 1; var currentIndex = null; var currentElement = null; while (minIndex <= maxIndex) { currentIndex = (minIndex + maxIndex) / 2 | 0; currentElement = list[currentIndex]; var comparisonResult = comparisonFunction(currentElement); if (comparisonResult > 0) { minIndex = currentIndex + 1; } else if (comparisonResult < 0) { maxIndex = currentIndex - 1; } else { return currentElement; } } return null; } }; module.exports = BinarySearch; },{}],21:[function(require,module,exports){ 'use strict'; Object.defineProperty(exports, '__esModule', { value: true }); function noop() {} var fakeLogger = { log: noop, warn: noop, info: noop, error: noop }; var exportedLogger = fakeLogger; var enableLogs = function enableLogs(debug) { if (debug === true || typeof debug === 'object') { exportedLogger.log = debug.log ? debug.log.bind(debug) : console.log.bind(console); exportedLogger.info = debug.info ? debug.info.bind(debug) : console.info.bind(console); exportedLogger.error = debug.error ? debug.error.bind(debug) : console.error.bind(console); exportedLogger.warn = debug.warn ? debug.warn.bind(debug) : console.warn.bind(console); // Some browsers don't allow to use bind on console object anyway // fallback to default if needed try { exportedLogger.log(); } catch (e) { exportedLogger.log = noop; exportedLogger.info = noop; exportedLogger.error = noop; exportedLogger.warn = noop; } } else { exportedLogger = fakeLogger; } }; exports.enableLogs = enableLogs; var logger = exportedLogger; exports.logger = logger; },{}],22:[function(require,module,exports){ 'use strict'; var URLHelper = { // build an absolute URL from a relative one using the provided baseURL // if relativeURL is an absolute URL it will be returned as is. buildAbsoluteURL: function buildAbsoluteURL(baseURL, relativeURL) { // remove any remaining space and CRLF relativeURL = relativeURL.trim(); if (/^[a-z]+:/i.test(relativeURL)) { // complete url, not relative return relativeURL; } var relativeURLQuery = null; var relativeURLHash = null; var relativeURLHashSplit = /^([^#]*)(.*)$/.exec(relativeURL); if (relativeURLHashSplit) { relativeURLHash = relativeURLHashSplit[2]; relativeURL = relativeURLHashSplit[1]; } var relativeURLQuerySplit = /^([^\?]*)(.*)$/.exec(relativeURL); if (relativeURLQuerySplit) { relativeURLQuery = relativeURLQuerySplit[2]; relativeURL = relativeURLQuerySplit[1]; } var baseURLHashSplit = /^([^#]*)(.*)$/.exec(baseURL); if (baseURLHashSplit) { baseURL = baseURLHashSplit[1]; } var baseURLQuerySplit = /^([^\?]*)(.*)$/.exec(baseURL); if (baseURLQuerySplit) { baseURL = baseURLQuerySplit[1]; } var baseURLDomainSplit = /^((([a-z]+):)?\/\/[a-z0-9\.-]+(:[0-9]+)?\/)(.*)$/i.exec(baseURL); var baseURLProtocol = baseURLDomainSplit[3]; var baseURLDomain = baseURLDomainSplit[1]; var baseURLPath = baseURLDomainSplit[5]; var builtURL = null; if (/^\/\//.test(relativeURL)) { builtURL = baseURLProtocol + '://' + URLHelper.buildAbsolutePath('', relativeURL.substring(2)); } else if (/^\//.test(relativeURL)) { builtURL = baseURLDomain + URLHelper.buildAbsolutePath('', relativeURL.substring(1)); } else { var newPath = URLHelper.buildAbsolutePath(baseURLPath, relativeURL); builtURL = baseURLDomain + newPath; } // put the query and hash parts back if (relativeURLQuery) { builtURL += relativeURLQuery; } if (relativeURLHash) { builtURL += relativeURLHash; } return builtURL; }, // build an absolute path using the provided basePath // adapted from https://developer.mozilla.org/en-US/docs/Web/API/document/cookie#Using_relative_URLs_in_the_path_parameter // this does not handle the case where relativePath is "/" or "//". These cases should be handled outside this. buildAbsolutePath: function buildAbsolutePath(basePath, relativePath) { var sRelPath = relativePath; var nUpLn, sDir = '', sPath = basePath.replace(/[^\/]*$/, sRelPath.replace(/(\/|^)(?:\.?\/+)+/g, '$1')); for (var nEnd, nStart = 0; nEnd = sPath.indexOf('/../', nStart), nEnd > -1; nStart = nEnd + nUpLn) { nUpLn = /^\/(?:\.\.\/)*/.exec(sPath.slice(nEnd))[0].length; sDir = (sDir + sPath.substring(nStart, nEnd)).replace(new RegExp('(?:\\\/+[^\\\/]*){0,' + (nUpLn - 1) / 3 + '}$'), '/'); } return sDir + sPath.substr(nStart); } }; module.exports = URLHelper; },{}],23:[function(require,module,exports){ /** * XHR based logger */ 'use strict'; Object.defineProperty(exports, '__esModule', { value: true }); var _createClass = (function () { function defineProperties(target, props) { for (var i = 0; i < props.length; i++) { var descriptor = props[i]; descriptor.enumerable = descriptor.enumerable || false; descriptor.configurable = true; if ('value' in descriptor) descriptor.writable = true; Object.defineProperty(target, descriptor.key, descriptor); } } return function (Constructor, protoProps, staticProps) { if (protoProps) defineProperties(Constructor.prototype, protoProps); if (staticProps) defineProperties(Constructor, staticProps); return Constructor; }; })(); function _classCallCheck(instance, Constructor) { if (!(instance instanceof Constructor)) { throw new TypeError('Cannot call a class as a function'); } } var _utilsLogger = require('../utils/logger'); var XhrLoader = (function () { function XhrLoader(config) { _classCallCheck(this, XhrLoader); if (config && config.xhrSetup) { this.xhrSetup = config.xhrSetup; } } _createClass(XhrLoader, [{ key: 'destroy', value: function destroy() { this.abort(); this.loader = null; } }, { key: 'abort', value: function abort() { if (this.loader && this.loader.readyState !== 4) { this.stats.aborted = true; this.loader.abort(); } if (this.timeoutHandle) { window.clearTimeout(this.timeoutHandle); } } }, { key: 'load', value: function load(url, responseType, onSuccess, onError, onTimeout, timeout, maxRetry, retryDelay) { var onProgress = arguments.length <= 8 || arguments[8] === undefined ? null : arguments[8]; var frag = arguments.length <= 9 || arguments[9] === undefined ? null : arguments[9]; this.url = url; if (frag && !isNaN(frag.byteRangeStartOffset) && !isNaN(frag.byteRangeEndOffset)) { this.byteRange = frag.byteRangeStartOffset + '-' + frag.byteRangeEndOffset; } this.responseType = responseType; this.onSuccess = onSuccess; this.onProgress = onProgress; this.onTimeout = onTimeout; this.onError = onError; this.stats = { trequest: performance.now(), retry: 0 }; this.timeout = timeout; this.maxRetry = maxRetry; this.retryDelay = retryDelay; this.timeoutHandle = window.setTimeout(this.loadtimeout.bind(this), timeout); this.loadInternal(); } }, { key: 'loadInternal', value: function loadInternal() { var xhr = this.loader = new XMLHttpRequest(); xhr.onload = this.loadsuccess.bind(this); xhr.onerror = this.loaderror.bind(this); xhr.onprogress = this.loadprogress.bind(this); xhr.open('GET', this.url, true); if (this.byteRange) { xhr.setRequestHeader('Range', 'bytes=' + this.byteRange); } xhr.responseType = this.responseType; this.stats.tfirst = null; this.stats.loaded = 0; if (this.xhrSetup) { this.xhrSetup(xhr); } xhr.send(); } }, { key: 'loadsuccess', value: function loadsuccess(event) { window.clearTimeout(this.timeoutHandle); this.stats.tload = performance.now(); this.onSuccess(event, this.stats); } }, { key: 'loaderror', value: function loaderror(event) { if (this.stats.retry < this.maxRetry) { _utilsLogger.logger.warn(event.type + ' while loading ' + this.url + ', retrying in ' + this.retryDelay + '...'); this.destroy(); window.setTimeout(this.loadInternal.bind(this), this.retryDelay); // exponential backoff this.retryDelay = Math.min(2 * this.retryDelay, 64000); this.stats.retry++; } else { window.clearTimeout(this.timeoutHandle); _utilsLogger.logger.error(event.type + ' while loading ' + this.url); this.onError(event); } } }, { key: 'loadtimeout', value: function loadtimeout(event) { _utilsLogger.logger.warn('timeout while loading ' + this.url); this.onTimeout(event, this.stats); } }, { key: 'loadprogress', value: function loadprogress(event) { var stats = this.stats; if (stats.tfirst === null) { stats.tfirst = performance.now(); } stats.loaded = event.loaded; if (this.onProgress) { this.onProgress(event, stats); } } }]); return XhrLoader; })(); exports['default'] = XhrLoader; module.exports = exports['default']; },{"../utils/logger":21}]},{},[14])(14) }); //# sourceMappingURL=hls.js.map