Bug 1148102: P6. Add WebMDemuxer object. r=jya

This commit is contained in:
Jan Gerber 2015-06-18 15:52:09 +02:00 коммит произвёл Jean-Yves Avenard
Родитель 02ddfcec8c
Коммит 6946238e5e
5 изменённых файлов: 1279 добавлений и 0 удалений

Просмотреть файл

@ -346,6 +346,50 @@ void WebMBufferedState::NotifyDataArrived(const unsigned char* aBuffer, uint32_t
}
}
void WebMBufferedState::Reset() {
mRangeParsers.Clear();
mTimeMapping.Clear();
}
void WebMBufferedState::UpdateIndex(const nsTArray<MediaByteRange>& aRanges, MediaResource* aResource)
{
for (uint32_t index = 0; index < aRanges.Length(); index++) {
const MediaByteRange& range = aRanges[index];
int64_t offset = range.mStart;
uint32_t length = range.mEnd - range.mStart;
uint32_t idx = mRangeParsers.IndexOfFirstElementGt(offset - 1);
if (!idx || !(mRangeParsers[idx-1] == offset)) {
// If the incoming data overlaps an already parsed range, adjust the
// buffer so that we only reparse the new data. It's also possible to
// have an overlap where the end of the incoming data is within an
// already parsed range, but we don't bother handling that other than by
// avoiding storing duplicate timecodes when the parser runs.
if (idx != mRangeParsers.Length() && mRangeParsers[idx].mStartOffset <= offset) {
// Complete overlap, skip parsing.
if (offset + length <= mRangeParsers[idx].mCurrentOffset) {
continue;
}
// Partial overlap, adjust the buffer to parse only the new data.
int64_t adjust = mRangeParsers[idx].mCurrentOffset - offset;
NS_ASSERTION(adjust >= 0, "Overlap detection bug.");
offset += adjust;
length -= uint32_t(adjust);
} else {
mRangeParsers.InsertElementAt(idx, WebMBufferedParser(offset));
if (idx) {
mRangeParsers[idx].SetTimecodeScale(mRangeParsers[0].GetTimecodeScale());
}
}
}
nsRefPtr<MediaByteBuffer> bytes = aResource->MediaReadAt(offset, length);
if(bytes) {
NotifyDataArrived(bytes->Elements(), bytes->Length(), offset);
}
}
}
int64_t WebMBufferedState::GetInitEndOffset()
{
if (mRangeParsers.IsEmpty()) {
@ -354,4 +398,37 @@ int64_t WebMBufferedState::GetInitEndOffset()
return mRangeParsers[0].mInitEndOffset;
}
bool WebMBufferedState::GetStartTime(uint64_t *aTime)
{
ReentrantMonitorAutoEnter mon(mReentrantMonitor);
if (mTimeMapping.IsEmpty()) {
return false;
}
uint32_t idx = mTimeMapping.IndexOfFirstElementGt(0, SyncOffsetComparator());
if (idx == mTimeMapping.Length()) {
return false;
}
*aTime = mTimeMapping[idx].mTimecode;
return true;
}
bool
WebMBufferedState::GetNextKeyframeTime(uint64_t aTime, uint64_t* aKeyframeTime)
{
ReentrantMonitorAutoEnter mon(mReentrantMonitor);
int64_t offset = 0;
bool rv = GetOffsetForTime(aTime, &offset);
if (!rv) {
return false;
}
uint32_t idx = mTimeMapping.IndexOfFirstElementGt(offset, SyncOffsetComparator());
if (idx == mTimeMapping.Length()) {
return false;
}
*aKeyframeTime = mTimeMapping[idx].mTimecode;
return true;
}
} // namespace mozilla

Просмотреть файл

@ -229,6 +229,8 @@ public:
}
void NotifyDataArrived(const unsigned char* aBuffer, uint32_t aLength, int64_t aOffset);
void Reset();
void UpdateIndex(const nsTArray<MediaByteRange>& aRanges, MediaResource* aResource);
bool CalculateBufferedForRange(int64_t aStartOffset, int64_t aEndOffset,
uint64_t* aStartTime, uint64_t* aEndTime);
@ -240,6 +242,12 @@ public:
// Returns end offset of init segment or -1 if none found.
int64_t GetInitEndOffset();
// Returns start time
bool GetStartTime(uint64_t *aTime);
// Returns keyframe for time
bool GetNextKeyframeTime(uint64_t aTime, uint64_t* aKeyframeTime);
private:
// Private destructor, to discourage deletion outside of Release():
~WebMBufferedState() {

Просмотреть файл

@ -0,0 +1,985 @@
/* -*- Mode: C++; tab-width: 2; indent-tabs-mode: nil; c-basic-offset: 2 -*- */
/* vim:set ts=2 sw=2 sts=2 et cindent: */
/* This Source Code Form is subject to the terms of the Mozilla Public
* License, v. 2.0. If a copy of the MPL was not distributed with this
* file, You can obtain one at http://mozilla.org/MPL/2.0/. */
#include "nsError.h"
#include "MediaDecoderStateMachine.h"
#include "AbstractMediaDecoder.h"
#include "MediaResource.h"
#include "WebMDemuxer.h"
#include "WebMBufferedParser.h"
#include "gfx2DGlue.h"
#include "mozilla/Preferences.h"
#include "SharedThreadPool.h"
#include "MediaDataDemuxer.h"
#include "nsAutoRef.h"
#include "NesteggPacketHolder.h"
#include <algorithm>
#define VPX_DONT_DEFINE_STDINT_TYPES
#include "vpx/vp8dx.h"
#include "vpx/vpx_decoder.h"
#define WEBM_DEBUG(arg, ...) MOZ_LOG(gMediaDecoderLog, mozilla::LogLevel::Debug, ("WebMDemuxer(%p)::%s: " arg, this, __func__, ##__VA_ARGS__))
namespace mozilla {
using namespace gfx;
extern PRLogModuleInfo* gMediaDecoderLog;
extern PRLogModuleInfo* gNesteggLog;
// Functions for reading and seeking using WebMDemuxer required for
// nestegg_io. The 'user data' passed to these functions is the demuxer
static int webmdemux_read(void* aBuffer, size_t aLength, void* aUserData)
{
MOZ_ASSERT(aUserData);
WebMDemuxer* demuxer = reinterpret_cast<WebMDemuxer*>(aUserData);
uint32_t bytes = 0;
bool eof = false;
char* p = static_cast<char*>(aBuffer);
nsresult rv = demuxer->Read(p, aLength, &bytes);
eof = bytes < aLength;
return NS_FAILED(rv) ? -1 : eof ? 0 : 1;
}
static int webmdemux_seek(int64_t aOffset, int aWhence, void* aUserData)
{
MOZ_ASSERT(aUserData);
WebMDemuxer* demuxer = reinterpret_cast<WebMDemuxer*>(aUserData);
nsresult rv = demuxer->Seek(aWhence, aOffset);
return NS_SUCCEEDED(rv) ? 0 : -1;
}
static int64_t webmdemux_tell(void* aUserData)
{
MOZ_ASSERT(aUserData);
WebMDemuxer* demuxer = reinterpret_cast<WebMDemuxer*>(aUserData);
return demuxer->Tell();
}
static void webmdemux_log(nestegg* aContext,
unsigned int aSeverity,
char const* aFormat, ...)
{
if (!MOZ_LOG_TEST(gNesteggLog, LogLevel::Debug)) {
return;
}
va_list args;
char msg[256];
const char* sevStr;
switch(aSeverity) {
case NESTEGG_LOG_DEBUG:
sevStr = "DBG";
break;
case NESTEGG_LOG_INFO:
sevStr = "INF";
break;
case NESTEGG_LOG_WARNING:
sevStr = "WRN";
break;
case NESTEGG_LOG_ERROR:
sevStr = "ERR";
break;
case NESTEGG_LOG_CRITICAL:
sevStr = "CRT";
break;
default:
sevStr = "UNK";
break;
}
va_start(args, aFormat);
PR_snprintf(msg, sizeof(msg), "%p [Nestegg-%s] ", aContext, sevStr);
PR_vsnprintf(msg+strlen(msg), sizeof(msg)-strlen(msg), aFormat, args);
MOZ_LOG(gNesteggLog, LogLevel::Debug, (msg));
va_end(args);
}
WebMDemuxer::WebMDemuxer(MediaResource* aResource)
: mResource(aResource)
, mBufferedState(nullptr)
, mInitData(nullptr)
, mContext(nullptr)
, mOffset(0)
, mVideoTrack(0)
, mAudioTrack(0)
, mSeekPreroll(0)
, mLastVideoFrameTime(0)
, mAudioCodec(-1)
, mVideoCodec(-1)
, mHasVideo(false)
, mHasAudio(false)
, mNeedReIndex(true)
{
if (!gNesteggLog) {
gNesteggLog = PR_NewLogModule("Nestegg");
}
}
WebMDemuxer::~WebMDemuxer()
{
Reset();
Cleanup();
}
nsRefPtr<WebMDemuxer::InitPromise>
WebMDemuxer::Init()
{
if (InitBufferedState() != NS_OK) {
return InitPromise::CreateAndReject(DemuxerFailureReason::WAITING_FOR_DATA, __func__);
}
if (ReadMetadata() != NS_OK) {
return InitPromise::CreateAndReject(DemuxerFailureReason::DEMUXER_ERROR, __func__);
}
if (!GetNumberTracks(TrackInfo::kAudioTrack) &&
!GetNumberTracks(TrackInfo::kVideoTrack)) {
return InitPromise::CreateAndReject(DemuxerFailureReason::DEMUXER_ERROR, __func__);
}
return InitPromise::CreateAndResolve(NS_OK, __func__);
}
nsresult
WebMDemuxer::InitBufferedState()
{
if(!mBufferedState) {
mBufferedState = new WebMBufferedState;
}
EnsureUpToDateIndex();
return NS_OK;
}
already_AddRefed<MediaDataDemuxer>
WebMDemuxer::Clone() const
{
nsRefPtr<WebMDemuxer> demuxer = new WebMDemuxer(mResource);
demuxer->mInitData = mInitData;
if (demuxer->InitBufferedState() != NS_OK ||
demuxer->ReadMetadata() != NS_OK) {
NS_WARNING("Couldn't recreate WebMDemuxer");
return nullptr;
}
return demuxer.forget();
}
bool
WebMDemuxer::HasTrackType(TrackInfo::TrackType aType) const
{
return !!GetNumberTracks(aType);
}
uint32_t
WebMDemuxer::GetNumberTracks(TrackInfo::TrackType aType) const
{
switch(aType) {
case TrackInfo::kAudioTrack:
return mHasAudio ? 1 : 0;
case TrackInfo::kVideoTrack:
return mHasVideo ? 1 : 0;
default:
return 0;
}
}
UniquePtr<TrackInfo>
WebMDemuxer::GetTrackInfo(TrackInfo::TrackType aType,
size_t aTrackNumber) const
{
switch(aType) {
case TrackInfo::kAudioTrack:
return mInfo.mAudio.Clone();
case TrackInfo::kVideoTrack:
return mInfo.mVideo.Clone();
default:
return nullptr;
}
}
already_AddRefed<MediaTrackDemuxer>
WebMDemuxer::GetTrackDemuxer(TrackInfo::TrackType aType, uint32_t aTrackNumber)
{
if (GetNumberTracks(aType) <= aTrackNumber) {
return nullptr;
}
nsRefPtr<WebMTrackDemuxer> e =
new WebMTrackDemuxer(this, aType, aTrackNumber);
mDemuxers.AppendElement(e);
return e.forget();
}
nsresult
WebMDemuxer::Reset()
{
mVideoPackets.Reset();
mAudioPackets.Reset();
return NS_OK;
}
void
WebMDemuxer::Cleanup()
{
if (mContext) {
nestegg_destroy(mContext);
mContext = nullptr;
}
mBufferedState = nullptr;
}
nsresult
WebMDemuxer::ReadMetadata()
{
nestegg_io io;
io.read = webmdemux_read;
io.seek = webmdemux_seek;
io.tell = webmdemux_tell;
io.userdata = this;
int64_t maxOffset = mBufferedState->GetInitEndOffset();
if (maxOffset == -1) {
maxOffset = mResource->GetLength();
}
int r = nestegg_init(&mContext, io, &webmdemux_log, maxOffset);
if (r == -1) {
return NS_ERROR_FAILURE;
}
unsigned int ntracks = 0;
r = nestegg_track_count(mContext, &ntracks);
if (r == -1) {
return NS_ERROR_FAILURE;
}
for (unsigned int track = 0; track < ntracks; ++track) {
int id = nestegg_track_codec_id(mContext, track);
if (id == -1) {
return NS_ERROR_FAILURE;
}
int type = nestegg_track_type(mContext, track);
if (type == NESTEGG_TRACK_VIDEO) {
nestegg_video_params params;
r = nestegg_track_video_params(mContext, track, &params);
if (r == -1) {
return NS_ERROR_FAILURE;
}
mVideoCodec = nestegg_track_codec_id(mContext, track);
switch(mVideoCodec) {
case NESTEGG_CODEC_VP8:
mInfo.mVideo.mMimeType = "video/webm; codecs=vp8";
break;
case NESTEGG_CODEC_VP9:
mInfo.mVideo.mMimeType = "video/webm; codecs=vp9";
break;
default:
NS_WARNING("Unknown WebM video codec");
return NS_ERROR_FAILURE;
}
// Picture region, taking into account cropping, before scaling
// to the display size.
unsigned int cropH = params.crop_right + params.crop_left;
unsigned int cropV = params.crop_bottom + params.crop_top;
nsIntRect pictureRect(params.crop_left,
params.crop_top,
params.width - cropH,
params.height - cropV);
// If the cropping data appears invalid then use the frame data
if (pictureRect.width <= 0 ||
pictureRect.height <= 0 ||
pictureRect.x < 0 ||
pictureRect.y < 0) {
pictureRect.x = 0;
pictureRect.y = 0;
pictureRect.width = params.width;
pictureRect.height = params.height;
}
// Validate the container-reported frame and pictureRect sizes. This
// ensures that our video frame creation code doesn't overflow.
nsIntSize displaySize(params.display_width, params.display_height);
nsIntSize frameSize(params.width, params.height);
if (!IsValidVideoRegion(frameSize, pictureRect, displaySize)) {
// Video track's frame sizes will overflow. Ignore the video track.
continue;
}
mVideoTrack = track;
mHasVideo = true;
mInfo.mVideo.mDisplay = displaySize;
mPicture = pictureRect;
switch (params.stereo_mode) {
case NESTEGG_VIDEO_MONO:
mInfo.mVideo.mStereoMode = StereoMode::MONO;
break;
case NESTEGG_VIDEO_STEREO_LEFT_RIGHT:
mInfo.mVideo.mStereoMode = StereoMode::LEFT_RIGHT;
break;
case NESTEGG_VIDEO_STEREO_BOTTOM_TOP:
mInfo.mVideo.mStereoMode = StereoMode::BOTTOM_TOP;
break;
case NESTEGG_VIDEO_STEREO_TOP_BOTTOM:
mInfo.mVideo.mStereoMode = StereoMode::TOP_BOTTOM;
break;
case NESTEGG_VIDEO_STEREO_RIGHT_LEFT:
mInfo.mVideo.mStereoMode = StereoMode::RIGHT_LEFT;
break;
}
uint64_t duration = 0;
r = nestegg_duration(mContext, &duration);
if (!r) {
mInfo.mVideo.mDuration = media::TimeUnit::FromNanoseconds(duration).ToMicroseconds();
}
} else if (type == NESTEGG_TRACK_AUDIO) {
nestegg_audio_params params;
r = nestegg_track_audio_params(mContext, track, &params);
if (r == -1) {
return NS_ERROR_FAILURE;
}
mAudioTrack = track;
mHasAudio = true;
mCodecDelay = media::TimeUnit::FromNanoseconds(params.codec_delay).ToMicroseconds();
mAudioCodec = nestegg_track_codec_id(mContext, track);
if (mAudioCodec == NESTEGG_CODEC_VORBIS) {
mInfo.mAudio.mMimeType = "audio/ogg; codecs=vorbis";
} else if (mAudioCodec == NESTEGG_CODEC_OPUS) {
mInfo.mAudio.mMimeType = "audio/ogg; codecs=opus";
uint8_t c[sizeof(uint64_t)];
BigEndian::writeUint64(&c[0], mCodecDelay);
mInfo.mAudio.mCodecSpecificConfig->AppendElements(&c[0], sizeof(uint64_t));
}
mSeekPreroll = params.seek_preroll;
mInfo.mAudio.mRate = params.rate;
mInfo.mAudio.mChannels = params.channels;
unsigned int nheaders = 0;
r = nestegg_track_codec_data_count(mContext, track, &nheaders);
if (r == -1) {
return NS_ERROR_FAILURE;
}
for (uint32_t header = 0; header < nheaders; ++header) {
unsigned char* data = 0;
size_t length = 0;
r = nestegg_track_codec_data(mContext, track, header, &data, &length);
if (r == -1) {
return NS_ERROR_FAILURE;
}
// Vorbis has 3 headers write length + data for each header
if (nheaders > 1) {
uint8_t c[2];
BigEndian::writeUint16(&c[0], length);
mInfo.mAudio.mCodecSpecificConfig->AppendElements(&c[0], 2);
}
mInfo.mAudio.mCodecSpecificConfig->AppendElements(data, length);
}
uint64_t duration = 0;
r = nestegg_duration(mContext, &duration);
if (!r) {
mInfo.mAudio.mDuration = media::TimeUnit::FromNanoseconds(duration).ToMicroseconds();
}
}
}
return NS_OK;
}
nsresult
WebMDemuxer::Read(char* aBuffer, uint32_t aCount, uint32_t* aBytes)
{
int64_t length = mResource->GetLength();
if (length >= 0 && aCount + mOffset > length) {
WEBM_DEBUG("requested to large amount, trying to get %ld bytes at %ld (length: %ld)", aCount, mOffset, length);
aCount = length - mOffset;
WEBM_DEBUG("will only return %ld", aCount);
}
nsRefPtr<MediaByteBuffer> bytes = mResource->MediaReadAt(mOffset, aCount);
if (!bytes) {
return NS_ERROR_FAILURE;
}
mOffset += bytes->Length();
*aBytes = bytes->Length();
memcpy(aBuffer, bytes->Elements(), bytes->Length());
return NS_OK;
}
nsresult
WebMDemuxer::Seek(int32_t aWhence, int64_t aOffset)
{
if (aWhence == SEEK_CUR) {
aOffset += mOffset;
} else if (aWhence == SEEK_END) {
int64_t length = mResource->GetLength();
if (length == -1 || length - aOffset < 0) {
return NS_ERROR_FAILURE;
}
aOffset = mResource->GetLength() - aOffset;
}
if (aOffset > mResource->GetLength()) {
return NS_ERROR_FAILURE;
}
mOffset = aOffset;
return NS_OK;
}
int64_t
WebMDemuxer::Tell()
{
return mOffset;
}
bool
WebMDemuxer::IsSeekable() const
{
return mContext && nestegg_has_cues(mContext);
}
void
WebMDemuxer::EnsureUpToDateIndex()
{
if (!mNeedReIndex) {
return;
}
AutoPinned<MediaResource> resource(mResource);
nsTArray<MediaByteRange> byteRanges;
nsresult rv = resource->GetCachedRanges(byteRanges);
if (NS_FAILED(rv)) {
return;
}
if (mInitData && mBufferedState->GetInitEndOffset() == -1) {
mBufferedState->NotifyDataArrived(mInitData->Elements(), mInitData->Length(), 0);
}
mBufferedState->UpdateIndex(byteRanges, mResource);
if (!mInitData && mBufferedState->GetInitEndOffset() != -1) {
mInitData = mResource->MediaReadAt(0, mBufferedState->GetInitEndOffset());
}
mNeedReIndex = false;
}
void
WebMDemuxer::NotifyDataArrived(uint32_t aLength, int64_t aOffset)
{
WEBM_DEBUG("length: %ld offset: %ld", aLength, aOffset);
mNeedReIndex = true;
}
void
WebMDemuxer::NotifyDataRemoved()
{
mBufferedState->Reset();
mNeedReIndex = true;
}
UniquePtr<EncryptionInfo>
WebMDemuxer::GetCrypto()
{
return nullptr;
}
bool
WebMDemuxer::GetNextPacket(TrackInfo::TrackType aType, MediaRawDataQueue *aSamples)
{
nsRefPtr<NesteggPacketHolder> holder(NextPacket(aType));
if (!holder) {
return false;
}
int r = 0;
unsigned int count = 0;
r = nestegg_packet_count(holder->Packet(), &count);
if (r == -1) {
return false;
}
int64_t tstamp = holder->Timestamp();
// The end time of this frame is the start time of the next frame. Fetch
// the timestamp of the next packet for this track. If we've reached the
// end of the resource, use the file's duration as the end time of this
// video frame.
int64_t next_tstamp = 0;
if (aType == TrackInfo::kAudioTrack) {
nsRefPtr<NesteggPacketHolder> next_holder(NextPacket(aType));
if (next_holder) {
next_tstamp = next_holder->Timestamp();
PushAudioPacket(next_holder);
} else {
next_tstamp = tstamp;
next_tstamp += tstamp - mLastAudioFrameTime;
}
mLastAudioFrameTime = tstamp;
} else if (aType == TrackInfo::kVideoTrack) {
nsRefPtr<NesteggPacketHolder> next_holder(NextPacket(aType));
if (next_holder) {
next_tstamp = next_holder->Timestamp();
PushVideoPacket(next_holder);
} else {
next_tstamp = tstamp;
next_tstamp += tstamp - mLastVideoFrameTime;
}
mLastVideoFrameTime = tstamp;
}
int64_t discardPadding = 0;
(void) nestegg_packet_discard_padding(holder->Packet(), &discardPadding);
for (uint32_t i = 0; i < count; ++i) {
unsigned char* data;
size_t length;
r = nestegg_packet_data(holder->Packet(), i, &data, &length);
if (r == -1) {
WEBM_DEBUG("nestegg_packet_data failed r=%d", r);
return false;
}
bool isKeyframe = false;
if (aType == TrackInfo::kAudioTrack) {
isKeyframe = true;
} else if (aType == TrackInfo::kVideoTrack) {
vpx_codec_stream_info_t si;
PodZero(&si);
si.sz = sizeof(si);
switch (mVideoCodec) {
case NESTEGG_CODEC_VP8:
vpx_codec_peek_stream_info(vpx_codec_vp8_dx(), data, length, &si);
break;
case NESTEGG_CODEC_VP9:
vpx_codec_peek_stream_info(vpx_codec_vp9_dx(), data, length, &si);
break;
}
isKeyframe = si.is_kf;
}
WEBM_DEBUG("push sample tstamp: %ld next_tstamp: %ld length: %ld kf: %d",
tstamp, next_tstamp, length, isKeyframe);
nsRefPtr<MediaRawData> sample = new MediaRawData(data, length);
sample->mTimecode = tstamp;
sample->mTime = tstamp;
sample->mDuration = next_tstamp - tstamp;
sample->mOffset = holder->Offset();
sample->mKeyframe = isKeyframe;
if (discardPadding) {
uint8_t c[8];
BigEndian::writeInt64(&c[0], discardPadding);
sample->mExtraData = new MediaByteBuffer;
sample->mExtraData->AppendElements(&c[0], 8);
}
aSamples->Push(sample);
}
return true;
}
nsRefPtr<NesteggPacketHolder>
WebMDemuxer::NextPacket(TrackInfo::TrackType aType)
{
bool isVideo = aType == TrackInfo::kVideoTrack;
// The packet queue that packets will be pushed on if they
// are not the type we are interested in.
WebMPacketQueue& otherPackets = isVideo ? mAudioPackets : mVideoPackets;
// The packet queue for the type that we are interested in.
WebMPacketQueue &packets = isVideo ? mVideoPackets : mAudioPackets;
// Flag to indicate that we do need to playback these types of
// packets.
bool hasType = isVideo ? mHasVideo : mHasAudio;
// Flag to indicate that we do need to playback the other type
// of track.
bool hasOtherType = isVideo ? mHasAudio : mHasVideo;
// Track we are interested in
uint32_t ourTrack = isVideo ? mVideoTrack : mAudioTrack;
// Value of other track
uint32_t otherTrack = isVideo ? mAudioTrack : mVideoTrack;
if (packets.GetSize() > 0) {
return packets.PopFront();
}
do {
nsRefPtr<NesteggPacketHolder> holder = DemuxPacket();
if (!holder) {
return nullptr;
}
if (hasOtherType && otherTrack == holder->Track()) {
// Save the packet for when we want these packets
otherPackets.Push(holder);
continue;
}
// The packet is for the track we want to play
if (hasType && ourTrack == holder->Track()) {
return holder;
}
} while (true);
}
nsRefPtr<NesteggPacketHolder>
WebMDemuxer::DemuxPacket()
{
nestegg_packet* packet;
int r = nestegg_read_packet(mContext, &packet);
if (r <= 0) {
return nullptr;
}
unsigned int track = 0;
r = nestegg_packet_track(packet, &track);
if (r == -1) {
return nullptr;
}
int64_t offset = Tell();
nsRefPtr<NesteggPacketHolder> holder = new NesteggPacketHolder();
if (!holder->Init(packet, offset, track, false)) {
return nullptr;
}
return holder;
}
int64_t
WebMDemuxer::GetNextKeyframeTime()
{
EnsureUpToDateIndex();
uint64_t keyframeTime;
uint64_t lastFrame =
media::TimeUnit::FromMicroseconds(mLastVideoFrameTime).ToNanoseconds();
if (!mBufferedState->GetNextKeyframeTime(lastFrame, &keyframeTime) ||
keyframeTime <= lastFrame) {
return -1;
}
return media::TimeUnit::FromNanoseconds(keyframeTime).ToMicroseconds();
}
void
WebMDemuxer::PushAudioPacket(NesteggPacketHolder* aItem)
{
mAudioPackets.PushFront(aItem);
}
void
WebMDemuxer::PushVideoPacket(NesteggPacketHolder* aItem)
{
mVideoPackets.PushFront(aItem);
}
nsresult
WebMDemuxer::SeekInternal(const media::TimeUnit& aTarget)
{
EnsureUpToDateIndex();
uint32_t trackToSeek = mHasVideo ? mVideoTrack : mAudioTrack;
uint64_t target = aTarget.ToNanoseconds();
if (NS_FAILED(Reset())) {
return NS_ERROR_FAILURE;
}
if (mSeekPreroll) {
uint64_t startTime = 0;
if (!mBufferedState->GetStartTime(&startTime)) {
startTime = 0;
}
WEBM_DEBUG("Seek Target: %f",
media::TimeUnit::FromNanoseconds(target).ToSeconds());
if (target < mSeekPreroll || target - mSeekPreroll < startTime) {
target = startTime;
} else {
target -= mSeekPreroll;
}
WEBM_DEBUG("SeekPreroll: %f StartTime: %f Adjusted Target: %f",
media::TimeUnit::FromNanoseconds(mSeekPreroll).ToSeconds(),
media::TimeUnit::FromNanoseconds(startTime).ToSeconds(),
media::TimeUnit::FromNanoseconds(target).ToSeconds());
}
int r = nestegg_track_seek(mContext, trackToSeek, target);
if (r == -1) {
WEBM_DEBUG("track_seek for track %u to %f failed, r=%d", trackToSeek,
media::TimeUnit::FromNanoseconds(target).ToSeconds(), r);
// Try seeking directly based on cluster information in memory.
int64_t offset = 0;
bool rv = mBufferedState->GetOffsetForTime(target, &offset);
if (!rv) {
WEBM_DEBUG("mBufferedState->GetOffsetForTime failed too");
return NS_ERROR_FAILURE;
}
r = nestegg_offset_seek(mContext, offset);
if (r == -1) {
WEBM_DEBUG("and nestegg_offset_seek to %" PRIu64 " failed", offset);
return NS_ERROR_FAILURE;
}
WEBM_DEBUG("got offset from buffered state: %" PRIu64 "", offset);
}
return NS_OK;
}
media::TimeIntervals
WebMDemuxer::GetBuffered()
{
EnsureUpToDateIndex();
AutoPinned<MediaResource> resource(mResource);
media::TimeIntervals buffered;
nsTArray<MediaByteRange> ranges;
nsresult rv = resource->GetCachedRanges(ranges);
if (NS_FAILED(rv)) {
return media::TimeIntervals();
}
uint64_t duration = 0;
uint64_t startOffset = 0;
if (!nestegg_duration(mContext, &duration)) {
if(mBufferedState->GetStartTime(&startOffset)) {
duration += startOffset;
}
WEBM_DEBUG("Duration: %f StartTime: %f",
media::TimeUnit::FromNanoseconds(duration).ToSeconds(),
media::TimeUnit::FromNanoseconds(startOffset).ToSeconds());
}
for (uint32_t index = 0; index < ranges.Length(); index++) {
uint64_t start, end;
bool rv = mBufferedState->CalculateBufferedForRange(ranges[index].mStart,
ranges[index].mEnd,
&start, &end);
if (rv) {
NS_ASSERTION(startOffset <= start,
"startOffset negative or larger than start time");
if (duration && end > duration) {
WEBM_DEBUG("limit range to duration, end: %f duration: %f",
media::TimeUnit::FromNanoseconds(end).ToSeconds(),
media::TimeUnit::FromNanoseconds(duration).ToSeconds());
end = duration;
}
media::TimeUnit startTime = media::TimeUnit::FromNanoseconds(start);
media::TimeUnit endTime = media::TimeUnit::FromNanoseconds(end);
WEBM_DEBUG("add range %f-%f", startTime.ToSeconds(), endTime.ToSeconds());
buffered += media::TimeInterval(startTime, endTime);
}
}
return buffered;
}
bool WebMDemuxer::GetOffsetForTime(uint64_t aTime, int64_t* aOffset)
{
EnsureUpToDateIndex();
return mBufferedState && mBufferedState->GetOffsetForTime(aTime, aOffset);
}
//WebMTrackDemuxer
WebMTrackDemuxer::WebMTrackDemuxer(WebMDemuxer* aParent,
TrackInfo::TrackType aType,
uint32_t aTrackNumber)
: mParent(aParent)
, mType(aType)
{
mInfo = mParent->GetTrackInfo(aType, aTrackNumber);
MOZ_ASSERT(mInfo);
}
WebMTrackDemuxer::~WebMTrackDemuxer()
{
mSamples.Reset();
}
UniquePtr<TrackInfo>
WebMTrackDemuxer::GetInfo() const
{
return mInfo->Clone();
}
nsRefPtr<WebMTrackDemuxer::SeekPromise>
WebMTrackDemuxer::Seek(media::TimeUnit aTime)
{
// Seeks to aTime. Upon success, SeekPromise will be resolved with the
// actual time seeked to. Typically the random access point time
media::TimeUnit seekTime = aTime;
mSamples.Reset();
mParent->SeekInternal(aTime);
mParent->GetNextPacket(mType, &mSamples);
// Check what time we actually seeked to.
if (mSamples.GetSize() > 0) {
nsRefPtr<MediaRawData> sample(mSamples.PopFront());
seekTime = media::TimeUnit::FromMicroseconds(sample->mTime);
mSamples.PushFront(sample);
}
SetNextKeyFrameTime();
return SeekPromise::CreateAndResolve(seekTime, __func__);
}
nsRefPtr<MediaRawData>
WebMTrackDemuxer::NextSample()
{
while (mSamples.GetSize() < 1 && mParent->GetNextPacket(mType, &mSamples)) {
}
if (mSamples.GetSize()) {
return mSamples.PopFront();
}
return nullptr;
}
nsRefPtr<WebMTrackDemuxer::SamplesPromise>
WebMTrackDemuxer::GetSamples(int32_t aNumSamples)
{
nsRefPtr<SamplesHolder> samples = new SamplesHolder;
if (!aNumSamples) {
return SamplesPromise::CreateAndReject(DemuxerFailureReason::DEMUXER_ERROR, __func__);
}
while (aNumSamples) {
nsRefPtr<MediaRawData> sample(NextSample());
if (!sample) {
break;
}
samples->mSamples.AppendElement(sample);
aNumSamples--;
}
if (samples->mSamples.IsEmpty()) {
return SamplesPromise::CreateAndReject(DemuxerFailureReason::END_OF_STREAM, __func__);
} else {
UpdateSamples(samples->mSamples);
return SamplesPromise::CreateAndResolve(samples, __func__);
}
}
void
WebMTrackDemuxer::SetNextKeyFrameTime()
{
int64_t frameTime = -1;
mNextKeyframeTime.reset();
if (mType == TrackInfo::kVideoTrack) {
MediaRawDataQueue skipSamplesQueue;
bool foundKeyframe = false;
while (!foundKeyframe && mSamples.GetSize()) {
nsRefPtr<MediaRawData> sample(mSamples.PopFront());
if (sample->mKeyframe) {
frameTime = sample->mTime;
foundKeyframe = true;
}
skipSamplesQueue.PushFront(sample);
}
while(skipSamplesQueue.GetSize()) {
mSamples.PushFront(skipSamplesQueue.PopFront());
}
if (frameTime == -1) {
frameTime = mParent->GetNextKeyframeTime();
}
}
if (frameTime != -1) {
mNextKeyframeTime.emplace(media::TimeUnit::FromMicroseconds(frameTime));
WEBM_DEBUG("Next Keyframe %f", mNextKeyframeTime.value().ToSeconds());
}
}
void
WebMTrackDemuxer::Reset()
{
mSamples.Reset();
media::TimeIntervals buffered = GetBuffered();
if (buffered.Length()) {
WEBM_DEBUG("Seek to start point: %f", buffered.Start(0).ToSeconds());
mParent->SeekInternal(buffered.Start(0));
SetNextKeyFrameTime();
} else {
mNextKeyframeTime.reset();
}
}
void
WebMTrackDemuxer::UpdateSamples(nsTArray<nsRefPtr<MediaRawData>>& aSamples)
{
if (mNextKeyframeTime.isNothing() ||
aSamples.LastElement()->mTime >= mNextKeyframeTime.value().ToMicroseconds()) {
SetNextKeyFrameTime();
}
}
nsresult
WebMTrackDemuxer::GetNextRandomAccessPoint(media::TimeUnit* aTime)
{
if (mNextKeyframeTime.isNothing()) {
// There's no next key frame.
*aTime =
media::TimeUnit::FromMicroseconds(std::numeric_limits<int64_t>::max());
} else {
*aTime = mNextKeyframeTime.ref();
}
return NS_OK;
}
nsRefPtr<WebMTrackDemuxer::SkipAccessPointPromise>
WebMTrackDemuxer::SkipToNextRandomAccessPoint(media::TimeUnit aTimeThreshold)
{
uint32_t parsed = 0;
bool found = false;
nsRefPtr<MediaRawData> sample;
WEBM_DEBUG("TimeThreshold: %f", aTimeThreshold.ToSeconds());
while (!found && (sample = NextSample())) {
parsed++;
if (sample->mKeyframe && sample->mTime >= aTimeThreshold.ToMicroseconds()) {
found = true;
mSamples.Reset();
mSamples.PushFront(sample);
}
}
SetNextKeyFrameTime();
if (found) {
WEBM_DEBUG("next sample: %f (parsed: %d)",
media::TimeUnit::FromMicroseconds(sample->mTime).ToSeconds(),
parsed);
return SkipAccessPointPromise::CreateAndResolve(parsed, __func__);
} else {
SkipFailureHolder failure(DemuxerFailureReason::END_OF_STREAM, parsed);
return SkipAccessPointPromise::CreateAndReject(Move(failure), __func__);
}
}
int64_t
WebMTrackDemuxer::GetEvictionOffset(media::TimeUnit aTime)
{
int64_t offset;
if (!mParent->GetOffsetForTime(aTime.ToNanoseconds(), &offset)) {
return 0;
}
return offset;
}
media::TimeIntervals
WebMTrackDemuxer::GetBuffered()
{
return mParent->GetBuffered();
}
void
WebMTrackDemuxer::BreakCycles()
{
mParent = nullptr;
}
#undef WEBM_DEBUG
} // namespace mozilla

Просмотреть файл

@ -0,0 +1,207 @@
/* -*- Mode: C++; tab-width: 2; indent-tabs-mode: nil; c-basic-offset: 2 -*- */
/* vim:set ts=2 sw=2 sts=2 et cindent: */
/* This Source Code Form is subject to the terms of the Mozilla Public
* License, v. 2.0. If a copy of the MPL was not distributed with this
* file, You can obtain one at http://mozilla.org/MPL/2.0/. */
#if !defined(WebMDemuxer_h_)
#define WebMDemuxer_h_
#include "nsTArray.h"
#include "MediaDataDemuxer.h"
typedef struct nestegg nestegg;
namespace mozilla {
class NesteggPacketHolder;
class WebMBufferedState;
class WebMPacketQueue;
// Queue for holding MediaRawData samples
class MediaRawDataQueue {
public:
uint32_t GetSize() {
return mQueue.size();
}
void Push(MediaRawData* aItem) {
mQueue.push_back(aItem);
}
void PushFront(MediaRawData* aItem) {
mQueue.push_front(aItem);
}
nsRefPtr<MediaRawData> PopFront() {
nsRefPtr<MediaRawData> result = mQueue.front();
mQueue.pop_front();
return result;
}
void Reset() {
while (!mQueue.empty()) {
mQueue.pop_front();
}
}
private:
std::deque<nsRefPtr<MediaRawData>> mQueue;
};
class WebMTrackDemuxer;
class WebMDemuxer : public MediaDataDemuxer
{
public:
explicit WebMDemuxer(MediaResource* aResource);
nsRefPtr<InitPromise> Init() override;
already_AddRefed<MediaDataDemuxer> Clone() const override;
bool HasTrackType(TrackInfo::TrackType aType) const override;
uint32_t GetNumberTracks(TrackInfo::TrackType aType) const override;
UniquePtr<TrackInfo> GetTrackInfo(TrackInfo::TrackType aType, size_t aTrackNumber) const;
already_AddRefed<MediaTrackDemuxer> GetTrackDemuxer(TrackInfo::TrackType aType,
uint32_t aTrackNumber) override;
bool IsSeekable() const override;
UniquePtr<EncryptionInfo> GetCrypto() override;
bool GetOffsetForTime(uint64_t aTime, int64_t* aOffset);
// Demux next WebM packet and append samples to MediaRawDataQueue
bool GetNextPacket(TrackInfo::TrackType aType, MediaRawDataQueue *aSamples);
nsresult Reset();
// Pushes a packet to the front of the audio packet queue.
virtual void PushAudioPacket(NesteggPacketHolder* aItem);
// Pushes a packet to the front of the video packet queue.
virtual void PushVideoPacket(NesteggPacketHolder* aItem);
nsresult Read(char* aBuffer, uint32_t aCount, uint32_t * aBytes);
nsresult Seek(int32_t aWhence, int64_t aOffset);
int64_t Tell();
private:
friend class WebMTrackDemuxer;
~WebMDemuxer();
void Cleanup();
nsresult InitBufferedState();
nsresult ReadMetadata();
void NotifyDataArrived(uint32_t aLength, int64_t aOffset) override;
void NotifyDataRemoved() override;
void EnsureUpToDateIndex();
media::TimeIntervals GetBuffered();
virtual nsresult SeekInternal(const media::TimeUnit& aTarget);
// Get the timestamp of the next keyframe
int64_t GetNextKeyframeTime();
// Read a packet from the nestegg file. Returns nullptr if all packets for
// the particular track have been read. Pass TrackInfo::kVideoTrack or
// TrackInfo::kVideoTrack to indicate the type of the packet we want to read.
nsRefPtr<NesteggPacketHolder> NextPacket(TrackInfo::TrackType aType);
// Internal method that demuxes the next packet from the stream. The caller
// is responsible for making sure it doesn't get lost.
nsRefPtr<NesteggPacketHolder> DemuxPacket();
nsRefPtr<MediaResource> mResource;
MediaInfo mInfo;
nsTArray<nsRefPtr<WebMTrackDemuxer>> mDemuxers;
// Parser state and computed offset-time mappings. Shared by multiple
// readers when decoder has been cloned. Main thread only.
nsRefPtr<WebMBufferedState> mBufferedState;
nsRefPtr<MediaByteBuffer> mInitData;
// libnestegg context for webm container.
// Access on reader's thread for main demuxer,
// or main thread for cloned demuxer
nestegg* mContext;
int64_t mOffset;
// Queue of video and audio packets that have been read but not decoded.
WebMPacketQueue mVideoPackets;
WebMPacketQueue mAudioPackets;
// Index of video and audio track to play
uint32_t mVideoTrack;
uint32_t mAudioTrack;
// Number of microseconds that must be discarded from the start of the Stream.
uint64_t mCodecDelay;
// Nanoseconds to discard after seeking.
uint64_t mSeekPreroll;
int64_t mLastAudioFrameTime;
// Calculate the frame duration from the last decodeable frame using the
// previous frame's timestamp. In NS.
int64_t mLastVideoFrameTime;
// Picture region, as relative to the initial frame size.
nsIntRect mPicture;
// Codec ID of audio track
int mAudioCodec;
// Codec ID of video track
int mVideoCodec;
// Booleans to indicate if we have audio and/or video data
bool mHasVideo;
bool mHasAudio;
bool mNeedReIndex;
};
class WebMTrackDemuxer : public MediaTrackDemuxer
{
public:
WebMTrackDemuxer(WebMDemuxer* aParent,
TrackInfo::TrackType aType,
uint32_t aTrackNumber);
UniquePtr<TrackInfo> GetInfo() const override;
nsRefPtr<SeekPromise> Seek(media::TimeUnit aTime) override;
nsRefPtr<SamplesPromise> GetSamples(int32_t aNumSamples = 1) override;
void Reset() override;
nsresult GetNextRandomAccessPoint(media::TimeUnit* aTime) override;
nsRefPtr<SkipAccessPointPromise> SkipToNextRandomAccessPoint(media::TimeUnit aTimeThreshold) override;
media::TimeIntervals GetBuffered() override;
int64_t GetEvictionOffset(media::TimeUnit aTime) override;
void BreakCycles() override;
private:
friend class WebMDemuxer;
~WebMTrackDemuxer();
void UpdateSamples(nsTArray<nsRefPtr<MediaRawData>>& aSamples);
void SetNextKeyFrameTime();
nsRefPtr<MediaRawData> NextSample ();
nsRefPtr<WebMDemuxer> mParent;
TrackInfo::TrackType mType;
UniquePtr<TrackInfo> mInfo;
Maybe<media::TimeUnit> mNextKeyframeTime;
// Queued samples extracted by the demuxer, but not yet returned.
MediaRawDataQueue mSamples;
};
} // namespace mozilla
#endif

Просмотреть файл

@ -10,6 +10,7 @@ EXPORTS += [
'SoftwareWebMVideoDecoder.h',
'WebMBufferedParser.h',
'WebMDecoder.h',
'WebMDemuxer.h',
'WebMReader.h',
]
@ -18,6 +19,7 @@ UNIFIED_SOURCES += [
'SoftwareWebMVideoDecoder.cpp',
'WebMBufferedParser.cpp',
'WebMDecoder.cpp',
'WebMDemuxer.cpp',
'WebMReader.cpp',
]