// Copyright 2011 The Chromium Authors. All rights reserved. // Use of this source code is governed by a BSD-style license that can be // found in the LICENSE file. #include "config.h" #if USE(ACCELERATED_COMPOSITING) #include "CCVideoLayerImpl.h" #include "CCIOSurfaceDrawQuad.h" #include "CCLayerTreeHostImpl.h" #include "CCProxy.h" #include "CCQuadSink.h" #include "CCResourceProvider.h" #include "CCStreamVideoDrawQuad.h" #include "CCTextureDrawQuad.h" #include "CCYUVVideoDrawQuad.h" #include "Extensions3DChromium.h" #include "GraphicsContext3D.h" #include "NotImplemented.h" #include namespace cc { CCVideoLayerImpl::CCVideoLayerImpl(int id, WebKit::WebVideoFrameProvider* provider) : CCLayerImpl(id) , m_provider(provider) , m_frame(0) , m_externalTextureResource(0) { // This matrix is the default transformation for stream textures, and flips on the Y axis. m_streamTextureMatrix = WebKit::WebTransformationMatrix( 1, 0, 0, 0, 0, -1, 0, 0, 0, 0, 1, 0, 0, 1, 0, 1); // This only happens during a commit on the compositor thread while the main // thread is blocked. That makes this a thread-safe call to set the video // frame provider client that does not require a lock. The same is true of // the call in the destructor. ASSERT(CCProxy::isMainThreadBlocked()); m_provider->setVideoFrameProviderClient(this); } CCVideoLayerImpl::~CCVideoLayerImpl() { // See comment in constructor for why this doesn't need a lock. ASSERT(CCProxy::isMainThreadBlocked()); if (m_provider) { m_provider->setVideoFrameProviderClient(0); m_provider = 0; } freePlaneData(layerTreeHostImpl()->resourceProvider()); #if !ASSERT_DISABLED for (unsigned i = 0; i < WebKit::WebVideoFrame::maxPlanes; ++i) ASSERT(!m_framePlanes[i].resourceId); ASSERT(!m_externalTextureResource); #endif } void CCVideoLayerImpl::stopUsingProvider() { // Block the provider from shutting down until this client is done // using the frame. MutexLocker locker(m_providerMutex); ASSERT(!m_frame); m_provider = 0; } // Convert WebKit::WebVideoFrame::Format to GraphicsContext3D's format enum values. static GC3Denum convertVFCFormatToGC3DFormat(const WebKit::WebVideoFrame& frame) { switch (frame.format()) { case WebKit::WebVideoFrame::FormatYV12: case WebKit::WebVideoFrame::FormatYV16: return GraphicsContext3D::LUMINANCE; case WebKit::WebVideoFrame::FormatNativeTexture: return frame.textureTarget(); case WebKit::WebVideoFrame::FormatInvalid: case WebKit::WebVideoFrame::FormatRGB32: case WebKit::WebVideoFrame::FormatEmpty: case WebKit::WebVideoFrame::FormatI420: notImplemented(); } return GraphicsContext3D::INVALID_VALUE; } void CCVideoLayerImpl::willDraw(CCResourceProvider* resourceProvider) { ASSERT(CCProxy::isImplThread()); CCLayerImpl::willDraw(resourceProvider); // Explicitly lock and unlock the provider mutex so it can be held from // willDraw to didDraw. Since the compositor thread is in the middle of // drawing, the layer will not be destroyed before didDraw is called. // Therefore, the only thing that will prevent this lock from being released // is the GPU process locking it. As the GPU process can't cause the // destruction of the provider (calling stopUsingProvider), holding this // lock should not cause a deadlock. m_providerMutex.lock(); willDrawInternal(resourceProvider); freeUnusedPlaneData(resourceProvider); if (!m_frame) m_providerMutex.unlock(); } void CCVideoLayerImpl::willDrawInternal(CCResourceProvider* resourceProvider) { ASSERT(CCProxy::isImplThread()); ASSERT(!m_externalTextureResource); if (!m_provider) { m_frame = 0; return; } m_frame = m_provider->getCurrentFrame(); if (!m_frame) return; m_format = convertVFCFormatToGC3DFormat(*m_frame); if (m_format == GraphicsContext3D::INVALID_VALUE) { m_provider->putCurrentFrame(m_frame); m_frame = 0; return; } if (m_frame->planes() > WebKit::WebVideoFrame::maxPlanes) { m_provider->putCurrentFrame(m_frame); m_frame = 0; return; } if (!allocatePlaneData(resourceProvider)) { m_provider->putCurrentFrame(m_frame); m_frame = 0; return; } if (!copyPlaneData(resourceProvider)) { m_provider->putCurrentFrame(m_frame); m_frame = 0; return; } if (m_format == GraphicsContext3D::TEXTURE_2D) m_externalTextureResource = resourceProvider->createResourceFromExternalTexture(m_frame->textureId()); } void CCVideoLayerImpl::appendQuads(CCQuadSink& quadSink, CCAppendQuadsData& appendQuadsData) { ASSERT(CCProxy::isImplThread()); if (!m_frame) return; CCSharedQuadState* sharedQuadState = quadSink.useSharedQuadState(createSharedQuadState()); appendDebugBorderQuad(quadSink, sharedQuadState, appendQuadsData); // FIXME: When we pass quads out of process, we need to double-buffer, or // otherwise synchonize use of all textures in the quad. IntRect quadRect(IntPoint(), contentBounds()); switch (m_format) { case GraphicsContext3D::LUMINANCE: { // YUV software decoder. const FramePlane& yPlane = m_framePlanes[WebKit::WebVideoFrame::yPlane]; const FramePlane& uPlane = m_framePlanes[WebKit::WebVideoFrame::uPlane]; const FramePlane& vPlane = m_framePlanes[WebKit::WebVideoFrame::vPlane]; OwnPtr yuvVideoQuad = CCYUVVideoDrawQuad::create(sharedQuadState, quadRect, yPlane, uPlane, vPlane); quadSink.append(yuvVideoQuad.release(), appendQuadsData); break; } case GraphicsContext3D::RGBA: { // RGBA software decoder. const FramePlane& plane = m_framePlanes[WebKit::WebVideoFrame::rgbPlane]; float widthScaleFactor = static_cast(plane.visibleSize.width()) / plane.size.width(); bool premultipliedAlpha = true; FloatRect uvRect(0, 0, widthScaleFactor, 1); bool flipped = false; OwnPtr textureQuad = CCTextureDrawQuad::create(sharedQuadState, quadRect, plane.resourceId, premultipliedAlpha, uvRect, flipped); quadSink.append(textureQuad.release(), appendQuadsData); break; } case GraphicsContext3D::TEXTURE_2D: { // NativeTexture hardware decoder. bool premultipliedAlpha = true; FloatRect uvRect(0, 0, 1, 1); bool flipped = false; OwnPtr textureQuad = CCTextureDrawQuad::create(sharedQuadState, quadRect, m_externalTextureResource, premultipliedAlpha, uvRect, flipped); quadSink.append(textureQuad.release(), appendQuadsData); break; } case Extensions3D::TEXTURE_RECTANGLE_ARB: { IntSize textureSize(m_frame->width(), m_frame->height()); OwnPtr ioSurfaceQuad = CCIOSurfaceDrawQuad::create(sharedQuadState, quadRect, textureSize, m_frame->textureId(), CCIOSurfaceDrawQuad::Unflipped); quadSink.append(ioSurfaceQuad.release(), appendQuadsData); break; } case Extensions3DChromium::GL_TEXTURE_EXTERNAL_OES: { // StreamTexture hardware decoder. OwnPtr streamVideoQuad = CCStreamVideoDrawQuad::create(sharedQuadState, quadRect, m_frame->textureId(), m_streamTextureMatrix); quadSink.append(streamVideoQuad.release(), appendQuadsData); break; } default: CRASH(); // Someone updated convertVFCFormatToGC3DFormat above but update this! } } void CCVideoLayerImpl::didDraw(CCResourceProvider* resourceProvider) { ASSERT(CCProxy::isImplThread()); CCLayerImpl::didDraw(resourceProvider); if (!m_frame) return; if (m_format == GraphicsContext3D::TEXTURE_2D) { ASSERT(m_externalTextureResource); // FIXME: the following assert will not be true when sending resources to a // parent compositor. We will probably need to hold on to m_frame for // longer, and have several "current frames" in the pipeline. ASSERT(!resourceProvider->inUseByConsumer(m_externalTextureResource)); resourceProvider->deleteResource(m_externalTextureResource); m_externalTextureResource = 0; } m_provider->putCurrentFrame(m_frame); m_frame = 0; m_providerMutex.unlock(); } static int videoFrameDimension(int originalDimension, unsigned plane, int format) { if (format == WebKit::WebVideoFrame::FormatYV12 && plane != WebKit::WebVideoFrame::yPlane) return originalDimension / 2; return originalDimension; } static bool hasPaddingBytes(const WebKit::WebVideoFrame& frame, unsigned plane) { return frame.stride(plane) > videoFrameDimension(frame.width(), plane, frame.format()); } IntSize CCVideoLayerImpl::computeVisibleSize(const WebKit::WebVideoFrame& frame, unsigned plane) { int visibleWidth = videoFrameDimension(frame.width(), plane, frame.format()); int originalWidth = visibleWidth; int visibleHeight = videoFrameDimension(frame.height(), plane, frame.format()); // When there are dead pixels at the edge of the texture, decrease // the frame width by 1 to prevent the rightmost pixels from // interpolating with the dead pixels. if (hasPaddingBytes(frame, plane)) --visibleWidth; // In YV12, every 2x2 square of Y values corresponds to one U and // one V value. If we decrease the width of the UV plane, we must decrease the // width of the Y texture by 2 for proper alignment. This must happen // always, even if Y's texture does not have padding bytes. if (plane == WebKit::WebVideoFrame::yPlane && frame.format() == WebKit::WebVideoFrame::FormatYV12) { if (hasPaddingBytes(frame, WebKit::WebVideoFrame::uPlane)) visibleWidth = originalWidth - 2; } return IntSize(visibleWidth, visibleHeight); } bool CCVideoLayerImpl::FramePlane::allocateData(CCResourceProvider* resourceProvider) { if (resourceId) return true; resourceId = resourceProvider->createResource(CCRenderer::ImplPool, size, format, CCResourceProvider::TextureUsageAny); return resourceId; } void CCVideoLayerImpl::FramePlane::freeData(CCResourceProvider* resourceProvider) { if (!resourceId) return; resourceProvider->deleteResource(resourceId); resourceId = 0; } bool CCVideoLayerImpl::allocatePlaneData(CCResourceProvider* resourceProvider) { int maxTextureSize = resourceProvider->maxTextureSize(); for (unsigned planeIndex = 0; planeIndex < m_frame->planes(); ++planeIndex) { CCVideoLayerImpl::FramePlane& plane = m_framePlanes[planeIndex]; IntSize requiredTextureSize(m_frame->stride(planeIndex), videoFrameDimension(m_frame->height(), planeIndex, m_frame->format())); // FIXME: Remove the test against maxTextureSize when tiled layers are implemented. if (requiredTextureSize.isZero() || requiredTextureSize.width() > maxTextureSize || requiredTextureSize.height() > maxTextureSize) return false; if (plane.size != requiredTextureSize || plane.format != m_format) { plane.freeData(resourceProvider); plane.size = requiredTextureSize; plane.format = m_format; } if (!plane.resourceId) { if (!plane.allocateData(resourceProvider)) return false; plane.visibleSize = computeVisibleSize(*m_frame, planeIndex); } } return true; } bool CCVideoLayerImpl::copyPlaneData(CCResourceProvider* resourceProvider) { size_t softwarePlaneCount = m_frame->planes(); if (!softwarePlaneCount) return true; for (size_t softwarePlaneIndex = 0; softwarePlaneIndex < softwarePlaneCount; ++softwarePlaneIndex) { CCVideoLayerImpl::FramePlane& plane = m_framePlanes[softwarePlaneIndex]; const uint8_t* softwarePlanePixels = static_cast(m_frame->data(softwarePlaneIndex)); IntRect planeRect(IntPoint(), plane.size); resourceProvider->upload(plane.resourceId, softwarePlanePixels, planeRect, planeRect, IntSize()); } return true; } void CCVideoLayerImpl::freePlaneData(CCResourceProvider* resourceProvider) { for (unsigned i = 0; i < WebKit::WebVideoFrame::maxPlanes; ++i) m_framePlanes[i].freeData(resourceProvider); } void CCVideoLayerImpl::freeUnusedPlaneData(CCResourceProvider* resourceProvider) { unsigned firstUnusedPlane = m_frame ? m_frame->planes() : 0; for (unsigned i = firstUnusedPlane; i < WebKit::WebVideoFrame::maxPlanes; ++i) m_framePlanes[i].freeData(resourceProvider); } void CCVideoLayerImpl::didReceiveFrame() { setNeedsRedraw(); } void CCVideoLayerImpl::didUpdateMatrix(const float matrix[16]) { m_streamTextureMatrix = WebKit::WebTransformationMatrix( matrix[0], matrix[1], matrix[2], matrix[3], matrix[4], matrix[5], matrix[6], matrix[7], matrix[8], matrix[9], matrix[10], matrix[11], matrix[12], matrix[13], matrix[14], matrix[15]); setNeedsRedraw(); } void CCVideoLayerImpl::didLoseContext() { freePlaneData(layerTreeHostImpl()->resourceProvider()); } void CCVideoLayerImpl::setNeedsRedraw() { layerTreeHostImpl()->setNeedsRedraw(); } void CCVideoLayerImpl::dumpLayerProperties(std::string* str, int indent) const { str->append(indentString(indent)); str->append("video layer\n"); CCLayerImpl::dumpLayerProperties(str, indent); } } #endif // USE(ACCELERATED_COMPOSITING)