// Copyright (c) 2012 The Chromium Authors. All rights reserved. // Use of this source code is governed by a BSD-style license that can be // found in the LICENSE file. #include "media/video/capture/win/video_capture_device_win.h" #include #include #include "base/command_line.h" #include "base/strings/string_util.h" #include "base/strings/sys_string_conversions.h" #include "base/win/metro.h" #include "base/win/scoped_co_mem.h" #include "base/win/scoped_variant.h" #include "media/base/media_switches.h" #include "media/video/capture/win/video_capture_device_mf_win.h" using base::win::ScopedCoMem; using base::win::ScopedComPtr; using base::win::ScopedVariant; namespace media { namespace { // Finds and creates a DirectShow Video Capture filter matching the device_name. HRESULT GetDeviceFilter(const VideoCaptureDevice::Name& device_name, IBaseFilter** filter) { DCHECK(filter); ScopedComPtr dev_enum; HRESULT hr = dev_enum.CreateInstance(CLSID_SystemDeviceEnum, NULL, CLSCTX_INPROC); if (FAILED(hr)) return hr; ScopedComPtr enum_moniker; hr = dev_enum->CreateClassEnumerator(CLSID_VideoInputDeviceCategory, enum_moniker.Receive(), 0); // CreateClassEnumerator returns S_FALSE on some Windows OS // when no camera exist. Therefore the FAILED macro can't be used. if (hr != S_OK) return NULL; ScopedComPtr moniker; ScopedComPtr capture_filter; DWORD fetched = 0; while (enum_moniker->Next(1, moniker.Receive(), &fetched) == S_OK) { ScopedComPtr prop_bag; hr = moniker->BindToStorage(0, 0, IID_IPropertyBag, prop_bag.ReceiveVoid()); if (FAILED(hr)) { moniker.Release(); continue; } // Find the description or friendly name. static const wchar_t* kPropertyNames[] = { L"DevicePath", L"Description", L"FriendlyName" }; ScopedVariant name; for (size_t i = 0; i < arraysize(kPropertyNames) && name.type() != VT_BSTR; ++i) { prop_bag->Read(kPropertyNames[i], name.Receive(), 0); } if (name.type() == VT_BSTR) { std::string device_path(base::SysWideToUTF8(V_BSTR(&name))); if (device_path.compare(device_name.id()) == 0) { // We have found the requested device hr = moniker->BindToObject(0, 0, IID_IBaseFilter, capture_filter.ReceiveVoid()); DVPLOG_IF(2, FAILED(hr)) << "Failed to bind camera filter."; break; } } moniker.Release(); } *filter = capture_filter.Detach(); if (!*filter && SUCCEEDED(hr)) hr = HRESULT_FROM_WIN32(ERROR_NOT_FOUND); return hr; } // Check if a Pin matches a category. bool PinMatchesCategory(IPin* pin, REFGUID category) { DCHECK(pin); bool found = false; ScopedComPtr ks_property; HRESULT hr = ks_property.QueryFrom(pin); if (SUCCEEDED(hr)) { GUID pin_category; DWORD return_value; hr = ks_property->Get(AMPROPSETID_Pin, AMPROPERTY_PIN_CATEGORY, NULL, 0, &pin_category, sizeof(pin_category), &return_value); if (SUCCEEDED(hr) && (return_value == sizeof(pin_category))) { found = (pin_category == category); } } return found; } // Finds a IPin on a IBaseFilter given the direction an category. HRESULT GetPin(IBaseFilter* filter, PIN_DIRECTION pin_dir, REFGUID category, IPin** pin) { DCHECK(pin); ScopedComPtr pin_emum; HRESULT hr = filter->EnumPins(pin_emum.Receive()); if (pin_emum == NULL) return hr; // Get first unconnected pin. hr = pin_emum->Reset(); // set to first pin while ((hr = pin_emum->Next(1, pin, NULL)) == S_OK) { PIN_DIRECTION this_pin_dir = static_cast(-1); hr = (*pin)->QueryDirection(&this_pin_dir); if (pin_dir == this_pin_dir) { if (category == GUID_NULL || PinMatchesCategory(*pin, category)) return S_OK; } (*pin)->Release(); } return E_FAIL; } // Release the format block for a media type. // http://msdn.microsoft.com/en-us/library/dd375432(VS.85).aspx void FreeMediaType(AM_MEDIA_TYPE* mt) { if (mt->cbFormat != 0) { CoTaskMemFree(mt->pbFormat); mt->cbFormat = 0; mt->pbFormat = NULL; } if (mt->pUnk != NULL) { NOTREACHED(); // pUnk should not be used. mt->pUnk->Release(); mt->pUnk = NULL; } } // Delete a media type structure that was allocated on the heap. // http://msdn.microsoft.com/en-us/library/dd375432(VS.85).aspx void DeleteMediaType(AM_MEDIA_TYPE* mt) { if (mt != NULL) { FreeMediaType(mt); CoTaskMemFree(mt); } } } // namespace // static void VideoCaptureDevice::GetDeviceNames(Names* device_names) { const CommandLine* cmd_line = CommandLine::ForCurrentProcess(); // Use Media Foundation for Metro processes (after and including Win8) // and DirectShow for any other platforms. if (base::win::IsMetroProcess() && !cmd_line->HasSwitch(switches::kForceDirectShowVideoCapture)) { VideoCaptureDeviceMFWin::GetDeviceNames(device_names); } else { VideoCaptureDeviceWin::GetDeviceNames(device_names); } } // static void VideoCaptureDevice::GetDeviceSupportedFormats(const Name& device, VideoCaptureFormats* formats) { NOTIMPLEMENTED(); } // static VideoCaptureDevice* VideoCaptureDevice::Create(const Name& device_name) { VideoCaptureDevice* ret = NULL; if (device_name.capture_api_type() == Name::MEDIA_FOUNDATION) { DCHECK(VideoCaptureDeviceMFWin::PlatformSupported()); scoped_ptr device( new VideoCaptureDeviceMFWin(device_name)); DVLOG(1) << " MediaFoundation Device: " << device_name.name(); if (device->Init()) ret = device.release(); } else if (device_name.capture_api_type() == Name::DIRECT_SHOW) { scoped_ptr device( new VideoCaptureDeviceWin(device_name)); DVLOG(1) << " DirectShow Device: " << device_name.name(); if (device->Init()) ret = device.release(); } else{ NOTREACHED() << " Couldn't recognize VideoCaptureDevice type"; } return ret; } // static void VideoCaptureDeviceWin::GetDeviceNames(Names* device_names) { DCHECK(device_names); ScopedComPtr dev_enum; HRESULT hr = dev_enum.CreateInstance(CLSID_SystemDeviceEnum, NULL, CLSCTX_INPROC); if (FAILED(hr)) return; ScopedComPtr enum_moniker; hr = dev_enum->CreateClassEnumerator(CLSID_VideoInputDeviceCategory, enum_moniker.Receive(), 0); // CreateClassEnumerator returns S_FALSE on some Windows OS // when no camera exist. Therefore the FAILED macro can't be used. if (hr != S_OK) return; device_names->clear(); // Name of a fake DirectShow filter that exist on computers with // GTalk installed. static const char kGoogleCameraAdapter[] = "google camera adapter"; // Enumerate all video capture devices. ScopedComPtr moniker; int index = 0; while (enum_moniker->Next(1, moniker.Receive(), NULL) == S_OK) { ScopedComPtr prop_bag; hr = moniker->BindToStorage(0, 0, IID_IPropertyBag, prop_bag.ReceiveVoid()); if (FAILED(hr)) { moniker.Release(); continue; } // Find the description or friendly name. ScopedVariant name; hr = prop_bag->Read(L"Description", name.Receive(), 0); if (FAILED(hr)) hr = prop_bag->Read(L"FriendlyName", name.Receive(), 0); if (SUCCEEDED(hr) && name.type() == VT_BSTR) { // Ignore all VFW drivers and the special Google Camera Adapter. // Google Camera Adapter is not a real DirectShow camera device. // VFW is very old Video for Windows drivers that can not be used. const wchar_t* str_ptr = V_BSTR(&name); const int name_length = arraysize(kGoogleCameraAdapter) - 1; if ((wcsstr(str_ptr, L"(VFW)") == NULL) && lstrlenW(str_ptr) < name_length || (!(LowerCaseEqualsASCII(str_ptr, str_ptr + name_length, kGoogleCameraAdapter)))) { std::string id; std::string device_name(base::SysWideToUTF8(str_ptr)); name.Reset(); hr = prop_bag->Read(L"DevicePath", name.Receive(), 0); if (FAILED(hr) || name.type() != VT_BSTR) { id = device_name; } else { DCHECK_EQ(name.type(), VT_BSTR); id = base::SysWideToUTF8(V_BSTR(&name)); } device_names->push_back(Name(device_name, id, Name::DIRECT_SHOW)); } } moniker.Release(); } } VideoCaptureDeviceWin::VideoCaptureDeviceWin(const Name& device_name) : device_name_(device_name), state_(kIdle) { DetachFromThread(); } VideoCaptureDeviceWin::~VideoCaptureDeviceWin() { DCHECK(CalledOnValidThread()); if (media_control_) media_control_->Stop(); if (graph_builder_) { if (sink_filter_) { graph_builder_->RemoveFilter(sink_filter_); sink_filter_ = NULL; } if (capture_filter_) graph_builder_->RemoveFilter(capture_filter_); if (mjpg_filter_) graph_builder_->RemoveFilter(mjpg_filter_); } } bool VideoCaptureDeviceWin::Init() { DCHECK(CalledOnValidThread()); HRESULT hr = GetDeviceFilter(device_name_, capture_filter_.Receive()); if (!capture_filter_) { DVLOG(2) << "Failed to create capture filter."; return false; } hr = GetPin(capture_filter_, PINDIR_OUTPUT, PIN_CATEGORY_CAPTURE, output_capture_pin_.Receive()); if (!output_capture_pin_) { DVLOG(2) << "Failed to get capture output pin"; return false; } // Create the sink filter used for receiving Captured frames. sink_filter_ = new SinkFilter(this); if (sink_filter_ == NULL) { DVLOG(2) << "Failed to create send filter"; return false; } input_sink_pin_ = sink_filter_->GetPin(0); hr = graph_builder_.CreateInstance(CLSID_FilterGraph, NULL, CLSCTX_INPROC_SERVER); if (FAILED(hr)) { DVLOG(2) << "Failed to create graph builder."; return false; } hr = graph_builder_.QueryInterface(media_control_.Receive()); if (FAILED(hr)) { DVLOG(2) << "Failed to create media control builder."; return false; } hr = graph_builder_->AddFilter(capture_filter_, NULL); if (FAILED(hr)) { DVLOG(2) << "Failed to add the capture device to the graph."; return false; } hr = graph_builder_->AddFilter(sink_filter_, NULL); if (FAILED(hr)) { DVLOG(2)<< "Failed to add the send filter to the graph."; return false; } return CreateCapabilityMap(); } void VideoCaptureDeviceWin::AllocateAndStart( const VideoCaptureParams& params, scoped_ptr client) { DCHECK(CalledOnValidThread()); if (state_ != kIdle) return; client_ = client.Pass(); // Get the camera capability that best match the requested resolution. const VideoCaptureCapabilityWin& found_capability = capabilities_.GetBestMatchedFormat( params.requested_format.frame_size.width(), params.requested_format.frame_size.height(), params.requested_format.frame_rate); VideoCaptureFormat format = found_capability.supported_format; // Reduce the frame rate if the requested frame rate is lower // than the capability. if (format.frame_rate > params.requested_format.frame_rate) format.frame_rate = params.requested_format.frame_rate; AM_MEDIA_TYPE* pmt = NULL; VIDEO_STREAM_CONFIG_CAPS caps; ScopedComPtr stream_config; HRESULT hr = output_capture_pin_.QueryInterface(stream_config.Receive()); if (FAILED(hr)) { SetErrorState("Can't get the Capture format settings"); return; } // Get the windows capability from the capture device. hr = stream_config->GetStreamCaps(found_capability.stream_index, &pmt, reinterpret_cast(&caps)); if (SUCCEEDED(hr)) { if (pmt->formattype == FORMAT_VideoInfo) { VIDEOINFOHEADER* h = reinterpret_cast(pmt->pbFormat); if (format.frame_rate > 0) h->AvgTimePerFrame = kSecondsToReferenceTime / format.frame_rate; } // Set the sink filter to request this format. sink_filter_->SetRequestedMediaFormat(format); // Order the capture device to use this format. hr = stream_config->SetFormat(pmt); } if (FAILED(hr)) SetErrorState("Failed to set capture device output format"); if (format.pixel_format == PIXEL_FORMAT_MJPEG && !mjpg_filter_.get()) { // Create MJPG filter if we need it. hr = mjpg_filter_.CreateInstance(CLSID_MjpegDec, NULL, CLSCTX_INPROC); if (SUCCEEDED(hr)) { GetPin(mjpg_filter_, PINDIR_INPUT, GUID_NULL, input_mjpg_pin_.Receive()); GetPin(mjpg_filter_, PINDIR_OUTPUT, GUID_NULL, output_mjpg_pin_.Receive()); hr = graph_builder_->AddFilter(mjpg_filter_, NULL); } if (FAILED(hr)) { mjpg_filter_.Release(); input_mjpg_pin_.Release(); output_mjpg_pin_.Release(); } } if (format.pixel_format == PIXEL_FORMAT_MJPEG && mjpg_filter_.get()) { // Connect the camera to the MJPEG decoder. hr = graph_builder_->ConnectDirect(output_capture_pin_, input_mjpg_pin_, NULL); // Connect the MJPEG filter to the Capture filter. hr += graph_builder_->ConnectDirect(output_mjpg_pin_, input_sink_pin_, NULL); } else { hr = graph_builder_->ConnectDirect(output_capture_pin_, input_sink_pin_, NULL); } if (FAILED(hr)) { SetErrorState("Failed to connect the Capture graph."); return; } hr = media_control_->Pause(); if (FAILED(hr)) { SetErrorState("Failed to Pause the Capture device. " "Is it already occupied?"); return; } // Get the format back from the sink filter after the filter have been // connected. capture_format_ = sink_filter_->ResultingFormat(); // Start capturing. hr = media_control_->Run(); if (FAILED(hr)) { SetErrorState("Failed to start the Capture device."); return; } state_ = kCapturing; } void VideoCaptureDeviceWin::StopAndDeAllocate() { DCHECK(CalledOnValidThread()); if (state_ != kCapturing) return; HRESULT hr = media_control_->Stop(); if (FAILED(hr)) { SetErrorState("Failed to stop the capture graph."); return; } graph_builder_->Disconnect(output_capture_pin_); graph_builder_->Disconnect(input_sink_pin_); // If the _mjpg filter exist disconnect it even if it has not been used. if (mjpg_filter_) { graph_builder_->Disconnect(input_mjpg_pin_); graph_builder_->Disconnect(output_mjpg_pin_); } if (FAILED(hr)) { SetErrorState("Failed to Stop the Capture device"); return; } client_.reset(); state_ = kIdle; } // Implements SinkFilterObserver::SinkFilterObserver. void VideoCaptureDeviceWin::FrameReceived(const uint8* buffer, int length) { client_->OnIncomingCapturedFrame( buffer, length, base::TimeTicks::Now(), 0, capture_format_); } bool VideoCaptureDeviceWin::CreateCapabilityMap() { DCHECK(CalledOnValidThread()); ScopedComPtr stream_config; HRESULT hr = output_capture_pin_.QueryInterface(stream_config.Receive()); if (FAILED(hr)) { DVLOG(2) << "Failed to get IAMStreamConfig interface from " "capture device"; return false; } // Get interface used for getting the frame rate. ScopedComPtr video_control; hr = capture_filter_.QueryInterface(video_control.Receive()); DVLOG_IF(2, FAILED(hr)) << "IAMVideoControl Interface NOT SUPPORTED"; AM_MEDIA_TYPE* media_type = NULL; VIDEO_STREAM_CONFIG_CAPS caps; int count, size; hr = stream_config->GetNumberOfCapabilities(&count, &size); if (FAILED(hr)) { DVLOG(2) << "Failed to GetNumberOfCapabilities"; return false; } for (int i = 0; i < count; ++i) { hr = stream_config->GetStreamCaps(i, &media_type, reinterpret_cast(&caps)); // GetStreamCaps() may return S_FALSE, so don't use FAILED() or SUCCEED() // macros here since they'll trigger incorrectly. if (hr != S_OK) { DVLOG(2) << "Failed to GetStreamCaps"; return false; } if (media_type->majortype == MEDIATYPE_Video && media_type->formattype == FORMAT_VideoInfo) { VideoCaptureCapabilityWin capability(i); VIDEOINFOHEADER* h = reinterpret_cast(media_type->pbFormat); capability.supported_format.frame_size.SetSize(h->bmiHeader.biWidth, h->bmiHeader.biHeight); // Try to get a better |time_per_frame| from IAMVideoControl. If not, use // the value from VIDEOINFOHEADER. REFERENCE_TIME time_per_frame = h->AvgTimePerFrame; if (video_control) { ScopedCoMem max_fps; LONG list_size = 0; SIZE size = {capability.supported_format.frame_size.width(), capability.supported_format.frame_size.height()}; // GetFrameRateList doesn't return max frame rate always // eg: Logitech Notebook. This may be due to a bug in that API // because GetFrameRateList array is reversed in the above camera. So // a util method written. Can't assume the first value will return // the max fps. hr = video_control->GetFrameRateList(output_capture_pin_, i, size, &list_size, &max_fps); // Sometimes |list_size| will be > 0, but max_fps will be NULL. Some // drivers may return an HRESULT of S_FALSE which SUCCEEDED() translates // into success, so explicitly check S_OK. See http://crbug.com/306237. if (hr == S_OK && list_size > 0 && max_fps) { time_per_frame = *std::min_element(max_fps.get(), max_fps.get() + list_size); } } capability.supported_format.frame_rate = (time_per_frame > 0) ? static_cast(kSecondsToReferenceTime / time_per_frame) : 0; // DirectShow works at the moment only on integer frame_rate but the // best capability matching class works on rational frame rates. capability.frame_rate_numerator = capability.supported_format.frame_rate; capability.frame_rate_denominator = 1; // We can't switch MEDIATYPE :~(. if (media_type->subtype == kMediaSubTypeI420) { capability.supported_format.pixel_format = PIXEL_FORMAT_I420; } else if (media_type->subtype == MEDIASUBTYPE_IYUV) { // This is identical to PIXEL_FORMAT_I420. capability.supported_format.pixel_format = PIXEL_FORMAT_I420; } else if (media_type->subtype == MEDIASUBTYPE_RGB24) { capability.supported_format.pixel_format = PIXEL_FORMAT_RGB24; } else if (media_type->subtype == MEDIASUBTYPE_YUY2) { capability.supported_format.pixel_format = PIXEL_FORMAT_YUY2; } else if (media_type->subtype == MEDIASUBTYPE_MJPG) { capability.supported_format.pixel_format = PIXEL_FORMAT_MJPEG; } else if (media_type->subtype == MEDIASUBTYPE_UYVY) { capability.supported_format.pixel_format = PIXEL_FORMAT_UYVY; } else if (media_type->subtype == MEDIASUBTYPE_ARGB32) { capability.supported_format.pixel_format = PIXEL_FORMAT_ARGB; } else { WCHAR guid_str[128]; StringFromGUID2(media_type->subtype, guid_str, arraysize(guid_str)); DVLOG(2) << "Device supports (also) an unknown media type " << guid_str; continue; } capabilities_.Add(capability); } DeleteMediaType(media_type); media_type = NULL; } return !capabilities_.empty(); } void VideoCaptureDeviceWin::SetErrorState(const std::string& reason) { DCHECK(CalledOnValidThread()); DVLOG(1) << reason; state_ = kError; client_->OnError(reason); } } // namespace media