// Copyright (c) 2012 The Chromium Authors. All rights reserved. // Use of this source code is governed by a BSD-style license that can be // found in the LICENSE file. #ifndef CONTENT_RENDERER_SPEECH_RECOGNITION_DISPATCHER_H_ #define CONTENT_RENDERER_SPEECH_RECOGNITION_DISPATCHER_H_ #include #include "base/basictypes.h" #include "base/memory/scoped_ptr.h" #include "base/memory/shared_memory.h" #include "base/sync_socket.h" #include "content/public/common/speech_recognition_result.h" #include "content/public/renderer/render_view_observer.h" #include "third_party/WebKit/public/platform/WebMediaStreamTrack.h" #include "third_party/WebKit/public/platform/WebVector.h" #include "third_party/WebKit/public/web/WebSpeechRecognitionHandle.h" #include "third_party/WebKit/public/web/WebSpeechRecognizer.h" namespace media { class AudioParameters; } namespace content { class RenderViewImpl; #if defined(ENABLE_WEBRTC) class SpeechRecognitionAudioSink; #endif struct SpeechRecognitionError; struct SpeechRecognitionResult; // SpeechRecognitionDispatcher is a delegate for methods used by WebKit for // scripted JS speech APIs. It's the complement of // SpeechRecognitionDispatcherHost (owned by RenderViewHost). class SpeechRecognitionDispatcher : public RenderViewObserver, public blink::WebSpeechRecognizer { public: explicit SpeechRecognitionDispatcher(RenderViewImpl* render_view); virtual ~SpeechRecognitionDispatcher(); // Aborts all speech recognitions. void AbortAllRecognitions(); private: // RenderViewObserver implementation. bool OnMessageReceived(const IPC::Message& message) override; // blink::WebSpeechRecognizer implementation. virtual void start(const blink::WebSpeechRecognitionHandle&, const blink::WebSpeechRecognitionParams&, blink::WebSpeechRecognizerClient*); virtual void stop(const blink::WebSpeechRecognitionHandle&, blink::WebSpeechRecognizerClient*); virtual void abort(const blink::WebSpeechRecognitionHandle&, blink::WebSpeechRecognizerClient*); void OnRecognitionStarted(int request_id); void OnAudioStarted(int request_id); void OnSoundStarted(int request_id); void OnSoundEnded(int request_id); void OnAudioEnded(int request_id); void OnErrorOccurred(int request_id, const SpeechRecognitionError& error); void OnRecognitionEnded(int request_id); void OnResultsRetrieved(int request_id, const SpeechRecognitionResults& result); void OnAudioReceiverReady(int session_id, const media::AudioParameters& params, const base::SharedMemoryHandle handle, const base::SyncSocket::TransitDescriptor socket); void ResetAudioSink(); int GetOrCreateIDForHandle(const blink::WebSpeechRecognitionHandle& handle); bool HandleExists(const blink::WebSpeechRecognitionHandle& handle); const blink::WebSpeechRecognitionHandle& GetHandleFromID(int handle_id); // The WebKit client class that we use to send events back to the JS world. blink::WebSpeechRecognizerClient* recognizer_client_; #if defined(ENABLE_WEBRTC) // Media stream audio track that the speech recognition connects to. // Accessed on the render thread. blink::WebMediaStreamTrack audio_track_; // Audio sink used to provide audio from the track. scoped_ptr speech_audio_sink_; #endif typedef std::map HandleMap; HandleMap handle_map_; int next_id_; DISALLOW_COPY_AND_ASSIGN(SpeechRecognitionDispatcher); }; } // namespace content #endif // CONTENT_RENDERER_SPEECH_RECOGNITION_DISPATCHER_H_