1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
274
275
276
277
278
279
280
281
282
283
284
285
286
287
288
289
290
291
292
293
294
295
296
297
298
299
300
301
302
303
304
305
306
307
308
309
310
311
312
313
314
315
316
317
318
319
320
321
322
323
324
325
326
327
328
329
330
331
332
333
334
335
336
337
338
339
340
341
342
343
344
345
346
347
348
349
350
351
352
353
354
355
356
357
358
359
360
361
362
363
364
365
366
367
368
369
370
371
372
373
374
375
376
|
// Copyright (c) 2012 The Chromium Authors. All rights reserved.
// Use of this source code is governed by a BSD-style license that can be
// found in the LICENSE file.
#include "content/renderer/media/webrtc_audio_capturer.h"
#include "base/bind.h"
#include "base/logging.h"
#include "base/metrics/histogram.h"
#include "base/string_util.h"
#include "content/common/child_process.h"
#include "content/renderer/media/audio_device_factory.h"
#include "content/renderer/media/webrtc_audio_capturer_sink_owner.h"
#include "content/renderer/media/webrtc_audio_device_impl.h"
#include "media/audio/audio_util.h"
#include "media/audio/sample_rates.h"
namespace content {
// Supported hardware sample rates for input and output sides.
#if defined(OS_WIN) || defined(OS_MACOSX)
// media::GetAudioInputHardwareSampleRate() asks the audio layer
// for its current sample rate (set by the user) on Windows and Mac OS X.
// The listed rates below adds restrictions and WebRtcAudioDeviceImpl::Init()
// will fail if the user selects any rate outside these ranges.
static int kValidInputRates[] = {96000, 48000, 44100, 32000, 16000, 8000};
#elif defined(OS_LINUX) || defined(OS_OPENBSD)
static int kValidInputRates[] = {48000, 44100};
#elif defined(OS_ANDROID)
static int kValidInputRates[] = {48000, 44100};
#else
static int kValidInputRates[] = {44100};
#endif
static int GetBufferSizeForSampleRate(int sample_rate) {
int buffer_size = 0;
#if defined(OS_WIN) || defined(OS_MACOSX)
// Use a buffer size of 10ms.
buffer_size = (sample_rate / 100);
#elif defined(OS_LINUX) || defined(OS_OPENBSD)
// Based on tests using the current ALSA implementation in Chrome, we have
// found that the best combination is 20ms on the input side and 10ms on the
// output side.
buffer_size = 2 * sample_rate / 100;
#elif defined(OS_ANDROID)
// TODO(leozwang): Tune and adjust buffer size on Android.
buffer_size = 2 * sample_rate / 100;
#endif
return buffer_size;
}
// This is a temporary audio buffer with parameters used to send data to
// callbacks.
class WebRtcAudioCapturer::ConfiguredBuffer :
public base::RefCounted<WebRtcAudioCapturer::ConfiguredBuffer> {
public:
ConfiguredBuffer() {}
bool Initialize(int sample_rate,
media::ChannelLayout channel_layout) {
int buffer_size = GetBufferSizeForSampleRate(sample_rate);
DVLOG(1) << "Using WebRTC input buffer size: " << buffer_size;
media::AudioParameters::Format format =
media::AudioParameters::AUDIO_PCM_LOW_LATENCY;
// bits_per_sample is always 16 for now.
int bits_per_sample = 16;
int channels = ChannelLayoutToChannelCount(channel_layout);
params_.Reset(format, channel_layout, channels, 0,
sample_rate, bits_per_sample, buffer_size);
buffer_.reset(new int16[params_.frames_per_buffer() * params_.channels()]);
return true;
}
int16* buffer() const { return buffer_.get(); }
const media::AudioParameters& params() const { return params_; }
private:
~ConfiguredBuffer() {}
friend class base::RefCounted<WebRtcAudioCapturer::ConfiguredBuffer>;
scoped_ptr<int16[]> buffer_;
// Cached values of utilized audio parameters.
media::AudioParameters params_;
};
// static
scoped_refptr<WebRtcAudioCapturer> WebRtcAudioCapturer::CreateCapturer() {
scoped_refptr<WebRtcAudioCapturer> capturer = new WebRtcAudioCapturer();
return capturer;
}
bool WebRtcAudioCapturer::Reconfigure(int sample_rate,
media::ChannelLayout channel_layout) {
scoped_refptr<ConfiguredBuffer> new_buffer(new ConfiguredBuffer());
if (!new_buffer->Initialize(sample_rate, channel_layout))
return false;
TrackList tracks;
{
base::AutoLock auto_lock(lock_);
buffer_ = new_buffer;
tracks = tracks_;
}
// Tell all audio_tracks which format we use.
for (TrackList::const_iterator it = tracks.begin();
it != tracks.end(); ++it)
(*it)->SetCaptureFormat(new_buffer->params());
return true;
}
bool WebRtcAudioCapturer::Initialize(int render_view_id,
media::ChannelLayout channel_layout,
int sample_rate,
int session_id) {
DCHECK(thread_checker_.CalledOnValidThread());
DVLOG(1) << "WebRtcAudioCapturer::Initialize()";
DVLOG(1) << "Audio input hardware channel layout: " << channel_layout;
UMA_HISTOGRAM_ENUMERATION("WebRTC.AudioInputChannelLayout",
channel_layout, media::CHANNEL_LAYOUT_MAX);
session_id_ = session_id;
// Verify that the reported input channel configuration is supported.
if (channel_layout != media::CHANNEL_LAYOUT_MONO &&
channel_layout != media::CHANNEL_LAYOUT_STEREO) {
DLOG(ERROR) << channel_layout
<< " is not a supported input channel configuration.";
return false;
}
DVLOG(1) << "Audio input hardware sample rate: " << sample_rate;
UMA_HISTOGRAM_ENUMERATION("WebRTC.AudioInputSampleRate",
sample_rate, media::kUnexpectedAudioSampleRate);
// Verify that the reported input hardware sample rate is supported
// on the current platform.
if (std::find(&kValidInputRates[0],
&kValidInputRates[0] + arraysize(kValidInputRates),
sample_rate) ==
&kValidInputRates[arraysize(kValidInputRates)]) {
DLOG(ERROR) << sample_rate << " is not a supported input rate.";
return false;
}
if (!Reconfigure(sample_rate, channel_layout))
return false;
// Create and configure the default audio capturing source. The |source_|
// will be overwritten if an external client later calls SetCapturerSource()
// providing an alternative media::AudioCapturerSource.
SetCapturerSource(AudioDeviceFactory::NewInputDevice(render_view_id),
channel_layout,
static_cast<float>(sample_rate));
return true;
}
WebRtcAudioCapturer::WebRtcAudioCapturer()
: source_(NULL),
running_(false),
agc_is_enabled_(false),
session_id_(0) {
DVLOG(1) << "WebRtcAudioCapturer::WebRtcAudioCapturer()";
}
WebRtcAudioCapturer::~WebRtcAudioCapturer() {
DCHECK(thread_checker_.CalledOnValidThread());
DCHECK(tracks_.empty());
DCHECK(!running_);
DVLOG(1) << "WebRtcAudioCapturer::~WebRtcAudioCapturer()";
}
void WebRtcAudioCapturer::AddSink(
WebRtcAudioCapturerSink* track) {
DCHECK(thread_checker_.CalledOnValidThread());
DCHECK(track);
DVLOG(1) << "WebRtcAudioCapturer::AddSink()";
base::AutoLock auto_lock(lock_);
// Verify that |track| is not already added to the list.
DCHECK(std::find_if(
tracks_.begin(), tracks_.end(),
WebRtcAudioCapturerSinkOwner::WrapsSink(track)) == tracks_.end());
if (buffer_.get()) {
track->SetCaptureFormat(buffer_->params());
} else {
DLOG(WARNING) << "The format of the capturer has not been correctly "
<< "initialized";
}
// Create (and add to the list) a new WebRtcAudioCapturerSinkOwner which owns
// the |track| and delagates all calls to the WebRtcAudioCapturerSink
// interface.
tracks_.push_back(new WebRtcAudioCapturerSinkOwner(track));
// TODO(xians): should we call SetCapturerFormat() to each track?
}
void WebRtcAudioCapturer::RemoveSink(
WebRtcAudioCapturerSink* track) {
DCHECK(thread_checker_.CalledOnValidThread());
DVLOG(1) << "WebRtcAudioCapturer::RemoveSink()";
base::AutoLock auto_lock(lock_);
// Get iterator to the first element for which WrapsSink(track) returns true.
TrackList::iterator it = std::find_if(
tracks_.begin(), tracks_.end(),
WebRtcAudioCapturerSinkOwner::WrapsSink(track));
if (it != tracks_.end()) {
// Clear the delegate to ensure that no more capture callbacks will
// be sent to this sink. Also avoids a possible crash which can happen
// if this method is called while capturing is active.
(*it)->Reset();
tracks_.erase(it);
}
}
void WebRtcAudioCapturer::SetCapturerSource(
const scoped_refptr<media::AudioCapturerSource>& source,
media::ChannelLayout channel_layout,
float sample_rate) {
DCHECK(thread_checker_.CalledOnValidThread());
DVLOG(1) << "SetCapturerSource(channel_layout=" << channel_layout << ","
<< "sample_rate=" << sample_rate << ")";
scoped_refptr<media::AudioCapturerSource> old_source;
scoped_refptr<ConfiguredBuffer> current_buffer;
{
base::AutoLock auto_lock(lock_);
if (source_.get() == source.get())
return;
source_.swap(old_source);
source_ = source;
current_buffer = buffer_;
// Reset the flag to allow calling Start() for the new source.
running_ = false;
}
const bool no_default_audio_source_exists = !current_buffer.get();
// Detach the old source from normal recording or perform first-time
// initialization if Initialize() has never been called. For the second
// case, the caller is not "taking over an ongoing session" but instead
// "taking control over a new session".
if (old_source.get() || no_default_audio_source_exists) {
DVLOG(1) << "New capture source will now be utilized.";
if (old_source.get())
old_source->Stop();
// Dispatch the new parameters both to the sink(s) and to the new source.
// The idea is to get rid of any dependency of the microphone parameters
// which would normally be used by default.
if (!Reconfigure(sample_rate, channel_layout)) {
return;
} else {
// The buffer has been reconfigured. Update |current_buffer|.
base::AutoLock auto_lock(lock_);
current_buffer = buffer_;
}
}
if (source.get()) {
// Make sure to grab the new parameters in case they were reconfigured.
source->Initialize(current_buffer->params(), this, session_id_);
}
}
void WebRtcAudioCapturer::Start() {
DVLOG(1) << "WebRtcAudioCapturer::Start()";
base::AutoLock auto_lock(lock_);
if (running_)
return;
// Start the data source, i.e., start capturing data from the current source.
// Note that, the source does not have to be a microphone.
if (source_.get()) {
// We need to set the AGC control before starting the stream.
source_->SetAutomaticGainControl(agc_is_enabled_);
source_->Start();
}
running_ = true;
}
void WebRtcAudioCapturer::Stop() {
DVLOG(1) << "WebRtcAudioCapturer::Stop()";
scoped_refptr<media::AudioCapturerSource> source;
{
base::AutoLock auto_lock(lock_);
if (!running_)
return;
source = source_;
running_ = false;
}
if (source.get())
source->Stop();
}
void WebRtcAudioCapturer::SetVolume(double volume) {
DVLOG(1) << "WebRtcAudioCapturer::SetVolume()";
base::AutoLock auto_lock(lock_);
if (source_.get())
source_->SetVolume(volume);
}
void WebRtcAudioCapturer::SetAutomaticGainControl(bool enable) {
base::AutoLock auto_lock(lock_);
// Store the setting since SetAutomaticGainControl() can be called before
// Initialize(), in this case stored setting will be applied in Start().
agc_is_enabled_ = enable;
if (source_.get())
source_->SetAutomaticGainControl(enable);
}
void WebRtcAudioCapturer::Capture(media::AudioBus* audio_source,
int audio_delay_milliseconds,
double volume) {
// This callback is driven by AudioInputDevice::AudioThreadCallback if
// |source_| is AudioInputDevice, otherwise it is driven by client's
// CaptureCallback.
TrackList tracks;
scoped_refptr<ConfiguredBuffer> buffer_ref_while_calling;
{
base::AutoLock auto_lock(lock_);
if (!running_)
return;
// Copy the stuff we will need to local variables. In particular, we grab
// a reference to the buffer so we can ensure it stays alive even if the
// buffer is reconfigured while we are calling back.
buffer_ref_while_calling = buffer_;
tracks = tracks_;
}
int bytes_per_sample =
buffer_ref_while_calling->params().bits_per_sample() / 8;
// Interleave, scale, and clip input to int and store result in
// a local byte buffer.
audio_source->ToInterleaved(audio_source->frames(), bytes_per_sample,
buffer_ref_while_calling->buffer());
// Feed the data to the tracks.
for (TrackList::const_iterator it = tracks.begin();
it != tracks.end();
++it) {
(*it)->CaptureData(buffer_ref_while_calling->buffer(),
audio_source->channels(), audio_source->frames(),
audio_delay_milliseconds, volume);
}
}
void WebRtcAudioCapturer::OnCaptureError() {
NOTIMPLEMENTED();
}
media::AudioParameters WebRtcAudioCapturer::audio_parameters() const {
base::AutoLock auto_lock(lock_);
// |buffer_| can be NULL when SetCapturerSource() or Initialize() has not
// been called.
return buffer_.get() ? buffer_->params() : media::AudioParameters();
}
} // namespace content
|