summaryrefslogtreecommitdiffstats
path: root/net/base
diff options
context:
space:
mode:
authorwillchan@chromium.org <willchan@chromium.org@0039d316-1c4b-4281-b951-d872f2087c98>2010-03-02 23:19:33 +0000
committerwillchan@chromium.org <willchan@chromium.org@0039d316-1c4b-4281-b951-d872f2087c98>2010-03-02 23:19:33 +0000
commitb1f031ddf0723e5a4811bf9fa6fdda68c97633de (patch)
tree81be42f2b7c8cbf8e2bb5aaea1f1b1a6213413ca /net/base
parent136ba26595ceafeb57a5c4d524caf7f87b95daf3 (diff)
downloadchromium_src-b1f031ddf0723e5a4811bf9fa6fdda68c97633de.zip
chromium_src-b1f031ddf0723e5a4811bf9fa6fdda68c97633de.tar.gz
chromium_src-b1f031ddf0723e5a4811bf9fa6fdda68c97633de.tar.bz2
Use CHECK_* in net
Review URL: http://codereview.chromium.org/660404 git-svn-id: svn://svn.chromium.org/chrome/trunk/src@40453 0039d316-1c4b-4281-b951-d872f2087c98
Diffstat (limited to 'net/base')
-rw-r--r--net/base/host_resolver_impl.cc6
-rw-r--r--net/base/mock_host_resolver.cc2
-rw-r--r--net/base/network_change_notifier_win.cc2
-rw-r--r--net/base/sdch_filter_unittest.cc20
-rw-r--r--net/base/upload_data_stream.cc3
5 files changed, 16 insertions, 17 deletions
diff --git a/net/base/host_resolver_impl.cc b/net/base/host_resolver_impl.cc
index 184324b..6af3079 100644
--- a/net/base/host_resolver_impl.cc
+++ b/net/base/host_resolver_impl.cc
@@ -442,7 +442,7 @@ class HostResolverImpl::JobPool {
// specific meaning of these parameters.
void SetConstraints(size_t max_outstanding_jobs,
size_t max_pending_requests) {
- CHECK(max_outstanding_jobs != 0u);
+ CHECK_NE(max_outstanding_jobs, 0u);
max_outstanding_jobs_ = max_outstanding_jobs;
max_pending_requests_ = max_pending_requests;
}
@@ -816,8 +816,8 @@ scoped_refptr<LoadLog> HostResolverImpl::GetRequestsTrace() {
void HostResolverImpl::SetPoolConstraints(JobPoolIndex pool_index,
size_t max_outstanding_jobs,
size_t max_pending_requests) {
- CHECK(pool_index >= 0);
- CHECK(pool_index < POOL_COUNT);
+ CHECK_GE(pool_index, 0);
+ CHECK_LT(pool_index, POOL_COUNT);
CHECK(jobs_.empty()) << "Can only set constraints during setup";
JobPool* pool = job_pools_[pool_index];
pool->SetConstraints(max_outstanding_jobs, max_pending_requests);
diff --git a/net/base/mock_host_resolver.cc b/net/base/mock_host_resolver.cc
index 9a2c916..239275e 100644
--- a/net/base/mock_host_resolver.cc
+++ b/net/base/mock_host_resolver.cc
@@ -224,7 +224,7 @@ ScopedDefaultHostResolverProc::ScopedDefaultHostResolverProc(
ScopedDefaultHostResolverProc::~ScopedDefaultHostResolverProc() {
HostResolverProc* old_proc = HostResolverProc::SetDefault(previous_proc_);
// The lifetimes of multiple instances must be nested.
- CHECK(old_proc == current_proc_);
+ CHECK_EQ(old_proc, current_proc_);
}
void ScopedDefaultHostResolverProc::Init(HostResolverProc* proc) {
diff --git a/net/base/network_change_notifier_win.cc b/net/base/network_change_notifier_win.cc
index d139eb0f..8ae39fb 100644
--- a/net/base/network_change_notifier_win.cc
+++ b/net/base/network_change_notifier_win.cc
@@ -51,7 +51,7 @@ NetworkChangeNotifierWin::Impl::~Impl() {
void NetworkChangeNotifierWin::Impl::WatchForAddressChange() {
HANDLE handle = NULL;
DWORD ret = NotifyAddrChange(&handle, &addr_overlapped_);
- CHECK(ret == ERROR_IO_PENDING);
+ CHECK_EQ(ERROR_IO_PENDING, ret);
addr_watcher_.StartWatching(addr_overlapped_.hEvent, this);
}
diff --git a/net/base/sdch_filter_unittest.cc b/net/base/sdch_filter_unittest.cc
index 99a3ffd..675a135 100644
--- a/net/base/sdch_filter_unittest.cc
+++ b/net/base/sdch_filter_unittest.cc
@@ -111,7 +111,7 @@ static bool FilterTestData(const std::string& source,
size_t input_block_length,
const size_t output_buffer_length,
Filter* filter, std::string* output) {
- CHECK(input_block_length > 0);
+ CHECK_GT(input_block_length, 0u);
Filter::FilterStatus status(Filter::FILTER_NEED_MORE_DATA);
size_t source_index = 0;
scoped_array<char> output_buffer(new char[output_buffer_length]);
@@ -375,8 +375,8 @@ TEST_F(SdchFilterTest, BasicBadDictionary) {
// list of dictionaries, so the filter should error out immediately.
std::string dictionary_hash_postfix("4abcd\0", 6);
- CHECK(dictionary_hash_postfix.size() <
- static_cast<size_t>(input_buffer_size));
+ CHECK_LT(dictionary_hash_postfix.size(),
+ static_cast<size_t>(input_buffer_size));
memcpy(input_buffer, dictionary_hash_postfix.data(),
dictionary_hash_postfix.size());
filter->FlushStreamBuffer(dictionary_hash_postfix.size());
@@ -792,7 +792,7 @@ static std::string gzip_compress(const std::string &input) {
8, // DEF_MEM_LEVEL
Z_DEFAULT_STRATEGY);
- CHECK(code == Z_OK);
+ CHECK_EQ(Z_OK, code);
// Fill in zlib control block
zlib_stream.next_in = bit_cast<Bytef*>(input.data());
@@ -819,7 +819,7 @@ static std::string gzip_compress(const std::string &input) {
// Header value we generate:
const char kGZipHeader[] = { '\037', '\213', '\010', '\000', '\000',
'\000', '\000', '\000', '\002', '\377' };
- CHECK(zlib_stream.avail_out > sizeof(kGZipHeader));
+ CHECK_GT(zlib_stream.avail_out, sizeof(kGZipHeader));
memcpy(zlib_stream.next_out, kGZipHeader, sizeof(kGZipHeader));
zlib_stream.next_out += sizeof(kGZipHeader);
zlib_stream.avail_out -= sizeof(kGZipHeader);
@@ -860,9 +860,9 @@ TEST_F(SdchFilterTest, FilterChaining) {
// First try with a large buffer (larger than test input, or compressed data).
const size_t kLargeInputBufferSize(1000); // Used internally in filters.
- CHECK(kLargeInputBufferSize > gzip_compressed_sdch.size());
- CHECK(kLargeInputBufferSize > sdch_compressed.size());
- CHECK(kLargeInputBufferSize > expanded_.size());
+ CHECK_GT(kLargeInputBufferSize, gzip_compressed_sdch.size());
+ CHECK_GT(kLargeInputBufferSize, sdch_compressed.size());
+ CHECK_GT(kLargeInputBufferSize, expanded_.size());
MockFilterContext filter_context(kLargeInputBufferSize);
filter_context.SetURL(url);
scoped_ptr<Filter> filter(Filter::Factory(filter_types, filter_context));
@@ -884,11 +884,11 @@ TEST_F(SdchFilterTest, FilterChaining) {
// Next try with a mid-sized internal buffer size.
const size_t kMidSizedInputBufferSize(100);
// Buffer should be big enough to swallow whole gzip content.
- CHECK(kMidSizedInputBufferSize > gzip_compressed_sdch.size());
+ CHECK_GT(kMidSizedInputBufferSize, gzip_compressed_sdch.size());
// Buffer should be small enough that entire SDCH content can't fit.
// We'll go even further, and force the chain to flush the buffer between the
// two filters more than once (that is why we multiply by 2).
- CHECK(kMidSizedInputBufferSize * 2 < sdch_compressed.size());
+ CHECK_LT(kMidSizedInputBufferSize * 2, sdch_compressed.size());
filter_context.SetBufferSize(kMidSizedInputBufferSize);
filter_context.SetURL(url);
filter.reset(Filter::Factory(filter_types, filter_context));
diff --git a/net/base/upload_data_stream.cc b/net/base/upload_data_stream.cc
index 221ef28..21f4ce8 100644
--- a/net/base/upload_data_stream.cc
+++ b/net/base/upload_data_stream.cc
@@ -27,8 +27,7 @@ UploadDataStream::~UploadDataStream() {
}
void UploadDataStream::DidConsume(size_t num_bytes) {
- // TODO(vandebo): Change back to a DCHECK when issue 27870 is resolved.
- CHECK(num_bytes <= buf_len_);
+ DCHECK_LE(num_bytes, buf_len_);
DCHECK(!eof_);
buf_len_ -= num_bytes;