diff options
author | eroman@chromium.org <eroman@chromium.org@0039d316-1c4b-4281-b951-d872f2087c98> | 2010-02-19 20:24:06 +0000 |
---|---|---|
committer | eroman@chromium.org <eroman@chromium.org@0039d316-1c4b-4281-b951-d872f2087c98> | 2010-02-19 20:24:06 +0000 |
commit | 7541206c7a5160f3489d563b97f1c841c853dc22 (patch) | |
tree | b69621265589060c0577559c23c86db4de667191 /net/proxy/proxy_bypass_rules_unittest.cc | |
parent | d68a04da3be6a4a5db3768f53b2b48735a6ec210 (diff) | |
download | chromium_src-7541206c7a5160f3489d563b97f1c841c853dc22.zip chromium_src-7541206c7a5160f3489d563b97f1c841c853dc22.tar.gz chromium_src-7541206c7a5160f3489d563b97f1c841c853dc22.tar.bz2 |
Split out the handling of proxy bypass rules into ProxyBypassRules. There are some pretty complicated rules, and this helps isolate that code and better test it.
This also lays a framework for addressing bug 9835 (IP/CIDR matching)
Lastly, adds support for the exclusion format ".domain" on all platforms, which is interpreted as "*.domain".
BUG=28112
TEST=ProxyBypassRulesTest.*
Review URL: http://codereview.chromium.org/601070
git-svn-id: svn://svn.chromium.org/chrome/trunk/src@39486 0039d316-1c4b-4281-b951-d872f2087c98
Diffstat (limited to 'net/proxy/proxy_bypass_rules_unittest.cc')
-rw-r--r-- | net/proxy/proxy_bypass_rules_unittest.cc | 286 |
1 files changed, 286 insertions, 0 deletions
diff --git a/net/proxy/proxy_bypass_rules_unittest.cc b/net/proxy/proxy_bypass_rules_unittest.cc new file mode 100644 index 0000000..1bf24cb --- /dev/null +++ b/net/proxy/proxy_bypass_rules_unittest.cc @@ -0,0 +1,286 @@ +// Copyright (c) 2010 The Chromium Authors. All rights reserved. +// Use of this source code is governed by a BSD-style license that can be +// found in the LICENSE file. + +#include "net/proxy/proxy_bypass_rules.h" + +#include "base/string_util.h" +#include "net/proxy/proxy_config_service_common_unittest.h" +#include "testing/gtest/include/gtest/gtest.h" + +namespace net { + +namespace { + +TEST(ProxyBypassRulesTest, ParseAndMatchBasicHost) { + ProxyBypassRules rules; + rules.ParseFromString("wWw.gOogle.com"); + ASSERT_EQ(1u, rules.rules().size()); + EXPECT_EQ("www.google.com", rules.rules()[0]->ToString()); + + // All of these match; port, scheme, and non-hostname components don't + // matter. + EXPECT_TRUE(rules.Matches(GURL("http://www.google.com"))); + EXPECT_TRUE(rules.Matches(GURL("ftp://www.google.com:99"))); + EXPECT_TRUE(rules.Matches(GURL("https://www.google.com:81"))); + + // Must be a strict host match to work. + EXPECT_FALSE(rules.Matches(GURL("http://foo.www.google.com"))); + EXPECT_FALSE(rules.Matches(GURL("http://xxx.google.com"))); + EXPECT_FALSE(rules.Matches(GURL("http://google.com"))); + EXPECT_FALSE(rules.Matches(GURL("http://www.google.com.baz.org"))); +} + +TEST(ProxyBypassRulesTest, ParseAndMatchBasicDomain) { + ProxyBypassRules rules; + rules.ParseFromString(".gOOgle.com"); + ASSERT_EQ(1u, rules.rules().size()); + // Note that we inferred this was an "ends with" test. + EXPECT_EQ("*.google.com", rules.rules()[0]->ToString()); + + // All of these match; port, scheme, and non-hostname components don't + // matter. + EXPECT_TRUE(rules.Matches(GURL("http://www.google.com"))); + EXPECT_TRUE(rules.Matches(GURL("ftp://www.google.com:99"))); + EXPECT_TRUE(rules.Matches(GURL("https://a.google.com:81"))); + EXPECT_TRUE(rules.Matches(GURL("http://foo.google.com/x/y?q"))); + EXPECT_TRUE(rules.Matches(GURL("http://foo:bar@baz.google.com#x"))); + + // Must be a strict "ends with" to work. + EXPECT_FALSE(rules.Matches(GURL("http://google.com"))); + EXPECT_FALSE(rules.Matches(GURL("http://foo.google.com.baz.org"))); +} + +TEST(ProxyBypassRulesTest, ParseAndMatchBasicDomainWithPort) { + ProxyBypassRules rules; + rules.ParseFromString("*.GOOGLE.com:80"); + ASSERT_EQ(1u, rules.rules().size()); + EXPECT_EQ("*.google.com:80", rules.rules()[0]->ToString()); + + // All of these match; scheme, and non-hostname components don't matter. + EXPECT_TRUE(rules.Matches(GURL("http://www.google.com"))); + EXPECT_TRUE(rules.Matches(GURL("ftp://www.google.com:80"))); + EXPECT_TRUE(rules.Matches(GURL("https://a.google.com:80?x"))); + + // Must be a strict "ends with" to work. + EXPECT_FALSE(rules.Matches(GURL("http://google.com"))); + EXPECT_FALSE(rules.Matches(GURL("http://foo.google.com.baz.org"))); + + // The ports must match. + EXPECT_FALSE(rules.Matches(GURL("http://www.google.com:90"))); + EXPECT_FALSE(rules.Matches(GURL("https://www.google.com"))); +} + +TEST(ProxyBypassRulesTest, MatchAll) { + ProxyBypassRules rules; + rules.ParseFromString("*"); + ASSERT_EQ(1u, rules.rules().size()); + EXPECT_EQ("*", rules.rules()[0]->ToString()); + + EXPECT_TRUE(rules.Matches(GURL("http://www.google.com"))); + EXPECT_TRUE(rules.Matches(GURL("ftp://www.foobar.com:99"))); + EXPECT_TRUE(rules.Matches(GURL("https://a.google.com:80?x"))); +} + +TEST(ProxyBypassRulesTest, WildcardAtStart) { + ProxyBypassRules rules; + rules.ParseFromString("*.org:443"); + ASSERT_EQ(1u, rules.rules().size()); + EXPECT_EQ("*.org:443", rules.rules()[0]->ToString()); + + EXPECT_TRUE(rules.Matches(GURL("http://www.google.org:443"))); + EXPECT_TRUE(rules.Matches(GURL("https://www.google.org"))); + + EXPECT_FALSE(rules.Matches(GURL("http://www.google.org"))); + EXPECT_FALSE(rules.Matches(GURL("https://www.google.com"))); + EXPECT_FALSE(rules.Matches(GURL("https://www.google.org.com"))); +} + +TEST(ProxyBypassRulesTest, IPV4Address) { + ProxyBypassRules rules; + rules.ParseFromString("192.168.1.1"); + ASSERT_EQ(1u, rules.rules().size()); + EXPECT_EQ("192.168.1.1", rules.rules()[0]->ToString()); + + EXPECT_TRUE(rules.Matches(GURL("http://192.168.1.1"))); + EXPECT_TRUE(rules.Matches(GURL("https://192.168.1.1:90"))); + + EXPECT_FALSE(rules.Matches(GURL("http://www.google.com"))); + EXPECT_FALSE(rules.Matches(GURL("http://sup.192.168.1.1"))); +} + +TEST(ProxyBypassRulesTest, IPV4AddressWithPort) { + ProxyBypassRules rules; + rules.ParseFromString("192.168.1.1:33"); + ASSERT_EQ(1u, rules.rules().size()); + EXPECT_EQ("192.168.1.1:33", rules.rules()[0]->ToString()); + + EXPECT_TRUE(rules.Matches(GURL("http://192.168.1.1:33"))); + + EXPECT_FALSE(rules.Matches(GURL("http://www.google.com"))); + EXPECT_FALSE(rules.Matches(GURL("http://192.168.1.1"))); + EXPECT_FALSE(rules.Matches(GURL("http://sup.192.168.1.1:33"))); +} + +TEST(ProxyBypassRulesTest, IPV6Address) { + ProxyBypassRules rules; + rules.ParseFromString("[3ffe:2a00:100:7031:0:0::1]"); + ASSERT_EQ(1u, rules.rules().size()); + // Note that we canonicalized the IP address. + EXPECT_EQ("[3ffe:2a00:100:7031::1]", rules.rules()[0]->ToString()); + + EXPECT_TRUE(rules.Matches(GURL("http://[3ffe:2a00:100:7031::1]"))); + EXPECT_TRUE(rules.Matches(GURL("http://[3ffe:2a00:100:7031::1]:33"))); + + EXPECT_FALSE(rules.Matches(GURL("http://www.google.com"))); + EXPECT_FALSE(rules.Matches(GURL("http://sup.192.168.1.1:33"))); +} + +TEST(ProxyBypassRulesTest, IPV6AddressWithPort) { + ProxyBypassRules rules; + rules.ParseFromString("[3ffe:2a00:100:7031::1]:33"); + ASSERT_EQ(1u, rules.rules().size()); + EXPECT_EQ("[3ffe:2a00:100:7031::1]:33", rules.rules()[0]->ToString()); + + EXPECT_TRUE(rules.Matches(GURL("http://[3ffe:2a00:100:7031::1]:33"))); + + EXPECT_FALSE(rules.Matches(GURL("http://[3ffe:2a00:100:7031::1]"))); + EXPECT_FALSE(rules.Matches(GURL("http://www.google.com"))); +} + +TEST(ProxyBypassRulesTest, HTTPOnly) { + ProxyBypassRules rules; + rules.ParseFromString("http://www.google.com"); + ASSERT_EQ(1u, rules.rules().size()); + EXPECT_EQ("http://www.google.com", rules.rules()[0]->ToString()); + + EXPECT_TRUE(rules.Matches(GURL("http://www.google.com/foo"))); + EXPECT_TRUE(rules.Matches(GURL("http://www.google.com:99"))); + + EXPECT_FALSE(rules.Matches(GURL("https://www.google.com"))); + EXPECT_FALSE(rules.Matches(GURL("ftp://www.google.com"))); + EXPECT_FALSE(rules.Matches(GURL("http://foo.www.google.com"))); + EXPECT_FALSE(rules.Matches(GURL("http://www.google.com.org"))); + EXPECT_FALSE(rules.Matches(GURL("https://www.google.com"))); +} + +TEST(ProxyBypassRulesTest, HTTPOnlyWithWildcard) { + ProxyBypassRules rules; + rules.ParseFromString("http://*www.google.com"); + ASSERT_EQ(1u, rules.rules().size()); + EXPECT_EQ("http://*www.google.com", rules.rules()[0]->ToString()); + + EXPECT_TRUE(rules.Matches(GURL("http://www.google.com/foo"))); + EXPECT_TRUE(rules.Matches(GURL("http://www.google.com:99"))); + EXPECT_TRUE(rules.Matches(GURL("http://foo.www.google.com"))); + + EXPECT_FALSE(rules.Matches(GURL("https://www.google.com"))); + EXPECT_FALSE(rules.Matches(GURL("ftp://www.google.com"))); + EXPECT_FALSE(rules.Matches(GURL("http://www.google.com.org"))); + EXPECT_FALSE(rules.Matches(GURL("https://www.google.com"))); +} + +TEST(ProxyBypassRulesTest, UseSuffixMatching) { + ProxyBypassRules rules; + rules.ParseFromStringUsingSuffixMatching( + "foo1.com, .foo2.com, 192.168.1.1, " + "*foobar.com:80, *.foo, http://baz, <local>"); + ASSERT_EQ(7u, rules.rules().size()); + EXPECT_EQ("*foo1.com", rules.rules()[0]->ToString()); + EXPECT_EQ("*.foo2.com", rules.rules()[1]->ToString()); + EXPECT_EQ("192.168.1.1", rules.rules()[2]->ToString()); + EXPECT_EQ("*foobar.com:80", rules.rules()[3]->ToString()); + EXPECT_EQ("*.foo", rules.rules()[4]->ToString()); + EXPECT_EQ("http://*baz", rules.rules()[5]->ToString()); + EXPECT_EQ("<local>", rules.rules()[6]->ToString()); + + EXPECT_TRUE(rules.Matches(GURL("http://foo1.com"))); + EXPECT_TRUE(rules.Matches(GURL("http://aaafoo1.com"))); + EXPECT_FALSE(rules.Matches(GURL("http://aaafoo1.com.net"))); +} + +TEST(ProxyBypassRulesTest, MultipleRules) { + ProxyBypassRules rules; + rules.ParseFromString(".google.com , .foobar.com:30"); + ASSERT_EQ(2u, rules.rules().size()); + + EXPECT_TRUE(rules.Matches(GURL("http://baz.google.com:40"))); + EXPECT_FALSE(rules.Matches(GURL("http://google.com:40"))); + EXPECT_TRUE(rules.Matches(GURL("http://bar.foobar.com:30"))); + EXPECT_FALSE(rules.Matches(GURL("http://bar.foobar.com"))); + EXPECT_FALSE(rules.Matches(GURL("http://bar.foobar.com:33"))); +} + +TEST(ProxyBypassRulesTest, BadInputs) { + ProxyBypassRules rules; + EXPECT_FALSE(rules.AddRuleFromString("://")); + EXPECT_FALSE(rules.AddRuleFromString(" ")); + EXPECT_FALSE(rules.AddRuleFromString("http://")); + EXPECT_FALSE(rules.AddRuleFromString("*.foo.com:-34")); + EXPECT_EQ(0u, rules.rules().size()); +} + +TEST(ProxyBypassRulesTest, Equals) { + ProxyBypassRules rules1; + ProxyBypassRules rules2; + + rules1.ParseFromString("foo1.com, .foo2.com"); + rules2.ParseFromString("foo1.com,.FOo2.com"); + + EXPECT_TRUE(rules1.Equals(rules2)); + EXPECT_TRUE(rules2.Equals(rules1)); + + rules1.ParseFromString(".foo2.com"); + rules2.ParseFromString("foo1.com,.FOo2.com"); + + EXPECT_FALSE(rules1.Equals(rules2)); + EXPECT_FALSE(rules2.Equals(rules1)); +} + +TEST(ProxyBypassRulesTest, BypassLocalNames) { + const struct { + const char* url; + bool expected_is_local; + } tests[] = { + // Single-component hostnames are considered local. + {"http://localhost/x", true}, + {"http://www", true}, + + // IPv4 loopback interface. + {"http://127.0.0.1/x", true}, + {"http://127.0.0.1:80/x", true}, + + // IPv6 loopback interface. + {"http://[::1]:80/x", true}, + {"http://[0:0::1]:6233/x", true}, + {"http://[0:0:0:0:0:0:0:1]/x", true}, + + // Non-local URLs. + {"http://foo.com/", false}, + {"http://localhost.i/", false}, + {"http://www.google.com/", false}, + {"http://192.168.0.1/", false}, + + // Try with different protocols. + {"ftp://127.0.0.1/x", true}, + {"ftp://foobar.com/x", false}, + + // This is a bit of a gray-area, but GURL does not strip trailing dots + // in host-names, so the following are considered non-local. + {"http://www./x", false}, + {"http://localhost./x", false}, + }; + + ProxyBypassRules rules; + rules.ParseFromString("<local>"); + + for (size_t i = 0; i < ARRAYSIZE_UNSAFE(tests); ++i) { + SCOPED_TRACE(StringPrintf( + "Test[%d]: %s", static_cast<int>(i), tests[i].url)); + EXPECT_EQ(tests[i].expected_is_local, rules.Matches(GURL(tests[i].url))); + } +} + +} // namespace + +} // namespace net |