summaryrefslogtreecommitdiffstats
path: root/base/process_win.cc
diff options
context:
space:
mode:
Diffstat (limited to 'base/process_win.cc')
-rw-r--r--base/process_win.cc110
1 files changed, 110 insertions, 0 deletions
diff --git a/base/process_win.cc b/base/process_win.cc
new file mode 100644
index 0000000..38fe63c
--- /dev/null
+++ b/base/process_win.cc
@@ -0,0 +1,110 @@
+// Copyright (c) 2006-2008 The Chromium Authors. All rights reserved.
+// Use of this source code is governed by a BSD-style license that can be
+// found in the LICENSE file.
+
+#include "base/process.h"
+#include "base/logging.h"
+#include "base/process_util.h"
+#include "base/scoped_ptr.h"
+
+bool Process::IsProcessBackgrounded() {
+ DCHECK(process_);
+ DWORD priority = GetPriorityClass(process_);
+ if (priority == 0)
+ return false; // Failure case.
+ return priority == BELOW_NORMAL_PRIORITY_CLASS;
+}
+
+bool Process::SetProcessBackgrounded(bool value) {
+ DCHECK(process_);
+ DWORD priority = value ? BELOW_NORMAL_PRIORITY_CLASS : NORMAL_PRIORITY_CLASS;
+ return (SetPriorityClass(process_, priority) != 0);
+}
+
+// According to MSDN, these are the default values which XP
+// uses to govern working set soft limits.
+// http://msdn.microsoft.com/en-us/library/ms686234.aspx
+static const int kWinDefaultMinSet = 50 * 4096;
+static const int kWinDefaultMaxSet = 345 * 4096;
+static const int kDampingFactor = 2;
+
+bool Process::ReduceWorkingSet() {
+ if (!process_)
+ return false;
+ // The idea here is that when we the process' working set has gone
+ // down, we want to release those pages to the OS quickly. However,
+ // when it is not going down, we want to be careful not to release
+ // too much back to the OS, as it could cause additional paging.
+
+ // We use a damping function to lessen the working set over time.
+ // As the process grows/shrinks, this algorithm will lag with
+ // working set reduction.
+ //
+ // The intended algorithm is:
+ // TargetWorkingSetSize = (LastWorkingSet/2 + CurrentWorkingSet) /2
+
+ scoped_ptr<process_util::ProcessMetrics> metrics(
+ process_util::ProcessMetrics::CreateProcessMetrics(process_));
+ process_util::WorkingSetKBytes working_set;
+ if (!metrics->GetWorkingSetKBytes(&working_set))
+ return false;
+
+
+ // We want to compute the amount of working set that the process
+ // needs to keep in memory. Since other processes contain the
+ // pages which are shared, we don't need to reserve them in our
+ // process, the system already has them tagged. Keep in mind, we
+ // don't get to control *which* pages get released, but if we
+ // assume reasonable distribution of pages, this should generally
+ // be the right value.
+ size_t current_working_set_size = working_set.priv +
+ working_set.shareable;
+
+ size_t max_size = current_working_set_size;
+ if (last_working_set_size_)
+ max_size = (max_size + last_working_set_size_) / 2; // Average.
+ max_size *= 1024; // Convert to KBytes.
+ last_working_set_size_ = current_working_set_size / kDampingFactor;
+
+ BOOL rv = SetProcessWorkingSetSize(process_, kWinDefaultMinSet, max_size);
+ return rv == TRUE;
+}
+
+bool Process::UnReduceWorkingSet() {
+ if (!process_)
+ return false;
+
+ if (!last_working_set_size_)
+ return true; // There was nothing to undo.
+
+ // We've had a reduced working set. Make sure we have lots of
+ // headroom now that we're active again.
+ size_t limit = last_working_set_size_ * kDampingFactor * 2 * 1024;
+ BOOL rv = SetProcessWorkingSetSize(process_, kWinDefaultMinSet, limit);
+ return rv == TRUE;
+}
+
+bool Process::EmptyWorkingSet() {
+ if (!process_)
+ return false;
+
+ BOOL rv = SetProcessWorkingSetSize(process_, -1, -1);
+ return rv == TRUE;
+}
+
+int32 Process::pid() const {
+ if (process_ == 0)
+ return 0;
+
+ return process_util::GetProcId(process_);
+}
+
+bool Process::is_current() const {
+ return process_ == GetCurrentProcess();
+}
+
+// static
+Process Process::Current() {
+ return Process(GetCurrentProcess());
+}
+