diff options
Diffstat (limited to 'chrome/browser/sync/syncable/syncable.h')
-rw-r--r-- | chrome/browser/sync/syncable/syncable.h | 1135 |
1 files changed, 1135 insertions, 0 deletions
diff --git a/chrome/browser/sync/syncable/syncable.h b/chrome/browser/sync/syncable/syncable.h new file mode 100644 index 0000000..291e798 --- /dev/null +++ b/chrome/browser/sync/syncable/syncable.h @@ -0,0 +1,1135 @@ +// Copyright (c) 2010 The Chromium Authors. All rights reserved. +// Use of this source code is governed by a BSD-style license that can be +// found in the LICENSE file. + +#ifndef CHROME_BROWSER_SYNC_SYNCABLE_SYNCABLE_H_ +#define CHROME_BROWSER_SYNC_SYNCABLE_SYNCABLE_H_ + +#include <algorithm> +#include <bitset> +#include <iosfwd> +#include <limits> +#include <map> +#include <set> +#include <string> +#include <vector> + +#include "base/atomicops.h" +#include "base/basictypes.h" +#include "base/file_path.h" +#include "base/gtest_prod_util.h" +#include "base/lock.h" +#include "base/time.h" +#include "chrome/browser/sync/protocol/sync.pb.h" +#include "chrome/browser/sync/syncable/blob.h" +#include "chrome/browser/sync/syncable/dir_open_result.h" +#include "chrome/browser/sync/syncable/directory_event.h" +#include "chrome/browser/sync/syncable/path_name_cmp.h" +#include "chrome/browser/sync/syncable/syncable_id.h" +#include "chrome/browser/sync/syncable/model_type.h" +#include "chrome/browser/sync/util/channel.h" +#include "chrome/browser/sync/util/dbgq.h" +#include "chrome/browser/sync/util/sync_types.h" +#include "chrome/common/deprecated/event_sys.h" + +struct PurgeInfo; + +namespace sync_api { +class ReadTransaction; +class WriteNode; +class ReadNode; +} + +namespace syncable { +class Entry; +} + +std::ostream& operator<<(std::ostream& s, const syncable::Entry& e); + +namespace syncable { + +class DirectoryBackingStore; + +static const int64 kInvalidMetaHandle = 0; + +enum { + BEGIN_FIELDS = 0, + INT64_FIELDS_BEGIN = BEGIN_FIELDS +}; + +enum MetahandleField { + // Primary key into the table. Keep this as a handle to the meta entry + // across transactions. + META_HANDLE = INT64_FIELDS_BEGIN +}; + +enum BaseVersion { + // After initial upload, the version is controlled by the server, and is + // increased whenever the data or metadata changes on the server. + BASE_VERSION = META_HANDLE + 1, +}; + +enum Int64Field { + SERVER_VERSION = BASE_VERSION + 1, + MTIME, + SERVER_MTIME, + CTIME, + SERVER_CTIME, + + // A numeric position value that indicates the relative ordering of + // this object among its siblings. + SERVER_POSITION_IN_PARENT, + + LOCAL_EXTERNAL_ID, // ID of an item in the external local storage that this + // entry is associated with. (such as bookmarks.js) + + INT64_FIELDS_END +}; + +enum { + INT64_FIELDS_COUNT = INT64_FIELDS_END, + ID_FIELDS_BEGIN = INT64_FIELDS_END, +}; + +enum IdField { + // Code in InitializeTables relies on ID being the first IdField value. + ID = ID_FIELDS_BEGIN, + PARENT_ID, + SERVER_PARENT_ID, + + PREV_ID, + NEXT_ID, + ID_FIELDS_END +}; + +enum { + ID_FIELDS_COUNT = ID_FIELDS_END - ID_FIELDS_BEGIN, + BIT_FIELDS_BEGIN = ID_FIELDS_END +}; + +enum IndexedBitField { + IS_UNSYNCED = BIT_FIELDS_BEGIN, + IS_UNAPPLIED_UPDATE, + INDEXED_BIT_FIELDS_END, +}; + +enum IsDelField { + IS_DEL = INDEXED_BIT_FIELDS_END, +}; + +enum BitField { + IS_DIR = IS_DEL + 1, + SERVER_IS_DIR, + SERVER_IS_DEL, + BIT_FIELDS_END +}; + +enum { + BIT_FIELDS_COUNT = BIT_FIELDS_END - BIT_FIELDS_BEGIN, + STRING_FIELDS_BEGIN = BIT_FIELDS_END +}; + +enum StringField { + // Name, will be truncated by server. Can be duplicated in a folder. + NON_UNIQUE_NAME = STRING_FIELDS_BEGIN, + // The server version of |NON_UNIQUE_NAME|. + SERVER_NON_UNIQUE_NAME, + + // A tag string which identifies this node as a particular top-level + // permanent object. The tag can be thought of as a unique key that + // identifies a singleton instance. + UNIQUE_SERVER_TAG, // Tagged by the server + UNIQUE_CLIENT_TAG, // Tagged by the client + STRING_FIELDS_END, +}; + +enum { + STRING_FIELDS_COUNT = STRING_FIELDS_END - STRING_FIELDS_BEGIN, + PROTO_FIELDS_BEGIN = STRING_FIELDS_END +}; + +// From looking at the sqlite3 docs, it's not directly stated, but it +// seems the overhead for storing a NULL blob is very small. +enum ProtoField { + SPECIFICS = PROTO_FIELDS_BEGIN, + SERVER_SPECIFICS, + PROTO_FIELDS_END, +}; + +enum { + PROTO_FIELDS_COUNT = PROTO_FIELDS_END - PROTO_FIELDS_BEGIN +}; + +enum { + FIELD_COUNT = PROTO_FIELDS_END, + // Past this point we have temporaries, stored in memory only. + BEGIN_TEMPS = PROTO_FIELDS_END, + BIT_TEMPS_BEGIN = BEGIN_TEMPS, +}; + +enum BitTemp { + SYNCING = BIT_TEMPS_BEGIN, + BIT_TEMPS_END, +}; + +enum { + BIT_TEMPS_COUNT = BIT_TEMPS_END - BIT_TEMPS_BEGIN +}; + +class BaseTransaction; +class WriteTransaction; +class ReadTransaction; +class Directory; +class ScopedDirLookup; + +// Instead of: +// Entry e = transaction.GetById(id); +// use: +// Entry e(transaction, GET_BY_ID, id); +// +// Why? The former would require a copy constructor, and it would be difficult +// to enforce that an entry never outlived its transaction if there were a copy +// constructor. +enum GetById { + GET_BY_ID +}; + +enum GetByClientTag { + GET_BY_CLIENT_TAG +}; + +enum GetByServerTag { + GET_BY_SERVER_TAG +}; + +enum GetByHandle { + GET_BY_HANDLE +}; + +enum Create { + CREATE +}; + +enum CreateNewUpdateItem { + CREATE_NEW_UPDATE_ITEM +}; + +typedef std::set<std::string> AttributeKeySet; + +typedef std::set<int64> MetahandleSet; + +// Why the singular enums? So the code compile-time dispatches instead of +// runtime dispatches as it would with a single enum and an if() statement. + +// The EntryKernel class contains the actual data for an entry. +struct EntryKernel { + private: + std::string string_fields[STRING_FIELDS_COUNT]; + sync_pb::EntitySpecifics specifics_fields[PROTO_FIELDS_COUNT]; + int64 int64_fields[INT64_FIELDS_COUNT]; + Id id_fields[ID_FIELDS_COUNT]; + std::bitset<BIT_FIELDS_COUNT> bit_fields; + std::bitset<BIT_TEMPS_COUNT> bit_temps; + + public: + EntryKernel() : dirty_(false) {} + + // Set the dirty bit, and optionally add this entry's metahandle to + // a provided index on dirty bits in |dirty_index|. Parameter may be null, + // and will result only in setting the dirty bit of this entry. + inline void mark_dirty(syncable::MetahandleSet* dirty_index) { + if (!dirty_ && dirty_index) { + DCHECK_NE(0, ref(META_HANDLE)); + dirty_index->insert(ref(META_HANDLE)); + } + dirty_ = true; + } + + // Clear the dirty bit, and optionally remove this entry's metahandle from + // a provided index on dirty bits in |dirty_index|. Parameter may be null, + // and will result only in clearing dirty bit of this entry. + inline void clear_dirty(syncable::MetahandleSet* dirty_index) { + if (dirty_ && dirty_index) { + DCHECK_NE(0, ref(META_HANDLE)); + dirty_index->erase(ref(META_HANDLE)); + } + dirty_ = false; + } + + inline bool is_dirty() const { + return dirty_; + } + + // Setters. + inline void put(MetahandleField field, int64 value) { + int64_fields[field - INT64_FIELDS_BEGIN] = value; + } + inline void put(Int64Field field, int64 value) { + int64_fields[field - INT64_FIELDS_BEGIN] = value; + } + inline void put(IdField field, const Id& value) { + id_fields[field - ID_FIELDS_BEGIN] = value; + } + inline void put(BaseVersion field, int64 value) { + int64_fields[field - INT64_FIELDS_BEGIN] = value; + } + inline void put(IndexedBitField field, bool value) { + bit_fields[field - BIT_FIELDS_BEGIN] = value; + } + inline void put(IsDelField field, bool value) { + bit_fields[field - BIT_FIELDS_BEGIN] = value; + } + inline void put(BitField field, bool value) { + bit_fields[field - BIT_FIELDS_BEGIN] = value; + } + inline void put(StringField field, const std::string& value) { + string_fields[field - STRING_FIELDS_BEGIN] = value; + } + inline void put(ProtoField field, const sync_pb::EntitySpecifics& value) { + specifics_fields[field - PROTO_FIELDS_BEGIN].CopyFrom(value); + } + inline void put(BitTemp field, bool value) { + bit_temps[field - BIT_TEMPS_BEGIN] = value; + } + + // Const ref getters. + inline int64 ref(MetahandleField field) const { + return int64_fields[field - INT64_FIELDS_BEGIN]; + } + inline int64 ref(Int64Field field) const { + return int64_fields[field - INT64_FIELDS_BEGIN]; + } + inline const Id& ref(IdField field) const { + return id_fields[field - ID_FIELDS_BEGIN]; + } + inline int64 ref(BaseVersion field) const { + return int64_fields[field - INT64_FIELDS_BEGIN]; + } + inline bool ref(IndexedBitField field) const { + return bit_fields[field - BIT_FIELDS_BEGIN]; + } + inline bool ref(IsDelField field) const { + return bit_fields[field - BIT_FIELDS_BEGIN]; + } + inline bool ref(BitField field) const { + return bit_fields[field - BIT_FIELDS_BEGIN]; + } + inline const std::string& ref(StringField field) const { + return string_fields[field - STRING_FIELDS_BEGIN]; + } + inline const sync_pb::EntitySpecifics& ref(ProtoField field) const { + return specifics_fields[field - PROTO_FIELDS_BEGIN]; + } + inline bool ref(BitTemp field) const { + return bit_temps[field - BIT_TEMPS_BEGIN]; + } + + // Non-const, mutable ref getters for object types only. + inline std::string& mutable_ref(StringField field) { + return string_fields[field - STRING_FIELDS_BEGIN]; + } + inline sync_pb::EntitySpecifics& mutable_ref(ProtoField field) { + return specifics_fields[field - PROTO_FIELDS_BEGIN]; + } + inline Id& mutable_ref(IdField field) { + return id_fields[field - ID_FIELDS_BEGIN]; + } + + private: + // Tracks whether this entry needs to be saved to the database. + bool dirty_; +}; + +// A read-only meta entry. +class Entry { + friend class Directory; + friend std::ostream& ::operator << (std::ostream& s, const Entry& e); + + public: + // After constructing, you must check good() to test whether the Get + // succeeded. + Entry(BaseTransaction* trans, GetByHandle, int64 handle); + Entry(BaseTransaction* trans, GetById, const Id& id); + Entry(BaseTransaction* trans, GetByServerTag, const std::string& tag); + Entry(BaseTransaction* trans, GetByClientTag, const std::string& tag); + + bool good() const { return 0 != kernel_; } + + BaseTransaction* trans() const { return basetrans_; } + + // Field accessors. + inline int64 Get(MetahandleField field) const { + DCHECK(kernel_); + return kernel_->ref(field); + } + inline Id Get(IdField field) const { + DCHECK(kernel_); + return kernel_->ref(field); + } + inline int64 Get(Int64Field field) const { + DCHECK(kernel_); + return kernel_->ref(field); + } + inline int64 Get(BaseVersion field) const { + DCHECK(kernel_); + return kernel_->ref(field); + } + inline bool Get(IndexedBitField field) const { + DCHECK(kernel_); + return kernel_->ref(field); + } + inline bool Get(IsDelField field) const { + DCHECK(kernel_); + return kernel_->ref(field); + } + inline bool Get(BitField field) const { + DCHECK(kernel_); + return kernel_->ref(field); + } + const std::string& Get(StringField field) const; + inline const sync_pb::EntitySpecifics& Get(ProtoField field) const { + DCHECK(kernel_); + return kernel_->ref(field); + } + inline bool Get(BitTemp field) const { + DCHECK(kernel_); + return kernel_->ref(field); + } + + ModelType GetServerModelType() const; + ModelType GetModelType() const; + + // If this returns false, we shouldn't bother maintaining + // a position value (sibling ordering) for this item. + bool ShouldMaintainPosition() const { + return GetModelType() == BOOKMARKS; + } + + inline bool ExistsOnClientBecauseNameIsNonEmpty() const { + DCHECK(kernel_); + return !kernel_->ref(NON_UNIQUE_NAME).empty(); + } + + inline bool IsRoot() const { + DCHECK(kernel_); + return kernel_->ref(ID).IsRoot(); + } + + Directory* dir() const; + + const EntryKernel GetKernelCopy() const { + return *kernel_; + } + + protected: // Don't allow creation on heap, except by sync API wrappers. + friend class sync_api::ReadNode; + void* operator new(size_t size) { return (::operator new)(size); } + + inline Entry(BaseTransaction* trans) : basetrans_(trans) { } + + protected: + + BaseTransaction* const basetrans_; + + EntryKernel* kernel_; + + DISALLOW_COPY_AND_ASSIGN(Entry); +}; + +// A mutable meta entry. Changes get committed to the database when the +// WriteTransaction is destroyed. +class MutableEntry : public Entry { + friend class WriteTransaction; + friend class Directory; + void Init(WriteTransaction* trans, const Id& parent_id, + const std::string& name); + public: + MutableEntry(WriteTransaction* trans, Create, const Id& parent_id, + const std::string& name); + MutableEntry(WriteTransaction* trans, CreateNewUpdateItem, const Id& id); + MutableEntry(WriteTransaction* trans, GetByHandle, int64); + MutableEntry(WriteTransaction* trans, GetById, const Id&); + MutableEntry(WriteTransaction* trans, GetByClientTag, const std::string& tag); + MutableEntry(WriteTransaction* trans, GetByServerTag, const std::string& tag); + + inline WriteTransaction* write_transaction() const { + return write_transaction_; + } + + // Field Accessors. Some of them trigger the re-indexing of the entry. + // Return true on success, return false on failure, which means + // that putting the value would have caused a duplicate in the index. + // TODO(chron): Remove some of these unecessary return values. + bool Put(Int64Field field, const int64& value); + bool Put(IdField field, const Id& value); + + // Do a simple property-only update if the PARENT_ID field. Use with caution. + // + // The normal Put(IS_PARENT) call will move the item to the front of the + // sibling order to maintain the linked list invariants when the parent + // changes. That's usually what you want to do, but it's inappropriate + // when the caller is trying to change the parent ID of a the whole set + // of children (e.g. because the ID changed during a commit). For those + // cases, there's this function. It will corrupt the sibling ordering + // if you're not careful. + void PutParentIdPropertyOnly(const Id& parent_id); + + bool Put(StringField field, const std::string& value); + bool Put(BaseVersion field, int64 value); + + bool Put(ProtoField field, const sync_pb::EntitySpecifics& value); + inline bool Put(BitField field, bool value) { + return PutField(field, value); + } + inline bool Put(IsDelField field, bool value) { + return PutIsDel(value); + } + bool Put(IndexedBitField field, bool value); + + // Sets the position of this item, and updates the entry kernels of the + // adjacent siblings so that list invariants are maintained. Returns false + // and fails if |predecessor_id| does not identify a sibling. Pass the root + // ID to put the node in first position. + bool PutPredecessor(const Id& predecessor_id); + + inline bool Put(BitTemp field, bool value) { + return PutTemp(field, value); + } + + protected: + syncable::MetahandleSet* GetDirtyIndexHelper(); + + template <typename FieldType, typename ValueType> + inline bool PutField(FieldType field, const ValueType& value) { + DCHECK(kernel_); + if (kernel_->ref(field) != value) { + kernel_->put(field, value); + kernel_->mark_dirty(GetDirtyIndexHelper()); + } + return true; + } + + template <typename TempType, typename ValueType> + inline bool PutTemp(TempType field, const ValueType& value) { + DCHECK(kernel_); + kernel_->put(field, value); + return true; + } + + bool PutIsDel(bool value); + + private: // Don't allow creation on heap, except by sync API wrappers. + friend class sync_api::WriteNode; + void* operator new(size_t size) { return (::operator new)(size); } + + bool PutImpl(StringField field, const std::string& value); + bool PutUniqueClientTag(const std::string& value); + + // Adjusts the successor and predecessor entries so that they no longer + // refer to this entry. + void UnlinkFromOrder(); + + // Kind of redundant. We should reduce the number of pointers + // floating around if at all possible. Could we store this in Directory? + // Scope: Set on construction, never changed after that. + WriteTransaction* const write_transaction_; + + protected: + MutableEntry(); + + DISALLOW_COPY_AND_ASSIGN(MutableEntry); +}; + +template <Int64Field field_index> +class SameField; +template <Int64Field field_index> +class HashField; +class LessParentIdAndHandle; +class LessMultiIncusionTargetAndMetahandle; +template <typename FieldType, FieldType field_index> +class LessField; +class LessEntryMetaHandles { + public: + inline bool operator()(const syncable::EntryKernel& a, + const syncable::EntryKernel& b) const { + return a.ref(META_HANDLE) < b.ref(META_HANDLE); + } +}; +typedef std::set<EntryKernel, LessEntryMetaHandles> OriginalEntries; + +// Represents one or more model types sharing an update timestamp, as is +// the case with a GetUpdates request. +struct MultiTypeTimeStamp { + syncable::ModelTypeBitSet data_types; + int64 timestamp; +}; + +// a WriteTransaction has a writer tag describing which body of code is doing +// the write. This is defined up here since DirectoryChangeEvent also contains +// one. +enum WriterTag { + INVALID, + SYNCER, + AUTHWATCHER, + UNITTEST, + VACUUM_AFTER_SAVE, + PURGE_ENTRIES, + SYNCAPI +}; + +// A separate Event type and channel for very frequent changes, caused +// by anything, not just the user. +struct DirectoryChangeEvent { + enum { + // Means listener should go through list of original entries and + // calculate what it needs to notify. It should *not* call any + // callbacks or attempt to lock anything because a + // WriteTransaction is being held until the listener returns. + CALCULATE_CHANGES, + // Means the WriteTransaction is ending, and this is the absolute + // last chance to perform any read operations in the current transaction. + // It is not recommended that the listener perform any writes. + TRANSACTION_ENDING, + // Means the WriteTransaction has been released and the listener + // can now take action on the changes it calculated. + TRANSACTION_COMPLETE, + // Channel is closing. + SHUTDOWN + } todo; + // These members are only valid for CALCULATE_CHANGES. + const OriginalEntries* originals; + BaseTransaction* trans; // This is valid also for TRANSACTION_ENDING + WriterTag writer; + typedef DirectoryChangeEvent EventType; + static inline bool IsChannelShutdownEvent(const EventType& e) { + return SHUTDOWN == e.todo; + } +}; + +// A list of metahandles whose metadata should not be purged. +typedef std::multiset<int64> Pegs; + +// The name Directory in this case means the entire directory +// structure within a single user account. +// +// Sqlite is a little goofy, in that each thread must access a database +// via its own handle. So, a Directory object should only be accessed +// from a single thread. Use DirectoryManager's Open() method to +// always get a directory that has been properly initialized on the +// current thread. +// +// The db is protected against concurrent modification by a reader/ +// writer lock, negotiated by the ReadTransaction and WriteTransaction +// friend classes. The in-memory indices are protected against +// concurrent modification by the kernel lock. +// +// All methods which require the reader/writer lock to be held either +// are protected and only called from friends in a transaction +// or are public and take a Transaction* argument. +// +// All methods which require the kernel lock to be already held take a +// ScopeKernelLock* argument. +// +// To prevent deadlock, the reader writer transaction lock must always +// be held before acquiring the kernel lock. +class ScopedKernelLock; +class IdFilter; +class DirectoryManager; +struct PathMatcher; + +class Directory { + friend class BaseTransaction; + friend class Entry; + friend class MutableEntry; + friend class ReadTransaction; + friend class ReadTransactionWithoutDB; + friend class ScopedKernelLock; + friend class ScopedKernelUnlock; + friend class WriteTransaction; + friend class SyncableDirectoryTest; + FRIEND_TEST_ALL_PREFIXES(SyncableDirectoryTest, + TakeSnapshotGetsAllDirtyHandlesTest); + FRIEND_TEST_ALL_PREFIXES(SyncableDirectoryTest, + TakeSnapshotGetsOnlyDirtyHandlesTest); + FRIEND_TEST_ALL_PREFIXES(SyncableDirectoryTest, TestPurgeEntriesWithTypeIn); + FRIEND_TEST_ALL_PREFIXES(SyncableDirectoryTest, + TakeSnapshotGetsMetahandlesToPurge); + + public: + class EventListenerHookup; + + // Various data that the Directory::Kernel we are backing (persisting data + // for) needs saved across runs of the application. + struct PersistedKernelInfo { + // Last sync timestamp fetched from the server. + int64 last_download_timestamp[MODEL_TYPE_COUNT]; + // true iff we ever reached the end of the changelog. + ModelTypeBitSet initial_sync_ended; + // The store birthday we were given by the server. Contents are opaque to + // the client. + std::string store_birthday; + // The next local ID that has not been used with this cache-GUID. + int64 next_id; + PersistedKernelInfo() : next_id(0) { + for (int i = 0; i < MODEL_TYPE_COUNT; ++i) { + last_download_timestamp[i] = 0; + } + } + }; + + // What the Directory needs on initialization to create itself and its Kernel. + // Filled by DirectoryBackingStore::Load. + struct KernelLoadInfo { + PersistedKernelInfo kernel_info; + std::string cache_guid; // Created on first initialization, never changes. + int64 max_metahandle; // Computed (using sql MAX aggregate) on init. + KernelLoadInfo() : max_metahandle(0) { + } + }; + + // The dirty/clean state of kernel fields backed by the share_info table. + // This is public so it can be used in SaveChangesSnapshot for persistence. + enum KernelShareInfoStatus { + KERNEL_SHARE_INFO_INVALID, + KERNEL_SHARE_INFO_VALID, + KERNEL_SHARE_INFO_DIRTY + }; + + // When the Directory is told to SaveChanges, a SaveChangesSnapshot is + // constructed and forms a consistent snapshot of what needs to be sent to + // the backing store. + struct SaveChangesSnapshot { + KernelShareInfoStatus kernel_info_status; + PersistedKernelInfo kernel_info; + OriginalEntries dirty_metas; + MetahandleSet metahandles_to_purge; + SaveChangesSnapshot() : kernel_info_status(KERNEL_SHARE_INFO_INVALID) { + } + }; + + Directory(); + virtual ~Directory(); + + DirOpenResult Open(const FilePath& file_path, const std::string& name); + + void Close(); + + int64 NextMetahandle(); + // Always returns a negative id. Positive client ids are generated + // by the server only. + Id NextId(); + + const FilePath& file_path() const { return kernel_->db_path; } + bool good() const { return NULL != store_; } + + // The download timestamp is an index into the server's list of changes for + // an account. We keep this for each datatype. It doesn't actually map + // to any time scale; its name is an historical artifact. + int64 last_download_timestamp(ModelType type) const; + void set_last_download_timestamp(ModelType type, int64 value); + + // Find the model type or model types which have the least timestamp, and + // return them along with the types having that timestamp. This is done + // with the intent of sequencing GetUpdates requests in an efficient way: + // bundling together requests that have the same timestamp, and requesting + // the oldest such bundles first in hopes that they might catch up to + // (and thus be merged with) the newer bundles. + MultiTypeTimeStamp GetTypesWithOldestLastDownloadTimestamp( + ModelTypeBitSet enabled_types) { + MultiTypeTimeStamp result; + result.timestamp = std::numeric_limits<int64>::max(); + for (int i = 0; i < MODEL_TYPE_COUNT; ++i) { + if (!enabled_types[i]) + continue; + int64 stamp = last_download_timestamp(ModelTypeFromInt(i)); + if (stamp < result.timestamp) { + result.data_types.reset(); + result.timestamp = stamp; + } + if (stamp == result.timestamp) + result.data_types.set(i); + } + return result; + } + + bool initial_sync_ended_for_type(ModelType type) const; + void set_initial_sync_ended_for_type(ModelType type, bool value); + + const std::string& name() const { return kernel_->name; } + + // (Account) Store birthday is opaque to the client, + // so we keep it in the format it is in the proto buffer + // in case we switch to a binary birthday later. + std::string store_birthday() const; + void set_store_birthday(std::string store_birthday); + + // Unique to each account / client pair. + std::string cache_guid() const; + + browser_sync::ChannelHookup<DirectoryChangeEvent>* AddChangeObserver( + browser_sync::ChannelEventHandler<DirectoryChangeEvent>* observer); + + protected: // for friends, mainly used by Entry constructors + EntryKernel* GetEntryByHandle(const int64 handle); + EntryKernel* GetEntryByHandle(const int64 metahandle, ScopedKernelLock* lock); + EntryKernel* GetEntryById(const Id& id); + EntryKernel* GetEntryByServerTag(const std::string& tag); + EntryKernel* GetEntryByClientTag(const std::string& tag); + EntryKernel* GetRootEntry(); + bool ReindexId(EntryKernel* const entry, const Id& new_id); + void ReindexParentId(EntryKernel* const entry, const Id& new_parent_id); + void ClearDirtyMetahandles(); + + // These don't do semantic checking. + // The semantic checking is implemented higher up. + void Undelete(EntryKernel* const entry); + void Delete(EntryKernel* const entry); + void UnlinkEntryFromOrder(EntryKernel* entry, + WriteTransaction* trans, + ScopedKernelLock* lock); + + // Overridden by tests. + virtual DirectoryBackingStore* CreateBackingStore( + const std::string& dir_name, + const FilePath& backing_filepath); + + private: + // These private versions expect the kernel lock to already be held + // before calling. + EntryKernel* GetEntryById(const Id& id, ScopedKernelLock* const lock); + + DirOpenResult OpenImpl(const FilePath& file_path, const std::string& name); + + struct DirectoryEventTraits { + typedef DirectoryEvent EventType; + static inline bool IsChannelShutdownEvent(const DirectoryEvent& event) { + return DIRECTORY_DESTROYED == event; + } + }; + public: + typedef EventChannel<DirectoryEventTraits, Lock> Channel; + typedef std::vector<int64> ChildHandles; + + // Returns the child meta handles for given parent id. + void GetChildHandles(BaseTransaction*, const Id& parent_id, + const std::string& path_spec, ChildHandles* result); + void GetChildHandles(BaseTransaction*, const Id& parent_id, + ChildHandles* result); + void GetChildHandlesImpl(BaseTransaction* trans, const Id& parent_id, + PathMatcher* matcher, ChildHandles* result); + + // Find the first or last child in the positional ordering under a parent, + // and return its id. Returns a root Id if parent has no children. + Id GetFirstChildId(BaseTransaction* trans, const Id& parent_id); + Id GetLastChildId(BaseTransaction* trans, const Id& parent_id); + + // SaveChanges works by taking a consistent snapshot of the current Directory + // state and indices (by deep copy) under a ReadTransaction, passing this + // snapshot to the backing store under no transaction, and finally cleaning + // up by either purging entries no longer needed (this part done under a + // WriteTransaction) or rolling back the dirty bits. It also uses + // internal locking to enforce SaveChanges operations are mutually exclusive. + // + // WARNING: THIS METHOD PERFORMS SYNCHRONOUS I/O VIA SQLITE. + bool SaveChanges(); + + // Returns the number of entities with the unsynced bit set. + int64 unsynced_entity_count() const; + + // Get GetUnsyncedMetaHandles should only be called after SaveChanges and + // before any new entries have been created. The intention is that the + // syncer should call it from its PerformSyncQueries member. + typedef std::vector<int64> UnsyncedMetaHandles; + void GetUnsyncedMetaHandles(BaseTransaction* trans, + UnsyncedMetaHandles* result); + + // Get all the metahandles for unapplied updates + typedef std::vector<int64> UnappliedUpdateMetaHandles; + void GetUnappliedUpdateMetaHandles(BaseTransaction* trans, + UnappliedUpdateMetaHandles* result); + + // Get the channel for post save notification, used by the syncer. + inline Channel* channel() const { + return kernel_->channel; + } + + // Checks tree metadata consistency. + // If full_scan is false, the function will avoid pulling any entries from the + // db and scan entries currently in ram. + // If full_scan is true, all entries will be pulled from the database. + // No return value, CHECKs will be triggered if we're given bad + // information. + void CheckTreeInvariants(syncable::BaseTransaction* trans, + bool full_scan); + + void CheckTreeInvariants(syncable::BaseTransaction* trans, + const OriginalEntries* originals); + + void CheckTreeInvariants(syncable::BaseTransaction* trans, + const MetahandleSet& handles, + const IdFilter& idfilter); + + // Purges all data associated with any entries whose ModelType or + // ServerModelType is found in |types|, from _both_ memory and disk. + // Only valid, "real" model types are allowed in |types| (see model_type.h + // for definitions). "Purge" is just meant to distinguish from "deleting" + // entries, which means something different in the syncable namespace. + // WARNING! This can be real slow, as it iterates over all entries. + // WARNING! Performs synchronous I/O. + virtual void PurgeEntriesWithTypeIn(const std::set<ModelType>& types); + + private: + // Helper to prime ids_index, parent_id_and_names_index, unsynced_metahandles + // and unapplied_metahandles from metahandles_index. + void InitializeIndices(); + + // Constructs a consistent snapshot of the current Directory state and + // indices (by deep copy) under a ReadTransaction for use in |snapshot|. + // See SaveChanges() for more information. + void TakeSnapshotForSaveChanges(SaveChangesSnapshot* snapshot); + + // Purges from memory any unused, safe to remove entries that were + // successfully deleted on disk as a result of the SaveChanges that processed + // |snapshot|. See SaveChanges() for more information. + void VacuumAfterSaveChanges(const SaveChangesSnapshot& snapshot); + + // Rolls back dirty bits in the event that the SaveChanges that + // processed |snapshot| failed, for example, due to no disk space. + void HandleSaveChangesFailure(const SaveChangesSnapshot& snapshot); + + // For new entry creation only + void InsertEntry(EntryKernel* entry, ScopedKernelLock* lock); + void InsertEntry(EntryKernel* entry); + + // Used by CheckTreeInvariants + void GetAllMetaHandles(BaseTransaction* trans, MetahandleSet* result); + bool SafeToPurgeFromMemory(const EntryKernel* const entry) const; + + // Helper method used to implement GetFirstChildId/GetLastChildId. + Id GetChildWithNullIdField(IdField field, + BaseTransaction* trans, + const Id& parent_id); + + // Internal setters that do not acquire a lock internally. These are unsafe + // on their own; caller must guarantee exclusive access manually by holding + // a ScopedKernelLock. + void set_initial_sync_ended_for_type_unsafe(ModelType type, bool x); + void set_last_download_timestamp_unsafe(ModelType model_type, int64 x); + + Directory& operator = (const Directory&); + + // TODO(sync): If lookups and inserts in these sets become + // the bottle-neck, then we can use hash-sets instead. But + // that will require using #ifdefs and compiler-specific code, + // so use standard sets for now. + public: + typedef std::set<EntryKernel*, LessField<MetahandleField, META_HANDLE> > + MetahandlesIndex; + typedef std::set<EntryKernel*, LessField<IdField, ID> > IdsIndex; + // All entries in memory must be in both the MetahandlesIndex and + // the IdsIndex, but only non-deleted entries will be the + // ParentIdChildIndex. + // This index contains EntryKernels ordered by parent ID and metahandle. + // It allows efficient lookup of the children of a given parent. + typedef std::set<EntryKernel*, LessParentIdAndHandle> ParentIdChildIndex; + + // Contains both deleted and existing entries with tags. + // We can't store only existing tags because the client would create + // items that had a duplicated ID in the end, resulting in a DB key + // violation. ID reassociation would fail after an attempted commit. + typedef std::set<EntryKernel*, + LessField<StringField, UNIQUE_CLIENT_TAG> > ClientTagIndex; + + protected: + // Used by tests. + void init_kernel(const std::string& name); + + private: + + struct Kernel { + Kernel(const FilePath& db_path, const std::string& name, + const KernelLoadInfo& info); + + ~Kernel(); + + void AddRef(); // For convenience. + void Release(); + + FilePath const db_path; + // TODO(timsteele): audit use of the member and remove if possible + volatile base::subtle::AtomicWord refcount; + + // Implements ReadTransaction / WriteTransaction using a simple lock. + Lock transaction_mutex; + + // The name of this directory. + std::string const name; + + // Protects all members below. + // The mutex effectively protects all the indices, but not the + // entries themselves. So once a pointer to an entry is pulled + // from the index, the mutex can be unlocked and entry read or written. + // + // Never hold the mutex and do anything with the database or any + // other buffered IO. Violating this rule will result in deadlock. + Lock mutex; + MetahandlesIndex* metahandles_index; // Entries indexed by metahandle + IdsIndex* ids_index; // Entries indexed by id + ParentIdChildIndex* parent_id_child_index; + ClientTagIndex* client_tag_index; + // So we don't have to create an EntryKernel every time we want to + // look something up in an index. Needle in haystack metaphor. + EntryKernel needle; + + // 3 in-memory indices on bits used extremely frequently by the syncer. + MetahandleSet* const unapplied_update_metahandles; + MetahandleSet* const unsynced_metahandles; + // Contains metahandles that are most likely dirty (though not + // necessarily). Dirtyness is confirmed in TakeSnapshotForSaveChanges(). + MetahandleSet* const dirty_metahandles; + + // When a purge takes place, we remove items from all our indices and stash + // them in here so that SaveChanges can persist their permanent deletion. + MetahandleSet* const metahandles_to_purge; + + // TODO(ncarter): Figure out what the hell this is, and comment it. + Channel* const channel; + + // The changes channel mutex is explicit because it must be locked + // while holding the transaction mutex and released after + // releasing the transaction mutex. + browser_sync::Channel<DirectoryChangeEvent> changes_channel; + + Lock changes_channel_mutex; + KernelShareInfoStatus info_status; + + // These 3 members are backed in the share_info table, and + // their state is marked by the flag above. + + // A structure containing the Directory state that is written back into the + // database on SaveChanges. + PersistedKernelInfo persisted_info; + + // A unique identifier for this account's cache db, used to generate + // unique server IDs. No need to lock, only written at init time. + std::string cache_guid; + + // It doesn't make sense for two threads to run SaveChanges at the same + // time; this mutex protects that activity. + Lock save_changes_mutex; + + // The next metahandle is protected by kernel mutex. + int64 next_metahandle; + + // Keep a history of recently flushed metahandles for debugging + // purposes. Protected by the save_changes_mutex. + DebugQueue<int64, 1000> flushed_metahandles; + }; + + Kernel* kernel_; + + DirectoryBackingStore* store_; +}; + +class ScopedKernelLock { + public: + explicit ScopedKernelLock(const Directory*); + ~ScopedKernelLock() {} + + AutoLock scoped_lock_; + Directory* const dir_; + DISALLOW_COPY_AND_ASSIGN(ScopedKernelLock); +}; + +// Transactions are now processed FIFO with a straight lock +class BaseTransaction { + friend class Entry; + public: + inline Directory* directory() const { return directory_; } + inline Id root_id() const { return Id(); } + + virtual ~BaseTransaction(); + + protected: + BaseTransaction(Directory* directory, const char* name, + const char* source_file, int line, WriterTag writer); + + void UnlockAndLog(OriginalEntries* entries); + + Directory* const directory_; + Directory::Kernel* const dirkernel_; // for brevity + const char* const name_; + base::TimeTicks time_acquired_; + const char* const source_file_; + const int line_; + WriterTag writer_; + + private: + void Lock(); + + DISALLOW_COPY_AND_ASSIGN(BaseTransaction); +}; + +// Locks db in constructor, unlocks in destructor. +class ReadTransaction : public BaseTransaction { + public: + ReadTransaction(Directory* directory, const char* source_file, + int line); + ReadTransaction(const ScopedDirLookup& scoped_dir, + const char* source_file, int line); + + virtual ~ReadTransaction(); + + protected: // Don't allow creation on heap, except by sync API wrapper. + friend class sync_api::ReadTransaction; + void* operator new(size_t size) { return (::operator new)(size); } + + DISALLOW_COPY_AND_ASSIGN(ReadTransaction); +}; + +// Locks db in constructor, unlocks in destructor. +class WriteTransaction : public BaseTransaction { + friend class MutableEntry; + public: + explicit WriteTransaction(Directory* directory, WriterTag writer, + const char* source_file, int line); + explicit WriteTransaction(const ScopedDirLookup& directory, + WriterTag writer, const char* source_file, + int line); + virtual ~WriteTransaction(); + + void SaveOriginal(EntryKernel* entry); + + protected: + // Before an entry gets modified, we copy the original into a list + // so that we can issue change notifications when the transaction + // is done. + OriginalEntries* const originals_; + + DISALLOW_COPY_AND_ASSIGN(WriteTransaction); +}; + +bool IsLegalNewParent(BaseTransaction* trans, const Id& id, const Id& parentid); +int ComparePathNames(const std::string& a, const std::string& b); + +// Exposed in header as this is used as a sqlite3 callback. +int ComparePathNames16(void*, int a_bytes, const void* a, int b_bytes, + const void* b); + +int64 Now(); + +// This function sets only the flags needed to get this entry to sync. +void MarkForSyncing(syncable::MutableEntry* e); + +// This is not a reset. It just sets the numeric fields which are not +// initialized by the constructor to zero. +void ZeroFields(EntryKernel* entry, int first_field); + +} // namespace syncable + +std::ostream& operator <<(std::ostream&, const syncable::Blob&); + +browser_sync::FastDump& operator << + (browser_sync::FastDump&, const syncable::Blob&); + +#endif // CHROME_BROWSER_SYNC_SYNCABLE_SYNCABLE_H_ |