/* This Source Code Form is subject to the terms of the Mozilla Public * License, v. 2.0. If a copy of the MPL was not distributed with this * file, You can obtain one at http://mozilla.org/MPL/2.0/. */ use std::{collections::HashMap, convert::TryFrom, fmt}; use dogear::{ debug, warn, AbortSignal, CompletionOps, Content, DeleteLocalItem, Guid, Item, Kind, MergedRoot, Tree, UploadItem, UploadTombstone, Validity, }; use nsstring::nsString; use storage::{Conn, Step}; use url::Url; use xpcom::interfaces::{mozISyncedBookmarksMerger, nsINavBookmarksService}; use crate::driver::{AbortController, Driver}; use crate::error::{Error, Result}; pub const LMANNO_FEEDURI: &'static str = "livemark/feedURI"; extern "C" { fn NS_NavBookmarksTotalSyncChanges() -> i64; } fn total_sync_changes() -> i64 { unsafe { NS_NavBookmarksTotalSyncChanges() } } // Return all the non-root-roots as a 'sql set' (ie, suitable for use in an // IN statement) fn user_roots_as_sql_set() -> String { format!( "('{0}', '{1}', '{2}', '{3}', '{4}')", dogear::MENU_GUID, dogear::MOBILE_GUID, dogear::TAGS_GUID, dogear::TOOLBAR_GUID, dogear::UNFILED_GUID ) } pub struct Store<'s> { db: &'s mut Conn, driver: &'s Driver, controller: &'s AbortController, /// The total Sync change count before merging. We store this before /// accessing Places, and compare the current and stored counts after /// opening our transaction. If they match, we can safely apply the /// tree. Otherwise, we bail and try merging again on the next sync. total_sync_changes: i64, local_time_millis: i64, remote_time_millis: i64, } impl<'s> Store<'s> { pub fn new( db: &'s mut Conn, driver: &'s Driver, controller: &'s AbortController, local_time_millis: i64, remote_time_millis: i64, ) -> Store<'s> { Store { db, driver, controller, total_sync_changes: total_sync_changes(), local_time_millis, remote_time_millis, } } /// Ensures that all local roots are parented correctly. /// /// The Places root can't be in another folder, or we'll recurse infinitely /// when we try to fetch the local tree. /// /// The five built-in roots should be under the Places root, or we'll build /// and sync an invalid tree (bug 1453994, bug 1472127). pub fn validate(&self) -> Result<()> { self.controller.err_if_aborted()?; let mut statement = self.db.prepare(format!( "SELECT NOT EXISTS( SELECT 1 FROM moz_bookmarks WHERE id = (SELECT parent FROM moz_bookmarks WHERE guid = '{root}') ) AND NOT EXISTS( SELECT 1 FROM moz_bookmarks b JOIN moz_bookmarks p ON p.id = b.parent WHERE b.guid IN {user_roots} AND p.guid <> '{root}' )", root = dogear::ROOT_GUID, user_roots = user_roots_as_sql_set(), ))?; let has_valid_roots = match statement.step()? { Some(row) => row.get_by_index::(0)? == 1, None => false, }; if has_valid_roots { Ok(()) } else { Err(Error::InvalidLocalRoots) } } /// Prepares the mirror database for a merge. pub fn prepare(&self) -> Result<()> { // Sync associates keywords with bookmarks, and doesn't sync POST data; // Places associates keywords with (URL, POST data) pairs, and multiple // bookmarks may have the same URL. When a keyword changes, clients // should reupload all bookmarks with the affected URL (see // `PlacesSyncUtils.bookmarks.addSyncChangesForBookmarksWithURL` and // bug 1328737). Just in case, we flag any remote bookmarks that have // different keywords for the same URL, or the same keyword for // different URLs, for reupload. self.controller.err_if_aborted()?; self.db.exec(format!( "UPDATE items SET validity = {} WHERE validity = {} AND ( urlId IN ( /* Same URL, different keywords. `COUNT` ignores NULLs, so we need to count them separately. This handles cases where a keyword was removed from one, but not all bookmarks with the same URL. */ SELECT urlId FROM items GROUP BY urlId HAVING COUNT(DISTINCT keyword) + COUNT(DISTINCT CASE WHEN keyword IS NULL THEN 1 END) > 1 ) OR keyword IN ( /* Different URLs, same keyword. Bookmarks with keywords but without URLs are already invalid, so we don't need to handle NULLs here. */ SELECT keyword FROM items WHERE keyword NOT NULL GROUP BY keyword HAVING COUNT(DISTINCT urlId) > 1 ) )", mozISyncedBookmarksMerger::VALIDITY_REUPLOAD, mozISyncedBookmarksMerger::VALIDITY_VALID, ))?; Ok(()) } /// Creates a local tree item from a row in the `localItems` CTE. fn local_row_to_item(&self, step: &Step) -> Result<(Item, Option)> { let raw_guid: nsString = step.get_by_name("guid")?; let guid = Guid::from_utf16(&*raw_guid)?; let raw_url_href: Option = step.get_by_name("url")?; let (url, validity) = match raw_url_href { // Local items might have (syntactically) invalid URLs, as in bug // 1615931. If we try to sync these items, other clients will flag // them as invalid (see `SyncedBookmarksMirror#storeRemote{Bookmark, // Query}`), delete them when merging, and upload tombstones for // them. We can avoid this extra round trip by flagging the local // item as invalid. If there's a corresponding remote item with a // valid URL, we'll replace the local item with it; if there isn't, // we'll delete the local item. Some(raw_url_href) => match Url::parse(&String::from_utf16(&*raw_url_href)?) { Ok(url) => (Some(url), Validity::Valid), Err(err) => { warn!( self.driver, "Failed to parse URL for local item {}: {}", guid, err ); (None, Validity::Replace) } }, None => (None, Validity::Valid), }; let typ: i64 = step.get_by_name("type")?; let kind = match u16::try_from(typ) { Ok(nsINavBookmarksService::TYPE_BOOKMARK) => match url.as_ref() { Some(u) if u.scheme() == "place" => Kind::Query, _ => Kind::Bookmark, }, Ok(nsINavBookmarksService::TYPE_FOLDER) => { let is_livemark: i64 = step.get_by_name("isLivemark")?; if is_livemark == 1 { Kind::Livemark } else { Kind::Folder } } Ok(nsINavBookmarksService::TYPE_SEPARATOR) => Kind::Separator, _ => return Err(Error::UnknownItemType(typ)), }; let mut item = Item::new(guid, kind); let local_modified: i64 = step.get_by_name_or_default("localModified"); item.age = (self.local_time_millis - local_modified).max(0); let sync_change_counter: i64 = step.get_by_name("syncChangeCounter")?; item.needs_merge = sync_change_counter > 0; item.validity = validity; let content = if item.validity == Validity::Replace || item.guid == dogear::ROOT_GUID { None } else { let sync_status: i64 = step.get_by_name("syncStatus")?; match u16::try_from(sync_status) { Ok(nsINavBookmarksService::SYNC_STATUS_NORMAL) => None, _ => match kind { Kind::Bookmark | Kind::Query => { let raw_title: nsString = step.get_by_name("title")?; let title = String::from_utf16(&*raw_title)?; url.map(|url| Content::Bookmark { title, url_href: url.into_string(), }) } Kind::Folder | Kind::Livemark => { let raw_title: nsString = step.get_by_name("title")?; let title = String::from_utf16(&*raw_title)?; Some(Content::Folder { title }) } Kind::Separator => Some(Content::Separator), }, } }; Ok((item, content)) } /// Creates a remote tree item from a row in `mirror.items`. fn remote_row_to_item(&self, step: &Step) -> Result<(Item, Option)> { let raw_guid: nsString = step.get_by_name("guid")?; let guid = Guid::from_utf16(&*raw_guid)?; let raw_kind: i64 = step.get_by_name("kind")?; let kind = Kind::from_column(raw_kind)?; let mut item = Item::new(guid, kind); let remote_modified: i64 = step.get_by_name("serverModified")?; item.age = (self.remote_time_millis - remote_modified).max(0); let needs_merge: i32 = step.get_by_name("needsMerge")?; item.needs_merge = needs_merge == 1; let raw_validity: i64 = step.get_by_name("validity")?; item.validity = Validity::from_column(raw_validity)?; let content = if item.validity == Validity::Replace || item.guid == dogear::ROOT_GUID || !item.needs_merge { None } else { match kind { Kind::Bookmark | Kind::Query => { let raw_title: nsString = step.get_by_name("title")?; let title = String::from_utf16(&*raw_title)?; let raw_url_href: Option = step.get_by_name("url")?; match raw_url_href { Some(raw_url_href) => { // Unlike for local items, we don't parse URLs for // remote items, since `storeRemote{Bookmark, // Query}` already parses and canonicalizes them // before inserting them into the mirror database. let url_href = String::from_utf16(&*raw_url_href)?; Some(Content::Bookmark { title, url_href }) } None => None, } } Kind::Folder | Kind::Livemark => { let raw_title: nsString = step.get_by_name("title")?; let title = String::from_utf16(&*raw_title)?; Some(Content::Folder { title }) } Kind::Separator => Some(Content::Separator), } }; Ok((item, content)) } } impl<'s> dogear::Store for Store<'s> { type Ok = ApplyStatus; type Error = Error; /// Builds a fully rooted, consistent tree from the items and tombstones in /// Places. fn fetch_local_tree(&self) -> Result { let mut root_statement = self.db.prepare(format!( "SELECT guid, type, syncChangeCounter, syncStatus, lastModified / 1000 AS localModified, NULL AS url, 0 AS isLivemark FROM moz_bookmarks WHERE guid = '{}'", dogear::ROOT_GUID ))?; let mut builder = match root_statement.step()? { Some(step) => { let (item, _) = self.local_row_to_item(&step)?; Tree::with_root(item) } None => return Err(Error::InvalidLocalRoots), }; // Add items and contents to the builder, keeping track of their // structure in a separate map. We can't call `p.by_structure(...)` // after adding the item, because this query might return rows for // children before their parents. This approach also lets us scan // `moz_bookmarks` once, using the index on `(b.parent, b.position)` // to avoid a temp B-tree for the `ORDER BY`. let mut child_guids_by_parent_guid: HashMap> = HashMap::new(); let mut items_statement = self.db.prepare(format!( "SELECT b.guid, p.guid AS parentGuid, b.type, b.syncChangeCounter, b.syncStatus, b.lastModified / 1000 AS localModified, IFNULL(b.title, '') AS title, (SELECT h.url FROM moz_places h WHERE h.id = b.fk) AS url, EXISTS(SELECT 1 FROM moz_items_annos a JOIN moz_anno_attributes n ON n.id = a.anno_attribute_id WHERE a.item_id = b.id AND n.name = '{}') AS isLivemark FROM moz_bookmarks b JOIN moz_bookmarks p ON p.id = b.parent WHERE b.guid <> '{}' ORDER BY b.parent, b.position", LMANNO_FEEDURI, dogear::ROOT_GUID, ))?; while let Some(step) = items_statement.step()? { self.controller.err_if_aborted()?; let (item, content) = self.local_row_to_item(&step)?; let raw_parent_guid: nsString = step.get_by_name("parentGuid")?; let parent_guid = Guid::from_utf16(&*raw_parent_guid)?; child_guids_by_parent_guid .entry(parent_guid) .or_default() .push(item.guid.clone()); let mut p = builder.item(item)?; if let Some(content) = content { p.content(content); } } // At this point, we've added entries for all items to the tree, so // we can add their structure info. for (parent_guid, child_guids) in &child_guids_by_parent_guid { for child_guid in child_guids { self.controller.err_if_aborted()?; builder.parent_for(child_guid).by_structure(parent_guid)?; } } let mut deletions_statement = self.db.prepare("SELECT guid FROM moz_bookmarks_deleted")?; while let Some(step) = deletions_statement.step()? { self.controller.err_if_aborted()?; let raw_guid: nsString = step.get_by_name("guid")?; let guid = Guid::from_utf16(&*raw_guid)?; builder.deletion(guid); } let tree = Tree::try_from(builder)?; Ok(tree) } /// Builds a fully rooted, consistent tree from the items and tombstones in the /// mirror. fn fetch_remote_tree(&self) -> Result { let mut root_statement = self.db.prepare(format!( "SELECT guid, serverModified, kind, needsMerge, validity FROM items WHERE guid = '{}'", dogear::ROOT_GUID, ))?; let mut builder = match root_statement.step()? { Some(step) => { let (item, _) = self.remote_row_to_item(&step)?; Tree::with_root(item) } None => return Err(Error::InvalidRemoteRoots), }; builder.reparent_orphans_to(&dogear::UNFILED_GUID); let mut items_statement = self.db.prepare(format!( "SELECT v.guid, v.parentGuid, v.serverModified, v.kind, IFNULL(v.title, '') AS title, v.needsMerge, v.validity, v.isDeleted, (SELECT u.url FROM urls u WHERE u.id = v.urlId) AS url FROM items v WHERE v.guid <> '{}' ORDER BY v.guid", dogear::ROOT_GUID, ))?; while let Some(step) = items_statement.step()? { self.controller.err_if_aborted()?; let is_deleted: i64 = step.get_by_name("isDeleted")?; if is_deleted == 1 { let needs_merge: i32 = step.get_by_name("needsMerge")?; if needs_merge == 0 { // Ignore already-merged tombstones. These aren't persisted // locally, so merging them is a no-op. continue; } let raw_guid: nsString = step.get_by_name("guid")?; let guid = Guid::from_utf16(&*raw_guid)?; builder.deletion(guid); } else { let (item, content) = self.remote_row_to_item(&step)?; let mut p = builder.item(item)?; if let Some(content) = content { p.content(content); } let raw_parent_guid: Option = step.get_by_name("parentGuid")?; if let Some(raw_parent_guid) = raw_parent_guid { p.by_parent_guid(Guid::from_utf16(&*raw_parent_guid)?)?; } } } let mut structure_statement = self.db.prepare(format!( "SELECT guid, parentGuid FROM structure WHERE guid <> '{}' ORDER BY parentGuid, position", dogear::ROOT_GUID, ))?; while let Some(step) = structure_statement.step()? { self.controller.err_if_aborted()?; let raw_guid: nsString = step.get_by_name("guid")?; let guid = Guid::from_utf16(&*raw_guid)?; let raw_parent_guid: nsString = step.get_by_name("parentGuid")?; let parent_guid = Guid::from_utf16(&*raw_parent_guid)?; builder.parent_for(&guid).by_children(&parent_guid)?; } let tree = Tree::try_from(builder)?; Ok(tree) } fn apply<'t>(&mut self, root: MergedRoot<'t>) -> Result { let ops = root.completion_ops_with_signal(self.controller)?; if ops.is_empty() { // If we don't have any items to apply, upload, or delete, // no need to open a transaction at all. return Ok(ApplyStatus::Skipped); } // Apply the merged tree and stage outgoing items. This transaction // blocks writes from the main connection until it's committed, so we // try to do as little work as possible within it. if self.db.transaction_in_progress()? { return Err(Error::StorageBusy); } let tx = self.db.transaction()?; if self.total_sync_changes != total_sync_changes() { return Err(Error::MergeConflict); } debug!(self.driver, "Updating local items in Places"); update_local_items_in_places( &tx, &self.driver, &self.controller, self.local_time_millis, &ops, )?; debug!(self.driver, "Staging items to upload"); stage_items_to_upload( &tx, &self.driver, &self.controller, &ops.upload_items, &ops.upload_tombstones, )?; cleanup(&tx)?; tx.commit()?; Ok(ApplyStatus::Merged) } } /// Builds a temporary table with the merge states of all nodes in the merged /// tree and updates Places to match the merged tree. /// /// Conceptually, we examine the merge state of each item, and either leave the /// item unchanged, upload the local side, apply the remote side, or apply and /// then reupload the remote side with a new structure. /// /// Note that we update Places and flag items *before* upload, while iOS /// updates the mirror *after* a successful upload. This simplifies our /// implementation, though we lose idempotent merges. If upload is interrupted, /// the next sync won't distinguish between new merge states from the previous /// sync, and local changes. fn update_local_items_in_places<'t>( db: &Conn, driver: &Driver, controller: &AbortController, local_time_millis: i64, ops: &CompletionOps<'t>, ) -> Result<()> { debug!( driver, "Cleaning up observer notifications left from last sync" ); controller.err_if_aborted()?; db.exec( "DELETE FROM itemsAdded; DELETE FROM guidsChanged; DELETE FROM itemsChanged; DELETE FROM itemsRemoved; DELETE FROM itemsMoved;", )?; // Places uses microsecond timestamps for dates added and last modified // times, rounded to the nearest millisecond. Using `now` for the local // time lets us set modified times deterministically for tests. let now = local_time_millis * 1000; // Insert URLs for new remote items into the `moz_places` table. We need to // do this before inserting new remote items, since we need Place IDs for // both old and new URLs. debug!(driver, "Inserting Places for new items"); for chunk in ops.apply_remote_items.chunks(db.variable_limit()?) { let mut statement = db.prepare(format!( "INSERT OR IGNORE INTO moz_places(url, url_hash, rev_host, hidden, frecency, guid) SELECT u.url, u.hash, u.revHost, 0, (CASE v.kind WHEN {} THEN 0 ELSE -1 END), IFNULL((SELECT h.guid FROM moz_places h WHERE h.url_hash = u.hash AND h.url = u.url), u.guid) FROM items v JOIN urls u ON u.id = v.urlId WHERE v.guid IN ({})", mozISyncedBookmarksMerger::KIND_QUERY, repeat_sql_vars(chunk.len()), ))?; for (index, op) in chunk.iter().enumerate() { controller.err_if_aborted()?; let remote_guid = nsString::from(&*op.remote_node().guid); statement.bind_by_index(index as u32, remote_guid)?; } statement.execute()?; } // Trigger frecency updates for all new origins. debug!(driver, "Updating origins for new URLs"); controller.err_if_aborted()?; db.exec("DELETE FROM moz_updateoriginsinsert_temp")?; // Build a table of new and updated items. debug!(driver, "Staging apply remote item ops"); for chunk in ops.apply_remote_items.chunks(db.variable_limit()? / 3) { // CTEs in `WITH` clauses aren't indexed, so this query needs a full // table scan on `ops`. But that's okay; a separate temp table for ops // would also need a full scan. Note that we need both the local _and_ // remote GUIDs here, because we haven't changed the local GUIDs yet. let mut statement = db.prepare(format!( "WITH ops(mergedGuid, localGuid, remoteGuid, level) AS (VALUES {}) INSERT INTO itemsToApply(mergedGuid, localId, remoteId, remoteGuid, newLevel, newType, localDateAddedMicroseconds, remoteDateAddedMicroseconds, lastModifiedMicroseconds, oldTitle, newTitle, oldPlaceId, newPlaceId, newKeyword) SELECT n.mergedGuid, b.id, v.id, v.guid, n.level, (CASE WHEN v.kind IN ({}, {}) THEN {} WHEN v.kind IN ({}, {}) THEN {} ELSE {} END), b.dateAdded, v.dateAdded * 1000, MAX(v.dateAdded * 1000, {}), b.title, v.title, b.fk, (SELECT h.id FROM moz_places h JOIN urls u ON u.hash = h.url_hash WHERE u.id = v.urlId AND u.url = h.url), v.keyword FROM ops n JOIN items v ON v.guid = n.remoteGuid LEFT JOIN moz_bookmarks b ON b.guid = n.localGuid", repeat_display(chunk.len(), ",", |index, f| { let op = &chunk[index]; write!(f, "(?, ?, ?, {})", op.level) }), mozISyncedBookmarksMerger::KIND_BOOKMARK, mozISyncedBookmarksMerger::KIND_QUERY, nsINavBookmarksService::TYPE_BOOKMARK, mozISyncedBookmarksMerger::KIND_FOLDER, mozISyncedBookmarksMerger::KIND_LIVEMARK, nsINavBookmarksService::TYPE_FOLDER, nsINavBookmarksService::TYPE_SEPARATOR, now, ))?; for (index, op) in chunk.iter().enumerate() { controller.err_if_aborted()?; let offset = (index * 3) as u32; // In most cases, the merged and remote GUIDs are the same for new // items. For updates, all three are typically the same. We could // try to avoid binding duplicates, but that complicates chunking, // and we don't expect many items to change after the first sync. let merged_guid = nsString::from(&*op.merged_node.guid); statement.bind_by_index(offset, merged_guid)?; let local_guid = op .merged_node .merge_state .local_node() .map(|node| nsString::from(&*node.guid)); statement.bind_by_index(offset + 1, local_guid)?; let remote_guid = nsString::from(&*op.remote_node().guid); statement.bind_by_index(offset + 2, remote_guid)?; } statement.execute()?; } debug!(driver, "Staging change GUID ops"); for chunk in ops.change_guids.chunks(db.variable_limit()? / 2) { let mut statement = db.prepare(format!( "INSERT INTO changeGuidOps(localGuid, mergedGuid, syncStatus, level, lastModifiedMicroseconds) VALUES {}", repeat_display(chunk.len(), ",", |index, f| { let op = &chunk[index]; // If only the local GUID changed, the item was deduped, so we // can mark it as syncing. Otherwise, we changed an invalid // GUID locally or remotely, so we leave its original sync // status in place until we've uploaded it. let sync_status = if op.merged_node.remote_guid_changed() { None } else { Some(nsINavBookmarksService::SYNC_STATUS_NORMAL) }; write!( f, "(?, ?, {}, {}, {})", NullableFragment(sync_status), op.level, now ) }) ))?; for (index, op) in chunk.iter().enumerate() { controller.err_if_aborted()?; let offset = (index * 2) as u32; let local_guid = nsString::from(&*op.local_node().guid); statement.bind_by_index(offset, local_guid)?; let merged_guid = nsString::from(&*op.merged_node.guid); statement.bind_by_index(offset + 1, merged_guid)?; } statement.execute()?; } debug!(driver, "Staging apply new local structure ops"); for chunk in ops .apply_new_local_structure .chunks(db.variable_limit()? / 2) { let mut statement = db.prepare(format!( "INSERT INTO applyNewLocalStructureOps(mergedGuid, mergedParentGuid, position, level, lastModifiedMicroseconds) VALUES {}", repeat_display(chunk.len(), ",", |index, f| { let op = &chunk[index]; write!(f, "(?, ?, {}, {}, {})", op.position, op.level, now) }) ))?; for (index, op) in chunk.iter().enumerate() { controller.err_if_aborted()?; let offset = (index * 2) as u32; let merged_guid = nsString::from(&*op.merged_node.guid); statement.bind_by_index(offset, merged_guid)?; let merged_parent_guid = nsString::from(&*op.merged_parent_node.guid); statement.bind_by_index(offset + 1, merged_parent_guid)?; } statement.execute()?; } debug!(driver, "Removing tombstones for revived items"); for chunk in ops.delete_local_tombstones.chunks(db.variable_limit()?) { let mut statement = db.prepare(format!( "DELETE FROM moz_bookmarks_deleted WHERE guid IN ({})", repeat_sql_vars(chunk.len()), ))?; for (index, op) in chunk.iter().enumerate() { controller.err_if_aborted()?; statement.bind_by_index(index as u32, nsString::from(&*op.guid().as_str()))?; } statement.execute()?; } debug!( driver, "Inserting new tombstones for non-syncable and invalid items" ); for chunk in ops.insert_local_tombstones.chunks(db.variable_limit()?) { let mut statement = db.prepare(format!( "INSERT INTO moz_bookmarks_deleted(guid, dateRemoved) VALUES {}", repeat_display(chunk.len(), ",", |_, f| write!(f, "(?, {})", now)), ))?; for (index, op) in chunk.iter().enumerate() { controller.err_if_aborted()?; statement.bind_by_index( index as u32, nsString::from(&*op.remote_node().guid.as_str()), )?; } statement.execute()?; } debug!(driver, "Removing local items"); for chunk in ops.delete_local_items.chunks(db.variable_limit()?) { remove_local_items(&db, driver, controller, chunk)?; } // Fires the `changeGuids` trigger. debug!(driver, "Changing GUIDs"); controller.err_if_aborted()?; db.exec("DELETE FROM changeGuidOps")?; debug!(driver, "Applying remote items"); apply_remote_items(db, driver, controller)?; // Trigger frecency updates for all affected origins. debug!(driver, "Updating origins for changed URLs"); controller.err_if_aborted()?; db.exec("DELETE FROM moz_updateoriginsupdate_temp")?; // Fires the `applyNewLocalStructure` trigger. debug!(driver, "Applying new local structure"); controller.err_if_aborted()?; db.exec("DELETE FROM applyNewLocalStructureOps")?; debug!( driver, "Resetting change counters for items that shouldn't be uploaded" ); for chunk in ops.set_local_merged.chunks(db.variable_limit()?) { let mut statement = db.prepare(format!( "UPDATE moz_bookmarks SET syncChangeCounter = 0 WHERE guid IN ({})", repeat_sql_vars(chunk.len()), ))?; for (index, op) in chunk.iter().enumerate() { controller.err_if_aborted()?; statement.bind_by_index(index as u32, nsString::from(&*op.merged_node.guid))?; } statement.execute()?; } debug!( driver, "Bumping change counters for items that should be uploaded" ); for chunk in ops.set_local_unmerged.chunks(db.variable_limit()?) { let mut statement = db.prepare(format!( "UPDATE moz_bookmarks SET syncChangeCounter = 1 WHERE guid IN ({})", repeat_sql_vars(chunk.len()), ))?; for (index, op) in chunk.iter().enumerate() { controller.err_if_aborted()?; statement.bind_by_index(index as u32, nsString::from(&*op.merged_node.guid))?; } statement.execute()?; } debug!(driver, "Flagging applied remote items as merged"); for chunk in ops.set_remote_merged.chunks(db.variable_limit()?) { let mut statement = db.prepare(format!( "UPDATE items SET needsMerge = 0 WHERE guid IN ({})", repeat_sql_vars(chunk.len()), ))?; for (index, op) in chunk.iter().enumerate() { controller.err_if_aborted()?; statement.bind_by_index(index as u32, nsString::from(op.guid().as_str()))?; } statement.execute()?; } Ok(()) } /// Upserts all new and updated items from the `itemsToApply` table into Places. fn apply_remote_items(db: &Conn, driver: &Driver, controller: &AbortController) -> Result<()> { debug!(driver, "Recording item added notifications for new items"); controller.err_if_aborted()?; db.exec( "INSERT INTO itemsAdded(guid, keywordChanged, level) SELECT n.mergedGuid, n.newKeyword NOT NULL OR EXISTS(SELECT 1 FROM moz_keywords k WHERE k.place_id = n.newPlaceId OR k.keyword = n.newKeyword), n.newLevel FROM itemsToApply n WHERE n.localId IS NULL", )?; debug!( driver, "Recording item changed notifications for existing items" ); controller.err_if_aborted()?; db.exec( "INSERT INTO itemsChanged(itemId, oldTitle, oldPlaceId, keywordChanged, level) SELECT n.localId, n.oldTitle, n.oldPlaceId, n.newKeyword NOT NULL OR EXISTS( SELECT 1 FROM moz_keywords k WHERE k.place_id IN (n.oldPlaceId, n.newPlaceId) OR k.keyword = n.newKeyword ), n.newLevel FROM itemsToApply n WHERE n.localId NOT NULL", )?; // Remove all keywords from old and new URLs, and remove new keywords from // all existing URLs. The `NOT NULL` conditions are important; they ensure // that SQLite uses our partial indexes, instead of a table scan. debug!(driver, "Removing old keywords"); controller.err_if_aborted()?; db.exec( "DELETE FROM moz_keywords WHERE place_id IN (SELECT oldPlaceId FROM itemsToApply WHERE oldPlaceId NOT NULL) OR place_id IN (SELECT newPlaceId FROM itemsToApply WHERE newPlaceId NOT NULL) OR keyword IN (SELECT newKeyword FROM itemsToApply WHERE newKeyword NOT NULL) ", )?; debug!(driver, "Removing old tags"); controller.err_if_aborted()?; db.exec( "DELETE FROM localTags WHERE placeId IN (SELECT oldPlaceId FROM itemsToApply WHERE oldPlaceId NOT NULL) OR placeId IN (SELECT newPlaceId FROM itemsToApply WHERE newPlaceId NOT NULL)", )?; // Insert and update items, using -1 for new items' parent IDs and // positions. We'll update these later, when we apply the new local // structure. This is a full table scan on `itemsToApply`. The no-op // `WHERE` clause is necessary to avoid a parsing ambiguity. debug!(driver, "Upserting new items"); controller.err_if_aborted()?; db.exec(format!( "INSERT INTO moz_bookmarks(id, guid, parent, position, type, fk, title, dateAdded, lastModified, syncStatus, syncChangeCounter) SELECT localId, mergedGuid, -1, -1, newType, newPlaceId, newTitle, /* Pick the older of the local and remote date added. We'll weakly reupload any items with an older local date. */ MIN(IFNULL(localDateAddedMicroseconds, remoteDateAddedMicroseconds), remoteDateAddedMicroseconds), /* The last modified date should always be newer than the date added, so we pick the newer of the two here. */ MAX(lastModifiedMicroseconds, remoteDateAddedMicroseconds), {syncStatusNormal}, 0 FROM itemsToApply WHERE 1 ON CONFLICT(id) DO UPDATE SET title = excluded.title, dateAdded = excluded.dateAdded, lastModified = excluded.lastModified, syncStatus = {syncStatusNormal}, /* It's important that we update the URL *after* removing old keywords and *before* inserting new ones, so that the above DELETEs select the correct affected items. */ fk = excluded.fk", syncStatusNormal = nsINavBookmarksService::SYNC_STATUS_NORMAL ))?; // The roots are never in `itemsToApply` but still need to (well, at least // *should*) have a syncStatus of NORMAL after a reconcilliation. The // ROOT_GUID doesn't matter in practice, but we include it to be consistent. db.exec(format!( "UPDATE moz_bookmarks SET syncStatus={syncStatusNormal} WHERE guid IN {user_roots} OR guid = '{root}' ", syncStatusNormal = nsINavBookmarksService::SYNC_STATUS_NORMAL, root = dogear::ROOT_GUID, user_roots = user_roots_as_sql_set(), ))?; // Flag frecencies for recalculation. This is a multi-index OR that uses the // `oldPlacesIds` and `newPlaceIds` partial indexes, since `<>` is only true // if both terms are not NULL. Without those constraints, the subqueries // would scan `itemsToApply` twice. The `oldPlaceId <> newPlaceId` and // `newPlaceId <> oldPlaceId` checks exclude items where the URL didn't // change; we don't need to recalculate their frecencies. debug!(driver, "Flagging frecencies for recalculation"); controller.err_if_aborted()?; db.exec( "UPDATE moz_places SET recalc_frecency = 1 WHERE frecency <> 0 AND ( id IN ( SELECT oldPlaceId FROM itemsToApply WHERE oldPlaceId <> newPlaceId ) OR id IN ( SELECT newPlaceId FROM itemsToApply WHERE newPlaceId <> oldPlaceId ) )", )?; debug!(driver, "Inserting new keywords for new URLs"); controller.err_if_aborted()?; db.exec( "INSERT OR IGNORE INTO moz_keywords(keyword, place_id, post_data) SELECT newKeyword, newPlaceId, '' FROM itemsToApply WHERE newKeyword NOT NULL", )?; debug!(driver, "Inserting new tags for new URLs"); controller.err_if_aborted()?; db.exec( "INSERT INTO localTags(tag, placeId, lastModifiedMicroseconds) SELECT t.tag, n.newPlaceId, n.lastModifiedMicroseconds FROM itemsToApply n JOIN tags t ON t.itemId = n.remoteId", )?; Ok(()) } /// Removes deleted local items from Places. fn remove_local_items( db: &Conn, driver: &Driver, controller: &AbortController, ops: &[DeleteLocalItem], ) -> Result<()> { debug!(driver, "Recording observer notifications for removed items"); let mut observer_statement = db.prepare(format!( "WITH ops(guid, level) AS (VALUES {}) INSERT INTO itemsRemoved(itemId, parentId, position, type, title, placeId, guid, parentGuid, level) SELECT b.id, b.parent, b.position, b.type, IFNULL(b.title, \"\"), b.fk, b.guid, p.guid, n.level FROM ops n JOIN moz_bookmarks b ON b.guid = n.guid JOIN moz_bookmarks p ON p.id = b.parent", repeat_display(ops.len(), ",", |index, f| { let op = &ops[index]; write!(f, "(?, {})", op.local_node().level()) }), ))?; for (index, op) in ops.iter().enumerate() { controller.err_if_aborted()?; observer_statement.bind_by_index( index as u32, nsString::from(&*op.local_node().guid.as_str()), )?; } observer_statement.execute()?; debug!(driver, "Recalculating frecencies for removed bookmark URLs"); let mut frecency_statement = db.prepare(format!( "UPDATE moz_places SET recalc_frecency = 1 WHERE id IN (SELECT b.fk FROM moz_bookmarks b WHERE b.guid IN ({})) AND frecency <> 0", repeat_sql_vars(ops.len()) ))?; for (index, op) in ops.iter().enumerate() { controller.err_if_aborted()?; frecency_statement.bind_by_index( index as u32, nsString::from(&*op.local_node().guid.as_str()), )?; } frecency_statement.execute()?; // This can be removed in bug 1460577. debug!(driver, "Removing annos for deleted items"); let mut annos_statement = db.prepare(format!( "DELETE FROM moz_items_annos WHERE item_id = (SELECT b.id FROM moz_bookmarks b WHERE b.guid IN ({}))", repeat_sql_vars(ops.len()), ))?; for (index, op) in ops.iter().enumerate() { controller.err_if_aborted()?; annos_statement.bind_by_index( index as u32, nsString::from(&*op.local_node().guid.as_str()), )?; } annos_statement.execute()?; debug!(driver, "Removing deleted items from Places"); let mut delete_statement = db.prepare(format!( "DELETE FROM moz_bookmarks WHERE guid IN ({})", repeat_sql_vars(ops.len()), ))?; for (index, op) in ops.iter().enumerate() { controller.err_if_aborted()?; delete_statement.bind_by_index( index as u32, nsString::from(&*op.local_node().guid.as_str()), )?; } delete_statement.execute()?; Ok(()) } /// Stores a snapshot of all locally changed items in a temporary table for /// upload. This is called from within the merge transaction, to ensure that /// changes made during the sync don't cause us to upload inconsistent records. /// /// For an example of why we use a temporary table instead of reading directly /// from Places, consider a user adding a bookmark, then changing its parent /// folder. We first add the bookmark to the default folder, bump the change /// counter of the new bookmark and the default folder, then trigger a sync. /// Depending on how quickly the user picks the new parent, we might upload /// a record for the default folder, commit the move, then upload the bookmark. /// We'll still upload the new parent on the next sync, but, in the meantime, /// we've introduced a parent-child disagreement. This can also happen if the /// user moves many items between folders. /// /// Conceptually, `itemsToUpload` is a transient "view" of locally changed /// items. The change counter in Places is the persistent record of items that /// we need to upload, so, if upload is interrupted or fails, we'll stage the /// items again on the next sync. fn stage_items_to_upload( db: &Conn, driver: &Driver, controller: &AbortController, upload_items: &[UploadItem], upload_tombstones: &[UploadTombstone], ) -> Result<()> { debug!(driver, "Cleaning up staged items left from last sync"); controller.err_if_aborted()?; db.exec("DELETE FROM itemsToUpload")?; // Stage remotely changed items with older local creation dates. These are // tracked "weakly": if the upload is interrupted or fails, we won't // reupload the record on the next sync. debug!(driver, "Staging items with older local dates added"); controller.err_if_aborted()?; db.exec(format!( "INSERT OR IGNORE INTO itemsToUpload(id, guid, syncChangeCounter, parentGuid, parentTitle, dateAdded, type, title, placeId, isQuery, url, keyword, position, tagFolderName) {} JOIN itemsToApply n ON n.mergedGuid = b.guid WHERE n.localDateAddedMicroseconds < n.remoteDateAddedMicroseconds", UploadItemsFragment("b"), ))?; debug!(driver, "Staging remaining locally changed items for upload"); for chunk in upload_items.chunks(db.variable_limit()?) { let mut statement = db.prepare(format!( "INSERT OR IGNORE INTO itemsToUpload(id, guid, syncChangeCounter, parentGuid, parentTitle, dateAdded, type, title, placeId, isQuery, url, keyword, position, tagFolderName) {} WHERE b.guid IN ({})", UploadItemsFragment("b"), repeat_sql_vars(chunk.len()), ))?; for (index, op) in chunk.iter().enumerate() { controller.err_if_aborted()?; statement.bind_by_index(index as u32, nsString::from(&*op.merged_node.guid))?; } statement.execute()?; } // Record the child GUIDs of locally changed folders, which we use to // populate the `children` array in the record. debug!(driver, "Staging structure to upload"); controller.err_if_aborted()?; db.exec( " INSERT INTO structureToUpload(guid, parentId, position) SELECT b.guid, b.parent, b.position FROM moz_bookmarks b JOIN itemsToUpload o ON o.id = b.parent", )?; // Stage tags for outgoing bookmarks. debug!(driver, "Staging tags to upload"); controller.err_if_aborted()?; db.exec( " INSERT OR IGNORE INTO tagsToUpload(id, tag) SELECT o.id, t.tag FROM localTags t JOIN itemsToUpload o ON o.placeId = t.placeId", )?; // Finally, stage tombstones for deleted items. Ignore conflicts if we have // tombstones for undeleted items; Places Maintenance should clean these up. debug!(driver, "Staging tombstones to upload"); for chunk in upload_tombstones.chunks(db.variable_limit()?) { let mut statement = db.prepare(format!( "INSERT OR IGNORE INTO itemsToUpload(guid, syncChangeCounter, isDeleted) VALUES {}", repeat_display(chunk.len(), ",", |_, f| write!(f, "(?, 1, 1)")) ))?; for (index, op) in chunk.iter().enumerate() { controller.err_if_aborted()?; statement.bind_by_index(index as u32, nsString::from(op.guid().as_str()))?; } statement.execute()?; } Ok(()) } fn cleanup(db: &Conn) -> Result<()> { db.exec("DELETE FROM itemsToApply")?; Ok(()) } /// Formats a list of binding parameters for inclusion in a SQL list. #[inline] fn repeat_sql_vars(count: usize) -> impl fmt::Display { repeat_display(count, ",", |_, f| write!(f, "?")) } /// Construct a `RepeatDisplay` that will repeatedly call `fmt_one` with a /// formatter `count` times, separated by `sep`. This is copied from the /// `sql_support` crate in `application-services`. #[inline] fn repeat_display<'a, F>(count: usize, sep: &'a str, fmt_one: F) -> RepeatDisplay<'a, F> where F: Fn(usize, &mut fmt::Formatter) -> fmt::Result, { RepeatDisplay { count, sep, fmt_one, } } /// Helper type for printing repeated strings more efficiently. #[derive(Debug, Clone)] struct RepeatDisplay<'a, F> { count: usize, sep: &'a str, fmt_one: F, } impl<'a, F> fmt::Display for RepeatDisplay<'a, F> where F: Fn(usize, &mut fmt::Formatter) -> fmt::Result, { fn fmt(&self, f: &mut fmt::Formatter) -> fmt::Result { for i in 0..self.count { if i != 0 { f.write_str(self.sep)?; } (self.fmt_one)(i, f)?; } Ok(()) } } /// Converts between a type `T` and its SQL representation. trait Column { fn from_column(raw: T) -> Result where Self: Sized; fn into_column(self) -> T; } impl Column for Kind { fn from_column(raw: i64) -> Result { Ok(match i16::try_from(raw) { Ok(mozISyncedBookmarksMerger::KIND_BOOKMARK) => Kind::Bookmark, Ok(mozISyncedBookmarksMerger::KIND_QUERY) => Kind::Query, Ok(mozISyncedBookmarksMerger::KIND_FOLDER) => Kind::Folder, Ok(mozISyncedBookmarksMerger::KIND_LIVEMARK) => Kind::Livemark, Ok(mozISyncedBookmarksMerger::KIND_SEPARATOR) => Kind::Separator, _ => return Err(Error::UnknownItemKind(raw)), }) } fn into_column(self) -> i64 { match self { Kind::Bookmark => mozISyncedBookmarksMerger::KIND_BOOKMARK as i64, Kind::Query => mozISyncedBookmarksMerger::KIND_QUERY as i64, Kind::Folder => mozISyncedBookmarksMerger::KIND_FOLDER as i64, Kind::Livemark => mozISyncedBookmarksMerger::KIND_LIVEMARK as i64, Kind::Separator => mozISyncedBookmarksMerger::KIND_SEPARATOR as i64, } } } impl Column for Validity { fn from_column(raw: i64) -> Result { Ok(match i16::try_from(raw) { Ok(mozISyncedBookmarksMerger::VALIDITY_VALID) => Validity::Valid, Ok(mozISyncedBookmarksMerger::VALIDITY_REUPLOAD) => Validity::Reupload, Ok(mozISyncedBookmarksMerger::VALIDITY_REPLACE) => Validity::Replace, _ => return Err(Error::UnknownItemValidity(raw).into()), }) } fn into_column(self) -> i64 { match self { Validity::Valid => mozISyncedBookmarksMerger::VALIDITY_VALID as i64, Validity::Reupload => mozISyncedBookmarksMerger::VALIDITY_REUPLOAD as i64, Validity::Replace => mozISyncedBookmarksMerger::VALIDITY_REPLACE as i64, } } } /// Formats an optional value so that it can be included in a SQL statement. struct NullableFragment(Option); impl fmt::Display for NullableFragment where T: fmt::Display, { fn fmt(&self, f: &mut fmt::Formatter) -> fmt::Result { match &self.0 { Some(v) => v.fmt(f), None => write!(f, "NULL"), } } } /// Formats a `SELECT` statement for staging local items in the `itemsToUpload` /// table. struct UploadItemsFragment(&'static str); impl fmt::Display for UploadItemsFragment { fn fmt(&self, f: &mut fmt::Formatter<'_>) -> fmt::Result { write!( f, "SELECT {0}.id, {0}.guid, {}.syncChangeCounter, p.guid AS parentGuid, p.title AS parentTitle, {0}.dateAdded / 1000 AS dateAdded, {0}.type, {0}.title, h.id AS placeId, IFNULL(substr(h.url, 1, 6) = 'place:', 0) AS isQuery, h.url, (SELECT keyword FROM moz_keywords WHERE place_id = h.id), {0}.position, (SELECT get_query_param(substr(url, 7), 'tag') WHERE substr(h.url, 1, 6) = 'place:') AS tagFolderName FROM moz_bookmarks {0} JOIN moz_bookmarks p ON p.id = {0}.parent LEFT JOIN moz_places h ON h.id = {0}.fk", self.0 ) } } pub enum ApplyStatus { Merged, Skipped, } impl From for bool { fn from(status: ApplyStatus) -> bool { match status { ApplyStatus::Merged => true, ApplyStatus::Skipped => false, } } }