summaryrefslogtreecommitdiffstats
path: root/toolkit/components/places/bookmark_sync
diff options
context:
space:
mode:
Diffstat (limited to 'toolkit/components/places/bookmark_sync')
-rw-r--r--toolkit/components/places/bookmark_sync/Cargo.toml20
-rw-r--r--toolkit/components/places/bookmark_sync/src/driver.rs259
-rw-r--r--toolkit/components/places/bookmark_sync/src/error.rs106
-rw-r--r--toolkit/components/places/bookmark_sync/src/lib.rs27
-rw-r--r--toolkit/components/places/bookmark_sync/src/merger.rs237
-rw-r--r--toolkit/components/places/bookmark_sync/src/store.rs1314
6 files changed, 1963 insertions, 0 deletions
diff --git a/toolkit/components/places/bookmark_sync/Cargo.toml b/toolkit/components/places/bookmark_sync/Cargo.toml
new file mode 100644
index 0000000000..71cd764c4b
--- /dev/null
+++ b/toolkit/components/places/bookmark_sync/Cargo.toml
@@ -0,0 +1,20 @@
+[package]
+name = "bookmark_sync"
+version = "0.1.0"
+authors = ["Lina Cambridge <lina@yakshaving.ninja>"]
+edition = "2018"
+license = "MPL-2.0"
+
+[dependencies]
+atomic_refcell = "0.1"
+dogear = "0.5.0"
+libc = "0.2"
+log = "0.4"
+cstr = "0.2"
+moz_task = { path = "../../../../xpcom/rust/moz_task" }
+nserror = { path = "../../../../xpcom/rust/nserror" }
+nsstring = { path = "../../../../xpcom/rust/nsstring" }
+storage = { path = "../../../../storage/rust" }
+storage_variant = { path = "../../../../storage/variant" }
+url = "2.0"
+xpcom = { path = "../../../../xpcom/rust/xpcom" }
diff --git a/toolkit/components/places/bookmark_sync/src/driver.rs b/toolkit/components/places/bookmark_sync/src/driver.rs
new file mode 100644
index 0000000000..43bb1f1f79
--- /dev/null
+++ b/toolkit/components/places/bookmark_sync/src/driver.rs
@@ -0,0 +1,259 @@
+/* This Source Code Form is subject to the terms of the Mozilla Public
+ * License, v. 2.0. If a copy of the MPL was not distributed with this
+ * file, You can obtain one at http://mozilla.org/MPL/2.0/. */
+
+use std::{
+ fmt::Write,
+ sync::atomic::{AtomicBool, Ordering},
+ time::Duration,
+};
+
+use dogear::{AbortSignal, Guid, ProblemCounts, StructureCounts, TelemetryEvent};
+use log::{Level, LevelFilter, Log, Metadata, Record};
+use moz_task::{Task, TaskRunnable, ThreadPtrHandle};
+use nserror::nsresult;
+use nsstring::{nsACString, nsCString, nsString};
+use storage_variant::HashPropertyBag;
+use xpcom::interfaces::{mozIServicesLogSink, mozISyncedBookmarksMirrorProgressListener};
+
+extern "C" {
+ fn NS_GeneratePlacesGUID(guid: *mut nsACString) -> nsresult;
+}
+
+fn generate_guid() -> Result<nsCString, nsresult> {
+ let mut guid = nsCString::new();
+ let rv = unsafe { NS_GeneratePlacesGUID(&mut *guid) };
+ if rv.succeeded() {
+ Ok(guid)
+ } else {
+ Err(rv)
+ }
+}
+
+/// An abort controller is used to abort merges running on the storage thread
+/// from the main thread. Its design is based on the DOM API of the same name.
+pub struct AbortController {
+ aborted: AtomicBool,
+}
+
+impl AbortController {
+ /// Signals the store to stop merging as soon as it can.
+ pub fn abort(&self) {
+ self.aborted.store(true, Ordering::Release)
+ }
+}
+
+impl Default for AbortController {
+ fn default() -> AbortController {
+ AbortController {
+ aborted: AtomicBool::new(false),
+ }
+ }
+}
+
+impl AbortSignal for AbortController {
+ fn aborted(&self) -> bool {
+ self.aborted.load(Ordering::Acquire)
+ }
+}
+
+/// The merger driver, created and used on the storage thread.
+pub struct Driver {
+ log: Logger,
+ progress: Option<ThreadPtrHandle<mozISyncedBookmarksMirrorProgressListener>>,
+}
+
+impl Driver {
+ #[inline]
+ pub fn new(
+ log: Logger,
+ progress: Option<ThreadPtrHandle<mozISyncedBookmarksMirrorProgressListener>>,
+ ) -> Driver {
+ Driver { log, progress }
+ }
+}
+
+impl dogear::Driver for Driver {
+ fn generate_new_guid(&self, invalid_guid: &Guid) -> dogear::Result<Guid> {
+ generate_guid()
+ .map_err(|_| dogear::ErrorKind::InvalidGuid(invalid_guid.clone()).into())
+ .and_then(|s| Guid::from_utf8(s.as_ref()))
+ }
+
+ #[inline]
+ fn max_log_level(&self) -> LevelFilter {
+ self.log.max_level
+ }
+
+ #[inline]
+ fn logger(&self) -> &dyn Log {
+ &self.log
+ }
+
+ fn record_telemetry_event(&self, event: TelemetryEvent) {
+ if let Some(ref progress) = self.progress {
+ let task = RecordTelemetryEventTask {
+ progress: progress.clone(),
+ event,
+ };
+ let _ = TaskRunnable::new(
+ "bookmark_sync::Driver::record_telemetry_event",
+ Box::new(task),
+ )
+ .and_then(|r| TaskRunnable::dispatch(r, progress.owning_thread()));
+ }
+ }
+}
+
+pub struct Logger {
+ pub max_level: LevelFilter,
+ logger: Option<ThreadPtrHandle<mozIServicesLogSink>>,
+}
+
+impl Logger {
+ #[inline]
+ pub fn new(
+ max_level: LevelFilter,
+ logger: Option<ThreadPtrHandle<mozIServicesLogSink>>,
+ ) -> Logger {
+ Logger { max_level, logger }
+ }
+}
+
+impl Log for Logger {
+ #[inline]
+ fn enabled(&self, meta: &Metadata) -> bool {
+ self.logger.is_some() && meta.level() <= self.max_level
+ }
+
+ fn log(&self, record: &Record) {
+ if !self.enabled(record.metadata()) {
+ return;
+ }
+ if let Some(logger) = &self.logger {
+ let mut message = nsString::new();
+ match write!(message, "{}", record.args()) {
+ Ok(_) => {
+ let task = LogTask {
+ logger: logger.clone(),
+ level: record.metadata().level(),
+ message,
+ };
+ let _ = TaskRunnable::new("bookmark_sync::Logger::log", Box::new(task))
+ .and_then(|r| TaskRunnable::dispatch(r, logger.owning_thread()));
+ }
+ Err(_) => {}
+ }
+ }
+ }
+
+ fn flush(&self) {}
+}
+
+/// Logs a message to the mirror logger. This task is created on the async
+/// thread, and dispatched to the main thread.
+struct LogTask {
+ logger: ThreadPtrHandle<mozIServicesLogSink>,
+ level: Level,
+ message: nsString,
+}
+
+impl Task for LogTask {
+ fn run(&self) {
+ let logger = self.logger.get().unwrap();
+ match self.level {
+ Level::Error => unsafe {
+ logger.Error(&*self.message);
+ },
+ Level::Warn => unsafe {
+ logger.Warn(&*self.message);
+ },
+ Level::Debug => unsafe {
+ logger.Debug(&*self.message);
+ },
+ Level::Trace => unsafe {
+ logger.Trace(&*self.message);
+ },
+ _ => {}
+ }
+ }
+
+ fn done(&self) -> Result<(), nsresult> {
+ Ok(())
+ }
+}
+
+/// Calls a progress listener callback for a merge telemetry event. This task is
+/// created on the async thread, and dispatched to the main thread.
+struct RecordTelemetryEventTask {
+ progress: ThreadPtrHandle<mozISyncedBookmarksMirrorProgressListener>,
+ event: TelemetryEvent,
+}
+
+impl Task for RecordTelemetryEventTask {
+ fn run(&self) {
+ let callback = self.progress.get().unwrap();
+ let _ = match &self.event {
+ TelemetryEvent::FetchLocalTree(stats) => unsafe {
+ callback.OnFetchLocalTree(
+ as_millis(stats.time),
+ stats.items as i64,
+ stats.deletions as i64,
+ problem_counts_to_bag(&stats.problems).bag().coerce(),
+ )
+ },
+ TelemetryEvent::FetchRemoteTree(stats) => unsafe {
+ callback.OnFetchRemoteTree(
+ as_millis(stats.time),
+ stats.items as i64,
+ stats.deletions as i64,
+ problem_counts_to_bag(&stats.problems).bag().coerce(),
+ )
+ },
+ TelemetryEvent::Merge(time, counts) => unsafe {
+ callback.OnMerge(
+ as_millis(*time),
+ structure_counts_to_bag(counts).bag().coerce(),
+ )
+ },
+ TelemetryEvent::Apply(time) => unsafe { callback.OnApply(as_millis(*time)) },
+ };
+ }
+
+ fn done(&self) -> std::result::Result<(), nsresult> {
+ Ok(())
+ }
+}
+
+fn as_millis(d: Duration) -> i64 {
+ d.as_secs() as i64 * 1000 + i64::from(d.subsec_millis())
+}
+
+fn problem_counts_to_bag(problems: &ProblemCounts) -> HashPropertyBag {
+ let mut bag = HashPropertyBag::new();
+ bag.set("orphans", problems.orphans as i64);
+ bag.set("misparentedRoots", problems.misparented_roots as i64);
+ bag.set(
+ "multipleParents",
+ problems.multiple_parents_by_children as i64,
+ );
+ bag.set("missingParents", problems.missing_parent_guids as i64);
+ bag.set("nonFolderParents", problems.non_folder_parent_guids as i64);
+ bag.set(
+ "parentChildDisagreements",
+ problems.parent_child_disagreements as i64,
+ );
+ bag.set("missingChildren", problems.missing_children as i64);
+ bag
+}
+
+fn structure_counts_to_bag(counts: &StructureCounts) -> HashPropertyBag {
+ let mut bag = HashPropertyBag::new();
+ bag.set("remoteRevives", counts.remote_revives as i64);
+ bag.set("localDeletes", counts.local_deletes as i64);
+ bag.set("localRevives", counts.local_revives as i64);
+ bag.set("remoteDeletes", counts.remote_deletes as i64);
+ bag.set("dupes", counts.dupes as i64);
+ bag.set("items", counts.merged_nodes as i64);
+ bag
+}
diff --git a/toolkit/components/places/bookmark_sync/src/error.rs b/toolkit/components/places/bookmark_sync/src/error.rs
new file mode 100644
index 0000000000..5ac7c5b62c
--- /dev/null
+++ b/toolkit/components/places/bookmark_sync/src/error.rs
@@ -0,0 +1,106 @@
+/* This Source Code Form is subject to the terms of the Mozilla Public
+ * License, v. 2.0. If a copy of the MPL was not distributed with this
+ * file, You can obtain one at http://mozilla.org/MPL/2.0/. */
+
+use std::{error, fmt, result, string::FromUtf16Error};
+
+use nserror::{
+ nsresult, NS_ERROR_ABORT, NS_ERROR_FAILURE, NS_ERROR_INVALID_ARG, NS_ERROR_STORAGE_BUSY,
+ NS_ERROR_UNEXPECTED,
+};
+
+pub type Result<T> = result::Result<T, Error>;
+
+#[derive(Debug)]
+pub enum Error {
+ Dogear(dogear::Error),
+ Storage(storage::Error),
+ InvalidLocalRoots,
+ InvalidRemoteRoots,
+ Nsresult(nsresult),
+ UnknownItemType(i64),
+ UnknownItemKind(i64),
+ MalformedString(Box<dyn error::Error + Send + Sync + 'static>),
+ MergeConflict,
+ StorageBusy,
+ UnknownItemValidity(i64),
+ DidNotRun,
+}
+
+impl error::Error for Error {
+ fn source(&self) -> Option<&(dyn error::Error + 'static)> {
+ match self {
+ Error::Dogear(err) => Some(err),
+ Error::Storage(err) => Some(err),
+ _ => None,
+ }
+ }
+}
+
+impl From<dogear::Error> for Error {
+ fn from(err: dogear::Error) -> Error {
+ Error::Dogear(err)
+ }
+}
+
+impl From<storage::Error> for Error {
+ fn from(err: storage::Error) -> Error {
+ Error::Storage(err)
+ }
+}
+
+impl From<nsresult> for Error {
+ fn from(result: nsresult) -> Error {
+ Error::Nsresult(result)
+ }
+}
+
+impl From<FromUtf16Error> for Error {
+ fn from(error: FromUtf16Error) -> Error {
+ Error::MalformedString(error.into())
+ }
+}
+
+impl From<Error> for nsresult {
+ fn from(error: Error) -> nsresult {
+ match error {
+ Error::Dogear(err) => match err.kind() {
+ dogear::ErrorKind::Abort => NS_ERROR_ABORT,
+ _ => NS_ERROR_FAILURE,
+ },
+ Error::InvalidLocalRoots | Error::InvalidRemoteRoots | Error::DidNotRun => {
+ NS_ERROR_UNEXPECTED
+ }
+ Error::Storage(err) => err.into(),
+ Error::Nsresult(result) => result.clone(),
+ Error::UnknownItemType(_)
+ | Error::UnknownItemKind(_)
+ | Error::MalformedString(_)
+ | Error::UnknownItemValidity(_) => NS_ERROR_INVALID_ARG,
+ Error::MergeConflict | Error::StorageBusy => NS_ERROR_STORAGE_BUSY,
+ }
+ }
+}
+
+impl fmt::Display for Error {
+ fn fmt(&self, f: &mut fmt::Formatter) -> fmt::Result {
+ match self {
+ Error::Dogear(err) => err.fmt(f),
+ Error::Storage(err) => err.fmt(f),
+ Error::InvalidLocalRoots => f.write_str("The Places roots are invalid"),
+ Error::InvalidRemoteRoots => {
+ f.write_str("The roots in the mirror database are invalid")
+ }
+ Error::Nsresult(result) => write!(f, "Operation failed with {}", result.error_name()),
+ Error::UnknownItemType(typ) => write!(f, "Unknown item type {} in Places", typ),
+ Error::UnknownItemKind(kind) => write!(f, "Unknown item kind {} in mirror", kind),
+ Error::MalformedString(err) => err.fmt(f),
+ Error::MergeConflict => f.write_str("Local tree changed during merge"),
+ Error::StorageBusy => f.write_str("The database is busy"),
+ Error::UnknownItemValidity(validity) => {
+ write!(f, "Unknown item validity {} in database", validity)
+ }
+ Error::DidNotRun => write!(f, "Failed to run merge on storage thread"),
+ }
+ }
+}
diff --git a/toolkit/components/places/bookmark_sync/src/lib.rs b/toolkit/components/places/bookmark_sync/src/lib.rs
new file mode 100644
index 0000000000..a0dadd51c2
--- /dev/null
+++ b/toolkit/components/places/bookmark_sync/src/lib.rs
@@ -0,0 +1,27 @@
+/* This Source Code Form is subject to the terms of the Mozilla Public
+ * License, v. 2.0. If a copy of the MPL was not distributed with this
+ * file, You can obtain one at http://mozilla.org/MPL/2.0/. */
+
+#![allow(non_snake_case)]
+
+#[macro_use]
+extern crate cstr;
+#[macro_use]
+extern crate xpcom;
+
+mod driver;
+mod error;
+mod merger;
+mod store;
+
+use xpcom::{interfaces::mozISyncedBookmarksMerger, RefPtr};
+
+use crate::merger::SyncedBookmarksMerger;
+
+#[no_mangle]
+pub unsafe extern "C" fn NS_NewSyncedBookmarksMerger(
+ result: *mut *const mozISyncedBookmarksMerger,
+) {
+ let merger = SyncedBookmarksMerger::new();
+ RefPtr::new(merger.coerce::<mozISyncedBookmarksMerger>()).forget(&mut *result);
+}
diff --git a/toolkit/components/places/bookmark_sync/src/merger.rs b/toolkit/components/places/bookmark_sync/src/merger.rs
new file mode 100644
index 0000000000..ca0fc17a75
--- /dev/null
+++ b/toolkit/components/places/bookmark_sync/src/merger.rs
@@ -0,0 +1,237 @@
+/* This Source Code Form is subject to the terms of the Mozilla Public
+ * License, v. 2.0. If a copy of the MPL was not distributed with this
+ * file, You can obtain one at http://mozilla.org/MPL/2.0/. */
+
+use std::{cell::RefCell, fmt::Write, mem, sync::Arc};
+
+use atomic_refcell::AtomicRefCell;
+use dogear::Store;
+use log::LevelFilter;
+use moz_task::{Task, TaskRunnable, ThreadPtrHandle, ThreadPtrHolder};
+use nserror::{nsresult, NS_ERROR_NOT_AVAILABLE, NS_OK};
+use nsstring::nsString;
+use storage::Conn;
+use xpcom::{
+ interfaces::{
+ mozIPlacesPendingOperation, mozIServicesLogSink, mozIStorageConnection,
+ mozISyncedBookmarksMirrorCallback, mozISyncedBookmarksMirrorProgressListener,
+ },
+ RefPtr, XpCom,
+};
+
+use crate::driver::{AbortController, Driver, Logger};
+use crate::error;
+use crate::store;
+
+#[xpcom(implement(mozISyncedBookmarksMerger), nonatomic)]
+pub struct SyncedBookmarksMerger {
+ db: RefCell<Option<Conn>>,
+ logger: RefCell<Option<RefPtr<mozIServicesLogSink>>>,
+}
+
+impl SyncedBookmarksMerger {
+ pub fn new() -> RefPtr<SyncedBookmarksMerger> {
+ SyncedBookmarksMerger::allocate(InitSyncedBookmarksMerger {
+ db: RefCell::default(),
+ logger: RefCell::default(),
+ })
+ }
+
+ xpcom_method!(get_db => GetDb() -> *const mozIStorageConnection);
+ fn get_db(&self) -> Result<RefPtr<mozIStorageConnection>, nsresult> {
+ self.db
+ .borrow()
+ .as_ref()
+ .map(|db| RefPtr::new(db.connection()))
+ .ok_or(NS_OK)
+ }
+
+ xpcom_method!(set_db => SetDb(connection: *const mozIStorageConnection));
+ fn set_db(&self, connection: Option<&mozIStorageConnection>) -> Result<(), nsresult> {
+ self.db
+ .replace(connection.map(|connection| Conn::wrap(RefPtr::new(connection))));
+ Ok(())
+ }
+
+ xpcom_method!(get_logger => GetLogger() -> *const mozIServicesLogSink);
+ fn get_logger(&self) -> Result<RefPtr<mozIServicesLogSink>, nsresult> {
+ match *self.logger.borrow() {
+ Some(ref logger) => Ok(logger.clone()),
+ None => Err(NS_OK),
+ }
+ }
+
+ xpcom_method!(set_logger => SetLogger(logger: *const mozIServicesLogSink));
+ fn set_logger(&self, logger: Option<&mozIServicesLogSink>) -> Result<(), nsresult> {
+ self.logger.replace(logger.map(RefPtr::new));
+ Ok(())
+ }
+
+ xpcom_method!(
+ merge => Merge(
+ local_time_seconds: i64,
+ remote_time_seconds: i64,
+ callback: *const mozISyncedBookmarksMirrorCallback
+ ) -> *const mozIPlacesPendingOperation
+ );
+ fn merge(
+ &self,
+ local_time_seconds: i64,
+ remote_time_seconds: i64,
+ callback: &mozISyncedBookmarksMirrorCallback,
+ ) -> Result<RefPtr<mozIPlacesPendingOperation>, nsresult> {
+ let callback = RefPtr::new(callback);
+ let db = match *self.db.borrow() {
+ Some(ref db) => db.clone(),
+ None => return Err(NS_ERROR_NOT_AVAILABLE),
+ };
+ let logger = &*self.logger.borrow();
+ let async_thread = db.thread()?;
+ let controller = Arc::new(AbortController::default());
+ let task = MergeTask::new(
+ &db,
+ Arc::clone(&controller),
+ logger.as_ref().cloned(),
+ local_time_seconds,
+ remote_time_seconds,
+ callback,
+ )?;
+ let runnable = TaskRunnable::new(
+ "bookmark_sync::SyncedBookmarksMerger::merge",
+ Box::new(task),
+ )?;
+ TaskRunnable::dispatch(runnable, &async_thread)?;
+ let op = MergeOp::new(controller);
+ Ok(RefPtr::new(op.coerce()))
+ }
+
+ xpcom_method!(reset => Reset());
+ fn reset(&self) -> Result<(), nsresult> {
+ mem::drop(self.db.borrow_mut().take());
+ mem::drop(self.logger.borrow_mut().take());
+ Ok(())
+ }
+}
+
+struct MergeTask {
+ db: Conn,
+ controller: Arc<AbortController>,
+ max_log_level: LevelFilter,
+ logger: Option<ThreadPtrHandle<mozIServicesLogSink>>,
+ local_time_millis: i64,
+ remote_time_millis: i64,
+ progress: Option<ThreadPtrHandle<mozISyncedBookmarksMirrorProgressListener>>,
+ callback: ThreadPtrHandle<mozISyncedBookmarksMirrorCallback>,
+ result: AtomicRefCell<error::Result<store::ApplyStatus>>,
+}
+
+impl MergeTask {
+ fn new(
+ db: &Conn,
+ controller: Arc<AbortController>,
+ logger: Option<RefPtr<mozIServicesLogSink>>,
+ local_time_seconds: i64,
+ remote_time_seconds: i64,
+ callback: RefPtr<mozISyncedBookmarksMirrorCallback>,
+ ) -> Result<MergeTask, nsresult> {
+ let max_log_level = logger
+ .as_ref()
+ .and_then(|logger| {
+ let mut level = 0i16;
+ unsafe { logger.GetMaxLevel(&mut level) }.to_result().ok()?;
+ Some(level)
+ })
+ .map(|level| match level {
+ mozIServicesLogSink::LEVEL_ERROR => LevelFilter::Error,
+ mozIServicesLogSink::LEVEL_WARN => LevelFilter::Warn,
+ mozIServicesLogSink::LEVEL_DEBUG => LevelFilter::Debug,
+ mozIServicesLogSink::LEVEL_TRACE => LevelFilter::Trace,
+ _ => LevelFilter::Off,
+ })
+ .unwrap_or(LevelFilter::Off);
+ let logger = match logger {
+ Some(logger) => Some(ThreadPtrHolder::new(cstr!("mozIServicesLogSink"), logger)?),
+ None => None,
+ };
+ let progress = callback
+ .query_interface::<mozISyncedBookmarksMirrorProgressListener>()
+ .and_then(|p| {
+ ThreadPtrHolder::new(cstr!("mozISyncedBookmarksMirrorProgressListener"), p).ok()
+ });
+ Ok(MergeTask {
+ db: db.clone(),
+ controller,
+ max_log_level,
+ logger,
+ local_time_millis: local_time_seconds * 1000,
+ remote_time_millis: remote_time_seconds * 1000,
+ progress,
+ callback: ThreadPtrHolder::new(cstr!("mozISyncedBookmarksMirrorCallback"), callback)?,
+ result: AtomicRefCell::new(Err(error::Error::DidNotRun)),
+ })
+ }
+
+ fn merge(&self) -> error::Result<store::ApplyStatus> {
+ let mut db = self.db.clone();
+ if db.transaction_in_progress()? {
+ // If a transaction is already open, we can avoid an unnecessary
+ // merge, since we won't be able to apply the merged tree back to
+ // Places. This is common, especially if the user makes lots of
+ // changes at once. In that case, our merge task might run in the
+ // middle of a `Sqlite.sys.mjs` transaction, and fail when we try to
+ // open our own transaction in `Store::apply`. Since the local
+ // tree might be in an inconsistent state, we can't safely update
+ // Places.
+ return Err(error::Error::StorageBusy);
+ }
+ let log = Logger::new(self.max_log_level, self.logger.clone());
+ let driver = Driver::new(log, self.progress.clone());
+ let mut store = store::Store::new(
+ &mut db,
+ &driver,
+ &self.controller,
+ self.local_time_millis,
+ self.remote_time_millis,
+ );
+ store.validate()?;
+ store.prepare()?;
+ let status = store.merge_with_driver(&driver, &*self.controller)?;
+ Ok(status)
+ }
+}
+
+impl Task for MergeTask {
+ fn run(&self) {
+ *self.result.borrow_mut() = self.merge();
+ }
+
+ fn done(&self) -> Result<(), nsresult> {
+ let callback = self.callback.get().unwrap();
+ match mem::replace(&mut *self.result.borrow_mut(), Err(error::Error::DidNotRun)) {
+ Ok(status) => unsafe { callback.HandleSuccess(status.into()) },
+ Err(err) => {
+ let mut message = nsString::new();
+ write!(message, "{}", err).unwrap();
+ unsafe { callback.HandleError(err.into(), &*message) }
+ }
+ }
+ .to_result()
+ }
+}
+
+#[xpcom(implement(mozIPlacesPendingOperation), atomic)]
+pub struct MergeOp {
+ controller: Arc<AbortController>,
+}
+
+impl MergeOp {
+ pub fn new(controller: Arc<AbortController>) -> RefPtr<MergeOp> {
+ MergeOp::allocate(InitMergeOp { controller })
+ }
+
+ xpcom_method!(cancel => Cancel());
+ fn cancel(&self) -> Result<(), nsresult> {
+ self.controller.abort();
+ Ok(())
+ }
+}
diff --git a/toolkit/components/places/bookmark_sync/src/store.rs b/toolkit/components/places/bookmark_sync/src/store.rs
new file mode 100644
index 0000000000..0d22e3efc9
--- /dev/null
+++ b/toolkit/components/places/bookmark_sync/src/store.rs
@@ -0,0 +1,1314 @@
+/* This Source Code Form is subject to the terms of the Mozilla Public
+ * License, v. 2.0. If a copy of the MPL was not distributed with this
+ * file, You can obtain one at http://mozilla.org/MPL/2.0/. */
+
+use std::{collections::HashMap, convert::TryFrom, fmt};
+
+use dogear::{
+ debug, warn, AbortSignal, CompletionOps, Content, DeleteLocalItem, Guid, Item, Kind,
+ MergedRoot, Tree, UploadItem, UploadTombstone, Validity,
+};
+use nsstring::nsString;
+use storage::{Conn, Step};
+use url::Url;
+use xpcom::interfaces::{mozISyncedBookmarksMerger, nsINavBookmarksService};
+
+use crate::driver::{AbortController, Driver};
+use crate::error::{Error, Result};
+
+pub const LMANNO_FEEDURI: &'static str = "livemark/feedURI";
+
+extern "C" {
+ fn NS_NavBookmarksTotalSyncChanges() -> i64;
+}
+
+fn total_sync_changes() -> i64 {
+ unsafe { NS_NavBookmarksTotalSyncChanges() }
+}
+
+// Return all the non-root-roots as a 'sql set' (ie, suitable for use in an
+// IN statement)
+fn user_roots_as_sql_set() -> String {
+ format!(
+ "('{0}', '{1}', '{2}', '{3}', '{4}')",
+ dogear::MENU_GUID,
+ dogear::MOBILE_GUID,
+ dogear::TAGS_GUID,
+ dogear::TOOLBAR_GUID,
+ dogear::UNFILED_GUID
+ )
+}
+
+pub struct Store<'s> {
+ db: &'s mut Conn,
+ driver: &'s Driver,
+ controller: &'s AbortController,
+
+ /// The total Sync change count before merging. We store this before
+ /// accessing Places, and compare the current and stored counts after
+ /// opening our transaction. If they match, we can safely apply the
+ /// tree. Otherwise, we bail and try merging again on the next sync.
+ total_sync_changes: i64,
+
+ local_time_millis: i64,
+ remote_time_millis: i64,
+}
+
+impl<'s> Store<'s> {
+ pub fn new(
+ db: &'s mut Conn,
+ driver: &'s Driver,
+ controller: &'s AbortController,
+ local_time_millis: i64,
+ remote_time_millis: i64,
+ ) -> Store<'s> {
+ Store {
+ db,
+ driver,
+ controller,
+ total_sync_changes: total_sync_changes(),
+ local_time_millis,
+ remote_time_millis,
+ }
+ }
+
+ /// Ensures that all local roots are parented correctly.
+ ///
+ /// The Places root can't be in another folder, or we'll recurse infinitely
+ /// when we try to fetch the local tree.
+ ///
+ /// The five built-in roots should be under the Places root, or we'll build
+ /// and sync an invalid tree (bug 1453994, bug 1472127).
+ pub fn validate(&self) -> Result<()> {
+ self.controller.err_if_aborted()?;
+ let mut statement = self.db.prepare(format!(
+ "SELECT NOT EXISTS(
+ SELECT 1 FROM moz_bookmarks
+ WHERE id = (SELECT parent FROM moz_bookmarks
+ WHERE guid = '{root}')
+ ) AND NOT EXISTS(
+ SELECT 1 FROM moz_bookmarks b
+ JOIN moz_bookmarks p ON p.id = b.parent
+ WHERE b.guid IN {user_roots} AND
+ p.guid <> '{root}'
+ )",
+ root = dogear::ROOT_GUID,
+ user_roots = user_roots_as_sql_set(),
+ ))?;
+ let has_valid_roots = match statement.step()? {
+ Some(row) => row.get_by_index::<i64>(0)? == 1,
+ None => false,
+ };
+ if has_valid_roots {
+ Ok(())
+ } else {
+ Err(Error::InvalidLocalRoots)
+ }
+ }
+
+ /// Prepares the mirror database for a merge.
+ pub fn prepare(&self) -> Result<()> {
+ // Sync associates keywords with bookmarks, and doesn't sync POST data;
+ // Places associates keywords with (URL, POST data) pairs, and multiple
+ // bookmarks may have the same URL. When a keyword changes, clients
+ // should reupload all bookmarks with the affected URL (see
+ // `PlacesSyncUtils.bookmarks.addSyncChangesForBookmarksWithURL` and
+ // bug 1328737). Just in case, we flag any remote bookmarks that have
+ // different keywords for the same URL, or the same keyword for
+ // different URLs, for reupload.
+ self.controller.err_if_aborted()?;
+ self.db.exec(format!(
+ "UPDATE items SET
+ validity = {}
+ WHERE validity = {} AND (
+ urlId IN (
+ /* Same URL, different keywords. `COUNT` ignores NULLs, so
+ we need to count them separately. This handles cases where
+ a keyword was removed from one, but not all bookmarks with
+ the same URL. */
+ SELECT urlId FROM items
+ GROUP BY urlId
+ HAVING COUNT(DISTINCT keyword) +
+ COUNT(DISTINCT CASE WHEN keyword IS NULL
+ THEN 1 END) > 1
+ ) OR keyword IN (
+ /* Different URLs, same keyword. Bookmarks with keywords but
+ without URLs are already invalid, so we don't need to handle
+ NULLs here. */
+ SELECT keyword FROM items
+ WHERE keyword NOT NULL
+ GROUP BY keyword
+ HAVING COUNT(DISTINCT urlId) > 1
+ )
+ )",
+ mozISyncedBookmarksMerger::VALIDITY_REUPLOAD,
+ mozISyncedBookmarksMerger::VALIDITY_VALID,
+ ))?;
+ Ok(())
+ }
+
+ /// Creates a local tree item from a row in the `localItems` CTE.
+ fn local_row_to_item(&self, step: &Step) -> Result<(Item, Option<Content>)> {
+ let raw_guid: nsString = step.get_by_name("guid")?;
+ let guid = Guid::from_utf16(&*raw_guid)?;
+
+ let raw_url_href: Option<nsString> = step.get_by_name("url")?;
+ let (url, validity) = match raw_url_href {
+ // Local items might have (syntactically) invalid URLs, as in bug
+ // 1615931. If we try to sync these items, other clients will flag
+ // them as invalid (see `SyncedBookmarksMirror#storeRemote{Bookmark,
+ // Query}`), delete them when merging, and upload tombstones for
+ // them. We can avoid this extra round trip by flagging the local
+ // item as invalid. If there's a corresponding remote item with a
+ // valid URL, we'll replace the local item with it; if there isn't,
+ // we'll delete the local item.
+ Some(raw_url_href) => match Url::parse(&String::from_utf16(&*raw_url_href)?) {
+ Ok(url) => (Some(url), Validity::Valid),
+ Err(err) => {
+ warn!(
+ self.driver,
+ "Failed to parse URL for local item {}: {}", guid, err
+ );
+ (None, Validity::Replace)
+ }
+ },
+ None => (None, Validity::Valid),
+ };
+
+ let typ: i64 = step.get_by_name("type")?;
+ let kind = match u16::try_from(typ) {
+ Ok(nsINavBookmarksService::TYPE_BOOKMARK) => match url.as_ref() {
+ Some(u) if u.scheme() == "place" => Kind::Query,
+ _ => Kind::Bookmark,
+ },
+ Ok(nsINavBookmarksService::TYPE_FOLDER) => {
+ let is_livemark: i64 = step.get_by_name("isLivemark")?;
+ if is_livemark == 1 {
+ Kind::Livemark
+ } else {
+ Kind::Folder
+ }
+ }
+ Ok(nsINavBookmarksService::TYPE_SEPARATOR) => Kind::Separator,
+ _ => return Err(Error::UnknownItemType(typ)),
+ };
+
+ let mut item = Item::new(guid, kind);
+
+ let local_modified: i64 = step.get_by_name_or_default("localModified");
+ item.age = (self.local_time_millis - local_modified).max(0);
+
+ let sync_change_counter: i64 = step.get_by_name("syncChangeCounter")?;
+ item.needs_merge = sync_change_counter > 0;
+
+ item.validity = validity;
+
+ let content = if item.validity == Validity::Replace || item.guid == dogear::ROOT_GUID {
+ None
+ } else {
+ let sync_status: i64 = step.get_by_name("syncStatus")?;
+ match u16::try_from(sync_status) {
+ Ok(nsINavBookmarksService::SYNC_STATUS_NORMAL) => None,
+ _ => match kind {
+ Kind::Bookmark | Kind::Query => {
+ let raw_title: nsString = step.get_by_name("title")?;
+ let title = String::from_utf16(&*raw_title)?;
+ url.map(|url| Content::Bookmark {
+ title,
+ url_href: url.into_string(),
+ })
+ }
+ Kind::Folder | Kind::Livemark => {
+ let raw_title: nsString = step.get_by_name("title")?;
+ let title = String::from_utf16(&*raw_title)?;
+ Some(Content::Folder { title })
+ }
+ Kind::Separator => Some(Content::Separator),
+ },
+ }
+ };
+
+ Ok((item, content))
+ }
+
+ /// Creates a remote tree item from a row in `mirror.items`.
+ fn remote_row_to_item(&self, step: &Step) -> Result<(Item, Option<Content>)> {
+ let raw_guid: nsString = step.get_by_name("guid")?;
+ let guid = Guid::from_utf16(&*raw_guid)?;
+
+ let raw_kind: i64 = step.get_by_name("kind")?;
+ let kind = Kind::from_column(raw_kind)?;
+
+ let mut item = Item::new(guid, kind);
+
+ let remote_modified: i64 = step.get_by_name("serverModified")?;
+ item.age = (self.remote_time_millis - remote_modified).max(0);
+
+ let needs_merge: i32 = step.get_by_name("needsMerge")?;
+ item.needs_merge = needs_merge == 1;
+
+ let raw_validity: i64 = step.get_by_name("validity")?;
+ item.validity = Validity::from_column(raw_validity)?;
+
+ let content = if item.validity == Validity::Replace
+ || item.guid == dogear::ROOT_GUID
+ || !item.needs_merge
+ {
+ None
+ } else {
+ match kind {
+ Kind::Bookmark | Kind::Query => {
+ let raw_title: nsString = step.get_by_name("title")?;
+ let title = String::from_utf16(&*raw_title)?;
+ let raw_url_href: Option<nsString> = step.get_by_name("url")?;
+ match raw_url_href {
+ Some(raw_url_href) => {
+ // Unlike for local items, we don't parse URLs for
+ // remote items, since `storeRemote{Bookmark,
+ // Query}` already parses and canonicalizes them
+ // before inserting them into the mirror database.
+ let url_href = String::from_utf16(&*raw_url_href)?;
+ Some(Content::Bookmark { title, url_href })
+ }
+ None => None,
+ }
+ }
+ Kind::Folder | Kind::Livemark => {
+ let raw_title: nsString = step.get_by_name("title")?;
+ let title = String::from_utf16(&*raw_title)?;
+ Some(Content::Folder { title })
+ }
+ Kind::Separator => Some(Content::Separator),
+ }
+ };
+
+ Ok((item, content))
+ }
+}
+
+impl<'s> dogear::Store for Store<'s> {
+ type Ok = ApplyStatus;
+ type Error = Error;
+
+ /// Builds a fully rooted, consistent tree from the items and tombstones in
+ /// Places.
+ fn fetch_local_tree(&self) -> Result<Tree> {
+ let mut root_statement = self.db.prepare(format!(
+ "SELECT guid, type, syncChangeCounter, syncStatus,
+ lastModified / 1000 AS localModified,
+ NULL AS url, 0 AS isLivemark
+ FROM moz_bookmarks
+ WHERE guid = '{}'",
+ dogear::ROOT_GUID
+ ))?;
+ let mut builder = match root_statement.step()? {
+ Some(step) => {
+ let (item, _) = self.local_row_to_item(&step)?;
+ Tree::with_root(item)
+ }
+ None => return Err(Error::InvalidLocalRoots),
+ };
+
+ // Add items and contents to the builder, keeping track of their
+ // structure in a separate map. We can't call `p.by_structure(...)`
+ // after adding the item, because this query might return rows for
+ // children before their parents. This approach also lets us scan
+ // `moz_bookmarks` once, using the index on `(b.parent, b.position)`
+ // to avoid a temp B-tree for the `ORDER BY`.
+ let mut child_guids_by_parent_guid: HashMap<Guid, Vec<Guid>> = HashMap::new();
+ let mut items_statement = self.db.prepare(format!(
+ "SELECT b.guid, p.guid AS parentGuid, b.type, b.syncChangeCounter,
+ b.syncStatus, b.lastModified / 1000 AS localModified,
+ IFNULL(b.title, '') AS title,
+ (SELECT h.url FROM moz_places h WHERE h.id = b.fk) AS url,
+ EXISTS(SELECT 1 FROM moz_items_annos a
+ JOIN moz_anno_attributes n ON n.id = a.anno_attribute_id
+ WHERE a.item_id = b.id AND
+ n.name = '{}') AS isLivemark
+ FROM moz_bookmarks b
+ JOIN moz_bookmarks p ON p.id = b.parent
+ WHERE b.guid <> '{}'
+ ORDER BY b.parent, b.position",
+ LMANNO_FEEDURI,
+ dogear::ROOT_GUID,
+ ))?;
+ while let Some(step) = items_statement.step()? {
+ self.controller.err_if_aborted()?;
+ let (item, content) = self.local_row_to_item(&step)?;
+
+ let raw_parent_guid: nsString = step.get_by_name("parentGuid")?;
+ let parent_guid = Guid::from_utf16(&*raw_parent_guid)?;
+ child_guids_by_parent_guid
+ .entry(parent_guid)
+ .or_default()
+ .push(item.guid.clone());
+
+ let mut p = builder.item(item)?;
+ if let Some(content) = content {
+ p.content(content);
+ }
+ }
+
+ // At this point, we've added entries for all items to the tree, so
+ // we can add their structure info.
+ for (parent_guid, child_guids) in &child_guids_by_parent_guid {
+ for child_guid in child_guids {
+ self.controller.err_if_aborted()?;
+ builder.parent_for(child_guid).by_structure(parent_guid)?;
+ }
+ }
+
+ let mut deletions_statement = self.db.prepare("SELECT guid FROM moz_bookmarks_deleted")?;
+ while let Some(step) = deletions_statement.step()? {
+ self.controller.err_if_aborted()?;
+ let raw_guid: nsString = step.get_by_name("guid")?;
+ let guid = Guid::from_utf16(&*raw_guid)?;
+ builder.deletion(guid);
+ }
+
+ let tree = Tree::try_from(builder)?;
+ Ok(tree)
+ }
+
+ /// Builds a fully rooted, consistent tree from the items and tombstones in the
+ /// mirror.
+ fn fetch_remote_tree(&self) -> Result<Tree> {
+ let mut root_statement = self.db.prepare(format!(
+ "SELECT guid, serverModified, kind, needsMerge, validity
+ FROM items
+ WHERE guid = '{}'",
+ dogear::ROOT_GUID,
+ ))?;
+ let mut builder = match root_statement.step()? {
+ Some(step) => {
+ let (item, _) = self.remote_row_to_item(&step)?;
+ Tree::with_root(item)
+ }
+ None => return Err(Error::InvalidRemoteRoots),
+ };
+ builder.reparent_orphans_to(&dogear::UNFILED_GUID);
+
+ let mut items_statement = self.db.prepare(format!(
+ "SELECT v.guid, v.parentGuid, v.serverModified, v.kind,
+ IFNULL(v.title, '') AS title, v.needsMerge, v.validity,
+ v.isDeleted,
+ (SELECT u.url FROM urls u
+ WHERE u.id = v.urlId) AS url
+ FROM items v
+ WHERE v.guid <> '{}'
+ ORDER BY v.guid",
+ dogear::ROOT_GUID,
+ ))?;
+ while let Some(step) = items_statement.step()? {
+ self.controller.err_if_aborted()?;
+
+ let is_deleted: i64 = step.get_by_name("isDeleted")?;
+ if is_deleted == 1 {
+ let needs_merge: i32 = step.get_by_name("needsMerge")?;
+ if needs_merge == 0 {
+ // Ignore already-merged tombstones. These aren't persisted
+ // locally, so merging them is a no-op.
+ continue;
+ }
+ let raw_guid: nsString = step.get_by_name("guid")?;
+ let guid = Guid::from_utf16(&*raw_guid)?;
+ builder.deletion(guid);
+ } else {
+ let (item, content) = self.remote_row_to_item(&step)?;
+ let mut p = builder.item(item)?;
+ if let Some(content) = content {
+ p.content(content);
+ }
+ let raw_parent_guid: Option<nsString> = step.get_by_name("parentGuid")?;
+ if let Some(raw_parent_guid) = raw_parent_guid {
+ p.by_parent_guid(Guid::from_utf16(&*raw_parent_guid)?)?;
+ }
+ }
+ }
+
+ let mut structure_statement = self.db.prepare(format!(
+ "SELECT guid, parentGuid FROM structure
+ WHERE guid <> '{}'
+ ORDER BY parentGuid, position",
+ dogear::ROOT_GUID,
+ ))?;
+ while let Some(step) = structure_statement.step()? {
+ self.controller.err_if_aborted()?;
+ let raw_guid: nsString = step.get_by_name("guid")?;
+ let guid = Guid::from_utf16(&*raw_guid)?;
+
+ let raw_parent_guid: nsString = step.get_by_name("parentGuid")?;
+ let parent_guid = Guid::from_utf16(&*raw_parent_guid)?;
+
+ builder.parent_for(&guid).by_children(&parent_guid)?;
+ }
+
+ let tree = Tree::try_from(builder)?;
+ Ok(tree)
+ }
+
+ fn apply<'t>(&mut self, root: MergedRoot<'t>) -> Result<ApplyStatus> {
+ let ops = root.completion_ops_with_signal(self.controller)?;
+
+ if ops.is_empty() {
+ // If we don't have any items to apply, upload, or delete,
+ // no need to open a transaction at all.
+ return Ok(ApplyStatus::Skipped);
+ }
+
+ // Apply the merged tree and stage outgoing items. This transaction
+ // blocks writes from the main connection until it's committed, so we
+ // try to do as little work as possible within it.
+ if self.db.transaction_in_progress()? {
+ return Err(Error::StorageBusy);
+ }
+ let tx = self.db.transaction()?;
+ if self.total_sync_changes != total_sync_changes() {
+ return Err(Error::MergeConflict);
+ }
+
+ debug!(self.driver, "Updating local items in Places");
+ update_local_items_in_places(
+ &tx,
+ &self.driver,
+ &self.controller,
+ self.local_time_millis,
+ &ops,
+ )?;
+
+ debug!(self.driver, "Staging items to upload");
+ stage_items_to_upload(
+ &tx,
+ &self.driver,
+ &self.controller,
+ &ops.upload_items,
+ &ops.upload_tombstones,
+ )?;
+
+ cleanup(&tx)?;
+ tx.commit()?;
+
+ Ok(ApplyStatus::Merged)
+ }
+}
+
+/// Builds a temporary table with the merge states of all nodes in the merged
+/// tree and updates Places to match the merged tree.
+///
+/// Conceptually, we examine the merge state of each item, and either leave the
+/// item unchanged, upload the local side, apply the remote side, or apply and
+/// then reupload the remote side with a new structure.
+///
+/// Note that we update Places and flag items *before* upload, while iOS
+/// updates the mirror *after* a successful upload. This simplifies our
+/// implementation, though we lose idempotent merges. If upload is interrupted,
+/// the next sync won't distinguish between new merge states from the previous
+/// sync, and local changes.
+fn update_local_items_in_places<'t>(
+ db: &Conn,
+ driver: &Driver,
+ controller: &AbortController,
+ local_time_millis: i64,
+ ops: &CompletionOps<'t>,
+) -> Result<()> {
+ debug!(
+ driver,
+ "Cleaning up observer notifications left from last sync"
+ );
+ controller.err_if_aborted()?;
+ db.exec(
+ "DELETE FROM itemsAdded;
+ DELETE FROM guidsChanged;
+ DELETE FROM itemsChanged;
+ DELETE FROM itemsRemoved;
+ DELETE FROM itemsMoved;",
+ )?;
+
+ // Places uses microsecond timestamps for dates added and last modified
+ // times, rounded to the nearest millisecond. Using `now` for the local
+ // time lets us set modified times deterministically for tests.
+ let now = local_time_millis * 1000;
+
+ // Insert URLs for new remote items into the `moz_places` table. We need to
+ // do this before inserting new remote items, since we need Place IDs for
+ // both old and new URLs.
+ debug!(driver, "Inserting Places for new items");
+ for chunk in ops.apply_remote_items.chunks(db.variable_limit()?) {
+ let mut statement = db.prepare(format!(
+ "INSERT OR IGNORE INTO moz_places(url, url_hash, rev_host, hidden,
+ frecency, guid)
+ SELECT u.url, u.hash, u.revHost, 0,
+ (CASE v.kind WHEN {} THEN 0 ELSE -1 END),
+ IFNULL((SELECT h.guid FROM moz_places h
+ WHERE h.url_hash = u.hash AND
+ h.url = u.url), u.guid)
+ FROM items v
+ JOIN urls u ON u.id = v.urlId
+ WHERE v.guid IN ({})",
+ mozISyncedBookmarksMerger::KIND_QUERY,
+ repeat_sql_vars(chunk.len()),
+ ))?;
+ for (index, op) in chunk.iter().enumerate() {
+ controller.err_if_aborted()?;
+ let remote_guid = nsString::from(&*op.remote_node().guid);
+ statement.bind_by_index(index as u32, remote_guid)?;
+ }
+ statement.execute()?;
+ }
+
+ // Trigger frecency updates for all new origins.
+ debug!(driver, "Updating origins for new URLs");
+ controller.err_if_aborted()?;
+ db.exec("DELETE FROM moz_updateoriginsinsert_temp")?;
+
+ // Build a table of new and updated items.
+ debug!(driver, "Staging apply remote item ops");
+ for chunk in ops.apply_remote_items.chunks(db.variable_limit()? / 3) {
+ // CTEs in `WITH` clauses aren't indexed, so this query needs a full
+ // table scan on `ops`. But that's okay; a separate temp table for ops
+ // would also need a full scan. Note that we need both the local _and_
+ // remote GUIDs here, because we haven't changed the local GUIDs yet.
+ let mut statement = db.prepare(format!(
+ "WITH
+ ops(mergedGuid, localGuid, remoteGuid, level) AS (VALUES {})
+ INSERT INTO itemsToApply(mergedGuid, localId, remoteId,
+ remoteGuid, newLevel,
+ newType,
+ localDateAddedMicroseconds,
+ remoteDateAddedMicroseconds,
+ lastModifiedMicroseconds,
+ oldTitle, newTitle, oldPlaceId,
+ newPlaceId,
+ newKeyword)
+ SELECT n.mergedGuid, b.id, v.id,
+ v.guid, n.level,
+ (CASE WHEN v.kind IN ({}, {}) THEN {}
+ WHEN v.kind IN ({}, {}) THEN {}
+ ELSE {}
+ END),
+ b.dateAdded,
+ v.dateAdded * 1000,
+ MAX(v.dateAdded * 1000, {}),
+ b.title, v.title, b.fk,
+ (SELECT h.id FROM moz_places h
+ JOIN urls u ON u.hash = h.url_hash
+ WHERE u.id = v.urlId AND
+ u.url = h.url),
+ v.keyword
+ FROM ops n
+ JOIN items v ON v.guid = n.remoteGuid
+ LEFT JOIN moz_bookmarks b ON b.guid = n.localGuid",
+ repeat_display(chunk.len(), ",", |index, f| {
+ let op = &chunk[index];
+ write!(f, "(?, ?, ?, {})", op.level)
+ }),
+ mozISyncedBookmarksMerger::KIND_BOOKMARK,
+ mozISyncedBookmarksMerger::KIND_QUERY,
+ nsINavBookmarksService::TYPE_BOOKMARK,
+ mozISyncedBookmarksMerger::KIND_FOLDER,
+ mozISyncedBookmarksMerger::KIND_LIVEMARK,
+ nsINavBookmarksService::TYPE_FOLDER,
+ nsINavBookmarksService::TYPE_SEPARATOR,
+ now,
+ ))?;
+ for (index, op) in chunk.iter().enumerate() {
+ controller.err_if_aborted()?;
+
+ let offset = (index * 3) as u32;
+
+ // In most cases, the merged and remote GUIDs are the same for new
+ // items. For updates, all three are typically the same. We could
+ // try to avoid binding duplicates, but that complicates chunking,
+ // and we don't expect many items to change after the first sync.
+ let merged_guid = nsString::from(&*op.merged_node.guid);
+ statement.bind_by_index(offset, merged_guid)?;
+
+ let local_guid = op
+ .merged_node
+ .merge_state
+ .local_node()
+ .map(|node| nsString::from(&*node.guid));
+ statement.bind_by_index(offset + 1, local_guid)?;
+
+ let remote_guid = nsString::from(&*op.remote_node().guid);
+ statement.bind_by_index(offset + 2, remote_guid)?;
+ }
+ statement.execute()?;
+ }
+
+ debug!(driver, "Staging change GUID ops");
+ for chunk in ops.change_guids.chunks(db.variable_limit()? / 2) {
+ let mut statement = db.prepare(format!(
+ "INSERT INTO changeGuidOps(localGuid, mergedGuid, syncStatus, level,
+ lastModifiedMicroseconds)
+ VALUES {}",
+ repeat_display(chunk.len(), ",", |index, f| {
+ let op = &chunk[index];
+ // If only the local GUID changed, the item was deduped, so we
+ // can mark it as syncing. Otherwise, we changed an invalid
+ // GUID locally or remotely, so we leave its original sync
+ // status in place until we've uploaded it.
+ let sync_status = if op.merged_node.remote_guid_changed() {
+ None
+ } else {
+ Some(nsINavBookmarksService::SYNC_STATUS_NORMAL)
+ };
+ write!(
+ f,
+ "(?, ?, {}, {}, {})",
+ NullableFragment(sync_status),
+ op.level,
+ now
+ )
+ })
+ ))?;
+ for (index, op) in chunk.iter().enumerate() {
+ controller.err_if_aborted()?;
+
+ let offset = (index * 2) as u32;
+
+ let local_guid = nsString::from(&*op.local_node().guid);
+ statement.bind_by_index(offset, local_guid)?;
+
+ let merged_guid = nsString::from(&*op.merged_node.guid);
+ statement.bind_by_index(offset + 1, merged_guid)?;
+ }
+ statement.execute()?;
+ }
+
+ debug!(driver, "Staging apply new local structure ops");
+ for chunk in ops
+ .apply_new_local_structure
+ .chunks(db.variable_limit()? / 2)
+ {
+ let mut statement = db.prepare(format!(
+ "INSERT INTO applyNewLocalStructureOps(mergedGuid, mergedParentGuid,
+ position, level,
+ lastModifiedMicroseconds)
+ VALUES {}",
+ repeat_display(chunk.len(), ",", |index, f| {
+ let op = &chunk[index];
+ write!(f, "(?, ?, {}, {}, {})", op.position, op.level, now)
+ })
+ ))?;
+ for (index, op) in chunk.iter().enumerate() {
+ controller.err_if_aborted()?;
+
+ let offset = (index * 2) as u32;
+
+ let merged_guid = nsString::from(&*op.merged_node.guid);
+ statement.bind_by_index(offset, merged_guid)?;
+
+ let merged_parent_guid = nsString::from(&*op.merged_parent_node.guid);
+ statement.bind_by_index(offset + 1, merged_parent_guid)?;
+ }
+ statement.execute()?;
+ }
+
+ debug!(driver, "Removing tombstones for revived items");
+ for chunk in ops.delete_local_tombstones.chunks(db.variable_limit()?) {
+ let mut statement = db.prepare(format!(
+ "DELETE FROM moz_bookmarks_deleted
+ WHERE guid IN ({})",
+ repeat_sql_vars(chunk.len()),
+ ))?;
+ for (index, op) in chunk.iter().enumerate() {
+ controller.err_if_aborted()?;
+ statement.bind_by_index(index as u32, nsString::from(&*op.guid().as_str()))?;
+ }
+ statement.execute()?;
+ }
+
+ debug!(
+ driver,
+ "Inserting new tombstones for non-syncable and invalid items"
+ );
+ for chunk in ops.insert_local_tombstones.chunks(db.variable_limit()?) {
+ let mut statement = db.prepare(format!(
+ "INSERT INTO moz_bookmarks_deleted(guid, dateRemoved)
+ VALUES {}",
+ repeat_display(chunk.len(), ",", |_, f| write!(f, "(?, {})", now)),
+ ))?;
+ for (index, op) in chunk.iter().enumerate() {
+ controller.err_if_aborted()?;
+ statement.bind_by_index(
+ index as u32,
+ nsString::from(&*op.remote_node().guid.as_str()),
+ )?;
+ }
+ statement.execute()?;
+ }
+
+ debug!(driver, "Removing local items");
+ for chunk in ops.delete_local_items.chunks(db.variable_limit()?) {
+ remove_local_items(&db, driver, controller, chunk)?;
+ }
+
+ // Fires the `changeGuids` trigger.
+ debug!(driver, "Changing GUIDs");
+ controller.err_if_aborted()?;
+ db.exec("DELETE FROM changeGuidOps")?;
+
+ debug!(driver, "Applying remote items");
+ apply_remote_items(db, driver, controller)?;
+
+ // Trigger frecency updates for all affected origins.
+ debug!(driver, "Updating origins for changed URLs");
+ controller.err_if_aborted()?;
+ db.exec("DELETE FROM moz_updateoriginsupdate_temp")?;
+
+ // Fires the `applyNewLocalStructure` trigger.
+ debug!(driver, "Applying new local structure");
+ controller.err_if_aborted()?;
+ db.exec("DELETE FROM applyNewLocalStructureOps")?;
+
+ debug!(
+ driver,
+ "Resetting change counters for items that shouldn't be uploaded"
+ );
+ for chunk in ops.set_local_merged.chunks(db.variable_limit()?) {
+ let mut statement = db.prepare(format!(
+ "UPDATE moz_bookmarks SET
+ syncChangeCounter = 0
+ WHERE guid IN ({})",
+ repeat_sql_vars(chunk.len()),
+ ))?;
+ for (index, op) in chunk.iter().enumerate() {
+ controller.err_if_aborted()?;
+ statement.bind_by_index(index as u32, nsString::from(&*op.merged_node.guid))?;
+ }
+ statement.execute()?;
+ }
+
+ debug!(
+ driver,
+ "Bumping change counters for items that should be uploaded"
+ );
+ for chunk in ops.set_local_unmerged.chunks(db.variable_limit()?) {
+ let mut statement = db.prepare(format!(
+ "UPDATE moz_bookmarks SET
+ syncChangeCounter = 1
+ WHERE guid IN ({})",
+ repeat_sql_vars(chunk.len()),
+ ))?;
+ for (index, op) in chunk.iter().enumerate() {
+ controller.err_if_aborted()?;
+ statement.bind_by_index(index as u32, nsString::from(&*op.merged_node.guid))?;
+ }
+ statement.execute()?;
+ }
+
+ debug!(driver, "Flagging applied remote items as merged");
+ for chunk in ops.set_remote_merged.chunks(db.variable_limit()?) {
+ let mut statement = db.prepare(format!(
+ "UPDATE items SET
+ needsMerge = 0
+ WHERE guid IN ({})",
+ repeat_sql_vars(chunk.len()),
+ ))?;
+ for (index, op) in chunk.iter().enumerate() {
+ controller.err_if_aborted()?;
+ statement.bind_by_index(index as u32, nsString::from(op.guid().as_str()))?;
+ }
+ statement.execute()?;
+ }
+
+ Ok(())
+}
+
+/// Upserts all new and updated items from the `itemsToApply` table into Places.
+fn apply_remote_items(db: &Conn, driver: &Driver, controller: &AbortController) -> Result<()> {
+ debug!(driver, "Recording item added notifications for new items");
+ controller.err_if_aborted()?;
+ db.exec(
+ "INSERT INTO itemsAdded(guid, keywordChanged, level)
+ SELECT n.mergedGuid, n.newKeyword NOT NULL OR
+ EXISTS(SELECT 1 FROM moz_keywords k
+ WHERE k.place_id = n.newPlaceId OR
+ k.keyword = n.newKeyword),
+ n.newLevel
+ FROM itemsToApply n
+ WHERE n.localId IS NULL",
+ )?;
+
+ debug!(
+ driver,
+ "Recording item changed notifications for existing items"
+ );
+ controller.err_if_aborted()?;
+ db.exec(
+ "INSERT INTO itemsChanged(itemId, oldTitle, oldPlaceId, keywordChanged,
+ level)
+ SELECT n.localId, n.oldTitle, n.oldPlaceId,
+ n.newKeyword NOT NULL OR EXISTS(
+ SELECT 1 FROM moz_keywords k
+ WHERE k.place_id IN (n.oldPlaceId, n.newPlaceId) OR
+ k.keyword = n.newKeyword
+ ),
+ n.newLevel
+ FROM itemsToApply n
+ WHERE n.localId NOT NULL",
+ )?;
+
+ // Remove all keywords from old and new URLs, and remove new keywords from
+ // all existing URLs. The `NOT NULL` conditions are important; they ensure
+ // that SQLite uses our partial indexes, instead of a table scan.
+ debug!(driver, "Removing old keywords");
+ controller.err_if_aborted()?;
+ db.exec(
+ "DELETE FROM moz_keywords
+ WHERE place_id IN (SELECT oldPlaceId FROM itemsToApply
+ WHERE oldPlaceId NOT NULL) OR
+ place_id IN (SELECT newPlaceId FROM itemsToApply
+ WHERE newPlaceId NOT NULL) OR
+ keyword IN (SELECT newKeyword FROM itemsToApply
+ WHERE newKeyword NOT NULL)
+ ",
+ )?;
+
+ debug!(driver, "Removing old tags");
+ controller.err_if_aborted()?;
+ db.exec(
+ "DELETE FROM localTags
+ WHERE placeId IN (SELECT oldPlaceId FROM itemsToApply
+ WHERE oldPlaceId NOT NULL) OR
+ placeId IN (SELECT newPlaceId FROM itemsToApply
+ WHERE newPlaceId NOT NULL)",
+ )?;
+
+ // Insert and update items, using -1 for new items' parent IDs and
+ // positions. We'll update these later, when we apply the new local
+ // structure. This is a full table scan on `itemsToApply`. The no-op
+ // `WHERE` clause is necessary to avoid a parsing ambiguity.
+ debug!(driver, "Upserting new items");
+ controller.err_if_aborted()?;
+ db.exec(format!(
+ "INSERT INTO moz_bookmarks(id, guid, parent, position, type, fk, title,
+ dateAdded,
+ lastModified,
+ syncStatus, syncChangeCounter)
+ SELECT localId, mergedGuid, -1, -1, newType, newPlaceId, newTitle,
+ /* Pick the older of the local and remote date added. We'll
+ weakly reupload any items with an older local date. */
+ MIN(IFNULL(localDateAddedMicroseconds,
+ remoteDateAddedMicroseconds),
+ remoteDateAddedMicroseconds),
+ /* The last modified date should always be newer than the date
+ added, so we pick the newer of the two here. */
+ MAX(lastModifiedMicroseconds, remoteDateAddedMicroseconds),
+ {syncStatusNormal}, 0
+ FROM itemsToApply
+ WHERE 1
+ ON CONFLICT(id) DO UPDATE SET
+ title = excluded.title,
+ dateAdded = excluded.dateAdded,
+ lastModified = excluded.lastModified,
+ syncStatus = {syncStatusNormal},
+ /* It's important that we update the URL *after* removing old keywords
+ and *before* inserting new ones, so that the above DELETEs select
+ the correct affected items. */
+ fk = excluded.fk",
+ syncStatusNormal = nsINavBookmarksService::SYNC_STATUS_NORMAL
+ ))?;
+ // The roots are never in `itemsToApply` but still need to (well, at least
+ // *should*) have a syncStatus of NORMAL after a reconcilliation. The
+ // ROOT_GUID doesn't matter in practice, but we include it to be consistent.
+ db.exec(format!(
+ "UPDATE moz_bookmarks SET
+ syncStatus={syncStatusNormal}
+ WHERE guid IN {user_roots} OR
+ guid = '{root}'
+ ",
+ syncStatusNormal = nsINavBookmarksService::SYNC_STATUS_NORMAL,
+ root = dogear::ROOT_GUID,
+ user_roots = user_roots_as_sql_set(),
+ ))?;
+
+ // Flag frecencies for recalculation. This is a multi-index OR that uses the
+ // `oldPlacesIds` and `newPlaceIds` partial indexes, since `<>` is only true
+ // if both terms are not NULL. Without those constraints, the subqueries
+ // would scan `itemsToApply` twice. The `oldPlaceId <> newPlaceId` and
+ // `newPlaceId <> oldPlaceId` checks exclude items where the URL didn't
+ // change; we don't need to recalculate their frecencies.
+ debug!(driver, "Flagging frecencies for recalculation");
+ controller.err_if_aborted()?;
+ db.exec(
+ "UPDATE moz_places SET
+ recalc_frecency = 1
+ WHERE frecency <> 0 AND (
+ id IN (
+ SELECT oldPlaceId FROM itemsToApply
+ WHERE oldPlaceId <> newPlaceId
+ ) OR id IN (
+ SELECT newPlaceId FROM itemsToApply
+ WHERE newPlaceId <> oldPlaceId
+ )
+ )",
+ )?;
+
+ debug!(driver, "Inserting new keywords for new URLs");
+ controller.err_if_aborted()?;
+ db.exec(
+ "INSERT OR IGNORE INTO moz_keywords(keyword, place_id, post_data)
+ SELECT newKeyword, newPlaceId, ''
+ FROM itemsToApply
+ WHERE newKeyword NOT NULL",
+ )?;
+
+ debug!(driver, "Inserting new tags for new URLs");
+ controller.err_if_aborted()?;
+ db.exec(
+ "INSERT INTO localTags(tag, placeId, lastModifiedMicroseconds)
+ SELECT t.tag, n.newPlaceId, n.lastModifiedMicroseconds
+ FROM itemsToApply n
+ JOIN tags t ON t.itemId = n.remoteId",
+ )?;
+
+ Ok(())
+}
+
+/// Removes deleted local items from Places.
+fn remove_local_items(
+ db: &Conn,
+ driver: &Driver,
+ controller: &AbortController,
+ ops: &[DeleteLocalItem],
+) -> Result<()> {
+ debug!(driver, "Recording observer notifications for removed items");
+ let mut observer_statement = db.prepare(format!(
+ "WITH
+ ops(guid, level) AS (VALUES {})
+ INSERT INTO itemsRemoved(itemId, parentId, position, type, title,
+ placeId, guid, parentGuid, level)
+ SELECT b.id, b.parent, b.position, b.type, IFNULL(b.title, \"\"), b.fk,
+ b.guid, p.guid, n.level
+ FROM ops n
+ JOIN moz_bookmarks b ON b.guid = n.guid
+ JOIN moz_bookmarks p ON p.id = b.parent",
+ repeat_display(ops.len(), ",", |index, f| {
+ let op = &ops[index];
+ write!(f, "(?, {})", op.local_node().level())
+ }),
+ ))?;
+ for (index, op) in ops.iter().enumerate() {
+ controller.err_if_aborted()?;
+ observer_statement.bind_by_index(
+ index as u32,
+ nsString::from(&*op.local_node().guid.as_str()),
+ )?;
+ }
+ observer_statement.execute()?;
+
+ debug!(driver, "Recalculating frecencies for removed bookmark URLs");
+ let mut frecency_statement = db.prepare(format!(
+ "UPDATE moz_places SET
+ recalc_frecency = 1
+ WHERE id IN (SELECT b.fk FROM moz_bookmarks b
+ WHERE b.guid IN ({})) AND
+ frecency <> 0",
+ repeat_sql_vars(ops.len())
+ ))?;
+ for (index, op) in ops.iter().enumerate() {
+ controller.err_if_aborted()?;
+ frecency_statement.bind_by_index(
+ index as u32,
+ nsString::from(&*op.local_node().guid.as_str()),
+ )?;
+ }
+ frecency_statement.execute()?;
+
+ // This can be removed in bug 1460577.
+ debug!(driver, "Removing annos for deleted items");
+ let mut annos_statement = db.prepare(format!(
+ "DELETE FROM moz_items_annos
+ WHERE item_id = (SELECT b.id FROM moz_bookmarks b
+ WHERE b.guid IN ({}))",
+ repeat_sql_vars(ops.len()),
+ ))?;
+ for (index, op) in ops.iter().enumerate() {
+ controller.err_if_aborted()?;
+ annos_statement.bind_by_index(
+ index as u32,
+ nsString::from(&*op.local_node().guid.as_str()),
+ )?;
+ }
+ annos_statement.execute()?;
+
+ debug!(driver, "Removing deleted items from Places");
+ let mut delete_statement = db.prepare(format!(
+ "DELETE FROM moz_bookmarks
+ WHERE guid IN ({})",
+ repeat_sql_vars(ops.len()),
+ ))?;
+ for (index, op) in ops.iter().enumerate() {
+ controller.err_if_aborted()?;
+ delete_statement.bind_by_index(
+ index as u32,
+ nsString::from(&*op.local_node().guid.as_str()),
+ )?;
+ }
+ delete_statement.execute()?;
+
+ Ok(())
+}
+
+/// Stores a snapshot of all locally changed items in a temporary table for
+/// upload. This is called from within the merge transaction, to ensure that
+/// changes made during the sync don't cause us to upload inconsistent records.
+///
+/// For an example of why we use a temporary table instead of reading directly
+/// from Places, consider a user adding a bookmark, then changing its parent
+/// folder. We first add the bookmark to the default folder, bump the change
+/// counter of the new bookmark and the default folder, then trigger a sync.
+/// Depending on how quickly the user picks the new parent, we might upload
+/// a record for the default folder, commit the move, then upload the bookmark.
+/// We'll still upload the new parent on the next sync, but, in the meantime,
+/// we've introduced a parent-child disagreement. This can also happen if the
+/// user moves many items between folders.
+///
+/// Conceptually, `itemsToUpload` is a transient "view" of locally changed
+/// items. The change counter in Places is the persistent record of items that
+/// we need to upload, so, if upload is interrupted or fails, we'll stage the
+/// items again on the next sync.
+fn stage_items_to_upload(
+ db: &Conn,
+ driver: &Driver,
+ controller: &AbortController,
+ upload_items: &[UploadItem],
+ upload_tombstones: &[UploadTombstone],
+) -> Result<()> {
+ debug!(driver, "Cleaning up staged items left from last sync");
+ controller.err_if_aborted()?;
+ db.exec("DELETE FROM itemsToUpload")?;
+
+ // Stage remotely changed items with older local creation dates. These are
+ // tracked "weakly": if the upload is interrupted or fails, we won't
+ // reupload the record on the next sync.
+ debug!(driver, "Staging items with older local dates added");
+ controller.err_if_aborted()?;
+ db.exec(format!(
+ "INSERT OR IGNORE INTO itemsToUpload(id, guid, syncChangeCounter,
+ parentGuid, parentTitle, dateAdded,
+ type, title, placeId, isQuery, url,
+ keyword, position, tagFolderName)
+ {}
+ JOIN itemsToApply n ON n.mergedGuid = b.guid
+ WHERE n.localDateAddedMicroseconds < n.remoteDateAddedMicroseconds",
+ UploadItemsFragment("b"),
+ ))?;
+
+ debug!(driver, "Staging remaining locally changed items for upload");
+ for chunk in upload_items.chunks(db.variable_limit()?) {
+ let mut statement = db.prepare(format!(
+ "INSERT OR IGNORE INTO itemsToUpload(id, guid, syncChangeCounter,
+ parentGuid, parentTitle,
+ dateAdded, type, title,
+ placeId, isQuery, url, keyword,
+ position, tagFolderName)
+ {}
+ WHERE b.guid IN ({})",
+ UploadItemsFragment("b"),
+ repeat_sql_vars(chunk.len()),
+ ))?;
+ for (index, op) in chunk.iter().enumerate() {
+ controller.err_if_aborted()?;
+ statement.bind_by_index(index as u32, nsString::from(&*op.merged_node.guid))?;
+ }
+ statement.execute()?;
+ }
+
+ // Record the child GUIDs of locally changed folders, which we use to
+ // populate the `children` array in the record.
+ debug!(driver, "Staging structure to upload");
+ controller.err_if_aborted()?;
+ db.exec(
+ "
+ INSERT INTO structureToUpload(guid, parentId, position)
+ SELECT b.guid, b.parent, b.position
+ FROM moz_bookmarks b
+ JOIN itemsToUpload o ON o.id = b.parent",
+ )?;
+
+ // Stage tags for outgoing bookmarks.
+ debug!(driver, "Staging tags to upload");
+ controller.err_if_aborted()?;
+ db.exec(
+ "
+ INSERT OR IGNORE INTO tagsToUpload(id, tag)
+ SELECT o.id, t.tag
+ FROM localTags t
+ JOIN itemsToUpload o ON o.placeId = t.placeId",
+ )?;
+
+ // Finally, stage tombstones for deleted items. Ignore conflicts if we have
+ // tombstones for undeleted items; Places Maintenance should clean these up.
+ debug!(driver, "Staging tombstones to upload");
+ for chunk in upload_tombstones.chunks(db.variable_limit()?) {
+ let mut statement = db.prepare(format!(
+ "INSERT OR IGNORE INTO itemsToUpload(guid, syncChangeCounter, isDeleted)
+ VALUES {}",
+ repeat_display(chunk.len(), ",", |_, f| write!(f, "(?, 1, 1)"))
+ ))?;
+ for (index, op) in chunk.iter().enumerate() {
+ controller.err_if_aborted()?;
+ statement.bind_by_index(index as u32, nsString::from(op.guid().as_str()))?;
+ }
+ statement.execute()?;
+ }
+
+ Ok(())
+}
+
+fn cleanup(db: &Conn) -> Result<()> {
+ db.exec("DELETE FROM itemsToApply")?;
+ Ok(())
+}
+
+/// Formats a list of binding parameters for inclusion in a SQL list.
+#[inline]
+fn repeat_sql_vars(count: usize) -> impl fmt::Display {
+ repeat_display(count, ",", |_, f| write!(f, "?"))
+}
+
+/// Construct a `RepeatDisplay` that will repeatedly call `fmt_one` with a
+/// formatter `count` times, separated by `sep`. This is copied from the
+/// `sql_support` crate in `application-services`.
+#[inline]
+fn repeat_display<'a, F>(count: usize, sep: &'a str, fmt_one: F) -> RepeatDisplay<'a, F>
+where
+ F: Fn(usize, &mut fmt::Formatter) -> fmt::Result,
+{
+ RepeatDisplay {
+ count,
+ sep,
+ fmt_one,
+ }
+}
+
+/// Helper type for printing repeated strings more efficiently.
+#[derive(Debug, Clone)]
+struct RepeatDisplay<'a, F> {
+ count: usize,
+ sep: &'a str,
+ fmt_one: F,
+}
+
+impl<'a, F> fmt::Display for RepeatDisplay<'a, F>
+where
+ F: Fn(usize, &mut fmt::Formatter) -> fmt::Result,
+{
+ fn fmt(&self, f: &mut fmt::Formatter) -> fmt::Result {
+ for i in 0..self.count {
+ if i != 0 {
+ f.write_str(self.sep)?;
+ }
+ (self.fmt_one)(i, f)?;
+ }
+ Ok(())
+ }
+}
+
+/// Converts between a type `T` and its SQL representation.
+trait Column<T> {
+ fn from_column(raw: T) -> Result<Self>
+ where
+ Self: Sized;
+ fn into_column(self) -> T;
+}
+
+impl Column<i64> for Kind {
+ fn from_column(raw: i64) -> Result<Kind> {
+ Ok(match i16::try_from(raw) {
+ Ok(mozISyncedBookmarksMerger::KIND_BOOKMARK) => Kind::Bookmark,
+ Ok(mozISyncedBookmarksMerger::KIND_QUERY) => Kind::Query,
+ Ok(mozISyncedBookmarksMerger::KIND_FOLDER) => Kind::Folder,
+ Ok(mozISyncedBookmarksMerger::KIND_LIVEMARK) => Kind::Livemark,
+ Ok(mozISyncedBookmarksMerger::KIND_SEPARATOR) => Kind::Separator,
+ _ => return Err(Error::UnknownItemKind(raw)),
+ })
+ }
+
+ fn into_column(self) -> i64 {
+ match self {
+ Kind::Bookmark => mozISyncedBookmarksMerger::KIND_BOOKMARK as i64,
+ Kind::Query => mozISyncedBookmarksMerger::KIND_QUERY as i64,
+ Kind::Folder => mozISyncedBookmarksMerger::KIND_FOLDER as i64,
+ Kind::Livemark => mozISyncedBookmarksMerger::KIND_LIVEMARK as i64,
+ Kind::Separator => mozISyncedBookmarksMerger::KIND_SEPARATOR as i64,
+ }
+ }
+}
+
+impl Column<i64> for Validity {
+ fn from_column(raw: i64) -> Result<Validity> {
+ Ok(match i16::try_from(raw) {
+ Ok(mozISyncedBookmarksMerger::VALIDITY_VALID) => Validity::Valid,
+ Ok(mozISyncedBookmarksMerger::VALIDITY_REUPLOAD) => Validity::Reupload,
+ Ok(mozISyncedBookmarksMerger::VALIDITY_REPLACE) => Validity::Replace,
+ _ => return Err(Error::UnknownItemValidity(raw).into()),
+ })
+ }
+
+ fn into_column(self) -> i64 {
+ match self {
+ Validity::Valid => mozISyncedBookmarksMerger::VALIDITY_VALID as i64,
+ Validity::Reupload => mozISyncedBookmarksMerger::VALIDITY_REUPLOAD as i64,
+ Validity::Replace => mozISyncedBookmarksMerger::VALIDITY_REPLACE as i64,
+ }
+ }
+}
+
+/// Formats an optional value so that it can be included in a SQL statement.
+struct NullableFragment<T>(Option<T>);
+
+impl<T> fmt::Display for NullableFragment<T>
+where
+ T: fmt::Display,
+{
+ fn fmt(&self, f: &mut fmt::Formatter) -> fmt::Result {
+ match &self.0 {
+ Some(v) => v.fmt(f),
+ None => write!(f, "NULL"),
+ }
+ }
+}
+
+/// Formats a `SELECT` statement for staging local items in the `itemsToUpload`
+/// table.
+struct UploadItemsFragment(&'static str);
+
+impl fmt::Display for UploadItemsFragment {
+ fn fmt(&self, f: &mut fmt::Formatter<'_>) -> fmt::Result {
+ write!(
+ f,
+ "SELECT {0}.id, {0}.guid, {}.syncChangeCounter,
+ p.guid AS parentGuid, p.title AS parentTitle,
+ {0}.dateAdded / 1000 AS dateAdded, {0}.type, {0}.title,
+ h.id AS placeId,
+ IFNULL(substr(h.url, 1, 6) = 'place:', 0) AS isQuery,
+ h.url,
+ (SELECT keyword FROM moz_keywords WHERE place_id = h.id),
+ {0}.position,
+ (SELECT get_query_param(substr(url, 7), 'tag')
+ WHERE substr(h.url, 1, 6) = 'place:') AS tagFolderName
+ FROM moz_bookmarks {0}
+ JOIN moz_bookmarks p ON p.id = {0}.parent
+ LEFT JOIN moz_places h ON h.id = {0}.fk",
+ self.0
+ )
+ }
+}
+
+pub enum ApplyStatus {
+ Merged,
+ Skipped,
+}
+
+impl From<ApplyStatus> for bool {
+ fn from(status: ApplyStatus) -> bool {
+ match status {
+ ApplyStatus::Merged => true,
+ ApplyStatus::Skipped => false,
+ }
+ }
+}