mirror of
https://github.com/chatmail/core.git
synced 2026-04-17 21:46:35 +03:00
Start of get support and create new module.
This commit is contained in:
140
src/imex.rs
140
src/imex.rs
@@ -8,8 +8,7 @@ use std::iter::FusedIterator;
|
||||
use std::path::{Path, PathBuf};
|
||||
|
||||
use ::pgp::types::KeyTrait;
|
||||
use anyhow::{anyhow, bail, ensure, format_err, Context as _, Result};
|
||||
use async_channel::Receiver;
|
||||
use anyhow::{bail, ensure, format_err, Context as _, Result};
|
||||
use futures::StreamExt;
|
||||
use futures_lite::FutureExt;
|
||||
use rand::{thread_rng, Rng};
|
||||
@@ -30,7 +29,6 @@ use crate::message::{Message, MsgId, Viewtype};
|
||||
use crate::mimeparser::SystemMessage;
|
||||
use crate::param::Param;
|
||||
use crate::pgp;
|
||||
use crate::qr::Qr;
|
||||
use crate::sql;
|
||||
use crate::stock_str;
|
||||
use crate::tools::{
|
||||
@@ -38,6 +36,10 @@ use crate::tools::{
|
||||
EmailAddress,
|
||||
};
|
||||
|
||||
mod transfer;
|
||||
|
||||
pub use transfer::{get_backup, BackupProvider};
|
||||
|
||||
// Name of the database file in the backup.
|
||||
const DBFILE_BACKUP_NAME: &str = "dc_database_backup.sqlite";
|
||||
const BLOBS_BACKUP_NAME: &str = "blobs_backup";
|
||||
@@ -755,138 +757,6 @@ async fn export_database(context: &Context, dest: &Path, passphrase: String) ->
|
||||
Ok(())
|
||||
}
|
||||
|
||||
#[derive(Debug)]
|
||||
pub struct BackupSender {
|
||||
/// A handle to the running provider.
|
||||
provider: sendme::provider::Provider,
|
||||
/// The ticket to retrieve the backup collection.
|
||||
ticket: sendme::provider::Ticket,
|
||||
/// Token holding the "ongoing" mutex. When this completes the provider should shut
|
||||
/// down.
|
||||
cancel_token: Receiver<()>,
|
||||
}
|
||||
|
||||
impl BackupSender {
|
||||
// - [x] check i/o is not running
|
||||
// - [x] check we have secret key
|
||||
// - [x] alloc ongoing
|
||||
// - [x] correctly cancel Provider when cancelled
|
||||
// - [x] create auth token
|
||||
// - [x] export backup with generated token as password
|
||||
// - needs a path to store the database
|
||||
// - [x] create the sendme database
|
||||
// - [x] start provider
|
||||
// - [ ] wait for one successful connection
|
||||
// - [ ] shutdown when a connection is closed, successful or not.
|
||||
//
|
||||
// - [ ] provide progress report
|
||||
/// Prepares for sending a backup to a second device.
|
||||
///
|
||||
/// Before calling this function all I/O must be stopped so that no changes to the blobs
|
||||
/// or database are happening, this is done by calling the `dc_accounts_stop_io` or
|
||||
/// `dc_stop_io` APIs first. TODO: Add the rust equivalents.
|
||||
///
|
||||
/// This will acquire the global "ongoing process" mutex. You must call
|
||||
/// [`BackupSender::join`] after creating this struct, otherwise this will not respect
|
||||
/// the possible cancellation of the "ongoing process".
|
||||
pub async fn perpare(context: &Context, dir: &Path) -> Result<Self> {
|
||||
ensure!(
|
||||
// TODO: Should we worry about path normalisation?
|
||||
dir != context.get_blobdir(),
|
||||
"Temporary database export directory should not be in blobdir"
|
||||
);
|
||||
e2ee::ensure_secret_key_exists(context)
|
||||
.await
|
||||
.context("Private key not available, aborting backup export")?;
|
||||
|
||||
// Acquire global "ongoing" mutex.
|
||||
let cancel_token = context.alloc_ongoing().await?;
|
||||
let res = tokio::select! {
|
||||
biased;
|
||||
res = Self::prepare_inner(context, dir) => {
|
||||
match res {
|
||||
Ok(slf) => {
|
||||
// TODO: maybe this is the wrong place to log this
|
||||
// TODO: Also needs to log progress somehow.
|
||||
info!(context, "Waiting for remote to connect");
|
||||
Ok(slf)
|
||||
},
|
||||
Err(err) => {
|
||||
error!(context, "Failed to set up second device setup: {:#}", err);
|
||||
Err(err)
|
||||
},
|
||||
}
|
||||
},
|
||||
_ = cancel_token.recv() => Err(format_err!("cancelled")),
|
||||
};
|
||||
let (provider, ticket) = match res {
|
||||
Ok((provider, ticket)) => (provider, ticket),
|
||||
Err(err) => {
|
||||
context.free_ongoing().await;
|
||||
return Err(err);
|
||||
}
|
||||
};
|
||||
Ok(Self {
|
||||
provider,
|
||||
ticket,
|
||||
cancel_token,
|
||||
})
|
||||
}
|
||||
|
||||
async fn prepare_inner(
|
||||
context: &Context,
|
||||
dir: &Path,
|
||||
) -> Result<(sendme::provider::Provider, sendme::provider::Ticket)> {
|
||||
// Generate the token up front: we also use it to encrypt the database.
|
||||
let token = sendme::protocol::AuthToken::generate();
|
||||
let dbfile = dir.join(DBFILE_BACKUP_NAME);
|
||||
export_database(context, &dbfile, token.to_string())
|
||||
.await
|
||||
.context("Database export failed")?;
|
||||
|
||||
// Now we can be sure IO is not running.
|
||||
let mut files = vec![sendme::provider::DataSource::from(dbfile)];
|
||||
let blobdir = BlobDirContents::new(context).await?;
|
||||
for blob in blobdir.iter() {
|
||||
files.push(blob.to_abs_path().into());
|
||||
}
|
||||
|
||||
// Start listening.
|
||||
let (db, hash) = sendme::provider::create_collection(files).await?;
|
||||
let provider = sendme::provider::Provider::builder(db)
|
||||
.auth_token(token)
|
||||
.spawn()?;
|
||||
let ticket = provider.ticket(hash);
|
||||
Ok((provider, ticket))
|
||||
}
|
||||
|
||||
pub fn qr(&self) -> Qr {
|
||||
Qr::Backup {
|
||||
ticket: self.ticket.clone(),
|
||||
}
|
||||
}
|
||||
|
||||
/// Wait for the backup sender to complete.
|
||||
///
|
||||
/// The sender completes when an authenticated client disconnects, whether the transfer
|
||||
/// was successful or not. When the ongoing task is cancelled the sender also completes
|
||||
/// with an error.
|
||||
///
|
||||
/// Note that this must be called and awaited for the ongoing cancellation to work.
|
||||
pub async fn join(self) -> Result<()> {
|
||||
// TODO: should wait for 1 transfer to complete or abort
|
||||
tokio::select! {
|
||||
biased;
|
||||
res = self.provider.join() => res.context("BackupSender failed"),
|
||||
_ = self.cancel_token.recv() => Err(anyhow!("BackupSender cancelled")),
|
||||
}
|
||||
}
|
||||
|
||||
pub fn abort(&self) {
|
||||
self.provider.abort()
|
||||
}
|
||||
}
|
||||
|
||||
/// All files in the blobdir.
|
||||
///
|
||||
/// This exists so we can have a [`BlobDirIter`] which needs something to own the data of
|
||||
|
||||
224
src/imex/transfer.rs
Normal file
224
src/imex/transfer.rs
Normal file
@@ -0,0 +1,224 @@
|
||||
//! Transfer a backup to an other device.
|
||||
//!
|
||||
//! This module provides support for using n0's sendme tool to initiate transfer of a backup
|
||||
//! to another device using a QR code.
|
||||
//!
|
||||
//! Using the sendme terminology there are two parties to this:
|
||||
//!
|
||||
//! - The *Provider*, which starts a server and listens for connections.
|
||||
//! - The *Getter*, which connects to the server and retrieves the data.
|
||||
//!
|
||||
//! Sendme is designed around the idea of verifying hashes, the downloads are verified as
|
||||
//! they are retrieved. The entire transfer is initiated by requesting the data of a single
|
||||
//! root hash.
|
||||
//!
|
||||
//! Both the provider and the getter are authenticated:
|
||||
//!
|
||||
//! - The provider is known by its *peer ID*.
|
||||
//! - The provider needs an *authentication token* from the getter before it accepts a
|
||||
//! connection.
|
||||
//!
|
||||
//! Both these are transferred in the QR code offered to the getter. This ensures that the
|
||||
//! getter can not connect to an impersonated provider and the provider does not offer the
|
||||
//! download to an impersonated getter.
|
||||
|
||||
use std::path::Path;
|
||||
|
||||
use crate::context::Context;
|
||||
use crate::e2ee;
|
||||
use crate::qr::Qr;
|
||||
use anyhow::{anyhow, bail, ensure, format_err, Context as _, Result};
|
||||
use async_channel::Receiver;
|
||||
use sendme::blobs::Collection;
|
||||
use sendme::get::{AsyncSliceDecoder, Hash, Options, ReceiveStream};
|
||||
use sendme::protocol::AuthToken;
|
||||
use sendme::provider::{DataSource, Provider, Ticket};
|
||||
use tokio::fs::File;
|
||||
use tokio::io::{self, BufWriter};
|
||||
|
||||
use super::{export_database, BlobDirContents, DBFILE_BACKUP_NAME};
|
||||
|
||||
/// Provide or send a backup of this device.
|
||||
///
|
||||
/// This creates a backup of the current device and starts a service which offers another
|
||||
/// device to download this backup.
|
||||
///
|
||||
/// This does not make a full backup on disk, only the SQLite database is created on disk,
|
||||
/// the blobs in the blob directory are not copied.
|
||||
#[derive(Debug)]
|
||||
pub struct BackupProvider {
|
||||
/// A handle to the running provider.
|
||||
provider: Provider,
|
||||
/// The ticket to retrieve the backup collection.
|
||||
ticket: Ticket,
|
||||
/// Token holding the "ongoing" mutex. When this completes the provider should shut
|
||||
/// down.
|
||||
cancel_token: Receiver<()>,
|
||||
}
|
||||
|
||||
impl BackupProvider {
|
||||
/// Prepares for sending a backup to a second device.
|
||||
///
|
||||
/// Before calling this function all I/O must be stopped so that no changes to the blobs
|
||||
/// or database are happening, this is done by calling the `dc_accounts_stop_io` or
|
||||
/// `dc_stop_io` APIs first. TODO: Add the rust equivalents.
|
||||
///
|
||||
/// This will acquire the global "ongoing process" mutex. You must call
|
||||
/// [`BackupSender::join`] after creating this struct, otherwise this will not respect
|
||||
/// the possible cancellation of the "ongoing process".
|
||||
pub async fn perpare(context: &Context, dir: &Path) -> Result<Self> {
|
||||
ensure!(
|
||||
// TODO: Should we worry about path normalisation?
|
||||
dir != context.get_blobdir(),
|
||||
"Temporary database export directory should not be in blobdir"
|
||||
);
|
||||
e2ee::ensure_secret_key_exists(context)
|
||||
.await
|
||||
.context("Private key not available, aborting backup export")?;
|
||||
|
||||
// Acquire global "ongoing" mutex.
|
||||
let cancel_token = context.alloc_ongoing().await?;
|
||||
let res = tokio::select! {
|
||||
biased;
|
||||
res = Self::prepare_inner(context, dir) => {
|
||||
match res {
|
||||
Ok(slf) => {
|
||||
// TODO: maybe this is the wrong place to log this
|
||||
// TODO: Also needs to log progress somehow.
|
||||
info!(context, "Waiting for remote to connect");
|
||||
Ok(slf)
|
||||
},
|
||||
Err(err) => {
|
||||
error!(context, "Failed to set up second device setup: {:#}", err);
|
||||
Err(err)
|
||||
},
|
||||
}
|
||||
},
|
||||
_ = cancel_token.recv() => Err(format_err!("cancelled")),
|
||||
};
|
||||
let (provider, ticket) = match res {
|
||||
Ok((provider, ticket)) => (provider, ticket),
|
||||
Err(err) => {
|
||||
context.free_ongoing().await;
|
||||
return Err(err);
|
||||
}
|
||||
};
|
||||
Ok(Self {
|
||||
provider,
|
||||
ticket,
|
||||
cancel_token,
|
||||
})
|
||||
}
|
||||
|
||||
async fn prepare_inner(context: &Context, dir: &Path) -> Result<(Provider, Ticket)> {
|
||||
// Generate the token up front: we also use it to encrypt the database.
|
||||
let token = AuthToken::generate();
|
||||
let dbfile = dir.join(DBFILE_BACKUP_NAME);
|
||||
export_database(context, &dbfile, token.to_string())
|
||||
.await
|
||||
.context("Database export failed")?;
|
||||
|
||||
// Now we can be sure IO is not running.
|
||||
let mut files = vec![DataSource::from(dbfile)];
|
||||
let blobdir = BlobDirContents::new(context).await?;
|
||||
for blob in blobdir.iter() {
|
||||
files.push(blob.to_abs_path().into());
|
||||
}
|
||||
|
||||
// Start listening.
|
||||
let (db, hash) = sendme::provider::create_collection(files).await?;
|
||||
let provider = Provider::builder(db).auth_token(token).spawn()?;
|
||||
let ticket = provider.ticket(hash);
|
||||
Ok((provider, ticket))
|
||||
}
|
||||
|
||||
pub fn qr(&self) -> Qr {
|
||||
Qr::Backup {
|
||||
ticket: self.ticket.clone(),
|
||||
}
|
||||
}
|
||||
|
||||
/// Wait for the backup sender to complete.
|
||||
///
|
||||
/// The sender completes when an authenticated client disconnects, whether the transfer
|
||||
/// was successful or not. When the ongoing task is cancelled the sender also completes
|
||||
/// with an error.
|
||||
///
|
||||
/// Note that this must be called and awaited for the ongoing cancellation to work.
|
||||
pub async fn join(self) -> Result<()> {
|
||||
// TODO: should wait for 1 transfer to complete or abort
|
||||
tokio::select! {
|
||||
biased;
|
||||
res = self.provider.join() => res.context("BackupSender failed"),
|
||||
_ = self.cancel_token.recv() => Err(anyhow!("BackupSender cancelled")),
|
||||
}
|
||||
}
|
||||
|
||||
pub fn abort(&self) {
|
||||
self.provider.abort()
|
||||
}
|
||||
}
|
||||
|
||||
/// Contacts a backup provider and receives the backup from it.
|
||||
///
|
||||
/// This uses a QR code to contact another instance of deltachat which is providing a backup
|
||||
/// using the [`BackupProvider`]. Once connected it will authenticate using the secrets in
|
||||
/// the QR code and retrieve the backup.
|
||||
///
|
||||
/// Using [`Qr`] as argument is a bit odd as it only accepts one specific variant of it. It
|
||||
/// does avoid having [`sendme::provider::Ticket`] in the primary API however, without
|
||||
/// having to revert to untyped bytes.
|
||||
pub async fn get_backup(context: &Context, qr: Qr) -> Result<()> {
|
||||
let Qr::Backup { ticket } = qr else {
|
||||
bail!("QR code for backup must be of type DCBACKUP");
|
||||
};
|
||||
ensure!(
|
||||
!context.is_configured().await?,
|
||||
"Cannot import backups to accounts in use."
|
||||
);
|
||||
ensure!(
|
||||
context.scheduler.read().await.is_none(),
|
||||
"cannot import backup, IO is running"
|
||||
);
|
||||
let opts = Options {
|
||||
addr: ticket.addr,
|
||||
peer_id: Some(ticket.peer),
|
||||
};
|
||||
let on_blob = |hash, reader, name| on_blob(context, hash, reader, name);
|
||||
sendme::get::run(
|
||||
ticket.hash,
|
||||
ticket.token,
|
||||
opts,
|
||||
on_connected,
|
||||
on_collection,
|
||||
on_blob,
|
||||
)
|
||||
.await?;
|
||||
|
||||
todo!();
|
||||
}
|
||||
|
||||
/// Get callback when the connection is established with the provider.
|
||||
async fn on_connected() -> Result<()> {
|
||||
Ok(())
|
||||
}
|
||||
|
||||
/// Get callback when a collection is received from the provider.
|
||||
async fn on_collection(_collection: Collection) -> Result<()> {
|
||||
Ok(())
|
||||
}
|
||||
|
||||
/// Get callback when a blob is received from the provider.
|
||||
async fn on_blob(
|
||||
context: &Context,
|
||||
_hash: Hash,
|
||||
mut reader: AsyncSliceDecoder<ReceiveStream>,
|
||||
name: String,
|
||||
) -> Result<AsyncSliceDecoder<ReceiveStream>> {
|
||||
ensure!(!name.is_empty(), "Received a nameless blob");
|
||||
let path = context.get_blobdir().join(name);
|
||||
let file = File::create(&path).await?;
|
||||
let mut file = BufWriter::with_capacity(128 * 1024, file);
|
||||
io::copy(&mut reader, &mut file).await?;
|
||||
Ok(reader)
|
||||
}
|
||||
Reference in New Issue
Block a user