// Copyright (c) 2022-2024 Niko Bonnieure, Par le Peuple, NextGraph.org developers // All rights reserved. // Licensed under the Apache License, Version 2.0 // // or the MIT license , // at your option. All files in the project carrying such // notice may not be copied, modified, or distributed except // according to those terms. //! App Protocol (between LocalBroker and Verifier) use lazy_static::lazy_static; use regex::Regex; use serde::{Deserialize, Serialize}; use ng_repo::errors::NgError; #[allow(unused_imports)] use ng_repo::log::*; use ng_repo::repo::CommitInfo; use ng_repo::types::*; use ng_repo::utils::{decode_digest, decode_key, decode_sym_key}; use ng_repo::utils::{decode_overlayid, display_timestamp_local}; use crate::types::*; lazy_static! { #[doc(hidden)] static ref RE_FILE_READ_CAP: Regex = Regex::new(r"^did:ng:j:([A-Za-z0-9-_]*):k:([A-Za-z0-9-_]*)$").unwrap(); #[doc(hidden)] static ref RE_REPO_O: Regex = Regex::new(r"^did:ng:o:([A-Za-z0-9-_]*)$").unwrap(); #[doc(hidden)] static ref RE_REPO: Regex = Regex::new(r"^did:ng:o:([A-Za-z0-9-_]*):v:([A-Za-z0-9-_]*)$").unwrap(); #[doc(hidden)] static ref RE_BRANCH: Regex = Regex::new(r"^did:ng:o:([A-Za-z0-9-_]*):v:([A-Za-z0-9-_]*):b:([A-Za-z0-9-_]*)$").unwrap(); #[doc(hidden)] static ref RE_NAMED_BRANCH_OR_COMMIT: Regex = Regex::new(r"^did:ng:o:([A-Za-z0-9-_]*):v:([A-Za-z0-9-_]*):a:([A-Za-z0-9-_%]*)$").unwrap(); //TODO: allow international chars. disallow digit as first char #[doc(hidden)] static ref RE_OBJECTS: Regex = Regex::new(r"^did:ng(?::o:([A-Za-z0-9-_]{44}))?:v:([A-Za-z0-9-_]{44})((?::c:[A-Za-z0-9-_]{44}:k:[A-Za-z0-9-_]{44})+)(?::s:([A-Za-z0-9-_]{44}):k:([A-Za-z0-9-_]{44}))?:l:([A-Za-z0-9-_]*)$").unwrap(); #[doc(hidden)] static ref RE_OBJECT_READ_CAPS: Regex = Regex::new(r":[cj]:([A-Za-z0-9-_]{44}):k:([A-Za-z0-9-_]{44})").unwrap(); } #[derive(Clone, Debug, Serialize, Deserialize)] pub enum AppFetchContentV0 { Get, // does not subscribe. Subscribe, Update, //Invoke, ReadQuery, WriteQuery, RdfDump, History, SignatureStatus, SignatureRequest, SignedSnapshotRequest, } impl AppFetchContentV0 { pub fn get_or_subscribe(subscribe: bool) -> Self { if !subscribe { AppFetchContentV0::Get } else { AppFetchContentV0::Subscribe } } } #[derive(Clone, Debug, Serialize, Deserialize)] pub enum NgAccessV0 { ReadCap(ReadCap), Token(Digest), #[serde(with = "serde_bytes")] ExtRequest(Vec), Key(BlockKey), Inbox(Digest), } #[derive(Clone, Debug, Serialize, Deserialize)] pub enum TargetBranchV0 { Chat, Stream, Comments, BackLinks, Context, BranchId(BranchId), Named(String), // branch or commit Commits(Vec), // only possible if access to their branch is given. must belong to the same branch. } impl TargetBranchV0 { pub fn is_valid_for_sparql_update(&self) -> bool { match self { Self::Commits(_) => false, _ => true, } } pub fn is_valid_for_discrete_update(&self) -> bool { match self { Self::BranchId(_) => true, //TODO: add Named(s) is s is a branch => true _ => false, } } pub fn branch_id(&self) -> &BranchId { match self { Self::BranchId(id) => id, _ => panic!("not a TargetBranchV0::BranchId"), } } } #[derive(Clone, Debug, Serialize, Deserialize)] pub enum NuriTargetV0 { UserSite, // targets the whole data set of the user PublicStore, ProtectedStore, PrivateStore, AllDialogs, Dialog(String), // shortname of a Dialog AllGroups, Group(String), // shortname of a Group Repo(RepoId), None, } impl NuriTargetV0 { pub fn is_valid_for_sparql_update(&self) -> bool { match self { Self::UserSite | Self::AllDialogs | Self::AllGroups => false, _ => true, } } pub fn is_valid_for_discrete_update(&self) -> bool { match self { Self::UserSite | Self::AllDialogs | Self::AllGroups | Self::None => false, _ => true, } } pub fn is_repo_id(&self) -> bool { match self { Self::Repo(_) => true, _ => false, } } pub fn repo_id(&self) -> &RepoId { match self { Self::Repo(id) => id, _ => panic!("not a NuriTargetV0::Repo"), } } } #[derive(Clone, Debug, Serialize, Deserialize, PartialEq, Eq)] pub struct CommitInfoJs { pub past: Vec, pub key: String, pub signature: Option, pub author: String, pub timestamp: String, pub final_consistency: bool, pub commit_type: CommitType, pub branch: Option, pub x: u32, pub y: u32, } impl From<&CommitInfo> for CommitInfoJs { fn from(info: &CommitInfo) -> Self { CommitInfoJs { past: info.past.iter().map(|objid| objid.to_string()).collect(), key: info.key.to_string(), signature: info.signature.as_ref().map(|s| NuriV0::signature_ref(&s)), author: info.author.clone(), timestamp: display_timestamp_local(info.timestamp), final_consistency: info.final_consistency, commit_type: info.commit_type.clone(), branch: info.branch.map(|b| b.to_string()), x: info.x, y: info.y, } } } const DID_PREFIX: &str = "did:ng"; #[derive(Clone, Debug, Serialize, Deserialize)] pub struct NuriV0 { pub identity: Option, // None for personal identity pub target: NuriTargetV0, pub entire_store: bool, // If it is a store, will include all the docs belonging to the store pub objects: Vec, // used only for FileGet. // cannot be used for queries. only to download an object (file,commit..) pub signature: Option, pub branch: Option, // if None, the main branch is chosen pub overlay: Option, pub access: Vec, pub topic: Option, pub locator: Option, } impl NuriV0 { pub fn new_empty() -> Self { NuriV0 { identity: None, target: NuriTargetV0::None, entire_store: false, objects: vec![], signature: None, branch: None, overlay: None, access: vec![], topic: None, locator: None, } } pub fn copy_target_from(&mut self, nuri: &NuriV0) { self.target = nuri.target.clone(); } pub fn commit_graph_name(commit_id: &ObjectId, overlay_id: &OverlayId) -> String { format!("{DID_PREFIX}:c:{commit_id}:v:{overlay_id}") } pub fn commit_graph_name_from_base64(commit_base64: &String, overlay_id: &OverlayId) -> String { format!("{DID_PREFIX}:c:{commit_base64}:v:{overlay_id}") } pub fn from_store_repo(store_repo: &StoreRepo) -> Self { NuriV0 { identity: None, target: NuriTargetV0::Repo(store_repo.repo_id().clone()), entire_store: false, objects: vec![], signature: None, branch: None, overlay: None, access: vec![], topic: None, locator: None, } } pub fn to_store_nuri_string(store_id: &RepoId) -> String { let overlay_id = OverlayId::outer(store_id); format!("o:{store_id}:v:{overlay_id}") } pub fn repo_graph_name(repo_id: &RepoId, overlay_id: &OverlayId) -> String { format!("{DID_PREFIX}:o:{repo_id}:v:{overlay_id}") } pub fn repo_skolem( repo_id: &RepoId, peer_id: &Vec, random: u128, ) -> Result { let mut arr = Vec::with_capacity(32); arr.extend_from_slice(peer_id); arr.extend_from_slice(&random.to_be_bytes()); let sko: SymKey = arr.as_slice().try_into()?; Ok(format!("{DID_PREFIX}:o:{repo_id}:u:{sko}")) } pub fn repo(&self) -> String { Self::repo_id(self.target.repo_id()) } pub fn repo_id(repo_id: &RepoId) -> String { format!("{DID_PREFIX}:o:{}", repo_id) } pub fn overlay_id(overlay_id: &OverlayId) -> String { format!("{DID_PREFIX}:v:{overlay_id}") } pub fn topic_id(topic_id: &TopicId) -> String { format!("{DID_PREFIX}:h:{topic_id}") } pub fn branch_id(branch_id: &BranchId) -> String { format!("{DID_PREFIX}:b:{branch_id}") } pub fn branch_id_from_base64(branch_base64: &String) -> String { format!("{DID_PREFIX}:b:{branch_base64}") } pub fn object_ref(obj_ref: &ObjectRef) -> String { format!("{DID_PREFIX}:{}", obj_ref.object_nuri()) } pub fn signature_ref(obj_ref: &ObjectRef) -> String { format!("s:{}:k:{}", obj_ref.id, obj_ref.key) } pub fn token(token: &Digest) -> String { format!("{DID_PREFIX}:n:{token}") } pub fn locator(locator: &Locator) -> String { format!("l:{locator}") } pub fn is_branch_identifier(&self) -> bool { self.locator.is_none() && self.topic.is_none() && self.access.is_empty() && self.overlay.as_ref().map_or(false, |o| o.is_outer()) && self .branch .as_ref() .map_or(true, |b| b.is_valid_for_sparql_update()) && self.objects.is_empty() && self.signature.is_none() && !self.entire_store && self.target.is_repo_id() } pub fn is_valid_for_sparql_update(&self) -> bool { self.objects.is_empty() && self.signature.is_none() && self.entire_store == false && self.target.is_valid_for_sparql_update() && self .branch .as_ref() .map_or(true, |b| b.is_valid_for_sparql_update()) } pub fn is_valid_for_discrete_update(&self) -> bool { self.objects.is_empty() && self.signature.is_none() && self.entire_store == false && self.target.is_valid_for_discrete_update() && self .branch .as_ref() .map_or(true, |b| b.is_valid_for_discrete_update()) } pub fn new_repo_target_from_string(repo_id_string: String) -> Result { let repo_id: RepoId = repo_id_string.as_str().try_into()?; Ok(Self { identity: None, target: NuriTargetV0::Repo(repo_id), entire_store: false, objects: vec![], signature: None, branch: None, overlay: None, access: vec![], topic: None, locator: None, }) } pub fn new_from_obj_ref(obj_ref: &ObjectRef) -> Self { Self { identity: None, target: NuriTargetV0::None, entire_store: false, objects: vec![obj_ref.clone()], signature: None, branch: None, overlay: None, access: vec![], topic: None, locator: None, } } pub fn new_private_store_target() -> Self { Self { identity: None, target: NuriTargetV0::PrivateStore, entire_store: false, objects: vec![], signature: None, branch: None, overlay: None, access: vec![], topic: None, locator: None, } } pub fn new_entire_user_site() -> Self { Self { identity: None, target: NuriTargetV0::UserSite, entire_store: false, objects: vec![], signature: None, branch: None, overlay: None, access: vec![], topic: None, locator: None, } } pub fn new_for_readcaps(from: &str) -> Result { let c = RE_OBJECTS.captures(from); if let Some(c) = c { let target = c.get(1).map_or(NuriTargetV0::None, |repo_match| { if let Ok(id) = decode_key(repo_match.as_str()) { NuriTargetV0::Repo(id) } else { NuriTargetV0::None } }); let overlay_id = decode_overlayid(c.get(2).ok_or(NgError::InvalidNuri)?.as_str())?; let read_caps = c.get(3).ok_or(NgError::InvalidNuri)?.as_str(); let sign_obj_id = c.get(4).map(|c| decode_digest(c.as_str())); let sign_obj_key = c.get(5).map(|c| decode_sym_key(c.as_str())); let locator = TryInto::::try_into(c.get(6).ok_or(NgError::InvalidNuri)?.as_str())?; let signature = if sign_obj_id.is_some() && sign_obj_key.is_some() { Some(ObjectRef::from_id_key( sign_obj_id.unwrap()?, sign_obj_key.unwrap()?, )) } else { None }; let objects = RE_OBJECT_READ_CAPS .captures_iter(read_caps) .map(|c| { Ok(ObjectRef::from_id_key( decode_digest(c.get(1).ok_or(NgError::InvalidNuri)?.as_str())?, decode_sym_key(c.get(2).ok_or(NgError::InvalidNuri)?.as_str())?, )) }) .collect::, NgError>>()?; if objects.len() < 1 { return Err(NgError::InvalidNuri); } Ok(Self { identity: None, target, entire_store: false, objects, signature, branch: None, overlay: Some(overlay_id.into()), access: vec![], topic: None, locator: Some(locator), }) } else { Err(NgError::InvalidNuri) } } pub fn new_from(from: &String) -> Result { let c = RE_REPO_O.captures(from); if c.is_some() && c.as_ref().unwrap().get(1).is_some() { let cap = c.unwrap(); let o = cap.get(1).unwrap().as_str(); let repo_id = decode_key(o)?; Ok(Self { identity: None, target: NuriTargetV0::Repo(repo_id), entire_store: false, objects: vec![], signature: None, branch: None, overlay: None, access: vec![], topic: None, locator: None, }) } else { let c = RE_FILE_READ_CAP.captures(from); if c.is_some() && c.as_ref().unwrap().get(1).is_some() && c.as_ref().unwrap().get(2).is_some() { let cap = c.unwrap(); let j = cap.get(1).unwrap().as_str(); let k = cap.get(2).unwrap().as_str(); let id = decode_digest(j)?; let key = decode_sym_key(k)?; Ok(Self { identity: None, target: NuriTargetV0::None, entire_store: false, objects: vec![ObjectRef::from_id_key(id, key)], signature: None, branch: None, overlay: None, access: vec![], topic: None, locator: None, }) } else { let c = RE_REPO.captures(from); if c.is_some() && c.as_ref().unwrap().get(1).is_some() && c.as_ref().unwrap().get(2).is_some() { let cap = c.unwrap(); let o = cap.get(1).unwrap().as_str(); let v = cap.get(2).unwrap().as_str(); let repo_id = decode_key(o)?; let overlay_id = decode_overlayid(v)?; Ok(Self { identity: None, target: NuriTargetV0::Repo(repo_id), entire_store: false, objects: vec![], signature: None, branch: None, overlay: Some(overlay_id.into()), access: vec![], topic: None, locator: None, }) } else { let c = RE_BRANCH.captures(from); if c.is_some() && c.as_ref().unwrap().get(1).is_some() && c.as_ref().unwrap().get(2).is_some() && c.as_ref().unwrap().get(3).is_some() { let cap = c.unwrap(); let o = cap.get(1).unwrap().as_str(); let v = cap.get(2).unwrap().as_str(); let b = cap.get(3).unwrap().as_str(); let repo_id = decode_key(o)?; let overlay_id = decode_overlayid(v)?; let branch_id = decode_key(b)?; Ok(Self { identity: None, target: NuriTargetV0::Repo(repo_id), entire_store: false, objects: vec![], signature: None, branch: Some(TargetBranchV0::BranchId(branch_id)), overlay: Some(overlay_id.into()), access: vec![], topic: None, locator: None, }) } else { Err(NgError::InvalidNuri) } } } } } } #[derive(Clone, Debug, Serialize, Deserialize)] pub enum AppRequestCommandV0 { Fetch(AppFetchContentV0), Pin, UnPin, Delete, Create, FileGet, // needs the Nuri of branch/doc/store AND ObjectId FilePut, // needs the Nuri of branch/doc/store } impl AppRequestCommandV0 { pub fn is_stream(&self) -> bool { match self { Self::Fetch(AppFetchContentV0::Subscribe) | Self::FileGet => true, Self::FilePut | Self::Create | Self::Delete | Self::UnPin | Self::Pin | Self::Fetch(_) => false, } } pub fn new_read_query() -> Self { AppRequestCommandV0::Fetch(AppFetchContentV0::ReadQuery) } pub fn new_write_query() -> Self { AppRequestCommandV0::Fetch(AppFetchContentV0::WriteQuery) } pub fn new_update() -> Self { AppRequestCommandV0::Fetch(AppFetchContentV0::Update) } pub fn new_rdf_dump() -> Self { AppRequestCommandV0::Fetch(AppFetchContentV0::RdfDump) } pub fn new_history() -> Self { AppRequestCommandV0::Fetch(AppFetchContentV0::History) } pub fn new_signature_status() -> Self { AppRequestCommandV0::Fetch(AppFetchContentV0::SignatureStatus) } pub fn new_signature_request() -> Self { AppRequestCommandV0::Fetch(AppFetchContentV0::SignatureRequest) } pub fn new_signed_snapshot_request() -> Self { AppRequestCommandV0::Fetch(AppFetchContentV0::SignedSnapshotRequest) } pub fn new_create() -> Self { AppRequestCommandV0::Create } } #[derive(Clone, Debug, Serialize, Deserialize)] pub struct AppRequestV0 { pub command: AppRequestCommandV0, pub nuri: NuriV0, pub payload: Option, pub session_id: u64, } #[derive(Clone, Debug, Serialize, Deserialize)] pub enum AppRequest { V0(AppRequestV0), } impl AppRequest { pub fn set_session_id(&mut self, session_id: u64) { match self { Self::V0(v0) => v0.session_id = session_id, } } pub fn session_id(&self) -> u64 { match self { Self::V0(v0) => v0.session_id, } } pub fn command(&self) -> &AppRequestCommandV0 { match self { Self::V0(v0) => &v0.command, } } pub fn new( command: AppRequestCommandV0, nuri: NuriV0, payload: Option, ) -> Self { AppRequest::V0(AppRequestV0 { command, nuri, payload, session_id: 0, }) } } #[derive(Clone, Debug, Serialize, Deserialize)] pub struct AppSessionStopV0 { pub session_id: u64, pub force_close: bool, } #[derive(Clone, Debug, Serialize, Deserialize)] pub enum AppSessionStop { V0(AppSessionStopV0), } impl AppSessionStop { pub fn session_id(&self) -> u64 { match self { Self::V0(v0) => v0.session_id, } } pub fn is_force_close(&self) -> bool { match self { Self::V0(v0) => v0.force_close, } } } #[derive(Clone, Debug, Serialize, Deserialize)] pub struct AppSessionStartV0 { pub session_id: u64, pub credentials: Option, pub user_id: UserId, pub detach: bool, } #[derive(Clone, Debug, Serialize, Deserialize)] pub enum AppSessionStart { V0(AppSessionStartV0), } #[derive(Clone, Debug, Serialize, Deserialize)] pub struct AppSessionStartResponseV0 { pub private_store: RepoId, pub protected_store: RepoId, pub public_store: RepoId, } #[derive(Clone, Debug, Serialize, Deserialize)] pub enum AppSessionStartResponse { V0(AppSessionStartResponseV0), } impl AppSessionStart { pub fn session_id(&self) -> u64 { match self { Self::V0(v0) => v0.session_id, } } pub fn credentials(&self) -> &Option { match self { Self::V0(v0) => &v0.credentials, } } pub fn user_id(&self) -> &UserId { match self { Self::V0(v0) => &v0.user_id, } } } #[derive(Clone, Debug, Serialize, Deserialize)] pub enum DocQuery { V0 { sparql: String, base: Option, }, } #[derive(Clone, Debug, Serialize, Deserialize)] pub struct GraphUpdate { // serialization of Vec #[serde(with = "serde_bytes")] pub inserts: Vec, // serialization of Vec #[serde(with = "serde_bytes")] pub removes: Vec, } #[derive(Clone, Debug, Serialize, Deserialize)] pub enum DiscreteUpdate { /// A yrs::Update #[serde(with = "serde_bytes")] YMap(Vec), #[serde(with = "serde_bytes")] YArray(Vec), #[serde(with = "serde_bytes")] YXml(Vec), #[serde(with = "serde_bytes")] YText(Vec), /// An automerge::Change.raw_bytes() #[serde(with = "serde_bytes")] Automerge(Vec), } impl DiscreteUpdate { pub fn from(crdt: String, update: Vec) -> Self { match crdt.as_str() { "YMap" => Self::YMap(update), "YArray" => Self::YArray(update), "YXml" => Self::YXml(update), "YText" => Self::YText(update), "Automerge" => Self::Automerge(update), _ => panic!("wrong crdt type"), } } } #[derive(Clone, Debug, Serialize, Deserialize)] pub struct DocUpdate { pub heads: Vec, pub graph: Option, pub discrete: Option, } #[derive(Clone, Debug, Serialize, Deserialize)] pub struct DocAddFile { pub filename: Option, pub object: ObjectRef, } #[derive(Clone, Debug, Serialize, Deserialize)] pub enum DocCreateDestination { Store, Stream, MagicCarpet, } impl DocCreateDestination { pub fn from(s: String) -> Result { Ok(match s.as_str() { "store" => Self::Store, "stream" => Self::Stream, "mc" => Self::MagicCarpet, _ => return Err(NgError::InvalidArgument), }) } } #[derive(Clone, Debug, Serialize, Deserialize)] pub struct DocCreate { pub store: StoreRepo, pub class: BranchCrdt, pub destination: DocCreateDestination, } #[derive(Clone, Debug, Serialize, Deserialize)] pub struct DocDelete { /// Nuri of doc to delete nuri: String, } #[derive(Clone, Debug, Serialize, Deserialize)] pub enum AppRequestPayloadV0 { Create(DocCreate), Query(DocQuery), Update(DocUpdate), AddFile(DocAddFile), //RemoveFile Delete(DocDelete), //Invoke(InvokeArguments), SmallFilePut(SmallFile), RandomAccessFilePut(String), // content_type (iana media type) RandomAccessFilePutChunk((u32, serde_bytes::ByteBuf)), // end the upload with an empty vec } #[derive(Clone, Debug, Serialize, Deserialize)] pub enum AppRequestPayload { V0(AppRequestPayloadV0), } impl AppRequestPayload { pub fn new_sparql_query(sparql: String, base: Option) -> Self { AppRequestPayload::V0(AppRequestPayloadV0::Query(DocQuery::V0 { sparql, base })) } pub fn new_discrete_update( head_strings: Vec, crdt: String, update: Vec, ) -> Result { let mut heads = Vec::with_capacity(head_strings.len()); for head in head_strings { heads.push(decode_digest(&head)?); } let discrete = Some(DiscreteUpdate::from(crdt, update)); Ok(AppRequestPayload::V0(AppRequestPayloadV0::Update( DocUpdate { heads, graph: None, discrete, }, ))) } } #[derive(Clone, Debug, Serialize, Deserialize)] pub enum DiscretePatch { /// A yrs::Update #[serde(with = "serde_bytes")] YMap(Vec), #[serde(with = "serde_bytes")] YArray(Vec), #[serde(with = "serde_bytes")] YXml(Vec), #[serde(with = "serde_bytes")] YText(Vec), /// An automerge::Change.raw_bytes() or a concatenation of several. #[serde(with = "serde_bytes")] Automerge(Vec), } #[derive(Clone, Debug, Serialize, Deserialize)] pub struct GraphPatch { // serialization of Vec #[serde(with = "serde_bytes")] pub inserts: Vec, // serialization of Vec #[serde(with = "serde_bytes")] pub removes: Vec, } #[derive(Clone, Debug, Serialize, Deserialize)] pub enum DiscreteState { /// A yrs::Update #[serde(with = "serde_bytes")] YMap(Vec), #[serde(with = "serde_bytes")] YArray(Vec), #[serde(with = "serde_bytes")] YXml(Vec), #[serde(with = "serde_bytes")] YText(Vec), // the output of Automerge::save() #[serde(with = "serde_bytes")] Automerge(Vec), } #[derive(Clone, Debug, Serialize, Deserialize)] pub struct GraphState { // serialization of Vec #[serde(with = "serde_bytes")] pub triples: Vec, } #[derive(Clone, Debug, Serialize, Deserialize)] pub struct AppState { pub heads: Vec, pub head_keys: Vec, pub graph: Option, // there is always a graph present in the branch. but it might not have been asked in the request pub discrete: Option, pub files: Vec, } #[derive(Clone, Debug, Serialize, Deserialize)] pub struct AppHistory { pub history: Vec<(ObjectId, CommitInfo)>, pub swimlane_state: Vec>, } #[derive(Clone, Debug, Serialize, Deserialize)] pub struct AppHistoryJs { pub history: Vec<(String, CommitInfoJs)>, pub swimlane_state: Vec>, } impl AppHistory { pub fn to_js(&self) -> AppHistoryJs { AppHistoryJs { history: Vec::from_iter( self.history .iter() .map(|(id, info)| (id.to_string(), info.into())), ), swimlane_state: Vec::from_iter( self.swimlane_state .iter() .map(|lane| lane.map_or(None, |b| Some(b.to_string()))), ), } } } #[derive(Clone, Debug, Serialize, Deserialize)] pub enum OtherPatch { FileAdd(FileName), FileRemove(ObjectId), AsyncSignature((String, Vec)), Snapshot(ObjectRef), Compact(ObjectRef), Other, } #[derive(Clone, Debug, Serialize, Deserialize)] pub struct AppPatch { pub commit_id: String, pub commit_info: CommitInfoJs, // or graph, or discrete, or both, or other. pub graph: Option, pub discrete: Option, pub other: Option, } #[derive(Clone, Debug, Serialize, Deserialize)] pub struct FileName { pub name: Option, pub reference: ObjectRef, pub nuri: String, } #[derive(Clone, Debug, Serialize, Deserialize)] pub struct FileMetaV0 { pub content_type: String, pub size: u64, } #[derive(Clone, Debug, Serialize, Deserialize)] pub struct AppTabStoreInfo { pub repo: Option, //+ pub overlay: Option, //+ pub has_outer: Option, pub store_type: Option, //+ pub readcap: Option, pub is_member: Option, pub inner: Option, pub title: Option, pub icon: Option, pub description: Option, } #[derive(Clone, Debug, Serialize, Deserialize)] pub struct AppTabDocInfo { pub nuri: Option, //+ pub is_store: Option, //+ pub is_member: Option, //+ pub title: Option, pub icon: Option, pub description: Option, pub authors: Option>, pub inbox: Option, pub can_edit: Option, //+ //TODO stream //TODO live_editors //TODO branches } #[derive(Clone, Debug, Serialize, Deserialize)] pub struct AppTabBranchInfo { pub id: Option, //+ pub readcap: Option, //+ pub comment_branch: Option, pub class: Option, //+ } #[derive(Clone, Debug, Serialize, Deserialize)] pub struct AppTabInfo { pub branch: Option, pub doc: Option, pub store: Option, } #[derive(Clone, Debug, Serialize, Deserialize)] pub enum AppResponseV0 { SessionStart(AppSessionStartResponse), TabInfo(AppTabInfo), State(AppState), Patch(AppPatch), History(AppHistory), SignatureStatus(Vec<(String, Option, bool)>), Text(String), //File(FileName), FileUploading(u32), FileUploaded(ObjectRef), #[serde(with = "serde_bytes")] FileBinary(Vec), FileMeta(FileMetaV0), #[serde(with = "serde_bytes")] QueryResult(Vec), // a serialized [SPARQL Query Results JSON Format](https://www.w3.org/TR/sparql11-results-json/) #[serde(with = "serde_bytes")] Graph(Vec), // a serde serialization of a list of triples. can be transformed on the client side to RDF-JS data model, or JSON-LD, or else (Turtle,...) http://rdf.js.org/data-model-spec/ Ok, True, False, Error(String), EndOfStream, Nuri(String), } #[derive(Clone, Debug, Serialize, Deserialize)] pub enum AppResponse { V0(AppResponseV0), } impl AppResponse { pub fn error(err: String) -> Self { AppResponse::V0(AppResponseV0::Error(err)) } pub fn ok() -> Self { AppResponse::V0(AppResponseV0::Ok) } }