mirror of
https://gitlab.com/veilid/veilid.git
synced 2024-10-01 01:26:08 -04:00
remove nodejs support
This commit is contained in:
parent
0e7f3e1c3c
commit
a54da97393
12
Cargo.lock
generated
12
Cargo.lock
generated
@ -2755,15 +2755,6 @@ dependencies = [
|
|||||||
"value-bag",
|
"value-bag",
|
||||||
]
|
]
|
||||||
|
|
||||||
[[package]]
|
|
||||||
name = "lru"
|
|
||||||
version = "0.8.1"
|
|
||||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
|
||||||
checksum = "b6e8aaa3f231bb4bd57b84b2d5dc3ae7f350265df8aa96492e0bc394a1571909"
|
|
||||||
dependencies = [
|
|
||||||
"hashbrown",
|
|
||||||
]
|
|
||||||
|
|
||||||
[[package]]
|
[[package]]
|
||||||
name = "lru-cache"
|
name = "lru-cache"
|
||||||
version = "0.1.2"
|
version = "0.1.2"
|
||||||
@ -5505,7 +5496,6 @@ dependencies = [
|
|||||||
"futures-util",
|
"futures-util",
|
||||||
"generic-array",
|
"generic-array",
|
||||||
"getrandom 0.2.8",
|
"getrandom 0.2.8",
|
||||||
"hashbrown",
|
|
||||||
"hashlink 0.8.1",
|
"hashlink 0.8.1",
|
||||||
"hex",
|
"hex",
|
||||||
"ifstructs",
|
"ifstructs",
|
||||||
@ -5515,11 +5505,11 @@ dependencies = [
|
|||||||
"js-sys",
|
"js-sys",
|
||||||
"json",
|
"json",
|
||||||
"keyring-manager",
|
"keyring-manager",
|
||||||
|
"keyvaluedb",
|
||||||
"keyvaluedb-sqlite",
|
"keyvaluedb-sqlite",
|
||||||
"keyvaluedb-web",
|
"keyvaluedb-web",
|
||||||
"lazy_static",
|
"lazy_static",
|
||||||
"libc",
|
"libc",
|
||||||
"lru",
|
|
||||||
"maplit",
|
"maplit",
|
||||||
"ndk 0.6.0",
|
"ndk 0.6.0",
|
||||||
"ndk-glue",
|
"ndk-glue",
|
||||||
|
@ -58,7 +58,7 @@ digest = "0.9.0"
|
|||||||
rtnetlink = { version = "^0", default-features = false, optional = true }
|
rtnetlink = { version = "^0", default-features = false, optional = true }
|
||||||
async-std-resolver = { version = "^0", optional = true }
|
async-std-resolver = { version = "^0", optional = true }
|
||||||
trust-dns-resolver = { version = "^0", optional = true }
|
trust-dns-resolver = { version = "^0", optional = true }
|
||||||
|
keyvaluedb = { path = "../external/keyvaluedb/keyvaluedb" }
|
||||||
|
|
||||||
# Dependencies for native builds only
|
# Dependencies for native builds only
|
||||||
# Linux, Windows, Mac, iOS, Android
|
# Linux, Windows, Mac, iOS, Android
|
||||||
@ -72,7 +72,6 @@ async-tungstenite = { version = "^0", features = ["async-tls"] }
|
|||||||
maplit = "^1"
|
maplit = "^1"
|
||||||
config = { version = "^0", features = ["yaml"] }
|
config = { version = "^0", features = ["yaml"] }
|
||||||
keyring-manager = { path = "../external/keyring-manager" }
|
keyring-manager = { path = "../external/keyring-manager" }
|
||||||
lru = "^0"
|
|
||||||
async-tls = "^0.11"
|
async-tls = "^0.11"
|
||||||
igd = { path = "../external/rust-igd" }
|
igd = { path = "../external/rust-igd" }
|
||||||
webpki = "^0"
|
webpki = "^0"
|
||||||
@ -96,8 +95,6 @@ nix = "^0"
|
|||||||
wasm-bindgen = "^0"
|
wasm-bindgen = "^0"
|
||||||
js-sys = "^0"
|
js-sys = "^0"
|
||||||
wasm-bindgen-futures = "^0"
|
wasm-bindgen-futures = "^0"
|
||||||
hashbrown = "^0"
|
|
||||||
lru = {version = "^0", features = ["hashbrown"] }
|
|
||||||
no-std-net = { path = "../external/no-std-net", features = ["serde"] }
|
no-std-net = { path = "../external/no-std-net", features = ["serde"] }
|
||||||
keyvaluedb-web = { path = "../external/keyvaluedb/keyvaluedb-web" }
|
keyvaluedb-web = { path = "../external/keyvaluedb/keyvaluedb-web" }
|
||||||
data-encoding = { version = "^2", default_features = false, features = ["alloc"] }
|
data-encoding = { version = "^2", default_features = false, features = ["alloc"] }
|
||||||
|
@ -119,12 +119,12 @@ impl Crypto {
|
|||||||
|
|
||||||
// load caches if they are valid for this node id
|
// load caches if they are valid for this node id
|
||||||
let mut db = table_store.open("crypto_caches", 1).await?;
|
let mut db = table_store.open("crypto_caches", 1).await?;
|
||||||
let caches_valid = match db.load(0, b"node_id").await? {
|
let caches_valid = match db.load(0, b"node_id")? {
|
||||||
Some(v) => v.as_slice() == node_id.bytes,
|
Some(v) => v.as_slice() == node_id.bytes,
|
||||||
None => false,
|
None => false,
|
||||||
};
|
};
|
||||||
if caches_valid {
|
if caches_valid {
|
||||||
if let Some(b) = db.load(0, b"dh_cache").await? {
|
if let Some(b) = db.load(0, b"dh_cache")? {
|
||||||
let mut inner = self.inner.lock();
|
let mut inner = self.inner.lock();
|
||||||
bytes_to_cache(&b, &mut inner.dh_cache);
|
bytes_to_cache(&b, &mut inner.dh_cache);
|
||||||
}
|
}
|
||||||
@ -132,7 +132,7 @@ impl Crypto {
|
|||||||
drop(db);
|
drop(db);
|
||||||
table_store.delete("crypto_caches").await?;
|
table_store.delete("crypto_caches").await?;
|
||||||
db = table_store.open("crypto_caches", 1).await?;
|
db = table_store.open("crypto_caches", 1).await?;
|
||||||
db.store(0, b"node_id", &node_id.bytes).await?;
|
db.store(0, b"node_id", &node_id.bytes)?;
|
||||||
}
|
}
|
||||||
|
|
||||||
// Schedule flushing
|
// Schedule flushing
|
||||||
@ -159,7 +159,7 @@ impl Crypto {
|
|||||||
};
|
};
|
||||||
|
|
||||||
let db = table_store.open("crypto_caches", 1).await?;
|
let db = table_store.open("crypto_caches", 1).await?;
|
||||||
db.store(0, b"dh_cache", &cache_bytes).await?;
|
db.store(0, b"dh_cache", &cache_bytes)?;
|
||||||
Ok(())
|
Ok(())
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -8,6 +8,8 @@ struct TableStoreInner {
|
|||||||
opened: BTreeMap<String, Weak<Mutex<TableDBInner>>>,
|
opened: BTreeMap<String, Weak<Mutex<TableDBInner>>>,
|
||||||
}
|
}
|
||||||
|
|
||||||
|
/// Veilid Table Storage
|
||||||
|
/// Database for storing key value pairs persistently across runs
|
||||||
#[derive(Clone)]
|
#[derive(Clone)]
|
||||||
pub struct TableStore {
|
pub struct TableStore {
|
||||||
config: VeilidConfig,
|
config: VeilidConfig,
|
||||||
@ -20,31 +22,38 @@ impl TableStore {
|
|||||||
opened: BTreeMap::new(),
|
opened: BTreeMap::new(),
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
pub fn new(config: VeilidConfig) -> Self {
|
pub(crate) fn new(config: VeilidConfig) -> Self {
|
||||||
Self {
|
Self {
|
||||||
config,
|
config,
|
||||||
inner: Arc::new(Mutex::new(Self::new_inner())),
|
inner: Arc::new(Mutex::new(Self::new_inner())),
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
pub async fn delete_all(&self) -> EyreResult<()> {
|
/// Delete all known tables
|
||||||
// Delete all known keys
|
pub async fn delete_all(&self) {
|
||||||
self.delete("crypto_caches").await?;
|
if let Err(e) = self.delete("crypto_caches").await {
|
||||||
|
error!("failed to delete 'crypto_caches': {}", e);
|
||||||
|
}
|
||||||
|
if let Err(e) = self.delete("RouteSpecStore").await {
|
||||||
|
error!("failed to delete 'RouteSpecStore': {}", e);
|
||||||
|
}
|
||||||
|
if let Err(e) = self.delete("routing_table").await {
|
||||||
|
error!("failed to delete 'routing_table': {}", e);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
pub(crate) async fn init(&self) -> EyreResult<()> {
|
||||||
Ok(())
|
Ok(())
|
||||||
}
|
}
|
||||||
|
|
||||||
pub async fn init(&self) -> EyreResult<()> {
|
pub(crate) async fn terminate(&self) {
|
||||||
Ok(())
|
|
||||||
}
|
|
||||||
|
|
||||||
pub async fn terminate(&self) {
|
|
||||||
assert!(
|
assert!(
|
||||||
self.inner.lock().opened.is_empty(),
|
self.inner.lock().opened.is_empty(),
|
||||||
"all open databases should have been closed"
|
"all open databases should have been closed"
|
||||||
);
|
);
|
||||||
}
|
}
|
||||||
|
|
||||||
pub fn on_table_db_drop(&self, table: String) {
|
pub(crate) fn on_table_db_drop(&self, table: String) {
|
||||||
let mut inner = self.inner.lock();
|
let mut inner = self.inner.lock();
|
||||||
if inner.opened.remove(&table).is_none() {
|
if inner.opened.remove(&table).is_none() {
|
||||||
unreachable!("should have removed an item");
|
unreachable!("should have removed an item");
|
||||||
@ -82,6 +91,8 @@ impl TableStore {
|
|||||||
})
|
})
|
||||||
}
|
}
|
||||||
|
|
||||||
|
/// Get or create a TableDB database table. If the column count is greater than an
|
||||||
|
/// existing TableDB's column count, the database will be upgraded to add the missing columns
|
||||||
pub async fn open(&self, name: &str, column_count: u32) -> EyreResult<TableDB> {
|
pub async fn open(&self, name: &str, column_count: u32) -> EyreResult<TableDB> {
|
||||||
let table_name = self.get_table_name(name)?;
|
let table_name = self.get_table_name(name)?;
|
||||||
|
|
||||||
@ -121,6 +132,7 @@ impl TableStore {
|
|||||||
Ok(table_db)
|
Ok(table_db)
|
||||||
}
|
}
|
||||||
|
|
||||||
|
/// Delete a TableDB table by name
|
||||||
pub async fn delete(&self, name: &str) -> EyreResult<bool> {
|
pub async fn delete(&self, name: &str) -> EyreResult<bool> {
|
||||||
let table_name = self.get_table_name(name)?;
|
let table_name = self.get_table_name(name)?;
|
||||||
|
|
||||||
|
@ -5,8 +5,10 @@ use serde::{Deserialize, Serialize};
|
|||||||
cfg_if! {
|
cfg_if! {
|
||||||
if #[cfg(target_arch = "wasm32")] {
|
if #[cfg(target_arch = "wasm32")] {
|
||||||
use keyvaluedb_web::*;
|
use keyvaluedb_web::*;
|
||||||
|
use keyvaluedb::*;
|
||||||
} else {
|
} else {
|
||||||
use keyvaluedb_sqlite::*;
|
use keyvaluedb_sqlite::*;
|
||||||
|
use keyvaluedb::*;
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -28,7 +30,7 @@ pub struct TableDB {
|
|||||||
}
|
}
|
||||||
|
|
||||||
impl TableDB {
|
impl TableDB {
|
||||||
pub fn new(table: String, table_store: TableStore, database: Database) -> Self {
|
pub(super) fn new(table: String, table_store: TableStore, database: Database) -> Self {
|
||||||
Self {
|
Self {
|
||||||
inner: Arc::new(Mutex::new(TableDBInner {
|
inner: Arc::new(Mutex::new(TableDBInner {
|
||||||
table,
|
table,
|
||||||
@ -38,22 +40,24 @@ impl TableDB {
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
pub fn try_new_from_weak_inner(weak_inner: Weak<Mutex<TableDBInner>>) -> Option<Self> {
|
pub(super) fn try_new_from_weak_inner(weak_inner: Weak<Mutex<TableDBInner>>) -> Option<Self> {
|
||||||
weak_inner.upgrade().map(|table_db_inner| Self {
|
weak_inner.upgrade().map(|table_db_inner| Self {
|
||||||
inner: table_db_inner,
|
inner: table_db_inner,
|
||||||
})
|
})
|
||||||
}
|
}
|
||||||
|
|
||||||
pub fn weak_inner(&self) -> Weak<Mutex<TableDBInner>> {
|
pub(super) fn weak_inner(&self) -> Weak<Mutex<TableDBInner>> {
|
||||||
Arc::downgrade(&self.inner)
|
Arc::downgrade(&self.inner)
|
||||||
}
|
}
|
||||||
|
|
||||||
pub async fn get_column_count(&self) -> EyreResult<u32> {
|
/// Get the total number of columns in the TableDB
|
||||||
|
pub fn get_column_count(&self) -> EyreResult<u32> {
|
||||||
let db = &self.inner.lock().database;
|
let db = &self.inner.lock().database;
|
||||||
db.num_columns().wrap_err("failed to get column count: {}")
|
db.num_columns().wrap_err("failed to get column count: {}")
|
||||||
}
|
}
|
||||||
|
|
||||||
pub async fn get_keys(&self, col: u32) -> EyreResult<Vec<Box<[u8]>>> {
|
/// Get the list of keys in a column of the TableDB
|
||||||
|
pub fn get_keys(&self, col: u32) -> EyreResult<Vec<Box<[u8]>>> {
|
||||||
let db = &self.inner.lock().database;
|
let db = &self.inner.lock().database;
|
||||||
let mut out: Vec<Box<[u8]>> = Vec::new();
|
let mut out: Vec<Box<[u8]>> = Vec::new();
|
||||||
db.iter(col, None, &mut |kv| {
|
db.iter(col, None, &mut |kv| {
|
||||||
@ -64,14 +68,25 @@ impl TableDB {
|
|||||||
Ok(out)
|
Ok(out)
|
||||||
}
|
}
|
||||||
|
|
||||||
pub async fn store(&self, col: u32, key: &[u8], value: &[u8]) -> EyreResult<()> {
|
/// Start a TableDB write transaction. The transaction object must be committed or rolled back before dropping.
|
||||||
|
pub fn transact<'a>(&'a self) -> TableDBTransaction<'a> {
|
||||||
|
let dbt = {
|
||||||
|
let db = &self.inner.lock().database;
|
||||||
|
db.transaction()
|
||||||
|
};
|
||||||
|
TableDBTransaction::new(self, dbt)
|
||||||
|
}
|
||||||
|
|
||||||
|
/// Store a key with a value in a column in the TableDB. Performs a single transaction immediately.
|
||||||
|
pub fn store(&self, col: u32, key: &[u8], value: &[u8]) -> EyreResult<()> {
|
||||||
let db = &self.inner.lock().database;
|
let db = &self.inner.lock().database;
|
||||||
let mut dbt = db.transaction();
|
let mut dbt = db.transaction();
|
||||||
dbt.put(col, key, value);
|
dbt.put(col, key, value);
|
||||||
db.write(dbt).wrap_err("failed to store key")
|
db.write(dbt).wrap_err("failed to store key")
|
||||||
}
|
}
|
||||||
|
|
||||||
pub async fn store_cbor<T>(&self, col: u32, key: &[u8], value: &T) -> EyreResult<()>
|
/// Store a key in CBOR format with a value in a column in the TableDB. Performs a single transaction immediately.
|
||||||
|
pub fn store_cbor<T>(&self, col: u32, key: &[u8], value: &T) -> EyreResult<()>
|
||||||
where
|
where
|
||||||
T: Serialize,
|
T: Serialize,
|
||||||
{
|
{
|
||||||
@ -83,12 +98,14 @@ impl TableDB {
|
|||||||
db.write(dbt).wrap_err("failed to store key")
|
db.write(dbt).wrap_err("failed to store key")
|
||||||
}
|
}
|
||||||
|
|
||||||
pub async fn load(&self, col: u32, key: &[u8]) -> EyreResult<Option<Vec<u8>>> {
|
/// Read a key from a column in the TableDB immediately.
|
||||||
|
pub fn load(&self, col: u32, key: &[u8]) -> EyreResult<Option<Vec<u8>>> {
|
||||||
let db = &self.inner.lock().database;
|
let db = &self.inner.lock().database;
|
||||||
db.get(col, key).wrap_err("failed to get key")
|
db.get(col, key).wrap_err("failed to get key")
|
||||||
}
|
}
|
||||||
|
|
||||||
pub async fn load_cbor<T>(&self, col: u32, key: &[u8]) -> EyreResult<Option<T>>
|
/// Read a key from a column in the TableDB immediately, in CBOR format.
|
||||||
|
pub fn load_cbor<T>(&self, col: u32, key: &[u8]) -> EyreResult<Option<T>>
|
||||||
where
|
where
|
||||||
T: for<'de> Deserialize<'de>,
|
T: for<'de> Deserialize<'de>,
|
||||||
{
|
{
|
||||||
@ -104,7 +121,8 @@ impl TableDB {
|
|||||||
Ok(Some(obj))
|
Ok(Some(obj))
|
||||||
}
|
}
|
||||||
|
|
||||||
pub async fn delete(&self, col: u32, key: &[u8]) -> EyreResult<bool> {
|
/// Delete key with from a column in the TableDB
|
||||||
|
pub fn delete(&self, col: u32, key: &[u8]) -> EyreResult<bool> {
|
||||||
let db = &self.inner.lock().database;
|
let db = &self.inner.lock().database;
|
||||||
let found = db.get(col, key).wrap_err("failed to get key")?;
|
let found = db.get(col, key).wrap_err("failed to get key")?;
|
||||||
match found {
|
match found {
|
||||||
@ -118,3 +136,66 @@ impl TableDB {
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
////////////////////////////////////////////////////////////////////////////////////////////////////////////////////////
|
||||||
|
|
||||||
|
/// A TableDB transaction
|
||||||
|
/// Atomically commits a group of writes or deletes to the TableDB
|
||||||
|
pub struct TableDBTransaction<'a> {
|
||||||
|
db: &'a TableDB,
|
||||||
|
dbt: Option<DBTransaction>,
|
||||||
|
_phantom: core::marker::PhantomData<&'a ()>,
|
||||||
|
}
|
||||||
|
|
||||||
|
impl<'a> TableDBTransaction<'a> {
|
||||||
|
fn new(db: &'a TableDB, dbt: DBTransaction) -> Self {
|
||||||
|
Self {
|
||||||
|
db,
|
||||||
|
dbt: Some(dbt),
|
||||||
|
_phantom: Default::default(),
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
/// Commit the transaction. Performs all actions atomically.
|
||||||
|
pub fn commit(mut self) -> EyreResult<()> {
|
||||||
|
self.db
|
||||||
|
.inner
|
||||||
|
.lock()
|
||||||
|
.database
|
||||||
|
.write(self.dbt.take().unwrap())
|
||||||
|
.wrap_err("commit failed")
|
||||||
|
}
|
||||||
|
|
||||||
|
/// Rollback the transaction. Does nothing to the TableDB.
|
||||||
|
pub fn rollback(mut self) {
|
||||||
|
self.dbt = None;
|
||||||
|
}
|
||||||
|
|
||||||
|
/// Store a key with a value in a column in the TableDB
|
||||||
|
pub fn store(&mut self, col: u32, key: &[u8], value: &[u8]) {
|
||||||
|
self.dbt.as_mut().unwrap().put(col, key, value);
|
||||||
|
}
|
||||||
|
|
||||||
|
/// Store a key in CBOR format with a value in a column in the TableDB
|
||||||
|
pub fn store_cbor<T>(&mut self, col: u32, key: &[u8], value: &T) -> EyreResult<()>
|
||||||
|
where
|
||||||
|
T: Serialize,
|
||||||
|
{
|
||||||
|
let v = serde_cbor::to_vec(value).wrap_err("couldn't store as CBOR")?;
|
||||||
|
self.dbt.as_mut().unwrap().put(col, key, v.as_slice());
|
||||||
|
Ok(())
|
||||||
|
}
|
||||||
|
|
||||||
|
/// Delete key with from a column in the TableDB
|
||||||
|
pub fn delete(&mut self, col: u32, key: &[u8]) {
|
||||||
|
self.dbt.as_mut().unwrap().delete(col, key);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
impl<'a> Drop for TableDBTransaction<'a> {
|
||||||
|
fn drop(&mut self) {
|
||||||
|
if self.dbt.is_some() {
|
||||||
|
warn!("Dropped transaction without commit or rollback");
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
@ -3,20 +3,11 @@ use crate::xx::*;
|
|||||||
use crate::*;
|
use crate::*;
|
||||||
use data_encoding::BASE64URL_NOPAD;
|
use data_encoding::BASE64URL_NOPAD;
|
||||||
use js_sys::*;
|
use js_sys::*;
|
||||||
|
use send_wrapper::*;
|
||||||
|
use serde::{Deserialize, Serialize};
|
||||||
use wasm_bindgen_futures::*;
|
use wasm_bindgen_futures::*;
|
||||||
use web_sys::*;
|
use web_sys::*;
|
||||||
|
|
||||||
#[wasm_bindgen]
|
|
||||||
extern "C" {
|
|
||||||
#[wasm_bindgen(catch, js_name = setPassword, js_namespace = ["global", "wasmhost", "keytar"])]
|
|
||||||
fn keytar_setPassword(service: &str, account: &str, password: &str)
|
|
||||||
-> Result<Promise, JsValue>;
|
|
||||||
#[wasm_bindgen(catch, js_name = getPassword, js_namespace = ["global", "wasmhost", "keytar"])]
|
|
||||||
fn keytar_getPassword(service: &str, account: &str) -> Result<Promise, JsValue>;
|
|
||||||
#[wasm_bindgen(catch, js_name = deletePassword, js_namespace = ["global", "wasmhost", "keytar"])]
|
|
||||||
fn keytar_deletePassword(service: &str, account: &str) -> Result<Promise, JsValue>;
|
|
||||||
}
|
|
||||||
|
|
||||||
#[derive(Clone)]
|
#[derive(Clone)]
|
||||||
pub struct ProtectedStore {
|
pub struct ProtectedStore {
|
||||||
config: VeilidConfig,
|
config: VeilidConfig,
|
||||||
@ -71,33 +62,9 @@ impl ProtectedStore {
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
#[instrument(level = "trace", skip(self, value), ret, err)]
|
//#[instrument(level = "trace", skip(self, value), ret, err)]
|
||||||
pub async fn save_user_secret_string(&self, key: &str, value: &str) -> EyreResult<bool> {
|
pub async fn save_user_secret_string(&self, key: &str, value: &str) -> EyreResult<bool> {
|
||||||
if is_nodejs() {
|
if is_browser() {
|
||||||
let prev = match JsFuture::from(
|
|
||||||
keytar_getPassword(self.keyring_name().as_str(), key)
|
|
||||||
.map_err(map_jsvalue_error)
|
|
||||||
.wrap_err("exception thrown")?,
|
|
||||||
)
|
|
||||||
.await
|
|
||||||
{
|
|
||||||
Ok(v) => v.is_truthy(),
|
|
||||||
Err(_) => false,
|
|
||||||
};
|
|
||||||
|
|
||||||
match JsFuture::from(
|
|
||||||
keytar_setPassword(self.keyring_name().as_str(), key, value)
|
|
||||||
.map_err(map_jsvalue_error)
|
|
||||||
.wrap_err("exception thrown")?,
|
|
||||||
)
|
|
||||||
.await
|
|
||||||
{
|
|
||||||
Ok(_) => {}
|
|
||||||
Err(_) => bail!("Failed to set password"),
|
|
||||||
}
|
|
||||||
|
|
||||||
Ok(prev)
|
|
||||||
} else if is_browser() {
|
|
||||||
let win = match window() {
|
let win = match window() {
|
||||||
Some(w) => w,
|
Some(w) => w,
|
||||||
None => {
|
None => {
|
||||||
@ -139,24 +106,7 @@ impl ProtectedStore {
|
|||||||
|
|
||||||
#[instrument(level = "trace", skip(self), err)]
|
#[instrument(level = "trace", skip(self), err)]
|
||||||
pub async fn load_user_secret_string(&self, key: &str) -> EyreResult<Option<String>> {
|
pub async fn load_user_secret_string(&self, key: &str) -> EyreResult<Option<String>> {
|
||||||
if is_nodejs() {
|
if is_browser() {
|
||||||
let prev = match JsFuture::from(
|
|
||||||
keytar_getPassword(self.keyring_name().as_str(), key)
|
|
||||||
.map_err(map_jsvalue_error)
|
|
||||||
.wrap_err("exception thrown")?,
|
|
||||||
)
|
|
||||||
.await
|
|
||||||
{
|
|
||||||
Ok(p) => p,
|
|
||||||
Err(_) => JsValue::UNDEFINED,
|
|
||||||
};
|
|
||||||
|
|
||||||
if prev.is_undefined() || prev.is_null() {
|
|
||||||
return Ok(None);
|
|
||||||
}
|
|
||||||
|
|
||||||
Ok(prev.as_string())
|
|
||||||
} else if is_browser() {
|
|
||||||
let win = match window() {
|
let win = match window() {
|
||||||
Some(w) => w,
|
Some(w) => w,
|
||||||
None => {
|
None => {
|
||||||
@ -252,18 +202,7 @@ impl ProtectedStore {
|
|||||||
|
|
||||||
#[instrument(level = "trace", skip(self), ret, err)]
|
#[instrument(level = "trace", skip(self), ret, err)]
|
||||||
pub async fn remove_user_secret(&self, key: &str) -> EyreResult<bool> {
|
pub async fn remove_user_secret(&self, key: &str) -> EyreResult<bool> {
|
||||||
if is_nodejs() {
|
if is_browser() {
|
||||||
match JsFuture::from(
|
|
||||||
keytar_deletePassword(self.keyring_name().as_str(), key)
|
|
||||||
.map_err(map_jsvalue_error)
|
|
||||||
.wrap_err("exception thrown")?,
|
|
||||||
)
|
|
||||||
.await
|
|
||||||
{
|
|
||||||
Ok(v) => Ok(v.is_truthy()),
|
|
||||||
Err(_) => bail!("Failed to delete"),
|
|
||||||
}
|
|
||||||
} else if is_browser() {
|
|
||||||
let win = match window() {
|
let win = match window() {
|
||||||
Some(w) => w,
|
Some(w) => w,
|
||||||
None => {
|
None => {
|
||||||
|
@ -19,10 +19,8 @@ extern "C" {
|
|||||||
pub fn get_timestamp() -> u64 {
|
pub fn get_timestamp() -> u64 {
|
||||||
if utils::is_browser() {
|
if utils::is_browser() {
|
||||||
return (Date::now() * 1000.0f64) as u64;
|
return (Date::now() * 1000.0f64) as u64;
|
||||||
} else if utils::is_nodejs() {
|
|
||||||
return (Date::now() * 1000.0f64) as u64;
|
|
||||||
} else {
|
} else {
|
||||||
panic!("WASM requires browser or nodejs environment");
|
panic!("WASM requires browser environment");
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -85,18 +83,22 @@ pub fn spawn<Out>(future: impl Future<Output = Out> + Send + 'static) -> MustJoi
|
|||||||
where
|
where
|
||||||
Out: Send + 'static,
|
Out: Send + 'static,
|
||||||
{
|
{
|
||||||
MustJoinHandle::new(Bindgen
|
MustJoinHandle::new(
|
||||||
.spawn_handle(future)
|
Bindgen
|
||||||
.expect("wasm-bindgen-futures spawn should never error out"))
|
.spawn_handle(future)
|
||||||
|
.expect("wasm-bindgen-futures spawn should never error out"),
|
||||||
|
)
|
||||||
}
|
}
|
||||||
|
|
||||||
pub fn spawn_local<Out>(future: impl Future<Output = Out> + 'static) -> MustJoinHandle<Out>
|
pub fn spawn_local<Out>(future: impl Future<Output = Out> + 'static) -> MustJoinHandle<Out>
|
||||||
where
|
where
|
||||||
Out: 'static,
|
Out: 'static,
|
||||||
{
|
{
|
||||||
MustJoinHandle::new(Bindgen
|
MustJoinHandle::new(
|
||||||
.spawn_handle_local(future)
|
Bindgen
|
||||||
.expect("wasm-bindgen-futures spawn_local should never error out"))
|
.spawn_handle_local(future)
|
||||||
|
.expect("wasm-bindgen-futures spawn_local should never error out"),
|
||||||
|
)
|
||||||
}
|
}
|
||||||
|
|
||||||
// pub fn spawn_with_local_set<Out>(
|
// pub fn spawn_with_local_set<Out>(
|
||||||
@ -114,10 +116,10 @@ where
|
|||||||
{
|
{
|
||||||
Bindgen
|
Bindgen
|
||||||
.spawn_handle_local(future)
|
.spawn_handle_local(future)
|
||||||
.expect("wasm-bindgen-futures spawn_local should never error out").detach()
|
.expect("wasm-bindgen-futures spawn_local should never error out")
|
||||||
|
.detach()
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
||||||
pub fn interval<F, FUT>(freq_ms: u32, callback: F) -> SendPinBoxFuture<()>
|
pub fn interval<F, FUT>(freq_ms: u32, callback: F) -> SendPinBoxFuture<()>
|
||||||
where
|
where
|
||||||
F: Fn() -> FUT + Send + Sync + 'static,
|
F: Fn() -> FUT + Send + Sync + 'static,
|
||||||
@ -160,12 +162,12 @@ pub async fn get_outbound_relay_peer() -> Option<crate::veilid_api::PeerInfo> {
|
|||||||
|
|
||||||
// pub async fn get_pwa_web_server_config() -> {
|
// pub async fn get_pwa_web_server_config() -> {
|
||||||
// if utils::is_browser() {
|
// if utils::is_browser() {
|
||||||
|
|
||||||
// let win = window().unwrap();
|
// let win = window().unwrap();
|
||||||
// let doc = win.document().unwrap();
|
// let doc = win.document().unwrap();
|
||||||
// let html_document = document.dyn_into::<web_sys::HtmlDocument>().unwrap();
|
// let html_document = document.dyn_into::<web_sys::HtmlDocument>().unwrap();
|
||||||
// let cookie = html_document.cookie().unwrap();
|
// let cookie = html_document.cookie().unwrap();
|
||||||
|
|
||||||
// // let wait_millis = if millis > u32::MAX {
|
// // let wait_millis = if millis > u32::MAX {
|
||||||
// // i32::MAX
|
// // i32::MAX
|
||||||
// // } else {
|
// // } else {
|
||||||
@ -177,22 +179,14 @@ pub async fn get_outbound_relay_peer() -> Option<crate::veilid_api::PeerInfo> {
|
|||||||
// // .unwrap();
|
// // .unwrap();
|
||||||
// // });
|
// // });
|
||||||
|
|
||||||
// // JsFuture::from(promise).await.unwrap();
|
|
||||||
// } else if utils::is_nodejs() {
|
|
||||||
// // let promise = Promise::new(&mut |yes, _| {
|
|
||||||
// // nodejs_global_set_timeout_with_callback_and_timeout_and_arguments_0(&yes, millis)
|
|
||||||
// // .unwrap();
|
|
||||||
// // });
|
|
||||||
|
|
||||||
// // JsFuture::from(promise).await.unwrap();
|
// // JsFuture::from(promise).await.unwrap();
|
||||||
// } else {
|
// } else {
|
||||||
// panic!("WASM requires browser or nodejs environment");
|
// panic!("WASM requires browser environment");
|
||||||
// }
|
// }
|
||||||
// }
|
// }
|
||||||
|
|
||||||
|
|
||||||
pub async fn txt_lookup<S: AsRef<str>>(_host: S) -> EyreResult<Vec<String>> {
|
pub async fn txt_lookup<S: AsRef<str>>(_host: S) -> EyreResult<Vec<String>> {
|
||||||
bail!("wasm does not support txt lookup")
|
bail!("wasm does not support txt lookup")
|
||||||
}
|
}
|
||||||
|
|
||||||
pub async fn ptr_lookup(_ip_addr: IpAddr) -> EyreResult<String> {
|
pub async fn ptr_lookup(_ip_addr: IpAddr) -> EyreResult<String> {
|
||||||
|
@ -22,7 +22,7 @@ impl TableStore {
|
|||||||
opened: BTreeMap::new(),
|
opened: BTreeMap::new(),
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
pub fn new(config: VeilidConfig) -> Self {
|
pub(crate) fn new(config: VeilidConfig) -> Self {
|
||||||
Self {
|
Self {
|
||||||
config,
|
config,
|
||||||
inner: Arc::new(Mutex::new(Self::new_inner())),
|
inner: Arc::new(Mutex::new(Self::new_inner())),
|
||||||
@ -30,12 +30,25 @@ impl TableStore {
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
pub async fn init(&self) -> EyreResult<()> {
|
/// Delete all known tables
|
||||||
|
pub async fn delete_all(&self) {
|
||||||
|
if let Err(e) = self.delete("crypto_caches").await {
|
||||||
|
error!("failed to delete 'crypto_caches': {}", e);
|
||||||
|
}
|
||||||
|
if let Err(e) = self.delete("RouteSpecStore").await {
|
||||||
|
error!("failed to delete 'RouteSpecStore': {}", e);
|
||||||
|
}
|
||||||
|
if let Err(e) = self.delete("routing_table").await {
|
||||||
|
error!("failed to delete 'routing_table': {}", e);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
pub(crate) async fn init(&self) -> EyreResult<()> {
|
||||||
let _async_guard = self.async_lock.lock().await;
|
let _async_guard = self.async_lock.lock().await;
|
||||||
Ok(())
|
Ok(())
|
||||||
}
|
}
|
||||||
|
|
||||||
pub async fn terminate(&self) {
|
pub(crate) async fn terminate(&self) {
|
||||||
let _async_guard = self.async_lock.lock().await;
|
let _async_guard = self.async_lock.lock().await;
|
||||||
assert!(
|
assert!(
|
||||||
self.inner.lock().opened.len() == 0,
|
self.inner.lock().opened.len() == 0,
|
||||||
@ -43,7 +56,7 @@ impl TableStore {
|
|||||||
);
|
);
|
||||||
}
|
}
|
||||||
|
|
||||||
pub fn on_table_db_drop(&self, table: String) {
|
pub(crate) fn on_table_db_drop(&self, table: String) {
|
||||||
let mut inner = self.inner.lock();
|
let mut inner = self.inner.lock();
|
||||||
match inner.opened.remove(&table) {
|
match inner.opened.remove(&table) {
|
||||||
Some(_) => (),
|
Some(_) => (),
|
||||||
@ -69,12 +82,14 @@ impl TableStore {
|
|||||||
})
|
})
|
||||||
}
|
}
|
||||||
|
|
||||||
|
/// Get or create a TableDB database table. If the column count is greater than an
|
||||||
|
/// existing TableDB's column count, the database will be upgraded to add the missing columns
|
||||||
pub async fn open(&self, name: &str, column_count: u32) -> EyreResult<TableDB> {
|
pub async fn open(&self, name: &str, column_count: u32) -> EyreResult<TableDB> {
|
||||||
let _async_guard = self.async_lock.lock().await;
|
let _async_guard = self.async_lock.lock().await;
|
||||||
let table_name = self.get_table_name(name)?;
|
let table_name = self.get_table_name(name)?;
|
||||||
|
|
||||||
{
|
{
|
||||||
let mut inner = self.inner.lock();
|
let mut inner = self.inner.lock();
|
||||||
if let Some(table_db_weak_inner) = inner.opened.get(&table_name) {
|
if let Some(table_db_weak_inner) = inner.opened.get(&table_name) {
|
||||||
match TableDB::try_new_from_weak_inner(table_db_weak_inner.clone()) {
|
match TableDB::try_new_from_weak_inner(table_db_weak_inner.clone()) {
|
||||||
Some(tdb) => {
|
Some(tdb) => {
|
||||||
@ -89,7 +104,10 @@ impl TableStore {
|
|||||||
let db = Database::open(table_name.clone(), column_count)
|
let db = Database::open(table_name.clone(), column_count)
|
||||||
.await
|
.await
|
||||||
.wrap_err("failed to open tabledb")?;
|
.wrap_err("failed to open tabledb")?;
|
||||||
info!("opened table store '{}' with table name '{:?}' with {} columns", name, table_name, column_count);
|
info!(
|
||||||
|
"opened table store '{}' with table name '{:?}' with {} columns",
|
||||||
|
name, table_name, column_count
|
||||||
|
);
|
||||||
|
|
||||||
let table_db = TableDB::new(table_name.clone(), self.clone(), db);
|
let table_db = TableDB::new(table_name.clone(), self.clone(), db);
|
||||||
|
|
||||||
@ -101,11 +119,12 @@ impl TableStore {
|
|||||||
Ok(table_db)
|
Ok(table_db)
|
||||||
}
|
}
|
||||||
|
|
||||||
|
/// Delete a TableDB table by name
|
||||||
pub async fn delete(&self, name: &str) -> EyreResult<bool> {
|
pub async fn delete(&self, name: &str) -> EyreResult<bool> {
|
||||||
let _async_guard = self.async_lock.lock().await;
|
let _async_guard = self.async_lock.lock().await;
|
||||||
trace!("TableStore::delete {}", name);
|
trace!("TableStore::delete {}", name);
|
||||||
let table_name = self.get_table_name(name)?;
|
let table_name = self.get_table_name(name)?;
|
||||||
|
|
||||||
{
|
{
|
||||||
let inner = self.inner.lock();
|
let inner = self.inner.lock();
|
||||||
if inner.opened.contains_key(&table_name) {
|
if inner.opened.contains_key(&table_name) {
|
||||||
@ -117,9 +136,7 @@ impl TableStore {
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
if utils::is_nodejs() {
|
if utils::is_browser() {
|
||||||
unimplemented!();
|
|
||||||
} else if utils::is_browser() {
|
|
||||||
let out = match Database::delete(table_name.clone()).await {
|
let out = match Database::delete(table_name.clone()).await {
|
||||||
Ok(_) => true,
|
Ok(_) => true,
|
||||||
Err(_) => false,
|
Err(_) => false,
|
||||||
|
@ -15,21 +15,6 @@ extern "C" {
|
|||||||
pub fn alert(s: &str);
|
pub fn alert(s: &str);
|
||||||
}
|
}
|
||||||
|
|
||||||
pub fn is_nodejs() -> bool {
|
|
||||||
static CACHE: AtomicI8 = AtomicI8::new(-1);
|
|
||||||
let cache = CACHE.load(Ordering::Relaxed);
|
|
||||||
if cache != -1 {
|
|
||||||
return cache != 0;
|
|
||||||
}
|
|
||||||
|
|
||||||
let res = js_sys::eval("process.release.name === 'node'")
|
|
||||||
.map(|res| res.is_truthy())
|
|
||||||
.unwrap_or_default();
|
|
||||||
|
|
||||||
CACHE.store(res as i8, Ordering::Relaxed);
|
|
||||||
res
|
|
||||||
}
|
|
||||||
|
|
||||||
pub fn is_browser() -> bool {
|
pub fn is_browser() -> bool {
|
||||||
static CACHE: AtomicI8 = AtomicI8::new(-1);
|
static CACHE: AtomicI8 = AtomicI8::new(-1);
|
||||||
let cache = CACHE.load(Ordering::Relaxed);
|
let cache = CACHE.load(Ordering::Relaxed);
|
||||||
@ -60,24 +45,6 @@ pub fn is_browser() -> bool {
|
|||||||
// res
|
// res
|
||||||
// }
|
// }
|
||||||
|
|
||||||
// pub fn node_require(module: &str) -> JsValue {
|
|
||||||
// if !is_nodejs() {
|
|
||||||
// return JsValue::UNDEFINED;
|
|
||||||
// }
|
|
||||||
|
|
||||||
// let mut home = env!("CARGO_MANIFEST_DIR");
|
|
||||||
// if home.len() == 0 {
|
|
||||||
// home = ".";
|
|
||||||
// }
|
|
||||||
|
|
||||||
// match js_sys::eval(format!("require(\"{}/{}\")", home, module).as_str()) {
|
|
||||||
// Ok(v) => v,
|
|
||||||
// Err(e) => {
|
|
||||||
// panic!("node_require failed: {:?}", e);
|
|
||||||
// }
|
|
||||||
// }
|
|
||||||
// }
|
|
||||||
|
|
||||||
#[derive(ThisError, Debug, Clone, Eq, PartialEq)]
|
#[derive(ThisError, Debug, Clone, Eq, PartialEq)]
|
||||||
#[error("JsValue error")]
|
#[error("JsValue error")]
|
||||||
pub struct JsValueError(String);
|
pub struct JsValueError(String);
|
||||||
|
@ -8,6 +8,8 @@ pub struct Bucket {
|
|||||||
}
|
}
|
||||||
pub(super) type EntriesIter<'a> = alloc::collections::btree_map::Iter<'a, DHTKey, Arc<BucketEntry>>;
|
pub(super) type EntriesIter<'a> = alloc::collections::btree_map::Iter<'a, DHTKey, Arc<BucketEntry>>;
|
||||||
|
|
||||||
|
type BucketData = (Vec<(DHTKey, Vec<u8>)>, Option<DHTKey>);
|
||||||
|
|
||||||
fn state_ordering(state: BucketEntryState) -> usize {
|
fn state_ordering(state: BucketEntryState) -> usize {
|
||||||
match state {
|
match state {
|
||||||
BucketEntryState::Dead => 0,
|
BucketEntryState::Dead => 0,
|
||||||
@ -25,6 +27,32 @@ impl Bucket {
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
pub(super) fn load_bucket(&mut self, data: &[u8]) -> EyreResult<()> {
|
||||||
|
let bucket_data: BucketData =
|
||||||
|
serde_cbor::from_slice::<BucketData>(data).wrap_err("failed to deserialize bucket")?;
|
||||||
|
|
||||||
|
for (k, d) in bucket_data.0 {
|
||||||
|
let entryinner = serde_cbor::from_slice::<BucketEntryInner>(&d)
|
||||||
|
.wrap_err("failed to deserialize bucket entry")?;
|
||||||
|
self.entries
|
||||||
|
.insert(k, Arc::new(BucketEntry::new_with_inner(entryinner)));
|
||||||
|
}
|
||||||
|
|
||||||
|
self.newest_entry = bucket_data.1;
|
||||||
|
|
||||||
|
Ok(())
|
||||||
|
}
|
||||||
|
pub(super) fn save_bucket(&self) -> EyreResult<Vec<u8>> {
|
||||||
|
let mut entry_vec = Vec::new();
|
||||||
|
for (k, v) in &self.entries {
|
||||||
|
let entry_bytes = v.with_mut_inner(|e| serde_cbor::to_vec(e))?;
|
||||||
|
entry_vec.push((*k, entry_bytes));
|
||||||
|
}
|
||||||
|
let bucket_data: BucketData = (entry_vec, self.newest_entry.clone());
|
||||||
|
let out = serde_cbor::to_vec(&bucket_data)?;
|
||||||
|
Ok(out)
|
||||||
|
}
|
||||||
|
|
||||||
pub(super) fn add_entry(&mut self, node_id: DHTKey) -> NodeRef {
|
pub(super) fn add_entry(&mut self, node_id: DHTKey) -> NodeRef {
|
||||||
log_rtab!("Node added: {}", node_id.encode());
|
log_rtab!("Node added: {}", node_id.encode());
|
||||||
|
|
||||||
|
@ -1,5 +1,6 @@
|
|||||||
use super::*;
|
use super::*;
|
||||||
use core::sync::atomic::{AtomicU32, Ordering};
|
use core::sync::atomic::{AtomicU32, Ordering};
|
||||||
|
use serde::{Deserialize, Serialize};
|
||||||
|
|
||||||
/// Reliable pings are done with increased spacing between pings
|
/// Reliable pings are done with increased spacing between pings
|
||||||
|
|
||||||
@ -42,7 +43,7 @@ pub enum BucketEntryState {
|
|||||||
struct LastConnectionKey(ProtocolType, AddressType);
|
struct LastConnectionKey(ProtocolType, AddressType);
|
||||||
|
|
||||||
/// Bucket entry information specific to the LocalNetwork RoutingDomain
|
/// Bucket entry information specific to the LocalNetwork RoutingDomain
|
||||||
#[derive(Debug)]
|
#[derive(Debug, Serialize, Deserialize)]
|
||||||
pub struct BucketEntryPublicInternet {
|
pub struct BucketEntryPublicInternet {
|
||||||
/// The PublicInternet node info
|
/// The PublicInternet node info
|
||||||
signed_node_info: Option<Box<SignedNodeInfo>>,
|
signed_node_info: Option<Box<SignedNodeInfo>>,
|
||||||
@ -53,7 +54,7 @@ pub struct BucketEntryPublicInternet {
|
|||||||
}
|
}
|
||||||
|
|
||||||
/// Bucket entry information specific to the LocalNetwork RoutingDomain
|
/// Bucket entry information specific to the LocalNetwork RoutingDomain
|
||||||
#[derive(Debug)]
|
#[derive(Debug, Serialize, Deserialize)]
|
||||||
pub struct BucketEntryLocalNetwork {
|
pub struct BucketEntryLocalNetwork {
|
||||||
/// The LocalNetwork node info
|
/// The LocalNetwork node info
|
||||||
signed_node_info: Option<Box<SignedNodeInfo>>,
|
signed_node_info: Option<Box<SignedNodeInfo>>,
|
||||||
@ -63,19 +64,24 @@ pub struct BucketEntryLocalNetwork {
|
|||||||
node_status: Option<LocalNetworkNodeStatus>,
|
node_status: Option<LocalNetworkNodeStatus>,
|
||||||
}
|
}
|
||||||
|
|
||||||
#[derive(Debug)]
|
#[derive(Debug, Serialize, Deserialize)]
|
||||||
pub struct BucketEntryInner {
|
pub struct BucketEntryInner {
|
||||||
min_max_version: Option<(u8, u8)>,
|
min_max_version: Option<(u8, u8)>,
|
||||||
updated_since_last_network_change: bool,
|
updated_since_last_network_change: bool,
|
||||||
|
#[serde(skip)]
|
||||||
last_connections: BTreeMap<LastConnectionKey, (ConnectionDescriptor, u64)>,
|
last_connections: BTreeMap<LastConnectionKey, (ConnectionDescriptor, u64)>,
|
||||||
public_internet: BucketEntryPublicInternet,
|
public_internet: BucketEntryPublicInternet,
|
||||||
local_network: BucketEntryLocalNetwork,
|
local_network: BucketEntryLocalNetwork,
|
||||||
peer_stats: PeerStats,
|
peer_stats: PeerStats,
|
||||||
|
#[serde(skip)]
|
||||||
latency_stats_accounting: LatencyStatsAccounting,
|
latency_stats_accounting: LatencyStatsAccounting,
|
||||||
|
#[serde(skip)]
|
||||||
transfer_stats_accounting: TransferStatsAccounting,
|
transfer_stats_accounting: TransferStatsAccounting,
|
||||||
#[cfg(feature = "tracking")]
|
#[cfg(feature = "tracking")]
|
||||||
|
#[serde(skip)]
|
||||||
next_track_id: usize,
|
next_track_id: usize,
|
||||||
#[cfg(feature = "tracking")]
|
#[cfg(feature = "tracking")]
|
||||||
|
#[serde(skip)]
|
||||||
node_ref_tracks: HashMap<usize, backtrace::Backtrace>,
|
node_ref_tracks: HashMap<usize, backtrace::Backtrace>,
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -657,6 +663,13 @@ impl BucketEntry {
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
pub(super) fn new_with_inner(inner: BucketEntryInner) -> Self {
|
||||||
|
Self {
|
||||||
|
ref_count: AtomicU32::new(0),
|
||||||
|
inner: RwLock::new(inner),
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
// Note, that this requires -also- holding the RoutingTable read lock, as an
|
// Note, that this requires -also- holding the RoutingTable read lock, as an
|
||||||
// immutable reference to RoutingTableInner must be passed in to get this
|
// immutable reference to RoutingTableInner must be passed in to get this
|
||||||
// This ensures that an operation on the routing table can not change entries
|
// This ensures that an operation on the routing table can not change entries
|
||||||
|
@ -154,6 +154,20 @@ impl RoutingTable {
|
|||||||
pub async fn init(&self) -> EyreResult<()> {
|
pub async fn init(&self) -> EyreResult<()> {
|
||||||
debug!("starting routing table init");
|
debug!("starting routing table init");
|
||||||
|
|
||||||
|
// Set up routing buckets
|
||||||
|
{
|
||||||
|
let mut inner = self.inner.write();
|
||||||
|
inner.init_buckets(self.clone());
|
||||||
|
}
|
||||||
|
|
||||||
|
// Load bucket entries from table db if possible
|
||||||
|
debug!("loading routing table entries");
|
||||||
|
if let Err(e) = self.load_buckets().await {
|
||||||
|
log_rtab!(warn "Error loading buckets from storage: {}. Resetting.", e);
|
||||||
|
let mut inner = self.inner.write();
|
||||||
|
inner.init_buckets(self.clone());
|
||||||
|
}
|
||||||
|
|
||||||
// Set up routespecstore
|
// Set up routespecstore
|
||||||
debug!("starting route spec store init");
|
debug!("starting route spec store init");
|
||||||
let route_spec_store = match RouteSpecStore::load(self.clone()).await {
|
let route_spec_store = match RouteSpecStore::load(self.clone()).await {
|
||||||
@ -165,10 +179,10 @@ impl RoutingTable {
|
|||||||
};
|
};
|
||||||
debug!("finished route spec store init");
|
debug!("finished route spec store init");
|
||||||
|
|
||||||
let mut inner = self.inner.write();
|
{
|
||||||
inner.init(self.clone())?;
|
let mut inner = self.inner.write();
|
||||||
|
inner.route_spec_store = Some(route_spec_store);
|
||||||
inner.route_spec_store = Some(route_spec_store);
|
}
|
||||||
|
|
||||||
debug!("finished routing table init");
|
debug!("finished routing table init");
|
||||||
Ok(())
|
Ok(())
|
||||||
@ -188,6 +202,12 @@ impl RoutingTable {
|
|||||||
error!("kick_buckets_task not stopped: {}", e);
|
error!("kick_buckets_task not stopped: {}", e);
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// Load bucket entries from table db if possible
|
||||||
|
debug!("saving routing table entries");
|
||||||
|
if let Err(e) = self.save_buckets().await {
|
||||||
|
error!("failed to save routing table entries: {}", e);
|
||||||
|
}
|
||||||
|
|
||||||
debug!("saving route spec store");
|
debug!("saving route spec store");
|
||||||
let rss = {
|
let rss = {
|
||||||
let mut inner = self.inner.write();
|
let mut inner = self.inner.write();
|
||||||
@ -201,12 +221,67 @@ impl RoutingTable {
|
|||||||
debug!("shutting down routing table");
|
debug!("shutting down routing table");
|
||||||
|
|
||||||
let mut inner = self.inner.write();
|
let mut inner = self.inner.write();
|
||||||
inner.terminate();
|
|
||||||
*inner = RoutingTableInner::new(self.unlocked_inner.clone());
|
*inner = RoutingTableInner::new(self.unlocked_inner.clone());
|
||||||
|
|
||||||
debug!("finished routing table terminate");
|
debug!("finished routing table terminate");
|
||||||
}
|
}
|
||||||
|
|
||||||
|
async fn save_buckets(&self) -> EyreResult<()> {
|
||||||
|
// Serialize all entries
|
||||||
|
let mut bucketvec: Vec<Vec<u8>> = Vec::new();
|
||||||
|
{
|
||||||
|
let inner = &*self.inner.read();
|
||||||
|
for bucket in &inner.buckets {
|
||||||
|
bucketvec.push(bucket.save_bucket()?)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
let table_store = self.network_manager().table_store();
|
||||||
|
let tdb = table_store.open("routing_table", 1).await?;
|
||||||
|
let bucket_count = bucketvec.len();
|
||||||
|
let mut dbx = tdb.transact();
|
||||||
|
if let Err(e) = dbx.store_cbor(0, b"bucket_count", &bucket_count) {
|
||||||
|
dbx.rollback();
|
||||||
|
return Err(e);
|
||||||
|
}
|
||||||
|
|
||||||
|
for (n, b) in bucketvec.iter().enumerate() {
|
||||||
|
dbx.store(0, format!("bucket_{}", n).as_bytes(), b)
|
||||||
|
}
|
||||||
|
dbx.commit()?;
|
||||||
|
Ok(())
|
||||||
|
}
|
||||||
|
|
||||||
|
async fn load_buckets(&self) -> EyreResult<()> {
|
||||||
|
// Deserialize all entries
|
||||||
|
let inner = &mut *self.inner.write();
|
||||||
|
|
||||||
|
let tstore = self.network_manager().table_store();
|
||||||
|
let tdb = tstore.open("routing_table", 1).await?;
|
||||||
|
let Some(bucket_count): Option<usize> = tdb.load_cbor(0, b"bucket_count")? else {
|
||||||
|
log_rtab!(debug "no bucket count in saved routing table");
|
||||||
|
return Ok(());
|
||||||
|
};
|
||||||
|
if bucket_count != inner.buckets.len() {
|
||||||
|
// Must have the same number of buckets
|
||||||
|
warn!("bucket count is different, not loading routing table");
|
||||||
|
return Ok(());
|
||||||
|
}
|
||||||
|
let mut bucketdata_vec: Vec<Vec<u8>> = Vec::new();
|
||||||
|
for n in 0..bucket_count {
|
||||||
|
let Some(bucketdata): Option<Vec<u8>> =
|
||||||
|
tdb.load(0, format!("bucket_{}", n).as_bytes())? else {
|
||||||
|
warn!("bucket data not loading, skipping loading routing table");
|
||||||
|
return Ok(());
|
||||||
|
};
|
||||||
|
bucketdata_vec.push(bucketdata);
|
||||||
|
}
|
||||||
|
for n in 0..bucket_count {
|
||||||
|
inner.buckets[n].load_bucket(&bucketdata_vec[n])?;
|
||||||
|
}
|
||||||
|
|
||||||
|
Ok(())
|
||||||
|
}
|
||||||
|
|
||||||
/// Set up the local network routing domain with our local routing table configuration
|
/// Set up the local network routing domain with our local routing table configuration
|
||||||
pub fn configure_local_network_routing_domain(&self, local_networks: Vec<(IpAddr, IpAddr)>) {
|
pub fn configure_local_network_routing_domain(&self, local_networks: Vec<(IpAddr, IpAddr)>) {
|
||||||
log_net!(debug "configure_local_network_routing_domain: {:#?}", local_networks);
|
log_net!(debug "configure_local_network_routing_domain: {:#?}", local_networks);
|
||||||
|
@ -225,7 +225,23 @@ impl RouteSpecStore {
|
|||||||
let table_store = routing_table.network_manager().table_store();
|
let table_store = routing_table.network_manager().table_store();
|
||||||
let rsstdb = table_store.open("RouteSpecStore", 1).await?;
|
let rsstdb = table_store.open("RouteSpecStore", 1).await?;
|
||||||
let mut content: RouteSpecStoreContent =
|
let mut content: RouteSpecStoreContent =
|
||||||
rsstdb.load_cbor(0, b"content").await?.unwrap_or_default();
|
rsstdb.load_cbor(0, b"content")?.unwrap_or_default();
|
||||||
|
|
||||||
|
// Look up all route hop noderefs since we can't serialize those
|
||||||
|
let mut dead_keys = Vec::new();
|
||||||
|
for (k, rsd) in &mut content.details {
|
||||||
|
for h in &rsd.hops {
|
||||||
|
let Some(nr) = routing_table.lookup_node_ref(*h) else {
|
||||||
|
dead_keys.push(*k);
|
||||||
|
break;
|
||||||
|
};
|
||||||
|
rsd.hop_node_refs.push(nr);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
for k in dead_keys {
|
||||||
|
log_rtab!(debug "no entry, killing off private route: {}", k.encode());
|
||||||
|
content.details.remove(&k);
|
||||||
|
}
|
||||||
|
|
||||||
// Load secrets from pstore
|
// Load secrets from pstore
|
||||||
let pstore = routing_table.network_manager().protected_store();
|
let pstore = routing_table.network_manager().protected_store();
|
||||||
@ -280,7 +296,7 @@ impl RouteSpecStore {
|
|||||||
.network_manager()
|
.network_manager()
|
||||||
.table_store();
|
.table_store();
|
||||||
let rsstdb = table_store.open("RouteSpecStore", 1).await?;
|
let rsstdb = table_store.open("RouteSpecStore", 1).await?;
|
||||||
rsstdb.store_cbor(0, b"content", &content).await?;
|
rsstdb.store_cbor(0, b"content", &content)?;
|
||||||
|
|
||||||
// // Keep secrets in protected store as well
|
// // Keep secrets in protected store as well
|
||||||
let pstore = self
|
let pstore = self
|
||||||
@ -1168,10 +1184,9 @@ impl RouteSpecStore {
|
|||||||
|
|
||||||
let mut buffer = vec![];
|
let mut buffer = vec![];
|
||||||
capnp::serialize_packed::write_message(&mut buffer, &pr_message)
|
capnp::serialize_packed::write_message(&mut buffer, &pr_message)
|
||||||
|
.map_err(RPCError::internal)
|
||||||
.wrap_err("failed to convert builder to vec")?;
|
.wrap_err("failed to convert builder to vec")?;
|
||||||
Ok(buffer)
|
Ok(buffer)
|
||||||
|
|
||||||
// builder_to_vec(pr_message).wrap_err("failed to convert builder to vec")
|
|
||||||
}
|
}
|
||||||
|
|
||||||
/// Convert binary blob to private route
|
/// Convert binary blob to private route
|
||||||
@ -1180,11 +1195,12 @@ impl RouteSpecStore {
|
|||||||
blob.as_slice(),
|
blob.as_slice(),
|
||||||
capnp::message::ReaderOptions::new(),
|
capnp::message::ReaderOptions::new(),
|
||||||
)
|
)
|
||||||
|
.map_err(RPCError::internal)
|
||||||
.wrap_err("failed to make message reader")?;
|
.wrap_err("failed to make message reader")?;
|
||||||
|
|
||||||
//let reader = ::capnp::message::Reader::new(RPCMessageData::new(blob), Default::default());
|
|
||||||
let pr_reader = reader
|
let pr_reader = reader
|
||||||
.get_root::<veilid_capnp::private_route::Reader>()
|
.get_root::<veilid_capnp::private_route::Reader>()
|
||||||
|
.map_err(RPCError::internal)
|
||||||
.wrap_err("failed to make reader for private_route")?;
|
.wrap_err("failed to make reader for private_route")?;
|
||||||
decode_private_route(&pr_reader).wrap_err("failed to decode private route")
|
decode_private_route(&pr_reader).wrap_err("failed to decode private route")
|
||||||
}
|
}
|
||||||
|
@ -323,19 +323,16 @@ impl RoutingTableInner {
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
pub fn init(&mut self, routing_table: RoutingTable) -> EyreResult<()> {
|
pub fn init_buckets(&mut self, routing_table: RoutingTable) {
|
||||||
// Size the buckets (one per bit)
|
// Size the buckets (one per bit)
|
||||||
|
self.buckets.clear();
|
||||||
self.buckets.reserve(DHT_KEY_LENGTH * 8);
|
self.buckets.reserve(DHT_KEY_LENGTH * 8);
|
||||||
for _ in 0..DHT_KEY_LENGTH * 8 {
|
for _ in 0..DHT_KEY_LENGTH * 8 {
|
||||||
let bucket = Bucket::new(routing_table.clone());
|
let bucket = Bucket::new(routing_table.clone());
|
||||||
self.buckets.push(bucket);
|
self.buckets.push(bucket);
|
||||||
}
|
}
|
||||||
|
|
||||||
Ok(())
|
|
||||||
}
|
}
|
||||||
|
|
||||||
pub fn terminate(&mut self) {}
|
|
||||||
|
|
||||||
pub fn configure_local_network_routing_domain(
|
pub fn configure_local_network_routing_domain(
|
||||||
&mut self,
|
&mut self,
|
||||||
local_networks: Vec<(IpAddr, IpAddr)>,
|
local_networks: Vec<(IpAddr, IpAddr)>,
|
||||||
|
@ -59,78 +59,67 @@ pub async fn test_store_delete_load(ts: TableStore) {
|
|||||||
);
|
);
|
||||||
|
|
||||||
assert_eq!(
|
assert_eq!(
|
||||||
db.load(0, b"foo").await.unwrap(),
|
db.load(0, b"foo").unwrap(),
|
||||||
None,
|
None,
|
||||||
"should not load missing key"
|
"should not load missing key"
|
||||||
);
|
);
|
||||||
assert!(
|
assert!(
|
||||||
db.store(1, b"foo", b"1234567890").await.is_ok(),
|
db.store(1, b"foo", b"1234567890").is_ok(),
|
||||||
"should store new key"
|
"should store new key"
|
||||||
);
|
);
|
||||||
assert_eq!(
|
assert_eq!(
|
||||||
db.load(0, b"foo").await.unwrap(),
|
db.load(0, b"foo").unwrap(),
|
||||||
None,
|
None,
|
||||||
"should not load missing key"
|
"should not load missing key"
|
||||||
);
|
);
|
||||||
assert_eq!(
|
assert_eq!(db.load(1, b"foo").unwrap(), Some(b"1234567890".to_vec()));
|
||||||
db.load(1, b"foo").await.unwrap(),
|
|
||||||
Some(b"1234567890".to_vec())
|
|
||||||
);
|
|
||||||
|
|
||||||
assert!(
|
assert!(
|
||||||
db.store(1, b"bar", b"FNORD").await.is_ok(),
|
db.store(1, b"bar", b"FNORD").is_ok(),
|
||||||
"should store new key"
|
"should store new key"
|
||||||
);
|
);
|
||||||
assert!(
|
assert!(
|
||||||
db.store(0, b"bar", b"ABCDEFGHIJKLMNOPQRSTUVWXYZ")
|
db.store(0, b"bar", b"ABCDEFGHIJKLMNOPQRSTUVWXYZ").is_ok(),
|
||||||
.await
|
|
||||||
.is_ok(),
|
|
||||||
"should store new key"
|
"should store new key"
|
||||||
);
|
);
|
||||||
assert!(
|
assert!(
|
||||||
db.store(2, b"bar", b"FNORD").await.is_ok(),
|
db.store(2, b"bar", b"FNORD").is_ok(),
|
||||||
"should store new key"
|
"should store new key"
|
||||||
);
|
);
|
||||||
assert!(
|
assert!(
|
||||||
db.store(2, b"baz", b"QWERTY").await.is_ok(),
|
db.store(2, b"baz", b"QWERTY").is_ok(),
|
||||||
"should store new key"
|
"should store new key"
|
||||||
);
|
);
|
||||||
assert!(
|
assert!(
|
||||||
db.store(2, b"bar", b"QWERTYUIOP").await.is_ok(),
|
db.store(2, b"bar", b"QWERTYUIOP").is_ok(),
|
||||||
"should store new key"
|
"should store new key"
|
||||||
);
|
);
|
||||||
|
|
||||||
assert_eq!(db.load(1, b"bar").await.unwrap(), Some(b"FNORD".to_vec()));
|
assert_eq!(db.load(1, b"bar").unwrap(), Some(b"FNORD".to_vec()));
|
||||||
assert_eq!(
|
assert_eq!(
|
||||||
db.load(0, b"bar").await.unwrap(),
|
db.load(0, b"bar").unwrap(),
|
||||||
Some(b"ABCDEFGHIJKLMNOPQRSTUVWXYZ".to_vec())
|
Some(b"ABCDEFGHIJKLMNOPQRSTUVWXYZ".to_vec())
|
||||||
);
|
);
|
||||||
assert_eq!(
|
assert_eq!(db.load(2, b"bar").unwrap(), Some(b"QWERTYUIOP".to_vec()));
|
||||||
db.load(2, b"bar").await.unwrap(),
|
assert_eq!(db.load(2, b"baz").unwrap(), Some(b"QWERTY".to_vec()));
|
||||||
Some(b"QWERTYUIOP".to_vec())
|
|
||||||
);
|
|
||||||
assert_eq!(db.load(2, b"baz").await.unwrap(), Some(b"QWERTY".to_vec()));
|
|
||||||
|
|
||||||
assert_eq!(db.delete(1, b"bar").await.unwrap(), true);
|
assert_eq!(db.delete(1, b"bar").unwrap(), true);
|
||||||
assert_eq!(db.delete(1, b"bar").await.unwrap(), false);
|
assert_eq!(db.delete(1, b"bar").unwrap(), false);
|
||||||
assert!(
|
assert!(
|
||||||
db.delete(4, b"bar").await.is_err(),
|
db.delete(4, b"bar").is_err(),
|
||||||
"can't delete from column that doesn't exist"
|
"can't delete from column that doesn't exist"
|
||||||
);
|
);
|
||||||
|
|
||||||
drop(db);
|
drop(db);
|
||||||
let db = ts.open("test", 3).await.expect("should have opened");
|
let db = ts.open("test", 3).await.expect("should have opened");
|
||||||
|
|
||||||
assert_eq!(db.load(1, b"bar").await.unwrap(), None);
|
assert_eq!(db.load(1, b"bar").unwrap(), None);
|
||||||
assert_eq!(
|
assert_eq!(
|
||||||
db.load(0, b"bar").await.unwrap(),
|
db.load(0, b"bar").unwrap(),
|
||||||
Some(b"ABCDEFGHIJKLMNOPQRSTUVWXYZ".to_vec())
|
Some(b"ABCDEFGHIJKLMNOPQRSTUVWXYZ".to_vec())
|
||||||
);
|
);
|
||||||
assert_eq!(
|
assert_eq!(db.load(2, b"bar").unwrap(), Some(b"QWERTYUIOP".to_vec()));
|
||||||
db.load(2, b"bar").await.unwrap(),
|
assert_eq!(db.load(2, b"baz").unwrap(), Some(b"QWERTY".to_vec()));
|
||||||
Some(b"QWERTYUIOP".to_vec())
|
|
||||||
);
|
|
||||||
assert_eq!(db.load(2, b"baz").await.unwrap(), Some(b"QWERTY".to_vec()));
|
|
||||||
}
|
}
|
||||||
|
|
||||||
pub async fn test_cbor(ts: TableStore) {
|
pub async fn test_cbor(ts: TableStore) {
|
||||||
@ -140,11 +129,11 @@ pub async fn test_cbor(ts: TableStore) {
|
|||||||
let db = ts.open("test", 3).await.expect("should have opened");
|
let db = ts.open("test", 3).await.expect("should have opened");
|
||||||
let (dht_key, _) = generate_secret();
|
let (dht_key, _) = generate_secret();
|
||||||
|
|
||||||
assert!(db.store_cbor(0, b"asdf", &dht_key).await.is_ok());
|
assert!(db.store_cbor(0, b"asdf", &dht_key).is_ok());
|
||||||
|
|
||||||
assert_eq!(db.load_cbor::<DHTKey>(0, b"qwer").await.unwrap(), None);
|
assert_eq!(db.load_cbor::<DHTKey>(0, b"qwer").unwrap(), None);
|
||||||
|
|
||||||
let d = match db.load_cbor::<DHTKey>(0, b"asdf").await {
|
let d = match db.load_cbor::<DHTKey>(0, b"asdf") {
|
||||||
Ok(x) => x,
|
Ok(x) => x,
|
||||||
Err(e) => {
|
Err(e) => {
|
||||||
panic!("couldn't decode cbor: {}", e);
|
panic!("couldn't decode cbor: {}", e);
|
||||||
@ -153,12 +142,12 @@ pub async fn test_cbor(ts: TableStore) {
|
|||||||
assert_eq!(d, Some(dht_key), "keys should be equal");
|
assert_eq!(d, Some(dht_key), "keys should be equal");
|
||||||
|
|
||||||
assert!(
|
assert!(
|
||||||
db.store(1, b"foo", b"1234567890").await.is_ok(),
|
db.store(1, b"foo", b"1234567890").is_ok(),
|
||||||
"should store new key"
|
"should store new key"
|
||||||
);
|
);
|
||||||
|
|
||||||
assert!(
|
assert!(
|
||||||
db.load_cbor::<DHTKey>(1, b"foo").await.is_err(),
|
db.load_cbor::<DHTKey>(1, b"foo").is_err(),
|
||||||
"should fail to load cbor"
|
"should fail to load cbor"
|
||||||
);
|
);
|
||||||
}
|
}
|
||||||
|
@ -203,11 +203,11 @@ fn config_callback(key: String) -> ConfigCallbackReturn {
|
|||||||
"network.routing_table.limit_attached_good" => Ok(Box::new(8u32)),
|
"network.routing_table.limit_attached_good" => Ok(Box::new(8u32)),
|
||||||
"network.routing_table.limit_attached_weak" => Ok(Box::new(4u32)),
|
"network.routing_table.limit_attached_weak" => Ok(Box::new(4u32)),
|
||||||
"network.rpc.concurrency" => Ok(Box::new(2u32)),
|
"network.rpc.concurrency" => Ok(Box::new(2u32)),
|
||||||
"network.rpc.queue_size" => Ok(Box::new(128u32)),
|
"network.rpc.queue_size" => Ok(Box::new(1024u32)),
|
||||||
"network.rpc.max_timestamp_behind_ms" => Ok(Box::new(Some(10_000u32))),
|
"network.rpc.max_timestamp_behind_ms" => Ok(Box::new(Some(10_000u32))),
|
||||||
"network.rpc.max_timestamp_ahead_ms" => Ok(Box::new(Some(10_000u32))),
|
"network.rpc.max_timestamp_ahead_ms" => Ok(Box::new(Some(10_000u32))),
|
||||||
"network.rpc.timeout_ms" => Ok(Box::new(10_000u32)),
|
"network.rpc.timeout_ms" => Ok(Box::new(10_000u32)),
|
||||||
"network.rpc.max_route_hop_count" => Ok(Box::new(7u8)),
|
"network.rpc.max_route_hop_count" => Ok(Box::new(4u8)),
|
||||||
"network.rpc.default_route_hop_count" => Ok(Box::new(2u8)),
|
"network.rpc.default_route_hop_count" => Ok(Box::new(2u8)),
|
||||||
"network.dht.resolve_node_timeout_ms" => Ok(Box::new(Option::<u32>::None)),
|
"network.dht.resolve_node_timeout_ms" => Ok(Box::new(Option::<u32>::None)),
|
||||||
"network.dht.resolve_node_count" => Ok(Box::new(20u32)),
|
"network.dht.resolve_node_count" => Ok(Box::new(20u32)),
|
||||||
@ -323,9 +323,9 @@ pub async fn test_config() {
|
|||||||
assert_eq!(inner.network.bootstrap, Vec::<String>::new());
|
assert_eq!(inner.network.bootstrap, Vec::<String>::new());
|
||||||
assert_eq!(inner.network.bootstrap_nodes, Vec::<String>::new());
|
assert_eq!(inner.network.bootstrap_nodes, Vec::<String>::new());
|
||||||
assert_eq!(inner.network.rpc.concurrency, 2u32);
|
assert_eq!(inner.network.rpc.concurrency, 2u32);
|
||||||
assert_eq!(inner.network.rpc.queue_size, 128u32);
|
assert_eq!(inner.network.rpc.queue_size, 1024u32);
|
||||||
assert_eq!(inner.network.rpc.timeout_ms, 10_000u32);
|
assert_eq!(inner.network.rpc.timeout_ms, 10_000u32);
|
||||||
assert_eq!(inner.network.rpc.max_route_hop_count, 7u8);
|
assert_eq!(inner.network.rpc.max_route_hop_count, 4u8);
|
||||||
assert_eq!(inner.network.rpc.default_route_hop_count, 2u8);
|
assert_eq!(inner.network.rpc.default_route_hop_count, 2u8);
|
||||||
assert_eq!(inner.network.routing_table.limit_over_attached, 64u32);
|
assert_eq!(inner.network.routing_table.limit_over_attached, 64u32);
|
||||||
assert_eq!(inner.network.routing_table.limit_fully_attached, 32u32);
|
assert_eq!(inner.network.routing_table.limit_fully_attached, 32u32);
|
||||||
|
@ -55,10 +55,7 @@ fn get_route_id(rss: RouteSpecStore) -> impl Fn(&str) -> Option<DHTKey> {
|
|||||||
}
|
}
|
||||||
|
|
||||||
fn get_safety_selection(text: &str, rss: RouteSpecStore) -> Option<SafetySelection> {
|
fn get_safety_selection(text: &str, rss: RouteSpecStore) -> Option<SafetySelection> {
|
||||||
if text.len() == 0 {
|
if text.len() != 0 && &text[0..1] == "-" {
|
||||||
return None;
|
|
||||||
}
|
|
||||||
if &text[0..1] == "-" {
|
|
||||||
// Unsafe
|
// Unsafe
|
||||||
let text = &text[1..];
|
let text = &text[1..];
|
||||||
let seq = get_sequencing(text).unwrap_or(Sequencing::NoPreference);
|
let seq = get_sequencing(text).unwrap_or(Sequencing::NoPreference);
|
||||||
|
@ -41,61 +41,42 @@ pub type PinBoxFutureLifetime<'a, T> = PinBox<dyn Future<Output = T> + 'a>;
|
|||||||
pub type SendPinBoxFuture<T> = PinBox<dyn Future<Output = T> + Send + 'static>;
|
pub type SendPinBoxFuture<T> = PinBox<dyn Future<Output = T> + Send + 'static>;
|
||||||
pub type SendPinBoxFutureLifetime<'a, T> = PinBox<dyn Future<Output = T> + Send + 'a>;
|
pub type SendPinBoxFutureLifetime<'a, T> = PinBox<dyn Future<Output = T> + Send + 'a>;
|
||||||
|
|
||||||
|
pub use std::borrow::{Cow, ToOwned};
|
||||||
|
pub use std::boxed::Box;
|
||||||
|
pub use std::cell::RefCell;
|
||||||
|
pub use std::cmp;
|
||||||
|
pub use std::collections::btree_map::BTreeMap;
|
||||||
|
pub use std::collections::btree_set::BTreeSet;
|
||||||
|
pub use std::collections::hash_map::HashMap;
|
||||||
|
pub use std::collections::hash_set::HashSet;
|
||||||
|
pub use std::collections::LinkedList;
|
||||||
|
pub use std::collections::VecDeque;
|
||||||
|
pub use std::convert::{TryFrom, TryInto};
|
||||||
|
pub use std::fmt;
|
||||||
|
pub use std::future::Future;
|
||||||
|
pub use std::mem;
|
||||||
|
pub use std::ops::{Fn, FnMut, FnOnce};
|
||||||
|
pub use std::pin::Pin;
|
||||||
|
pub use std::rc::Rc;
|
||||||
|
pub use std::string::String;
|
||||||
|
pub use std::sync::atomic::{AtomicBool, Ordering};
|
||||||
|
pub use std::sync::{Arc, Weak};
|
||||||
|
pub use std::task;
|
||||||
|
pub use std::time::Duration;
|
||||||
|
pub use std::vec::Vec;
|
||||||
|
|
||||||
cfg_if! {
|
cfg_if! {
|
||||||
if #[cfg(target_arch = "wasm32")] {
|
if #[cfg(target_arch = "wasm32")] {
|
||||||
pub use alloc::string::String;
|
|
||||||
pub use alloc::vec::Vec;
|
|
||||||
pub use alloc::collections::LinkedList;
|
|
||||||
pub use alloc::collections::VecDeque;
|
|
||||||
pub use alloc::collections::btree_map::BTreeMap;
|
|
||||||
pub use alloc::collections::btree_set::BTreeSet;
|
|
||||||
pub use hashbrown::hash_map::HashMap;
|
|
||||||
pub use hashbrown::hash_set::HashSet;
|
|
||||||
pub use alloc::boxed::Box;
|
|
||||||
pub use alloc::borrow::{Cow, ToOwned};
|
|
||||||
pub use wasm_bindgen::prelude::*;
|
pub use wasm_bindgen::prelude::*;
|
||||||
pub use core::cmp;
|
|
||||||
pub use core::convert::{TryFrom, TryInto};
|
|
||||||
pub use core::mem;
|
|
||||||
pub use core::fmt;
|
|
||||||
pub use alloc::rc::Rc;
|
|
||||||
pub use core::cell::RefCell;
|
|
||||||
pub use core::task;
|
|
||||||
pub use core::future::Future;
|
|
||||||
pub use core::time::Duration;
|
|
||||||
pub use core::pin::Pin;
|
|
||||||
pub use core::sync::atomic::{Ordering, AtomicBool};
|
|
||||||
pub use alloc::sync::{Arc, Weak};
|
|
||||||
pub use core::ops::{FnOnce, FnMut, Fn};
|
|
||||||
pub use async_lock::Mutex as AsyncMutex;
|
pub use async_lock::Mutex as AsyncMutex;
|
||||||
pub use async_lock::MutexGuard as AsyncMutexGuard;
|
pub use async_lock::MutexGuard as AsyncMutexGuard;
|
||||||
pub use async_lock::MutexGuardArc as AsyncMutexGuardArc;
|
pub use async_lock::MutexGuardArc as AsyncMutexGuardArc;
|
||||||
pub use no_std_net::{ SocketAddr, SocketAddrV4, SocketAddrV6, ToSocketAddrs, IpAddr, Ipv4Addr, Ipv6Addr };
|
|
||||||
pub use async_executors::JoinHandle as LowLevelJoinHandle;
|
pub use async_executors::JoinHandle as LowLevelJoinHandle;
|
||||||
|
|
||||||
|
pub use no_std_net::{ SocketAddr, SocketAddrV4, SocketAddrV6, ToSocketAddrs, IpAddr, Ipv4Addr, Ipv6Addr };
|
||||||
} else {
|
} else {
|
||||||
pub use std::string::String;
|
|
||||||
pub use std::vec::Vec;
|
|
||||||
pub use std::collections::LinkedList;
|
|
||||||
pub use std::collections::VecDeque;
|
|
||||||
pub use std::collections::btree_map::BTreeMap;
|
|
||||||
pub use std::collections::btree_set::BTreeSet;
|
|
||||||
pub use std::collections::hash_map::HashMap;
|
|
||||||
pub use std::collections::hash_set::HashSet;
|
|
||||||
pub use std::boxed::Box;
|
|
||||||
pub use std::borrow::{Cow, ToOwned};
|
|
||||||
pub use std::cmp;
|
|
||||||
pub use std::convert::{TryFrom, TryInto};
|
|
||||||
pub use std::mem;
|
|
||||||
pub use std::fmt;
|
|
||||||
pub use std::sync::atomic::{Ordering, AtomicBool};
|
|
||||||
pub use std::sync::{Arc, Weak};
|
|
||||||
pub use std::rc::Rc;
|
|
||||||
pub use std::cell::RefCell;
|
|
||||||
pub use std::task;
|
|
||||||
pub use std::future::Future;
|
|
||||||
pub use std::time::Duration;
|
|
||||||
pub use std::pin::Pin;
|
|
||||||
pub use std::ops::{FnOnce, FnMut, Fn};
|
|
||||||
cfg_if! {
|
cfg_if! {
|
||||||
if #[cfg(feature="rt-async-std")] {
|
if #[cfg(feature="rt-async-std")] {
|
||||||
pub use async_std::sync::Mutex as AsyncMutex;
|
pub use async_std::sync::Mutex as AsyncMutex;
|
||||||
|
Loading…
Reference in New Issue
Block a user