test work

This commit is contained in:
John Smith 2023-06-18 18:47:39 -04:00
parent b8e5039251
commit b6e055e47d
21 changed files with 664 additions and 171 deletions

View File

@ -44,7 +44,14 @@ impl Network {
// Spawn a local async task for each socket // Spawn a local async task for each socket
let mut protocol_handlers_unordered = FuturesUnordered::new(); let mut protocol_handlers_unordered = FuturesUnordered::new();
let network_manager = this.network_manager(); let network_manager = this.network_manager();
let stop_token = this.inner.lock().stop_source.as_ref().unwrap().token(); let stop_token = {
let inner = this.inner.lock();
if inner.stop_source.is_none() {
log_net!(debug "exiting UDP listener before it starts because we encountered an error");
return;
}
inner.stop_source.as_ref().unwrap().token()
};
for ph in protocol_handlers { for ph in protocol_handlers {
let network_manager = network_manager.clone(); let network_manager = network_manager.clone();

View File

@ -1234,9 +1234,10 @@ impl RoutingTableInner {
let kind = node_id.kind; let kind = node_id.kind;
let mut closest_nodes_locked: Vec<NodeRefLocked> = closest_nodes let mut closest_nodes_locked: Vec<NodeRefLocked> = closest_nodes
.iter() .iter()
.filter_map(|x| { .filter_map(|nr| {
if x.node_ids().kinds().contains(&kind) { let nr_locked = nr.locked(self);
Some(x.locked(self)) if nr_locked.node_ids().kinds().contains(&kind) {
Some(nr_locked)
} else { } else {
None None
} }

View File

@ -184,17 +184,22 @@ where
} }
// Delete record // Delete record
rt_xact.delete(0, &k.bytes()); if let Err(e) = rt_xact.delete(0, &k.bytes()) {
log_stor!(error "record could not be deleted: {}", e);
}
// Delete subkeys // Delete subkeys
let subkey_count = v.subkey_count() as u32; let stored_subkeys = v.stored_subkeys();
for sk in 0..subkey_count { for sk in stored_subkeys.iter() {
// From table // From table
let stk = SubkeyTableKey { let stk = SubkeyTableKey {
key: k.key, key: k.key,
subkey: sk, subkey: sk,
}; };
st_xact.delete(0, &stk.bytes()); let stkb = stk.bytes();
if let Err(e) = st_xact.delete(0, &stkb) {
log_stor!(error "subkey could not be deleted: {}", e);
}
// From cache // From cache
self.remove_from_subkey_cache(stk); self.remove_from_subkey_cache(stk);
@ -355,8 +360,8 @@ where
want_descriptor: bool, want_descriptor: bool,
) -> VeilidAPIResult<Option<SubkeyResult>> { ) -> VeilidAPIResult<Option<SubkeyResult>> {
// record from index // record from index
let Some((subkey_count, opt_descriptor)) = self.with_record(key, |record| { let Some((subkey_count, has_subkey, opt_descriptor)) = self.with_record(key, |record| {
(record.subkey_count(), if want_descriptor { (record.subkey_count(), record.stored_subkeys().contains(subkey), if want_descriptor {
Some(record.descriptor().clone()) Some(record.descriptor().clone())
} else { } else {
None None
@ -371,6 +376,15 @@ where
apibail_invalid_argument!("subkey out of range", "subkey", subkey); apibail_invalid_argument!("subkey out of range", "subkey", subkey);
} }
// See if we have this subkey stored
if !has_subkey {
// If not, return no value but maybe with descriptor
return Ok(Some(SubkeyResult {
value: None,
descriptor: opt_descriptor,
}));
}
// Get subkey table // Get subkey table
let Some(subkey_table) = self.subkey_table.clone() else { let Some(subkey_table) = self.subkey_table.clone() else {
apibail_internal!("record store not initialized"); apibail_internal!("record store not initialized");
@ -386,28 +400,23 @@ where
descriptor: opt_descriptor, descriptor: opt_descriptor,
})); }));
} }
// If not in cache, try to pull from table store // If not in cache, try to pull from table store if it is in our stored subkey set
if let Some(record_data) = subkey_table let Some(record_data) = subkey_table
.load_rkyv::<RecordData>(0, &stk.bytes()) .load_rkyv::<RecordData>(0, &stk.bytes())
.await .await
.map_err(VeilidAPIError::internal)? .map_err(VeilidAPIError::internal)? else {
{ apibail_internal!("failed to get subkey that was stored");
let out = record_data.signed_value_data().clone(); };
// Add to cache, do nothing with lru out let out = record_data.signed_value_data().clone();
self.add_to_subkey_cache(stk, record_data);
return Ok(Some(SubkeyResult { // Add to cache, do nothing with lru out
value: Some(out), self.add_to_subkey_cache(stk, record_data);
descriptor: opt_descriptor,
}));
};
// Record was available, but subkey was not found, maybe descriptor gets returned return Ok(Some(SubkeyResult {
Ok(Some(SubkeyResult { value: Some(out),
value: None,
descriptor: opt_descriptor, descriptor: opt_descriptor,
})) }));
} }
pub async fn set_subkey( pub async fn set_subkey(
@ -492,6 +501,7 @@ where
// Update record // Update record
self.with_record_mut(key, |record| { self.with_record_mut(key, |record| {
record.store_subkey(subkey);
record.set_record_data_size(new_record_data_size); record.set_record_data_size(new_record_data_size);
}) })
.expect("record should still be here"); .expect("record should still be here");
@ -522,10 +532,11 @@ where
out += "Record Index:\n"; out += "Record Index:\n";
for (rik, rec) in &self.record_index { for (rik, rec) in &self.record_index {
out += &format!( out += &format!(
" {} @ {} len={}\n", " {} @ {} len={} subkeys={}\n",
rik.key.to_string(), rik.key.to_string(),
rec.last_touched().as_u64(), rec.last_touched().as_u64(),
rec.record_data_size() rec.record_data_size(),
rec.stored_subkeys(),
); );
} }
out += &format!("Subkey Cache Count: {}\n", self.subkey_cache.len()); out += &format!("Subkey Cache Count: {}\n", self.subkey_cache.len());

View File

@ -12,6 +12,7 @@ where
{ {
descriptor: SignedValueDescriptor, descriptor: SignedValueDescriptor,
subkey_count: usize, subkey_count: usize,
stored_subkeys: ValueSubkeyRangeSet,
last_touched_ts: Timestamp, last_touched_ts: Timestamp,
record_data_size: usize, record_data_size: usize,
detail: D, detail: D,
@ -33,6 +34,7 @@ where
Ok(Self { Ok(Self {
descriptor, descriptor,
subkey_count, subkey_count,
stored_subkeys: ValueSubkeyRangeSet::new(),
last_touched_ts: cur_ts, last_touched_ts: cur_ts,
record_data_size: 0, record_data_size: 0,
detail, detail,
@ -50,6 +52,13 @@ where
self.subkey_count self.subkey_count
} }
pub fn stored_subkeys(&self) -> &ValueSubkeyRangeSet {
&self.stored_subkeys
}
pub fn store_subkey(&mut self, subkey: ValueSubkey) {
self.stored_subkeys.insert(subkey);
}
pub fn touch(&mut self, cur_ts: Timestamp) { pub fn touch(&mut self, cur_ts: Timestamp) {
self.last_touched_ts = cur_ts self.last_touched_ts = cur_ts
} }

View File

@ -45,6 +45,7 @@ impl Drop for TableDBUnlockedInner {
#[derive(Debug, Clone)] #[derive(Debug, Clone)]
pub struct TableDB { pub struct TableDB {
opened_column_count: u32,
unlocked_inner: Arc<TableDBUnlockedInner>, unlocked_inner: Arc<TableDBUnlockedInner>,
} }
@ -56,11 +57,13 @@ impl TableDB {
database: Database, database: Database,
encryption_key: Option<TypedSharedSecret>, encryption_key: Option<TypedSharedSecret>,
decryption_key: Option<TypedSharedSecret>, decryption_key: Option<TypedSharedSecret>,
opened_column_count: u32,
) -> Self { ) -> Self {
let encrypt_info = encryption_key.map(|ek| CryptInfo::new(crypto.clone(), ek)); let encrypt_info = encryption_key.map(|ek| CryptInfo::new(crypto.clone(), ek));
let decrypt_info = decryption_key.map(|dk| CryptInfo::new(crypto.clone(), dk)); let decrypt_info = decryption_key.map(|dk| CryptInfo::new(crypto.clone(), dk));
Self { Self {
opened_column_count,
unlocked_inner: Arc::new(TableDBUnlockedInner { unlocked_inner: Arc::new(TableDBUnlockedInner {
table, table,
table_store, table_store,
@ -71,8 +74,12 @@ impl TableDB {
} }
} }
pub(super) fn try_new_from_weak_inner(weak_inner: Weak<TableDBUnlockedInner>) -> Option<Self> { pub(super) fn try_new_from_weak_inner(
weak_inner: Weak<TableDBUnlockedInner>,
opened_column_count: u32,
) -> Option<Self> {
weak_inner.upgrade().map(|table_db_unlocked_inner| Self { weak_inner.upgrade().map(|table_db_unlocked_inner| Self {
opened_column_count,
unlocked_inner: table_db_unlocked_inner, unlocked_inner: table_db_unlocked_inner,
}) })
} }
@ -82,6 +89,7 @@ impl TableDB {
} }
/// Get the total number of columns in the TableDB /// Get the total number of columns in the TableDB
/// Not the number of columns that were opened, rather the total number that could be opened
pub fn get_column_count(&self) -> VeilidAPIResult<u32> { pub fn get_column_count(&self) -> VeilidAPIResult<u32> {
let db = &self.unlocked_inner.database; let db = &self.unlocked_inner.database;
db.num_columns().map_err(VeilidAPIError::from) db.num_columns().map_err(VeilidAPIError::from)
@ -144,8 +152,14 @@ impl TableDB {
} }
} }
/// Get the list of keys in a column of the TableDB /// Get the list of keys in a column of the TableDAB
pub async fn get_keys(&self, col: u32) -> VeilidAPIResult<Vec<Vec<u8>>> { pub async fn get_keys(&self, col: u32) -> VeilidAPIResult<Vec<Vec<u8>>> {
if col >= self.opened_column_count {
apibail_generic!(format!(
"Column exceeds opened column count {} >= {}",
col, self.opened_column_count
));
}
let db = self.unlocked_inner.database.clone(); let db = self.unlocked_inner.database.clone();
let mut out = Vec::new(); let mut out = Vec::new();
db.iter_keys(col, None, |k| { db.iter_keys(col, None, |k| {
@ -165,6 +179,12 @@ impl TableDB {
/// Store a key with a value in a column in the TableDB. Performs a single transaction immediately. /// Store a key with a value in a column in the TableDB. Performs a single transaction immediately.
pub async fn store(&self, col: u32, key: &[u8], value: &[u8]) -> VeilidAPIResult<()> { pub async fn store(&self, col: u32, key: &[u8], value: &[u8]) -> VeilidAPIResult<()> {
if col >= self.opened_column_count {
apibail_generic!(format!(
"Column exceeds opened column count {} >= {}",
col, self.opened_column_count
));
}
let db = self.unlocked_inner.database.clone(); let db = self.unlocked_inner.database.clone();
let mut dbt = db.transaction(); let mut dbt = db.transaction();
dbt.put( dbt.put(
@ -195,6 +215,12 @@ impl TableDB {
/// Read a key from a column in the TableDB immediately. /// Read a key from a column in the TableDB immediately.
pub async fn load(&self, col: u32, key: &[u8]) -> VeilidAPIResult<Option<Vec<u8>>> { pub async fn load(&self, col: u32, key: &[u8]) -> VeilidAPIResult<Option<Vec<u8>>> {
if col >= self.opened_column_count {
apibail_generic!(format!(
"Column exceeds opened column count {} >= {}",
col, self.opened_column_count
));
}
let db = self.unlocked_inner.database.clone(); let db = self.unlocked_inner.database.clone();
let key = self.maybe_encrypt(key, true); let key = self.maybe_encrypt(key, true);
Ok(db Ok(db
@ -233,6 +259,12 @@ impl TableDB {
/// Delete key with from a column in the TableDB /// Delete key with from a column in the TableDB
pub async fn delete(&self, col: u32, key: &[u8]) -> VeilidAPIResult<Option<Vec<u8>>> { pub async fn delete(&self, col: u32, key: &[u8]) -> VeilidAPIResult<Option<Vec<u8>>> {
if col >= self.opened_column_count {
apibail_generic!(format!(
"Column exceeds opened column count {} >= {}",
col, self.opened_column_count
));
}
let key = self.maybe_encrypt(key, true); let key = self.maybe_encrypt(key, true);
let db = self.unlocked_inner.database.clone(); let db = self.unlocked_inner.database.clone();
@ -330,11 +362,19 @@ impl TableDBTransaction {
} }
/// Store a key with a value in a column in the TableDB /// Store a key with a value in a column in the TableDB
pub fn store(&self, col: u32, key: &[u8], value: &[u8]) { pub fn store(&self, col: u32, key: &[u8], value: &[u8]) -> VeilidAPIResult<()> {
if col >= self.db.opened_column_count {
apibail_generic!(format!(
"Column exceeds opened column count {} >= {}",
col, self.db.opened_column_count
));
}
let key = self.db.maybe_encrypt(key, true); let key = self.db.maybe_encrypt(key, true);
let value = self.db.maybe_encrypt(value, false); let value = self.db.maybe_encrypt(value, false);
let mut inner = self.inner.lock(); let mut inner = self.inner.lock();
inner.dbt.as_mut().unwrap().put_owned(col, key, value); inner.dbt.as_mut().unwrap().put_owned(col, key, value);
Ok(())
} }
/// Store a key in rkyv format with a value in a column in the TableDB /// Store a key in rkyv format with a value in a column in the TableDB
@ -343,12 +383,7 @@ impl TableDBTransaction {
T: RkyvSerialize<DefaultVeilidRkyvSerializer>, T: RkyvSerialize<DefaultVeilidRkyvSerializer>,
{ {
let value = to_rkyv(value)?; let value = to_rkyv(value)?;
let key = self.db.maybe_encrypt(key, true); self.store(col, key, &value)
let value = self.db.maybe_encrypt(&value, false);
let mut inner = self.inner.lock();
inner.dbt.as_mut().unwrap().put_owned(col, key, value);
Ok(())
} }
/// Store a key in rkyv format with a value in a column in the TableDB /// Store a key in rkyv format with a value in a column in the TableDB
@ -357,19 +392,22 @@ impl TableDBTransaction {
T: serde::Serialize, T: serde::Serialize,
{ {
let value = serde_json::to_vec(value).map_err(VeilidAPIError::internal)?; let value = serde_json::to_vec(value).map_err(VeilidAPIError::internal)?;
let key = self.db.maybe_encrypt(key, true); self.store(col, key, &value)
let value = self.db.maybe_encrypt(&value, false);
let mut inner = self.inner.lock();
inner.dbt.as_mut().unwrap().put_owned(col, key, value);
Ok(())
} }
/// Delete key with from a column in the TableDB /// Delete key with from a column in the TableDB
pub fn delete(&self, col: u32, key: &[u8]) { pub fn delete(&self, col: u32, key: &[u8]) -> VeilidAPIResult<()> {
if col >= self.db.opened_column_count {
apibail_generic!(format!(
"Column exceeds opened column count {} >= {}",
col, self.db.opened_column_count
));
}
let key = self.db.maybe_encrypt(key, true); let key = self.db.maybe_encrypt(key, true);
let mut inner = self.inner.lock(); let mut inner = self.inner.lock();
inner.dbt.as_mut().unwrap().delete_owned(col, key); inner.dbt.as_mut().unwrap().delete_owned(col, key);
Ok(())
} }
} }

View File

@ -428,6 +428,7 @@ impl TableStore {
} }
pub(crate) fn on_table_db_drop(&self, table: String) { pub(crate) fn on_table_db_drop(&self, table: String) {
log_rtab!(debug "dropping table db: {}", table);
let mut inner = self.inner.lock(); let mut inner = self.inner.lock();
if inner.opened.remove(&table).is_none() { if inner.opened.remove(&table).is_none() {
unreachable!("should have removed an item"); unreachable!("should have removed an item");
@ -449,12 +450,21 @@ impl TableStore {
let table_name = self.name_get_or_create(name).await?; let table_name = self.name_get_or_create(name).await?;
// See if this table is already opened // See if this table is already opened, if so the column count must be the same
{ {
let mut inner = self.inner.lock(); let mut inner = self.inner.lock();
if let Some(table_db_weak_inner) = inner.opened.get(&table_name) { if let Some(table_db_weak_inner) = inner.opened.get(&table_name) {
match TableDB::try_new_from_weak_inner(table_db_weak_inner.clone()) { match TableDB::try_new_from_weak_inner(table_db_weak_inner.clone(), column_count) {
Some(tdb) => { Some(tdb) => {
// Ensure column count isnt bigger
let existing_col_count = tdb.get_column_count()?;
if column_count > existing_col_count {
return Err(VeilidAPIError::generic(format!(
"database must be closed before increasing column count {} -> {}",
existing_col_count, column_count,
)));
}
return Ok(tdb); return Ok(tdb);
} }
None => { None => {
@ -465,7 +475,7 @@ impl TableStore {
} }
// Open table db using platform-specific driver // Open table db using platform-specific driver
let db = match self let mut db = match self
.table_store_driver .table_store_driver
.open(&table_name, column_count) .open(&table_name, column_count)
.await .await
@ -481,6 +491,24 @@ impl TableStore {
// Flush table names to disk // Flush table names to disk
self.flush().await; self.flush().await;
// If more columns are available, open the low level db with the max column count but restrict the tabledb object to the number requested
let existing_col_count = db.num_columns().map_err(VeilidAPIError::from)?;
if existing_col_count > column_count {
drop(db);
db = match self
.table_store_driver
.open(&table_name, existing_col_count)
.await
{
Ok(db) => db,
Err(e) => {
self.name_delete(name).await.expect("cleanup failed");
self.flush().await;
return Err(e);
}
};
}
// Wrap low-level Database in TableDB object // Wrap low-level Database in TableDB object
let mut inner = self.inner.lock(); let mut inner = self.inner.lock();
let table_db = TableDB::new( let table_db = TableDB::new(
@ -490,6 +518,7 @@ impl TableStore {
db, db,
inner.encryption_key.clone(), inner.encryption_key.clone(),
inner.encryption_key.clone(), inner.encryption_key.clone(),
column_count,
); );
// Keep track of opened DBs // Keep track of opened DBs

View File

@ -132,6 +132,41 @@ pub async fn test_store_delete_load(ts: TableStore) {
assert_eq!(db.load(2, b"baz").await.unwrap(), Some(b"QWERTY".to_vec())); assert_eq!(db.load(2, b"baz").await.unwrap(), Some(b"QWERTY".to_vec()));
} }
pub async fn test_transaction(ts: TableStore) {
trace!("test_transaction");
let _ = ts.delete("test");
let db = ts.open("test", 3).await.expect("should have opened");
assert!(
ts.delete("test").await.is_err(),
"should fail because file is opened"
);
let tx = db.transact();
assert!(tx.store(0, b"aaa", b"a-value").is_ok());
assert!(tx.store_json(1, b"bbb", &"b-value".to_owned()).is_ok());
assert!(tx.store_rkyv(2, b"ccc", &"c-value".to_owned()).is_ok());
assert!(tx.store(3, b"ddd", b"d-value").is_err());
assert!(tx.store(0, b"ddd", b"d-value").is_ok());
assert!(tx.delete(0, b"ddd").is_ok());
assert!(tx.commit().await.is_ok());
let tx = db.transact();
assert!(tx.delete(2, b"ccc").is_ok());
tx.rollback();
assert_eq!(db.load(0, b"aaa").await, Ok(Some(b"a-value".to_vec())));
assert_eq!(
db.load_json::<String>(1, b"bbb").await,
Ok(Some("b-value".to_owned()))
);
assert_eq!(
db.load_rkyv::<String>(2, b"ccc").await,
Ok(Some("c-value".to_owned()))
);
assert_eq!(db.load(0, b"ddd").await, Ok(None));
}
pub async fn test_rkyv(vcrypto: CryptoSystemVersion, ts: TableStore) { pub async fn test_rkyv(vcrypto: CryptoSystemVersion, ts: TableStore) {
trace!("test_rkyv"); trace!("test_rkyv");
@ -268,6 +303,7 @@ pub async fn test_all() {
test_protect_unprotect(vcrypto.clone(), ts.clone()).await; test_protect_unprotect(vcrypto.clone(), ts.clone()).await;
test_delete_open_delete(ts.clone()).await; test_delete_open_delete(ts.clone()).await;
test_store_delete_load(ts.clone()).await; test_store_delete_load(ts.clone()).await;
test_transaction(ts.clone()).await;
test_rkyv(vcrypto.clone(), ts.clone()).await; test_rkyv(vcrypto.clone(), ts.clone()).await;
test_json(vcrypto, ts.clone()).await; test_json(vcrypto, ts.clone()).await;
let _ = ts.delete("test").await; let _ = ts.delete("test").await;

View File

@ -405,12 +405,15 @@ impl JsonRequestProcessor {
TableDbTransactionResponseOp::Rollback {} TableDbTransactionResponseOp::Rollback {}
} }
TableDbTransactionRequestOp::Store { col, key, value } => { TableDbTransactionRequestOp::Store { col, key, value } => {
table_db_transaction.store(col, &key, &value); TableDbTransactionResponseOp::Store {
TableDbTransactionResponseOp::Store {} result: to_json_api_result(table_db_transaction.store(col, &key, &value)),
}
} }
TableDbTransactionRequestOp::Delete { col, key } => { TableDbTransactionRequestOp::Delete { col, key } => {
table_db_transaction.delete(col, &key); TableDbTransactionResponseOp::Delete {
TableDbTransactionResponseOp::Delete {} result: to_json_api_result(table_db_transaction.delete(col, &key)),
}
} }
}; };
TableDbTransactionResponse { TableDbTransactionResponse {

View File

@ -124,6 +124,12 @@ pub enum TableDbTransactionResponseOp {
result: ApiResult<()>, result: ApiResult<()>,
}, },
Rollback {}, Rollback {},
Store {}, Store {
Delete {}, #[serde(flatten)]
result: ApiResult<()>,
},
Delete {
#[serde(flatten)]
result: ApiResult<()>,
},
} }

View File

@ -4,7 +4,6 @@ use range_set_blaze::*;
#[derive( #[derive(
Clone, Clone,
Debug,
Default, Default,
PartialOrd, PartialOrd,
PartialEq, PartialEq,
@ -55,3 +54,15 @@ impl DerefMut for ValueSubkeyRangeSet {
&mut self.data &mut self.data
} }
} }
impl fmt::Debug for ValueSubkeyRangeSet {
fn fmt(&self, f: &mut fmt::Formatter<'_>) -> fmt::Result {
write!(f, "{:?}", self.data)
}
}
impl fmt::Display for ValueSubkeyRangeSet {
fn fmt(&self, f: &mut fmt::Formatter<'_>) -> fmt::Result {
write!(f, "{}", self.data)
}
}

View File

@ -884,7 +884,7 @@ pub extern "C" fn table_db_transaction_store(port: i64, id: u32, col: u32, key:
tdbt.clone() tdbt.clone()
}; };
tdbt.store(col, &key, &value); tdbt.store(col, &key, &value)?;
APIRESULT_VOID APIRESULT_VOID
}); });
} }
@ -908,7 +908,7 @@ pub extern "C" fn table_db_transaction_delete(port: i64, id: u32, col: u32, key:
tdbt.clone() tdbt.clone()
}; };
tdbt.delete(col, &key); tdbt.delete(col, &key)?;
APIRESULT_VOID APIRESULT_VOID
}); });
} }

View File

@ -9,12 +9,12 @@ from .conftest import simple_update_callback
@pytest.mark.asyncio @pytest.mark.asyncio
async def test_connect(api_connection): async def test_connect(api_connection: veilid.VeilidAPI):
pass pass
@pytest.mark.asyncio @pytest.mark.asyncio
async def test_get_node_id(api_connection): async def test_get_node_id(api_connection: veilid.VeilidAPI):
state = await api_connection.get_state() state = await api_connection.get_state()
node_ids = state.config.config.network.routing_table.node_id node_ids = state.config.config.network.routing_table.node_id
@ -35,7 +35,7 @@ async def test_fail_connect():
@pytest.mark.asyncio @pytest.mark.asyncio
async def test_version(api_connection): async def test_version(api_connection: veilid.VeilidAPI):
v = await api_connection.veilid_version() v = await api_connection.veilid_version()
print(f"veilid_version: {v.__dict__}") print(f"veilid_version: {v.__dict__}")
assert v.__dict__.keys() >= {"_major", "_minor", "_patch"} assert v.__dict__.keys() >= {"_major", "_minor", "_patch"}

View File

@ -3,29 +3,26 @@
import pytest import pytest
import veilid import veilid
from veilid.api import CryptoSystem from veilid.api import CryptoSystem
import gc
@pytest.mark.asyncio @pytest.mark.asyncio
async def test_best_crypto_system(api_connection): async def test_best_crypto_system(api_connection: veilid.VeilidAPI):
bcs: CryptoSystem = await api_connection.best_crypto_system() cs: CryptoSystem = await api_connection.best_crypto_system()
async with cs:
assert await bcs.default_salt_length() == 16 assert await cs.default_salt_length() == 16
@pytest.mark.asyncio @pytest.mark.asyncio
async def test_get_crypto_system(api_connection): async def test_get_crypto_system(api_connection: veilid.VeilidAPI):
cs: CryptoSystem = await api_connection.get_crypto_system( cs: CryptoSystem = await api_connection.get_crypto_system(
veilid.CryptoKind.CRYPTO_KIND_VLD0 veilid.CryptoKind.CRYPTO_KIND_VLD0
) )
async with cs:
assert await cs.default_salt_length() == 16 assert await cs.default_salt_length() == 16
# clean up handle early
del cs
@pytest.mark.asyncio @pytest.mark.asyncio
async def test_get_crypto_system_invalid(api_connection): async def test_get_crypto_system_invalid(api_connection: veilid.VeilidAPI):
with pytest.raises(veilid.VeilidAPIErrorInvalidArgument) as exc: with pytest.raises(veilid.VeilidAPIErrorInvalidArgument) as exc:
await api_connection.get_crypto_system(veilid.CryptoKind.CRYPTO_KIND_NONE) await api_connection.get_crypto_system(veilid.CryptoKind.CRYPTO_KIND_NONE)
@ -35,15 +32,17 @@ async def test_get_crypto_system_invalid(api_connection):
@pytest.mark.asyncio @pytest.mark.asyncio
async def test_hash_and_verify_password(api_connection): async def test_hash_and_verify_password(api_connection: veilid.VeilidAPI):
bcs = await api_connection.best_crypto_system() cs = await api_connection.best_crypto_system()
nonce = await bcs.random_nonce() async with cs:
salt = nonce.to_bytes() nonce = await cs.random_nonce()
salt = nonce.to_bytes()
# Password match # Password match
phash = await bcs.hash_password(b"abc123", salt) phash = await cs.hash_password(b"abc123", salt)
assert await bcs.verify_password(b"abc123", phash) assert await cs.verify_password(b"abc123", phash)
# Password mismatch
phash2 = await cs.hash_password(b"abc1234", salt)
assert not await cs.verify_password(b"abc12345", phash)
# Password mismatch
phash2 = await bcs.hash_password(b"abc1234", salt)
assert not await bcs.verify_password(b"abc12345", phash)

View File

@ -0,0 +1,51 @@
# Routing context veilid tests
import veilid
import pytest
import asyncio
import json
from . import *
##################################################################
BOGUS_KEY = veilid.TypedKey.from_value(veilid.CryptoKind.CRYPTO_KIND_VLD0, veilid.PublicKey.from_bytes(b' '))
# @pytest.mark.asyncio
# async def test_get_dht_value_unopened(api_connection: veilid.VeilidAPI):
# rc = await api_connection.new_routing_context()
# async with rc:
# with pytest.raises(veilid.VeilidAPIError):
# out = await rc.get_dht_value(BOGUS_KEY, veilid.ValueSubkey(0), False)
# @pytest.mark.asyncio
# async def test_open_dht_record_nonexistent_no_writer(api_connection: veilid.VeilidAPI):
# rc = await api_connection.new_routing_context()
# async with rc:
# with pytest.raises(veilid.VeilidAPIError):
# out = await rc.open_dht_record(BOGUS_KEY, None)
# @pytest.mark.asyncio
# async def test_close_dht_record_nonexistent(api_connection: veilid.VeilidAPI):
# rc = await api_connection.new_routing_context()
# async with rc:
# with pytest.raises(veilid.VeilidAPIError):
# await rc.close_dht_record(BOGUS_KEY)
# @pytest.mark.asyncio
# async def test_delete_dht_record_nonexistent(api_connection: veilid.VeilidAPI):
# rc = await api_connection.new_routing_context()
# async with rc:
# with pytest.raises(veilid.VeilidAPIError):
# await rc.delete_dht_record(BOGUS_KEY)
# @pytest.mark.asyncio
# async def test_create_delete_dht_record_simple(api_connection: veilid.VeilidAPI):
# rc = await api_connection.new_routing_context()
# async with rc:
# rec = await rc.create_dht_record(veilid.CryptoKind.CRYPTO_KIND_VLD0, veilid.DHTSchema.dflt(1))
# await rc.close_dht_record(rec.key)
# await rc.delete_dht_record(rec.key)
# xxx make tests for tabledb api first
# xxx then make a test that creates a record, stores it in a table
# xxx then make another test that gets the keys from the table and closes/deletes them

View File

@ -12,11 +12,15 @@ from .conftest import server_info
@pytest.mark.asyncio @pytest.mark.asyncio
async def test_routing_contexts(api_connection): async def test_routing_contexts(api_connection: veilid.VeilidAPI):
rc = await api_connection.new_routing_context() rc = await api_connection.new_routing_context()
rcp = await rc.with_privacy() async with rc:
rcps = await rcp.with_sequencing(veilid.Sequencing.ENSURE_ORDERED) rcp = await rc.with_privacy(release = False)
await rcps.with_custom_privacy(veilid.Stability.RELIABLE) async with rcp:
rcps = await rcp.with_sequencing(veilid.Sequencing.ENSURE_ORDERED, release = False)
async with rcps:
rcpscp = await rcps.with_custom_privacy(veilid.Stability.RELIABLE, release = False)
await rcpscp.release()
@pytest.mark.asyncio @pytest.mark.asyncio
@ -38,24 +42,25 @@ async def test_routing_context_app_message_loopback():
# make a routing context that uses a safety route # make a routing context that uses a safety route
rc = await (await api.new_routing_context()).with_privacy() rc = await (await api.new_routing_context()).with_privacy()
async with rc:
# make a new local private route
prl, blob = await api.new_private_route()
# make a new local private route # import it as a remote route as well so we can send to it
prl, blob = await api.new_private_route() prr = await api.import_remote_private_route(blob)
# import it as a remote route as well so we can send to it # send an app message to our own private route
prr = await api.import_remote_private_route(blob) message = b"abcd1234"
await rc.app_message(prr, message)
# send an app message to our own private route # we should get the same message back
message = b"abcd1234" update: veilid.VeilidUpdate = await asyncio.wait_for(
await rc.app_message(prr, message) app_message_queue.get(), timeout=10
)
# we should get the same message back assert isinstance(update.detail, veilid.VeilidAppMessage)
update: veilid.VeilidUpdate = await asyncio.wait_for( assert update.detail.message == message
app_message_queue.get(), timeout=10
)
assert isinstance(update.detail, veilid.VeilidAppMessage)
assert update.detail.message == message
@pytest.mark.asyncio @pytest.mark.asyncio
@ -74,37 +79,33 @@ async def test_routing_context_app_call_loopback():
# make a routing context that uses a safety route # make a routing context that uses a safety route
rc = await (await api.new_routing_context()).with_privacy() rc = await (await api.new_routing_context()).with_privacy()
async with rc:
# make a new local private route
prl, blob = await api.new_private_route()
# make a new local private route # import it as a remote route as well so we can send to it
prl, blob = await api.new_private_route() prr = await api.import_remote_private_route(blob)
# import it as a remote route as well so we can send to it # send an app message to our own private route
prr = await api.import_remote_private_route(blob) request = b"abcd1234"
app_call_task = asyncio.create_task(
rc.app_call(prr, request), name="app call task"
)
# send an app message to our own private route # we should get the same request back
request = b"abcd1234" update: veilid.VeilidUpdate = await asyncio.wait_for(
app_call_task = asyncio.create_task( app_call_queue.get(), timeout=10
rc.app_call(prr, request), name="app call task" )
) appcall = update.detail
# we should get the same request back assert isinstance(appcall, veilid.VeilidAppCall)
update: veilid.VeilidUpdate = await asyncio.wait_for( assert appcall.message == request
app_call_queue.get(), timeout=10
)
appcall = update.detail
assert isinstance(appcall, veilid.VeilidAppCall) # now we reply to the request
assert appcall.message == request reply = b"qwer5678"
await api.app_call_reply(appcall.call_id, reply)
# now we reply to the request # now we should get the reply from the call
reply = b"qwer5678" result = await app_call_task
# TK: OperationId use to be a subclass of `int`. When I wrapped `appcall.call_id` in int(), assert result == reply
# this failed JSON schema validation, which defines `call_id` as a string. Maybe that was a
# typo, and OperationId is really *supposed* to be a str? Alternatively, perhaps the
# signature of `app_call_reply` is wrong and it's supposed to take a type other than
# OperationId?
await api.app_call_reply(OperationId(appcall.call_id), reply)
# now we should get the reply from the call
result = await app_call_task
assert result == reply

View File

@ -0,0 +1,127 @@
# TableDB veilid tests
import pytest
import veilid
from veilid.api import CryptoSystem
TEST_DB = "__pytest_db"
TEST_NONEXISTENT_DB = "__pytest_nonexistent_db"
@pytest.mark.asyncio
async def test_delete_table_db_nonexistent(api_connection: veilid.VeilidAPI):
deleted = await api_connection.delete_table_db(TEST_NONEXISTENT_DB)
assert not deleted
@pytest.mark.asyncio
async def test_open_delete_table_db(api_connection: veilid.VeilidAPI):
# delete test db if it exists
await api_connection.delete_table_db(TEST_DB)
tdb = await api_connection.open_table_db(TEST_DB, 1)
async with tdb:
# delete should fail since it is still open
with pytest.raises(veilid.VeilidAPIErrorGeneric) as exc:
await api_connection.delete_table_db(TEST_DB)
# drop the db
# now delete should succeed
deleted = await api_connection.delete_table_db(TEST_DB)
assert deleted
@pytest.mark.asyncio
async def test_open_twice_table_db(api_connection: veilid.VeilidAPI):
# delete test db if it exists
await api_connection.delete_table_db(TEST_DB)
tdb = await api_connection.open_table_db(TEST_DB, 1)
tdb2 = await api_connection.open_table_db(TEST_DB, 1)
# delete should fail because open
with pytest.raises(veilid.VeilidAPIErrorGeneric) as exc:
await api_connection.delete_table_db(TEST_DB)
await tdb.release()
# delete should fail because open
with pytest.raises(veilid.VeilidAPIErrorGeneric) as exc:
await api_connection.delete_table_db(TEST_DB)
await tdb2.release()
# delete should now succeed
deleted = await api_connection.delete_table_db(TEST_DB)
assert deleted
@pytest.mark.asyncio
async def test_open_twice_table_db_store_load(api_connection: veilid.VeilidAPI):
# delete test db if it exists
await api_connection.delete_table_db(TEST_DB)
tdb = await api_connection.open_table_db(TEST_DB, 1)
async with tdb:
tdb2 = await api_connection.open_table_db(TEST_DB, 1)
async with tdb2:
# store into first db copy
await tdb.store(b"asdf", b"1234")
# load from second db copy
assert await tdb.load(b"asdf") == b"1234"
# delete should now succeed
deleted = await api_connection.delete_table_db(TEST_DB)
assert deleted
@pytest.mark.asyncio
async def test_open_twice_table_db_store_delete_load(api_connection: veilid.VeilidAPI):
# delete test db if it exists
await api_connection.delete_table_db(TEST_DB)
tdb = await api_connection.open_table_db(TEST_DB, 1)
async with tdb:
tdb2 = await api_connection.open_table_db(TEST_DB, 1)
async with tdb2:
# store into first db copy
await tdb.store(b"asdf", b"1234")
# delete from second db copy and clean up
await tdb2.delete(b"asdf")
# load from first db copy
assert await tdb.load(b"asdf") == None
# delete should now succeed
deleted = await api_connection.delete_table_db(TEST_DB)
assert deleted
@pytest.mark.asyncio
async def test_resize_table_db(api_connection: veilid.VeilidAPI):
# delete test db if it exists
await api_connection.delete_table_db(TEST_DB)
tdb = await api_connection.open_table_db(TEST_DB, 1)
async with tdb:
# reopen the db with more columns should fail if it is already open
with pytest.raises(veilid.VeilidAPIErrorGeneric) as exc:
await api_connection.open_table_db(TEST_DB, 2)
tdb2 = await api_connection.open_table_db(TEST_DB, 2)
async with tdb2:
# write something to second column
await tdb2.store(b"qwer", b"5678", col = 1)
# reopen the db with fewer columns
tdb = await api_connection.open_table_db(TEST_DB, 1)
async with tdb:
# Should fail access to second column
with pytest.raises(veilid.VeilidAPIErrorGeneric) as exc:
await tdb.load(b"qwer", col = 1)
# Should succeed with access to second column
assert await tdb2.load(b"qwer", col = 1) == b"5678"
# now delete should succeed
deleted = await api_connection.delete_table_db(TEST_DB)
assert deleted

View File

@ -7,15 +7,19 @@ from .state import VeilidState
class RoutingContext(ABC): class RoutingContext(ABC):
@abstractmethod @abstractmethod
async def with_privacy(self) -> Self: async def release(self):
pass pass
@abstractmethod @abstractmethod
async def with_custom_privacy(self, stability: types.Stability) -> Self: async def with_privacy(self, release = True) -> Self:
pass pass
@abstractmethod @abstractmethod
async def with_sequencing(self, sequencing: types.Sequencing) -> Self: async def with_custom_privacy(self, stability: types.Stability, release = True) -> Self:
pass
@abstractmethod
async def with_sequencing(self, sequencing: types.Sequencing, release = True) -> Self:
pass pass
@abstractmethod @abstractmethod
@ -89,21 +93,25 @@ class TableDbTransaction(ABC):
pass pass
@abstractmethod @abstractmethod
async def store(self, col: int, key: bytes, value: bytes): async def store(self, key: bytes, value: bytes, col: int = 0):
pass pass
@abstractmethod @abstractmethod
async def delete(self, col: int, key: bytes): async def delete(self, key: bytes, col: int = 0):
pass pass
class TableDb(ABC): class TableDb(ABC):
@abstractmethod
async def release(self):
pass
@abstractmethod @abstractmethod
async def get_column_count(self) -> int: async def get_column_count(self) -> int:
pass pass
@abstractmethod @abstractmethod
async def get_keys(self, col: int) -> list[bytes]: async def get_keys(self, col: int = 0) -> list[bytes]:
pass pass
@abstractmethod @abstractmethod
@ -111,19 +119,23 @@ class TableDb(ABC):
pass pass
@abstractmethod @abstractmethod
async def store(self, col: int, key: bytes, value: bytes): async def store(self, key: bytes, value: bytes, col: int = 0):
pass pass
@abstractmethod @abstractmethod
async def load(self, col: int, key: bytes) -> Optional[bytes]: async def load(self, key: bytes, col: int = 0) -> Optional[bytes]:
pass pass
@abstractmethod @abstractmethod
async def delete(self, col: int, key: bytes) -> Optional[bytes]: async def delete(self, key: bytes, col: int = 0) -> Optional[bytes]:
pass pass
class CryptoSystem(ABC): class CryptoSystem(ABC):
@abstractmethod
async def release(self):
pass
@abstractmethod @abstractmethod
async def cached_dh( async def cached_dh(
self, key: types.PublicKey, secret: types.SecretKey self, key: types.PublicKey, secret: types.SecretKey
@ -284,7 +296,7 @@ class VeilidAPI(ABC):
pass pass
@abstractmethod @abstractmethod
async def delete_table_db(self, name: str): async def delete_table_db(self, name: str) -> bool:
pass pass
@abstractmethod @abstractmethod

View File

@ -320,7 +320,7 @@ class _JsonVeilidAPI(VeilidAPI):
) )
return _JsonTableDb(self, db_id) return _JsonTableDb(self, db_id)
async def delete_table_db(self, name: str): async def delete_table_db(self, name: str) -> bool:
return raise_api_result( return raise_api_result(
await self.send_ndjson_request(Operation.DELETE_TABLE_DB, name=name) await self.send_ndjson_request(Operation.DELETE_TABLE_DB, name=name)
) )
@ -411,19 +411,44 @@ def validate_rc_op(request: dict, response: dict):
class _JsonRoutingContext(RoutingContext): class _JsonRoutingContext(RoutingContext):
api: _JsonVeilidAPI api: _JsonVeilidAPI
rc_id: int rc_id: int
done: bool
def __init__(self, api: _JsonVeilidAPI, rc_id: int): def __init__(self, api: _JsonVeilidAPI, rc_id: int):
self.api = api self.api = api
self.rc_id = rc_id self.rc_id = rc_id
self.done = False
def __del__(self): def __del__(self):
self.api.send_one_way_ndjson_request( if not self.done:
Operation.ROUTING_CONTEXT, # attempt to clean up server-side anyway
rc_id=self.rc_id, self.api.send_one_way_ndjson_request(
rc_op=RoutingContextOperation.RELEASE, Operation.ROUTING_CONTEXT,
) rc_id=self.rc_id,
rc_op=RoutingContextOperation.RELEASE
)
# complain
raise AssertionError("Should have released routing context before dropping object")
async def with_privacy(self) -> Self: async def __aenter__(self) -> Self:
return self
async def __aexit__(self, *excinfo):
if not self.done:
await self.release()
async def release(self):
if self.done:
return
await self.api.send_ndjson_request(
Operation.ROUTING_CONTEXT,
validate=validate_rc_op,
rc_id=self.rc_id,
rc_op=RoutingContextOperation.RELEASE
)
self.done = True
async def with_privacy(self, release = True) -> Self:
new_rc_id = raise_api_result( new_rc_id = raise_api_result(
await self.api.send_ndjson_request( await self.api.send_ndjson_request(
Operation.ROUTING_CONTEXT, Operation.ROUTING_CONTEXT,
@ -432,9 +457,11 @@ class _JsonRoutingContext(RoutingContext):
rc_op=RoutingContextOperation.WITH_PRIVACY, rc_op=RoutingContextOperation.WITH_PRIVACY,
) )
) )
if release:
await self.release()
return self.__class__(self.api, new_rc_id) return self.__class__(self.api, new_rc_id)
async def with_custom_privacy(self, stability: Stability) -> Self: async def with_custom_privacy(self, stability: Stability, release = True) -> Self:
new_rc_id = raise_api_result( new_rc_id = raise_api_result(
await self.api.send_ndjson_request( await self.api.send_ndjson_request(
Operation.ROUTING_CONTEXT, Operation.ROUTING_CONTEXT,
@ -444,9 +471,11 @@ class _JsonRoutingContext(RoutingContext):
stability=stability, stability=stability,
) )
) )
if release:
await self.release()
return self.__class__(self.api, new_rc_id) return self.__class__(self.api, new_rc_id)
async def with_sequencing(self, sequencing: Sequencing) -> Self: async def with_sequencing(self, sequencing: Sequencing, release = True) -> Self:
new_rc_id = raise_api_result( new_rc_id = raise_api_result(
await self.api.send_ndjson_request( await self.api.send_ndjson_request(
Operation.ROUTING_CONTEXT, Operation.ROUTING_CONTEXT,
@ -456,6 +485,8 @@ class _JsonRoutingContext(RoutingContext):
sequencing=sequencing, sequencing=sequencing,
) )
) )
if release:
await self.release()
return self.__class__(self.api, new_rc_id) return self.__class__(self.api, new_rc_id)
async def app_call(self, target: TypedKey | RouteId, request: bytes) -> bytes: async def app_call(self, target: TypedKey | RouteId, request: bytes) -> bytes:
@ -627,9 +658,27 @@ class _JsonTableDbTransaction(TableDbTransaction):
def __del__(self): def __del__(self):
if not self.done: if not self.done:
raise AssertionError("Should have committed or rolled back transaction") # attempt to clean up server-side anyway
self.api.send_one_way_ndjson_request(
Operation.TABLE_DB_TRANSACTION,
tx_id=self.tx_id,
tx_op=TableDbTransactionOperation.ROLLBACK,
)
# complain
raise AssertionError("Should have committed or rolled back transaction before dropping object")
async def __aenter__(self) -> Self:
return self
async def __aexit__(self, *excinfo):
if not self.done:
await self.rollback()
async def commit(self): async def commit(self):
if self.done:
raise AssertionError("Transaction is already done")
raise_api_result( raise_api_result(
await self.api.send_ndjson_request( await self.api.send_ndjson_request(
Operation.TABLE_DB_TRANSACTION, Operation.TABLE_DB_TRANSACTION,
@ -641,17 +690,17 @@ class _JsonTableDbTransaction(TableDbTransaction):
self.done = True self.done = True
async def rollback(self): async def rollback(self):
raise_api_result( if self.done:
await self.api.send_ndjson_request( raise AssertionError("Transaction is already done")
Operation.TABLE_DB_TRANSACTION, await self.api.send_ndjson_request(
validate=validate_tx_op, Operation.TABLE_DB_TRANSACTION,
tx_id=self.tx_id, validate=validate_tx_op,
tx_op=TableDbTransactionOperation.ROLLBACK, tx_id=self.tx_id,
) tx_op=TableDbTransactionOperation.ROLLBACK,
) )
self.done = True self.done = True
async def store(self, col: int, key: bytes, value: bytes): async def store(self, key: bytes, value: bytes, col: int = 0):
await self.api.send_ndjson_request( await self.api.send_ndjson_request(
Operation.TABLE_DB_TRANSACTION, Operation.TABLE_DB_TRANSACTION,
validate=validate_tx_op, validate=validate_tx_op,
@ -662,7 +711,7 @@ class _JsonTableDbTransaction(TableDbTransaction):
value=value, value=value,
) )
async def delete(self, col: int, key: bytes): async def delete(self, key: bytes, col: int = 0):
await self.api.send_ndjson_request( await self.api.send_ndjson_request(
Operation.TABLE_DB_TRANSACTION, Operation.TABLE_DB_TRANSACTION,
validate=validate_tx_op, validate=validate_tx_op,
@ -684,15 +733,44 @@ def validate_db_op(request: dict, response: dict):
class _JsonTableDb(TableDb): class _JsonTableDb(TableDb):
api: _JsonVeilidAPI api: _JsonVeilidAPI
db_id: int db_id: int
done: bool
def __init__(self, api: _JsonVeilidAPI, db_id: int): def __init__(self, api: _JsonVeilidAPI, db_id: int):
self.api = api self.api = api
self.db_id = db_id self.db_id = db_id
self.done = False
def __del__(self): def __del__(self):
self.api.send_one_way_ndjson_request( if not self.done:
Operation.TABLE_DB, db_id=self.db_id, rc_op=TableDbOperation.RELEASE
# attempt to clean up server-side anyway
self.api.send_one_way_ndjson_request(
Operation.TABLE_DB,
db_id=self.db_id,
db_op=TableDbOperation.RELEASE
)
# complain
raise AssertionError("Should have released table db before dropping object")
async def __aenter__(self) -> Self:
return self
async def __aexit__(self, *excinfo):
if not self.done:
await self.release()
async def release(self):
if self.done:
return
await self.api.send_ndjson_request(
Operation.TABLE_DB,
validate=validate_db_op,
db_id=self.db_id,
db_op=TableDbOperation.RELEASE
) )
self.done = True
async def get_column_count(self) -> int: async def get_column_count(self) -> int:
return raise_api_result( return raise_api_result(
@ -704,7 +782,7 @@ class _JsonTableDb(TableDb):
) )
) )
async def get_keys(self, col: int) -> list[bytes]: async def get_keys(self, col: int = 0) -> list[bytes]:
return list( return list(
map( map(
lambda x: urlsafe_b64decode_no_pad(x), lambda x: urlsafe_b64decode_no_pad(x),
@ -731,7 +809,7 @@ class _JsonTableDb(TableDb):
) )
return _JsonTableDbTransaction(self.api, tx_id) return _JsonTableDbTransaction(self.api, tx_id)
async def store(self, col: int, key: bytes, value: bytes): async def store(self, key: bytes, value: bytes, col: int = 0):
return raise_api_result( return raise_api_result(
await self.api.send_ndjson_request( await self.api.send_ndjson_request(
Operation.TABLE_DB, Operation.TABLE_DB,
@ -744,7 +822,7 @@ class _JsonTableDb(TableDb):
) )
) )
async def load(self, col: int, key: bytes) -> Optional[bytes]: async def load(self, key: bytes, col: int = 0) -> Optional[bytes]:
res = raise_api_result( res = raise_api_result(
await self.api.send_ndjson_request( await self.api.send_ndjson_request(
Operation.TABLE_DB, Operation.TABLE_DB,
@ -757,7 +835,7 @@ class _JsonTableDb(TableDb):
) )
return None if res is None else urlsafe_b64decode_no_pad(res) return None if res is None else urlsafe_b64decode_no_pad(res)
async def delete(self, col: int, key: bytes) -> Optional[bytes]: async def delete(self, key: bytes, col: int = 0) -> Optional[bytes]:
res = raise_api_result( res = raise_api_result(
await self.api.send_ndjson_request( await self.api.send_ndjson_request(
Operation.TABLE_DB, Operation.TABLE_DB,
@ -782,17 +860,43 @@ def validate_cs_op(request: dict, response: dict):
class _JsonCryptoSystem(CryptoSystem): class _JsonCryptoSystem(CryptoSystem):
api: _JsonVeilidAPI api: _JsonVeilidAPI
cs_id: int cs_id: int
done: bool
def __init__(self, api: _JsonVeilidAPI, cs_id: int): def __init__(self, api: _JsonVeilidAPI, cs_id: int):
self.api = api self.api = api
self.cs_id = cs_id self.cs_id = cs_id
self.done = False
def __del__(self): def __del__(self):
self.api.send_one_way_ndjson_request( if not self.done:
# attempt to clean up server-side anyway
self.api.send_one_way_ndjson_request(
Operation.CRYPTO_SYSTEM,
cs_id=self.cs_id,
cs_op=CryptoSystemOperation.RELEASE
)
# complain
raise AssertionError("Should have released crypto system before dropping object")
async def __aenter__(self) -> Self:
return self
async def __aexit__(self, *excinfo):
if not self.done:
await self.release()
async def release(self):
if self.done:
return
await self.api.send_ndjson_request(
Operation.CRYPTO_SYSTEM, Operation.CRYPTO_SYSTEM,
validate=validate_cs_op,
cs_id=self.cs_id, cs_id=self.cs_id,
cs_op=CryptoSystemOperation.RELEASE, cs_op=CryptoSystemOperation.RELEASE
) )
self.done = True
async def cached_dh(self, key: PublicKey, secret: SecretKey) -> SharedSecret: async def cached_dh(self, key: PublicKey, secret: SecretKey) -> SharedSecret:
return SharedSecret( return SharedSecret(

View File

@ -1319,6 +1319,30 @@
}, },
{ {
"type": "object", "type": "object",
"anyOf": [
{
"type": "object",
"required": [
"value"
],
"properties": {
"value": {
"type": "null"
}
}
},
{
"type": "object",
"required": [
"error"
],
"properties": {
"error": {
"$ref": "#/definitions/VeilidAPIError"
}
}
}
],
"required": [ "required": [
"tx_op" "tx_op"
], ],
@ -1333,6 +1357,30 @@
}, },
{ {
"type": "object", "type": "object",
"anyOf": [
{
"type": "object",
"required": [
"value"
],
"properties": {
"value": {
"type": "null"
}
}
},
{
"type": "object",
"required": [
"error"
],
"properties": {
"error": {
"$ref": "#/definitions/VeilidAPIError"
}
}
}
],
"required": [ "required": [
"tx_op" "tx_op"
], ],

View File

@ -3,7 +3,7 @@ from typing import Optional, Self
from .config import VeilidConfig from .config import VeilidConfig
from .types import (ByteCount, RouteId, Timestamp, TimestampDuration, TypedKey, from .types import (ByteCount, RouteId, Timestamp, TimestampDuration, TypedKey,
ValueData, ValueSubkey, VeilidLogLevel, ValueData, ValueSubkey, VeilidLogLevel, OperationId,
urlsafe_b64decode_no_pad) urlsafe_b64decode_no_pad)
@ -309,9 +309,9 @@ class VeilidAppMessage:
class VeilidAppCall: class VeilidAppCall:
sender: Optional[TypedKey] sender: Optional[TypedKey]
message: bytes message: bytes
call_id: str call_id: OperationId
def __init__(self, sender: Optional[TypedKey], message: bytes, call_id: str): def __init__(self, sender: Optional[TypedKey], message: bytes, call_id: OperationId):
self.sender = sender self.sender = sender
self.message = message self.message = message
self.call_id = call_id self.call_id = call_id
@ -322,7 +322,7 @@ class VeilidAppCall:
return cls( return cls(
None if j["sender"] is None else TypedKey(j["sender"]), None if j["sender"] is None else TypedKey(j["sender"]),
urlsafe_b64decode_no_pad(j["message"]), urlsafe_b64decode_no_pad(j["message"]),
j["call_id"], OperationId(j["call_id"]),
) )

View File

@ -817,7 +817,7 @@ pub fn table_db_transaction_store(id: u32, col: u32, key: String, value: String)
tdbt.clone() tdbt.clone()
}; };
tdbt.store(col, &key, &value); tdbt.store(col, &key, &value)?;
APIRESULT_UNDEFINED APIRESULT_UNDEFINED
}) })
} }
@ -836,7 +836,7 @@ pub fn table_db_transaction_delete(id: u32, col: u32, key: String) -> Promise {
tdbt.clone() tdbt.clone()
}; };
tdbt.delete(col, &key); tdbt.delete(col, &key)?;
APIRESULT_UNDEFINED APIRESULT_UNDEFINED
}) })
} }