1 Commits

Author SHA1 Message Date
CleverWild
5a34463228 security(server): bind grant revocation state (revoked_at) to integrity hash
Some checks failed
ci/woodpecker/pr/server-audit Pipeline was successful
ci/woodpecker/pr/server-lint Pipeline was successful
ci/woodpecker/pr/server-vet Pipeline failed
ci/woodpecker/pr/server-test Pipeline was successful
2026-04-08 12:09:54 +02:00
69 changed files with 1001 additions and 1490 deletions

View File

@@ -100,27 +100,6 @@ diesel migration generate <name> --migration-dir crates/arbiter-server/migration
diesel migration run --migration-dir crates/arbiter-server/migrations diesel migration run --migration-dir crates/arbiter-server/migrations
``` ```
### Code Conventions
**`#[must_use]` Attribute:**
Apply the `#[must_use]` attribute to return types of functions where the return value is critical and should not be accidentally ignored. This is commonly used for:
- Methods that return `bool` indicating success/failure or validation state
- Any function where ignoring the return value indicates a logic error
Do not apply `#[must_use]` redundantly to items (types or functions) that are already annotated with `#[must_use]`.
Example:
```rust
#[must_use]
pub fn verify(&self, nonce: i32, context: &[u8], signature: &Signature) -> bool {
// verification logic
}
```
This forces callers to either use the return value or explicitly ignore it with `let _ = ...;`, preventing silent failures.
## User Agent (Flutter + Rinf at `useragent/`) ## User Agent (Flutter + Rinf at `useragent/`)
The Flutter app uses [Rinf](https://rinf.cunarist.org) to call Rust code. The Rust logic lives in `useragent/native/hub/` as a separate crate that uses `arbiter-useragent` for the gRPC client. The Flutter app uses [Rinf](https://rinf.cunarist.org) to call Rust code. The Rust logic lives in `useragent/native/hub/` as a separate crate that uses `arbiter-useragent` for the gRPC client.

View File

@@ -100,27 +100,6 @@ diesel migration generate <name> --migration-dir crates/arbiter-server/migration
diesel migration run --migration-dir crates/arbiter-server/migrations diesel migration run --migration-dir crates/arbiter-server/migrations
``` ```
### Code Conventions
**`#[must_use]` Attribute:**
Apply the `#[must_use]` attribute to return types of functions where the return value is critical and should not be accidentally ignored. This is commonly used for:
- Methods that return `bool` indicating success/failure or validation state
- Any function where ignoring the return value indicates a logic error
Do not apply `#[must_use]` redundantly to items (types or functions) that are already annotated with `#[must_use]`.
Example:
```rust
#[must_use]
pub fn verify(&self, nonce: i32, context: &[u8], signature: &Signature) -> bool {
// verification logic
}
```
This forces callers to either use the return value or explicitly ignore it with `let _ = ...;`, preventing silent failures.
## User Agent (Flutter + Rinf at `useragent/`) ## User Agent (Flutter + Rinf at `useragent/`)
The Flutter app uses [Rinf](https://rinf.cunarist.org) to call Rust code. The Rust logic lives in `useragent/native/hub/` as a separate crate that uses `arbiter-useragent` for the gRPC client. The Flutter app uses [Rinf](https://rinf.cunarist.org) to call Rust code. The Rust logic lives in `useragent/native/hub/` as a separate crate that uses `arbiter-useragent` for the gRPC client.

676
server/Cargo.lock generated

File diff suppressed because it is too large Load Diff

View File

@@ -4,169 +4,47 @@ members = [
] ]
resolver = "3" resolver = "3"
[workspace.lints.clippy]
disallowed-methods = "deny"
[workspace.dependencies] [workspace.dependencies]
alloy = "1.7.3" tonic = { version = "0.14.5", features = [
async-trait = "0.1.89" "deflate",
base64 = "0.22.1" "gzip",
chrono = { version = "0.4.44", features = ["serde"] } "tls-connect-info",
ed25519-dalek = { version = "3.0.0-pre.6", features = ["rand_core"] } "zstd",
futures = "0.3.32" ] }
hmac = "0.12.1"
k256 = { version = "0.13.4", features = ["ecdsa", "pkcs8"] }
kameo = { version = "0.20.0", git = "https://github.com/tqwewe/kameo" } # hold this until new patch version is released
miette = { version = "7.6.0", features = ["fancy", "serde"] }
ml-dsa = { version = "0.1.0-rc.8", features = ["zeroize"] }
mutants = "0.0.4"
prost = "0.14.3"
prost-types = { version = "0.14.3", features = ["chrono"] }
rand = "0.10.0"
rcgen = { version = "0.14.7", features = [ "aws_lc_rs", "pem", "x509-parser", "zeroize" ], default-features = false }
rsa = { version = "0.9", features = ["sha2"] }
rstest = "0.26.1"
rustls = { version = "0.23.37", features = ["aws-lc-rs", "logging", "prefer-post-quantum", "std"], default-features = false }
rustls-pki-types = "1.14.0"
sha2 = "0.10"
smlang = "0.8.0"
spki = "0.7"
thiserror = "2.0.18"
tokio = { version = "1.50.0", features = ["full"] }
tokio-stream = { version = "0.1.18", features = ["full"] }
tonic = { version = "0.14.5", features = [ "deflate", "gzip", "tls-connect-info", "zstd" ] }
tracing = "0.1.44" tracing = "0.1.44"
tokio = { version = "1.50.0", features = ["full"] }
ed25519-dalek = { version = "3.0.0-pre.6", features = ["rand_core"] }
chrono = { version = "0.4.44", features = ["serde"] }
rand = "0.10.0"
rustls = { version = "0.23.37", features = ["aws-lc-rs", "logging", "prefer-post-quantum", "std"], default-features = false }
smlang = "0.8.0"
thiserror = "2.0.18"
async-trait = "0.1.89"
futures = "0.3.32"
tokio-stream = { version = "0.1.18", features = ["full"] }
kameo = "0.19.2"
prost-types = { version = "0.14.3", features = ["chrono"] }
x25519-dalek = { version = "2.0.1", features = ["getrandom"] } x25519-dalek = { version = "2.0.1", features = ["getrandom"] }
rstest = "0.26.1"
[workspace.lints.rust] rustls-pki-types = "1.14.0"
missing_unsafe_on_extern = "deny" alloy = "1.7.3"
unsafe_attr_outside_unsafe = "deny" rcgen = { version = "0.14.7", features = [
unsafe_op_in_unsafe_fn = "deny" "aws_lc_rs",
unstable_features = "deny" "pem",
"x509-parser",
deprecated_safe_2024 = "warn" "zeroize",
ffi_unwind_calls = "warn" ], default-features = false }
linker_messages = "warn" k256 = { version = "0.13.4", features = ["ecdsa", "pkcs8"] }
rsa = { version = "0.9", features = ["sha2"] }
elided_lifetimes_in_paths = "warn" sha2 = "0.10"
explicit_outlives_requirements = "warn" spki = "0.7"
impl-trait-overcaptures = "warn" prost = "0.14.3"
impl-trait-redundant-captures = "warn" miette = { version = "7.6.0", features = ["fancy", "serde"] }
redundant_lifetimes = "warn" mutants = "0.0.4"
single_use_lifetimes = "warn" ml-dsa = { version = "0.1.0-rc.8", features = ["zeroize"] }
unused_lifetimes = "warn" base64 = "0.22.1"
hmac = "0.12.1"
macro_use_extern_crate = "warn"
redundant_imports = "warn"
unused_import_braces = "warn"
unused_macro_rules = "warn"
unused_qualifications = "warn"
unit_bindings = "warn"
# missing_docs = "warn" # ENABLE BY THE FIRST MAJOR VERSION!!
unnameable_types = "warn"
variant_size_differences = "warn"
[workspace.lints.clippy]
derive_partial_eq_without_eq = "allow"
future_not_send = "allow"
inconsistent_struct_constructor = "allow"
inline_always = "allow"
missing_errors_doc = "allow"
missing_fields_in_debug = "allow"
missing_panics_doc = "allow"
must_use_candidate = "allow"
needless_pass_by_ref_mut = "allow"
pub_underscore_fields = "allow"
redundant_pub_crate = "allow"
uninhabited_references = "allow" # safe with unsafe_code = "forbid" and standard uninhabited pattern (match *self {})
# restriction lints
alloc_instead_of_core = "warn"
allow_attributes_without_reason = "warn"
as_conversions = "warn"
assertions_on_result_states = "warn"
cfg_not_test = "warn"
clone_on_ref_ptr = "warn"
cognitive_complexity = "warn"
create_dir = "warn"
dbg_macro = "warn"
decimal_literal_representation = "warn"
default_union_representation = "warn"
deref_by_slicing = "warn"
disallowed_script_idents = "warn"
doc_include_without_cfg = "warn"
empty_drop = "warn"
empty_enum_variants_with_brackets = "warn"
empty_structs_with_brackets = "warn"
error_impl_error = "warn"
exit = "warn"
filetype_is_file = "warn"
float_arithmetic = "warn"
float_cmp_const = "warn"
fn_to_numeric_cast_any = "warn"
get_unwrap = "warn"
if_then_some_else_none = "warn"
indexing_slicing = "warn"
infinite_loop = "warn"
inline_asm_x86_att_syntax = "warn"
inline_asm_x86_intel_syntax = "warn"
integer_division = "warn"
large_include_file = "warn"
lossy_float_literal = "warn"
map_with_unused_argument_over_ranges = "warn"
mem_forget = "warn"
missing_assert_message = "warn"
mixed_read_write_in_expression = "warn"
modulo_arithmetic = "warn"
multiple_unsafe_ops_per_block = "warn"
mutex_atomic = "warn"
mutex_integer = "warn"
needless_raw_strings = "warn"
non_ascii_literal = "warn"
non_zero_suggestions = "warn"
pathbuf_init_then_push = "warn"
pointer_format = "warn"
precedence_bits = "warn"
pub_without_shorthand = "warn"
rc_buffer = "warn"
rc_mutex = "warn"
redundant_test_prefix = "warn"
redundant_type_annotations = "warn"
ref_patterns = "warn"
renamed_function_params = "warn"
rest_pat_in_fully_bound_structs = "warn"
return_and_then = "warn"
semicolon_inside_block = "warn"
str_to_string = "warn"
string_add = "warn"
string_lit_chars_any = "warn"
string_slice = "warn"
suspicious_xor_used_as_pow = "warn"
try_err = "warn"
undocumented_unsafe_blocks = "warn"
uninlined_format_args = "warn"
unnecessary_safety_comment = "warn"
unnecessary_safety_doc = "warn"
unnecessary_self_imports = "warn"
unneeded_field_pattern = "warn"
unused_result_ok = "warn"
verbose_file_reads = "warn"
# cargo lints
negative_feature_names = "warn"
redundant_feature_names = "warn"
wildcard_dependencies = "warn"
# ENABLE BY THE FIRST MAJOR VERSION!!
# todo = "warn"
# unimplemented = "warn"
# panic = "warn"
# panic_in_result_fn = "warn"
#
# cargo_common_metadata = "warn"
# multiple_crate_versions = "warn" # a controversial option since it's really difficult to maintain
disallowed_methods = "deny"
nursery = { level = "warn", priority = -1 }
pedantic = { level = "warn", priority = -1 }

View File

@@ -7,22 +7,3 @@ disallowed-methods = [
{ path = "rsa::traits::Decryptor::decrypt", reason = "RSA decryption is forbidden (RUSTSEC-2023-0071 Marvin Attack). This blocks decrypt() on rsa::{pkcs1v15,oaep}::DecryptingKey." }, { path = "rsa::traits::Decryptor::decrypt", reason = "RSA decryption is forbidden (RUSTSEC-2023-0071 Marvin Attack). This blocks decrypt() on rsa::{pkcs1v15,oaep}::DecryptingKey." },
{ path = "rsa::traits::RandomizedDecryptor::decrypt_with_rng", reason = "RSA decryption is forbidden (RUSTSEC-2023-0071 Marvin Attack). This blocks decrypt_with_rng() on rsa::{pkcs1v15,oaep}::DecryptingKey." }, { path = "rsa::traits::RandomizedDecryptor::decrypt_with_rng", reason = "RSA decryption is forbidden (RUSTSEC-2023-0071 Marvin Attack). This blocks decrypt_with_rng() on rsa::{pkcs1v15,oaep}::DecryptingKey." },
] ]
allow-indexing-slicing-in-tests = true
allow-panic-in-tests = true
check-inconsistent-struct-field-initializers = true
suppress-restriction-lint-in-const = true
allow-renamed-params-for = [
"core::convert::From",
"core::convert::TryFrom",
"core::str::FromStr",
"kameo::actor::Actor",
]
module-items-ordered-within-groupings = ["UPPER_SNAKE_CASE"]
source-item-ordering = ["enum"]
trait-assoc-item-kinds-order = [
"const",
"type",
"fn",
] # community tested standard

View File

@@ -23,20 +23,20 @@ use crate::{
#[derive(Debug, thiserror::Error)] #[derive(Debug, thiserror::Error)]
pub enum AuthError { pub enum AuthError {
#[error("Client approval denied by User Agent")]
ApprovalDenied,
#[error("Auth challenge was not returned by server")] #[error("Auth challenge was not returned by server")]
MissingAuthChallenge, MissingAuthChallenge,
#[error("Client approval denied by User Agent")]
ApprovalDenied,
#[error("No User Agents online to approve client")] #[error("No User Agents online to approve client")]
NoUserAgentsOnline, NoUserAgentsOnline,
#[error("Signing key storage error")]
Storage(#[from] StorageError),
#[error("Unexpected auth response payload")] #[error("Unexpected auth response payload")]
UnexpectedAuthResponse, UnexpectedAuthResponse,
#[error("Signing key storage error")]
Storage(#[from] StorageError),
} }
fn map_auth_result(code: i32) -> AuthError { fn map_auth_result(code: i32) -> AuthError {
@@ -55,7 +55,7 @@ async fn send_auth_challenge_request(
transport: &mut ClientTransport, transport: &mut ClientTransport,
metadata: ClientMetadata, metadata: ClientMetadata,
key: &SigningKey, key: &SigningKey,
) -> Result<(), AuthError> { ) -> std::result::Result<(), AuthError> {
transport transport
.send(ClientRequest { .send(ClientRequest {
request_id: next_request_id(), request_id: next_request_id(),
@@ -76,7 +76,7 @@ async fn send_auth_challenge_request(
async fn receive_auth_challenge( async fn receive_auth_challenge(
transport: &mut ClientTransport, transport: &mut ClientTransport,
) -> Result<AuthChallenge, AuthError> { ) -> std::result::Result<AuthChallenge, AuthError> {
let response = transport let response = transport
.recv() .recv()
.await .await
@@ -97,7 +97,7 @@ async fn send_auth_challenge_solution(
transport: &mut ClientTransport, transport: &mut ClientTransport,
key: &SigningKey, key: &SigningKey,
challenge: AuthChallenge, challenge: AuthChallenge,
) -> Result<(), AuthError> { ) -> std::result::Result<(), AuthError> {
let challenge_payload = format_challenge(challenge.nonce, &challenge.pubkey); let challenge_payload = format_challenge(challenge.nonce, &challenge.pubkey);
let signature = key let signature = key
.sign_message(&challenge_payload, CLIENT_CONTEXT) .sign_message(&challenge_payload, CLIENT_CONTEXT)
@@ -117,7 +117,9 @@ async fn send_auth_challenge_solution(
.map_err(|_| AuthError::UnexpectedAuthResponse) .map_err(|_| AuthError::UnexpectedAuthResponse)
} }
async fn receive_auth_confirmation(transport: &mut ClientTransport) -> Result<(), AuthError> { async fn receive_auth_confirmation(
transport: &mut ClientTransport,
) -> std::result::Result<(), AuthError> {
let response = transport let response = transport
.recv() .recv()
.await .await
@@ -138,11 +140,11 @@ async fn receive_auth_confirmation(transport: &mut ClientTransport) -> Result<()
} }
} }
pub async fn authenticate( pub(crate) async fn authenticate(
transport: &mut ClientTransport, transport: &mut ClientTransport,
metadata: ClientMetadata, metadata: ClientMetadata,
key: &SigningKey, key: &SigningKey,
) -> Result<(), AuthError> { ) -> std::result::Result<(), AuthError> {
send_auth_challenge_request(transport, metadata, key).await?; send_auth_challenge_request(transport, metadata, key).await?;
let challenge = receive_auth_challenge(transport).await?; let challenge = receive_auth_challenge(transport).await?;
send_auth_challenge_solution(transport, key, challenge).await?; send_auth_challenge_solution(transport, key, challenge).await?;

View File

@@ -29,16 +29,16 @@ async fn main() {
} }
}; };
println!("{url:#?}"); println!("{:#?}", url);
let metadata = ClientMetadata { let metadata = ClientMetadata {
name: "arbiter-client test_connect".to_owned(), name: "arbiter-client test_connect".to_string(),
description: Some("Manual connection smoke test".to_owned()), description: Some("Manual connection smoke test".to_string()),
version: Some(env!("CARGO_PKG_VERSION").to_owned()), version: Some(env!("CARGO_PKG_VERSION").to_string()),
}; };
match ArbiterClient::connect(url, metadata).await { match ArbiterClient::connect(url, metadata).await {
Ok(_) => println!("Connected and authenticated successfully."), Ok(_) => println!("Connected and authenticated successfully."),
Err(err) => eprintln!("Failed to connect: {err:#?}"), Err(err) => eprintln!("Failed to connect: {:#?}", err),
} }
} }

View File

@@ -18,39 +18,33 @@ use crate::{
use crate::wallets::evm::ArbiterEvmWallet; use crate::wallets::evm::ArbiterEvmWallet;
#[derive(Debug, thiserror::Error)] #[derive(Debug, thiserror::Error)]
pub enum ArbiterClientError { pub enum Error {
#[error("Authentication error")] #[error("gRPC error")]
Authentication(#[from] AuthError), Grpc(#[from] tonic::Status),
#[error("Could not establish connection")] #[error("Could not establish connection")]
Connection(#[from] tonic::transport::Error), Connection(#[from] tonic::transport::Error),
#[error("gRPC error")] #[error("Invalid server URI")]
Grpc(#[from] tonic::Status), InvalidUri(#[from] http::uri::InvalidUri),
#[error("Invalid CA certificate")] #[error("Invalid CA certificate")]
InvalidCaCert(#[from] webpki::Error), InvalidCaCert(#[from] webpki::Error),
#[error("Invalid server URI")] #[error("Authentication error")]
InvalidUri(#[from] http::uri::InvalidUri), Authentication(#[from] AuthError),
#[error("Storage error")] #[error("Storage error")]
Storage(#[from] StorageError), Storage(#[from] StorageError),
} }
pub struct ArbiterClient { pub struct ArbiterClient {
#[expect( #[allow(dead_code)]
dead_code,
reason = "transport will be used in future methods for sending requests and receiving responses"
)]
transport: Arc<Mutex<ClientTransport>>, transport: Arc<Mutex<ClientTransport>>,
} }
impl ArbiterClient { impl ArbiterClient {
pub async fn connect( pub async fn connect(url: ArbiterUrl, metadata: ClientMetadata) -> Result<Self, Error> {
url: ArbiterUrl,
metadata: ClientMetadata,
) -> Result<Self, ArbiterClientError> {
let storage = FileSigningKeyStorage::from_default_location()?; let storage = FileSigningKeyStorage::from_default_location()?;
Self::connect_with_storage(url, metadata, &storage).await Self::connect_with_storage(url, metadata, &storage).await
} }
@@ -59,7 +53,7 @@ impl ArbiterClient {
url: ArbiterUrl, url: ArbiterUrl,
metadata: ClientMetadata, metadata: ClientMetadata,
storage: &S, storage: &S,
) -> Result<Self, ArbiterClientError> { ) -> Result<Self, Error> {
let key = storage.load_or_create()?; let key = storage.load_or_create()?;
Self::connect_with_key(url, metadata, key).await Self::connect_with_key(url, metadata, key).await
} }
@@ -68,7 +62,7 @@ impl ArbiterClient {
url: ArbiterUrl, url: ArbiterUrl,
metadata: ClientMetadata, metadata: ClientMetadata,
key: SigningKey, key: SigningKey,
) -> Result<Self, ArbiterClientError> { ) -> Result<Self, Error> {
let anchor = webpki::anchor_from_trusted_cert(&url.ca_cert)?.to_owned(); let anchor = webpki::anchor_from_trusted_cert(&url.ca_cert)?.to_owned();
let tls = ClientTlsConfig::new().trust_anchor(anchor); let tls = ClientTlsConfig::new().trust_anchor(anchor);
@@ -95,8 +89,7 @@ impl ArbiterClient {
} }
#[cfg(feature = "evm")] #[cfg(feature = "evm")]
#[expect(clippy::unused_async, reason = "false positive")] pub async fn evm_wallets(&self) -> Result<Vec<ArbiterEvmWallet>, Error> {
pub async fn evm_wallets(&self) -> Result<Vec<ArbiterEvmWallet>, ArbiterClientError> {
todo!("fetch EVM wallet list from server") todo!("fetch EVM wallet list from server")
} }
} }

View File

@@ -5,7 +5,7 @@ mod transport;
pub mod wallets; pub mod wallets;
pub use auth::AuthError; pub use auth::AuthError;
pub use client::{ArbiterClient, ArbiterClientError}; pub use client::{ArbiterClient, Error};
pub use storage::{FileSigningKeyStorage, SigningKeyStorage, StorageError}; pub use storage::{FileSigningKeyStorage, SigningKeyStorage, StorageError};
#[cfg(feature = "evm")] #[cfg(feature = "evm")]

View File

@@ -4,15 +4,15 @@ use std::path::{Path, PathBuf};
#[derive(Debug, thiserror::Error)] #[derive(Debug, thiserror::Error)]
pub enum StorageError { pub enum StorageError {
#[error("Invalid signing key length in storage: expected {expected} bytes, got {actual} bytes")]
InvalidKeyLength { expected: usize, actual: usize },
#[error("I/O error")] #[error("I/O error")]
Io(#[from] std::io::Error), Io(#[from] std::io::Error),
#[error("Invalid signing key length in storage: expected {expected} bytes, got {actual} bytes")]
InvalidKeyLength { expected: usize, actual: usize },
} }
pub trait SigningKeyStorage { pub trait SigningKeyStorage {
fn load_or_create(&self) -> Result<SigningKey, StorageError>; fn load_or_create(&self) -> std::result::Result<SigningKey, StorageError>;
} }
#[derive(Debug, Clone)] #[derive(Debug, Clone)]
@@ -27,11 +27,11 @@ impl FileSigningKeyStorage {
Self { path: path.into() } Self { path: path.into() }
} }
pub fn from_default_location() -> Result<Self, StorageError> { pub fn from_default_location() -> std::result::Result<Self, StorageError> {
Ok(Self::new(home_path()?.join(Self::DEFAULT_FILE_NAME))) Ok(Self::new(home_path()?.join(Self::DEFAULT_FILE_NAME)))
} }
fn read_key(path: &Path) -> Result<SigningKey, StorageError> { fn read_key(path: &Path) -> std::result::Result<SigningKey, StorageError> {
let bytes = std::fs::read(path)?; let bytes = std::fs::read(path)?;
let raw: [u8; 32] = let raw: [u8; 32] =
bytes bytes
@@ -45,7 +45,7 @@ impl FileSigningKeyStorage {
} }
impl SigningKeyStorage for FileSigningKeyStorage { impl SigningKeyStorage for FileSigningKeyStorage {
fn load_or_create(&self) -> Result<SigningKey, StorageError> { fn load_or_create(&self) -> std::result::Result<SigningKey, StorageError> {
if let Some(parent) = self.path.parent() { if let Some(parent) = self.path.parent() {
std::fs::create_dir_all(parent)?; std::fs::create_dir_all(parent)?;
} }
@@ -125,7 +125,7 @@ mod tests {
assert_eq!(expected, 32); assert_eq!(expected, 32);
assert_eq!(actual, 31); assert_eq!(actual, 31);
} }
other @ StorageError::Io(_) => panic!("unexpected error: {other:?}"), other => panic!("unexpected error: {other:?}"),
} }
std::fs::remove_file(path).expect("temp key file should be removable"); std::fs::remove_file(path).expect("temp key file should be removable");

View File

@@ -2,15 +2,15 @@ use arbiter_proto::proto::client::{ClientRequest, ClientResponse};
use std::sync::atomic::{AtomicI32, Ordering}; use std::sync::atomic::{AtomicI32, Ordering};
use tokio::sync::mpsc; use tokio::sync::mpsc;
pub const BUFFER_LENGTH: usize = 16; pub(crate) const BUFFER_LENGTH: usize = 16;
static NEXT_REQUEST_ID: AtomicI32 = AtomicI32::new(1); static NEXT_REQUEST_ID: AtomicI32 = AtomicI32::new(1);
pub fn next_request_id() -> i32 { pub(crate) fn next_request_id() -> i32 {
NEXT_REQUEST_ID.fetch_add(1, Ordering::Relaxed) NEXT_REQUEST_ID.fetch_add(1, Ordering::Relaxed)
} }
#[derive(Debug, thiserror::Error)] #[derive(Debug, thiserror::Error)]
pub enum ClientSignError { pub(crate) enum ClientSignError {
#[error("Transport channel closed")] #[error("Transport channel closed")]
ChannelClosed, ChannelClosed,
@@ -18,7 +18,7 @@ pub enum ClientSignError {
ConnectionClosed, ConnectionClosed,
} }
pub struct ClientTransport { pub(crate) struct ClientTransport {
pub(crate) sender: mpsc::Sender<ClientRequest>, pub(crate) sender: mpsc::Sender<ClientRequest>,
pub(crate) receiver: tonic::Streaming<ClientResponse>, pub(crate) receiver: tonic::Streaming<ClientResponse>,
} }
@@ -27,17 +27,18 @@ impl ClientTransport {
pub(crate) async fn send( pub(crate) async fn send(
&mut self, &mut self,
request: ClientRequest, request: ClientRequest,
) -> Result<(), ClientSignError> { ) -> std::result::Result<(), ClientSignError> {
self.sender self.sender
.send(request) .send(request)
.await .await
.map_err(|_| ClientSignError::ChannelClosed) .map_err(|_| ClientSignError::ChannelClosed)
} }
pub(crate) async fn recv(&mut self) -> Result<ClientResponse, ClientSignError> { pub(crate) async fn recv(&mut self) -> std::result::Result<ClientResponse, ClientSignError> {
match self.receiver.message().await { match self.receiver.message().await {
Ok(Some(resp)) => Ok(resp), Ok(Some(resp)) => Ok(resp),
Ok(None) | Err(_) => Err(ClientSignError::ConnectionClosed), Ok(None) => Err(ClientSignError::ConnectionClosed),
Err(_) => Err(ClientSignError::ConnectionClosed),
} }
} }
} }

View File

@@ -59,11 +59,7 @@ pub struct ArbiterEvmWallet {
} }
impl ArbiterEvmWallet { impl ArbiterEvmWallet {
#[expect( pub(crate) fn new(transport: Arc<Mutex<ClientTransport>>, address: Address) -> Self {
dead_code,
reason = "new will be used in future methods for creating wallets with different parameters"
)]
pub(crate) const fn new(transport: Arc<Mutex<ClientTransport>>, address: Address) -> Self {
Self { Self {
transport, transport,
address, address,
@@ -71,12 +67,11 @@ impl ArbiterEvmWallet {
} }
} }
pub const fn address(&self) -> Address { pub fn address(&self) -> Address {
self.address self.address
} }
#[must_use] pub fn with_chain_id(mut self, chain_id: ChainId) -> Self {
pub const fn with_chain_id(mut self, chain_id: ChainId) -> Self {
self.chain_id = Some(chain_id); self.chain_id = Some(chain_id);
self self
} }
@@ -151,7 +146,6 @@ impl TxSigner<Signature> for ArbiterEvmWallet {
.recv() .recv()
.await .await
.map_err(|_| Error::other("failed to receive evm sign transaction response"))?; .map_err(|_| Error::other("failed to receive evm sign transaction response"))?;
drop(transport);
if response.request_id != Some(request_id) { if response.request_id != Some(request_id) {
return Err(Error::other( return Err(Error::other(

View File

@@ -35,7 +35,6 @@ impl PublicKey {
self.0.encode().0.to_vec() self.0.encode().0.to_vec()
} }
#[must_use]
pub fn verify(&self, nonce: i32, context: &[u8], signature: &Signature) -> bool { pub fn verify(&self, nonce: i32, context: &[u8], signature: &Signature) -> bool {
self.0.verify_with_context( self.0.verify_with_context(
&format_challenge(nonce, &self.to_bytes()), &format_challenge(nonce, &self.to_bytes()),

View File

@@ -49,7 +49,7 @@ impl<T: Hashable + PartialOrd> Hashable for Vec<T> {
} }
} }
impl<T: Hashable + PartialOrd, S: std::hash::BuildHasher> Hashable for HashSet<T, S> { impl<T: Hashable + PartialOrd> Hashable for HashSet<T> {
fn hash<H: Digest>(&self, hasher: &mut H) { fn hash<H: Digest>(&self, hasher: &mut H) {
let ref_sorted = { let ref_sorted = {
let mut sorted = self.iter().collect::<Vec<_>>(); let mut sorted = self.iter().collect::<Vec<_>>();

View File

@@ -29,7 +29,7 @@ pub trait SafeCellHandle<T> {
let mut cell = Self::new(T::default()); let mut cell = Self::new(T::default());
{ {
let mut handle = cell.write(); let mut handle = cell.write();
f(&mut *handle); f(handle.deref_mut());
} }
cell cell
} }

View File

@@ -20,7 +20,7 @@ pub(crate) fn derive(input: &DeriveInput) -> TokenStream {
} }
} }
fn hashable_struct(input: &DeriveInput, struct_data: &DataStruct) -> TokenStream { fn hashable_struct(input: &DeriveInput, struct_data: &syn::DataStruct) -> TokenStream {
let ident = &input.ident; let ident = &input.ident;
let hashable_trait = HASHABLE_TRAIT_PATH.to_path(); let hashable_trait = HASHABLE_TRAIT_PATH.to_path();
let hmac_digest = HMAC_DIGEST_PATH.to_path(); let hmac_digest = HMAC_DIGEST_PATH.to_path();

View File

@@ -1,24 +1,19 @@
pub(crate) struct ToPath(pub &'static str); pub struct ToPath(pub &'static str);
impl ToPath { impl ToPath {
pub(crate) fn to_path(&self) -> syn::Path { pub fn to_path(&self) -> syn::Path {
syn::parse_str(self.0).expect("Invalid path") syn::parse_str(self.0).expect("Invalid path")
} }
} }
macro_rules! ensure_path { macro_rules! ensure_path {
($path:path as $name:ident) => { ($path:path) => {{
const _: () = { #[cfg(test)]
#[cfg(test)] #[expect(unused_imports)]
#[expect( use $path as _;
unused_imports, ToPath(stringify!($path))
reason = "Ensures the path is valid and will cause a compile error if not" }};
)]
use $path as _;
};
pub(crate) const $name: ToPath = ToPath(stringify!($path));
};
} }
ensure_path!(::arbiter_crypto::hashing::Hashable as HASHABLE_TRAIT_PATH); pub const HASHABLE_TRAIT_PATH: ToPath = ensure_path!(::arbiter_crypto::hashing::Hashable);
ensure_path!(::arbiter_crypto::hashing::Digest as HMAC_DIGEST_PATH); pub const HMAC_DIGEST_PATH: ToPath = ensure_path!(::arbiter_crypto::hashing::Digest);

View File

@@ -105,7 +105,7 @@ mod tests {
#[rstest] #[rstest]
fn parsing_correctness( fn test_parsing_correctness(
#[values("127.0.0.1", "localhost", "192.168.1.1", "some.domain.com")] host: &str, #[values("127.0.0.1", "localhost", "192.168.1.1", "some.domain.com")] host: &str,
#[values(None, Some("token123".to_string()))] bootstrap_token: Option<String>, #[values(None, Some("token123".to_string()))] bootstrap_token: Option<String>,

View File

@@ -13,8 +13,8 @@ const TOKEN_LENGTH: usize = 64;
pub async fn generate_token() -> Result<String, std::io::Error> { pub async fn generate_token() -> Result<String, std::io::Error> {
let rng: StdRng = make_rng(); let rng: StdRng = make_rng();
let token = rng.sample_iter(Alphanumeric).take(TOKEN_LENGTH).fold( let token: String = rng.sample_iter(Alphanumeric).take(TOKEN_LENGTH).fold(
String::default(), Default::default(),
|mut accum, char| { |mut accum, char| {
accum += char.to_string().as_str(); accum += char.to_string().as_str();
accum accum
@@ -27,15 +27,15 @@ pub async fn generate_token() -> Result<String, std::io::Error> {
} }
#[derive(Error, Debug)] #[derive(Error, Debug)]
pub enum BootstrappError { pub enum Error {
#[error("Database error: {0}")] #[error("Database error: {0}")]
Database(#[from] db::PoolError), Database(#[from] db::PoolError),
#[error("I/O error: {0}")]
Io(#[from] std::io::Error),
#[error("Database query error: {0}")] #[error("Database query error: {0}")]
Query(#[from] diesel::result::Error), Query(#[from] diesel::result::Error),
#[error("I/O error: {0}")]
Io(#[from] std::io::Error),
} }
#[derive(Actor)] #[derive(Actor)]
@@ -44,7 +44,7 @@ pub struct Bootstrapper {
} }
impl Bootstrapper { impl Bootstrapper {
pub async fn new(db: &DatabasePool) -> Result<Self, BootstrappError> { pub async fn new(db: &DatabasePool) -> Result<Self, Error> {
let row_count: i64 = { let row_count: i64 = {
let mut conn = db.get().await?; let mut conn = db.get().await?;
@@ -69,13 +69,16 @@ impl Bootstrapper {
impl Bootstrapper { impl Bootstrapper {
#[message] #[message]
pub fn is_correct_token(&self, token: String) -> bool { pub fn is_correct_token(&self, token: String) -> bool {
self.token.as_ref().is_some_and(|expected| { match &self.token {
let expected_bytes = expected.as_bytes(); Some(expected) => {
let token_bytes = token.as_bytes(); let expected_bytes = expected.as_bytes();
let token_bytes = token.as_bytes();
let choice = expected_bytes.ct_eq(token_bytes); let choice = expected_bytes.ct_eq(token_bytes);
bool::from(choice) bool::from(choice)
}) }
None => false,
}
} }
#[message] #[message]

View File

@@ -1,7 +1,6 @@
use arbiter_crypto::authn::{self, CLIENT_CONTEXT}; use arbiter_crypto::authn::{self, CLIENT_CONTEXT};
use arbiter_proto::{ use arbiter_proto::{
ClientMetadata, ClientMetadata,
proto::client::auth::{AuthChallenge as ProtoAuthChallenge, AuthResult as ProtoAuthResult},
transport::{Bi, expect_message}, transport::{Bi, expect_message},
}; };
use chrono::Utc; use chrono::Utc;
@@ -28,60 +27,34 @@ use crate::{
}; };
#[derive(thiserror::Error, Debug, Clone, PartialEq, Eq)] #[derive(thiserror::Error, Debug, Clone, PartialEq, Eq)]
pub enum ClientAuthError { pub enum Error {
#[error("Client approval request failed")]
ApproveError(#[from] ApproveError),
#[error("Database operation failed")]
DatabaseOperationFailed,
#[error("Database pool unavailable")] #[error("Database pool unavailable")]
DatabasePoolUnavailable, DatabasePoolUnavailable,
#[error("Database operation failed")]
DatabaseOperationFailed,
#[error("Integrity check failed")] #[error("Integrity check failed")]
IntegrityCheckFailed, IntegrityCheckFailed,
#[error("Invalid challenge solution")] #[error("Invalid challenge solution")]
InvalidChallengeSolution, InvalidChallengeSolution,
#[error("Client approval request failed")]
ApproveError(#[from] ApproveError),
#[error("Transport error")] #[error("Transport error")]
Transport, Transport,
} }
impl From<diesel::result::Error> for ClientAuthError { impl From<diesel::result::Error> for Error {
fn from(e: diesel::result::Error) -> Self { fn from(e: diesel::result::Error) -> Self {
error!(?e, "Database error"); error!(?e, "Database error");
Self::DatabaseOperationFailed Self::DatabaseOperationFailed
} }
} }
impl From<ClientAuthError> for arbiter_proto::proto::client::auth::AuthResult {
fn from(value: ClientAuthError) -> Self {
match value {
ClientAuthError::ApproveError(e) => match e {
ApproveError::Denied => Self::ApprovalDenied,
ApproveError::Internal => Self::Internal,
ApproveError::Upstream(flow_coordinator::ApprovalError::NoUserAgentsConnected) => {
Self::NoUserAgentsOnline
} // ApproveError::Upstream(_) => Self::Internal,
},
ClientAuthError::DatabaseOperationFailed
| ClientAuthError::DatabasePoolUnavailable
| ClientAuthError::IntegrityCheckFailed
| ClientAuthError::Transport => Self::Internal,
ClientAuthError::InvalidChallengeSolution => Self::InvalidSignature,
}
}
}
#[derive(thiserror::Error, Debug, Clone, PartialEq, Eq)] #[derive(thiserror::Error, Debug, Clone, PartialEq, Eq)]
pub enum ApproveError { pub enum ApproveError {
#[error("Client connection denied by user agents")]
Denied,
#[error("Internal error")] #[error("Internal error")]
Internal, Internal,
#[error("Client connection denied by user agents")]
Denied,
#[error("Upstream error: {0}")] #[error("Upstream error: {0}")]
Upstream(flow_coordinator::ApprovalError), Upstream(flow_coordinator::ApprovalError),
} }
@@ -106,28 +79,16 @@ pub enum Outbound {
AuthSuccess, AuthSuccess,
} }
impl From<Outbound> for arbiter_proto::proto::client::auth::response::Payload {
fn from(value: Outbound) -> Self {
match value {
Outbound::AuthChallenge { pubkey, nonce } => Self::Challenge(ProtoAuthChallenge {
pubkey: pubkey.to_bytes(),
nonce,
}),
Outbound::AuthSuccess => Self::Result(ProtoAuthResult::Success.into()),
}
}
}
/// Returns the current nonce and client ID for a registered client. /// Returns the current nonce and client ID for a registered client.
/// Returns `None` if the pubkey is not registered. /// Returns `None` if the pubkey is not registered.
async fn get_current_nonce_and_id( async fn get_current_nonce_and_id(
db: &db::DatabasePool, db: &db::DatabasePool,
pubkey: &authn::PublicKey, pubkey: &authn::PublicKey,
) -> Result<Option<(i32, i32)>, ClientAuthError> { ) -> Result<Option<(i32, i32)>, Error> {
let pubkey_bytes = pubkey.to_bytes(); let pubkey_bytes = pubkey.to_bytes();
let mut conn = db.get().await.map_err(|e| { let mut conn = db.get().await.map_err(|e| {
error!(error = ?e, "Database pool error"); error!(error = ?e, "Database pool error");
ClientAuthError::DatabasePoolUnavailable Error::DatabasePoolUnavailable
})?; })?;
program_client::table program_client::table
.filter(program_client::public_key.eq(&pubkey_bytes)) .filter(program_client::public_key.eq(&pubkey_bytes))
@@ -137,7 +98,7 @@ async fn get_current_nonce_and_id(
.optional() .optional()
.map_err(|e| { .map_err(|e| {
error!(error = ?e, "Database error"); error!(error = ?e, "Database error");
ClientAuthError::DatabaseOperationFailed Error::DatabaseOperationFailed
}) })
} }
@@ -145,15 +106,15 @@ async fn verify_integrity(
db: &db::DatabasePool, db: &db::DatabasePool,
keyholder: &ActorRef<KeyHolder>, keyholder: &ActorRef<KeyHolder>,
pubkey: &authn::PublicKey, pubkey: &authn::PublicKey,
) -> Result<(), ClientAuthError> { ) -> Result<(), Error> {
let mut db_conn = db.get().await.map_err(|e| { let mut db_conn = db.get().await.map_err(|e| {
error!(error = ?e, "Database pool error"); error!(error = ?e, "Database pool error");
ClientAuthError::DatabasePoolUnavailable Error::DatabasePoolUnavailable
})?; })?;
let (id, nonce) = get_current_nonce_and_id(db, pubkey).await?.ok_or_else(|| { let (id, nonce) = get_current_nonce_and_id(db, pubkey).await?.ok_or_else(|| {
error!("Client not found during integrity verification"); error!("Client not found during integrity verification");
ClientAuthError::DatabaseOperationFailed Error::DatabaseOperationFailed
})?; })?;
let attestation = integrity::verify_entity( let attestation = integrity::verify_entity(
@@ -168,12 +129,12 @@ async fn verify_integrity(
.await .await
.map_err(|e| { .map_err(|e| {
error!(?e, "Integrity verification failed"); error!(?e, "Integrity verification failed");
ClientAuthError::IntegrityCheckFailed Error::IntegrityCheckFailed
})?; })?;
if attestation != AttestationStatus::Attested { if attestation != AttestationStatus::Attested {
error!("Integrity attestation unavailable for client {id}"); error!("Integrity attestation unavailable for client {id}");
return Err(ClientAuthError::IntegrityCheckFailed); return Err(Error::IntegrityCheckFailed);
} }
Ok(()) Ok(())
@@ -185,13 +146,13 @@ async fn create_nonce(
db: &db::DatabasePool, db: &db::DatabasePool,
keyholder: &ActorRef<KeyHolder>, keyholder: &ActorRef<KeyHolder>,
pubkey: &authn::PublicKey, pubkey: &authn::PublicKey,
) -> Result<i32, ClientAuthError> { ) -> Result<i32, Error> {
let pubkey_bytes = pubkey.to_bytes(); let pubkey_bytes = pubkey.to_bytes();
let pubkey = pubkey.clone(); let pubkey = pubkey.clone();
let mut conn = db.get().await.map_err(|e| { let mut conn = db.get().await.map_err(|e| {
error!(error = ?e, "Database pool error"); error!(error = ?e, "Database pool error");
ClientAuthError::DatabasePoolUnavailable Error::DatabasePoolUnavailable
})?; })?;
conn.exclusive_transaction(|conn| { conn.exclusive_transaction(|conn| {
@@ -217,7 +178,7 @@ async fn create_nonce(
.await .await
.map_err(|e| { .map_err(|e| {
error!(?e, "Integrity sign failed after nonce update"); error!(?e, "Integrity sign failed after nonce update");
ClientAuthError::DatabaseOperationFailed Error::DatabaseOperationFailed
})?; })?;
Ok(new_nonce) Ok(new_nonce)
@@ -229,7 +190,7 @@ async fn create_nonce(
async fn approve_new_client( async fn approve_new_client(
actors: &crate::actors::GlobalActors, actors: &crate::actors::GlobalActors,
profile: ClientProfile, profile: ClientProfile,
) -> Result<(), ClientAuthError> { ) -> Result<(), Error> {
let result = actors let result = actors
.flow_coordinator .flow_coordinator
.ask(RequestClientApproval { client: profile }) .ask(RequestClientApproval { client: profile })
@@ -237,14 +198,14 @@ async fn approve_new_client(
match result { match result {
Ok(true) => Ok(()), Ok(true) => Ok(()),
Ok(false) => Err(ClientAuthError::ApproveError(ApproveError::Denied)), Ok(false) => Err(Error::ApproveError(ApproveError::Denied)),
Err(SendError::HandlerError(e)) => { Err(SendError::HandlerError(e)) => {
error!(error = ?e, "Approval upstream error"); error!(error = ?e, "Approval upstream error");
Err(ClientAuthError::ApproveError(ApproveError::Upstream(e))) Err(Error::ApproveError(ApproveError::Upstream(e)))
} }
Err(e) => { Err(e) => {
error!(error = ?e, "Approval request to flow coordinator failed"); error!(error = ?e, "Approval request to flow coordinator failed");
Err(ClientAuthError::ApproveError(ApproveError::Internal)) Err(Error::ApproveError(ApproveError::Internal))
} }
} }
} }
@@ -254,14 +215,14 @@ async fn insert_client(
keyholder: &ActorRef<KeyHolder>, keyholder: &ActorRef<KeyHolder>,
pubkey: &authn::PublicKey, pubkey: &authn::PublicKey,
metadata: &ClientMetadata, metadata: &ClientMetadata,
) -> Result<i32, ClientAuthError> { ) -> Result<i32, Error> {
use crate::db::schema::client_metadata; use crate::db::schema::{client_metadata, program_client};
let pubkey = pubkey.clone(); let pubkey = pubkey.clone();
let metadata = metadata.clone(); let metadata = metadata.clone();
let mut conn = db.get().await.map_err(|e| { let mut conn = db.get().await.map_err(|e| {
error!(error = ?e, "Database pool error"); error!(error = ?e, "Database pool error");
ClientAuthError::DatabasePoolUnavailable Error::DatabasePoolUnavailable
})?; })?;
conn.exclusive_transaction(|conn| { conn.exclusive_transaction(|conn| {
@@ -303,7 +264,7 @@ async fn insert_client(
.await .await
.map_err(|e| { .map_err(|e| {
error!(error = ?e, "Failed to sign integrity tag for new client key"); error!(error = ?e, "Failed to sign integrity tag for new client key");
ClientAuthError::DatabaseOperationFailed Error::DatabaseOperationFailed
})?; })?;
Ok(client_id) Ok(client_id)
@@ -316,14 +277,14 @@ async fn sync_client_metadata(
db: &db::DatabasePool, db: &db::DatabasePool,
client_id: i32, client_id: i32,
metadata: &ClientMetadata, metadata: &ClientMetadata,
) -> Result<(), ClientAuthError> { ) -> Result<(), Error> {
use crate::db::schema::{client_metadata, client_metadata_history}; use crate::db::schema::{client_metadata, client_metadata_history};
let now = SqliteTimestamp(Utc::now()); let now = SqliteTimestamp(Utc::now());
let mut conn = db.get().await.map_err(|e| { let mut conn = db.get().await.map_err(|e| {
error!(error = ?e, "Database pool error"); error!(error = ?e, "Database pool error");
ClientAuthError::DatabasePoolUnavailable Error::DatabasePoolUnavailable
})?; })?;
conn.exclusive_transaction(|conn| { conn.exclusive_transaction(|conn| {
@@ -379,7 +340,7 @@ async fn sync_client_metadata(
.await .await
.map_err(|e| { .map_err(|e| {
error!(error = ?e, "Database error"); error!(error = ?e, "Database error");
ClientAuthError::DatabaseOperationFailed Error::DatabaseOperationFailed
}) })
} }
@@ -387,9 +348,9 @@ async fn challenge_client<T>(
transport: &mut T, transport: &mut T,
pubkey: authn::PublicKey, pubkey: authn::PublicKey,
nonce: i32, nonce: i32,
) -> Result<(), ClientAuthError> ) -> Result<(), Error>
where where
T: Bi<Inbound, Result<Outbound, ClientAuthError>> + ?Sized, T: Bi<Inbound, Result<Outbound, Error>> + ?Sized,
{ {
transport transport
.send(Ok(Outbound::AuthChallenge { .send(Ok(Outbound::AuthChallenge {
@@ -399,51 +360,51 @@ where
.await .await
.map_err(|e| { .map_err(|e| {
error!(error = ?e, "Failed to send auth challenge"); error!(error = ?e, "Failed to send auth challenge");
ClientAuthError::Transport Error::Transport
})?; })?;
let signature = expect_message(transport, |req: Inbound| match req { let signature = expect_message(transport, |req: Inbound| match req {
Inbound::AuthChallengeSolution { signature } => Some(signature), Inbound::AuthChallengeSolution { signature } => Some(signature),
Inbound::AuthChallengeRequest { .. } => None, _ => None,
}) })
.await .await
.map_err(|e| { .map_err(|e| {
error!(error = ?e, "Failed to receive challenge solution"); error!(error = ?e, "Failed to receive challenge solution");
ClientAuthError::Transport Error::Transport
})?; })?;
if !pubkey.verify(nonce, CLIENT_CONTEXT, &signature) { if !pubkey.verify(nonce, CLIENT_CONTEXT, &signature) {
error!("Challenge solution verification failed"); error!("Challenge solution verification failed");
return Err(ClientAuthError::InvalidChallengeSolution); return Err(Error::InvalidChallengeSolution);
} }
Ok(()) Ok(())
} }
pub async fn authenticate<T>( pub async fn authenticate<T>(props: &mut ClientConnection, transport: &mut T) -> Result<i32, Error>
props: &mut ClientConnection,
transport: &mut T,
) -> Result<i32, ClientAuthError>
where where
T: Bi<Inbound, Result<Outbound, ClientAuthError>> + Send + ?Sized, T: Bi<Inbound, Result<Outbound, Error>> + Send + ?Sized,
{ {
let Some(Inbound::AuthChallengeRequest { pubkey, metadata }) = transport.recv().await else { let Some(Inbound::AuthChallengeRequest { pubkey, metadata }) = transport.recv().await else {
return Err(ClientAuthError::Transport); return Err(Error::Transport);
}; };
let client_id = if let Some((id, _)) = get_current_nonce_and_id(&props.db, &pubkey).await? { let client_id = match get_current_nonce_and_id(&props.db, &pubkey).await? {
verify_integrity(&props.db, &props.actors.key_holder, &pubkey).await?; Some((id, _)) => {
id verify_integrity(&props.db, &props.actors.key_holder, &pubkey).await?;
} else { id
approve_new_client( }
&props.actors, None => {
ClientProfile { approve_new_client(
pubkey: pubkey.clone(), &props.actors,
metadata: metadata.clone(), ClientProfile {
}, pubkey: pubkey.clone(),
) metadata: metadata.clone(),
.await?; },
insert_client(&props.db, &props.actors.key_holder, &pubkey, &metadata).await? )
.await?;
insert_client(&props.db, &props.actors.key_holder, &pubkey, &metadata).await?
}
}; };
sync_client_metadata(&props.db, client_id, &metadata).await?; sync_client_metadata(&props.db, client_id, &metadata).await?;
@@ -455,7 +416,7 @@ where
.await .await
.map_err(|e| { .map_err(|e| {
error!(error = ?e, "Failed to send auth success"); error!(error = ?e, "Failed to send auth success");
ClientAuthError::Transport Error::Transport
})?; })?;
Ok(client_id) Ok(client_id)

View File

@@ -31,7 +31,7 @@ pub struct ClientConnection {
} }
impl ClientConnection { impl ClientConnection {
pub const fn new(db: db::DatabasePool, actors: GlobalActors) -> Self { pub fn new(db: db::DatabasePool, actors: GlobalActors) -> Self {
Self { db, actors } Self { db, actors }
} }
} }
@@ -41,10 +41,10 @@ pub mod session;
pub async fn connect_client<T>(mut props: ClientConnection, transport: &mut T) pub async fn connect_client<T>(mut props: ClientConnection, transport: &mut T)
where where
T: Bi<auth::Inbound, Result<auth::Outbound, auth::ClientAuthError>> + Send + ?Sized, T: Bi<auth::Inbound, Result<auth::Outbound, auth::Error>> + Send + ?Sized,
{ {
let fut = auth::authenticate(&mut props, transport); let fut = auth::authenticate(&mut props, transport);
println!("authenticate future size: {}", size_of_val(&fut)); println!("authenticate future size: {}", std::mem::size_of_val(&fut));
match fut.await { match fut.await {
Ok(client_id) => { Ok(client_id) => {
ClientSession::spawn(ClientSession::new(props, client_id)); ClientSession::spawn(ClientSession::new(props, client_id));

View File

@@ -21,7 +21,7 @@ pub struct ClientSession {
} }
impl ClientSession { impl ClientSession {
pub(crate) const fn new(props: ClientConnection, client_id: i32) -> Self { pub(crate) fn new(props: ClientConnection, client_id: i32) -> Self {
Self { props, client_id } Self { props, client_id }
} }
} }
@@ -29,16 +29,14 @@ impl ClientSession {
#[messages] #[messages]
impl ClientSession { impl ClientSession {
#[message] #[message]
pub(crate) async fn handle_query_vault_state( pub(crate) async fn handle_query_vault_state(&mut self) -> Result<KeyHolderState, Error> {
&mut self,
) -> Result<KeyHolderState, ClientSessionError> {
use crate::actors::keyholder::GetState; use crate::actors::keyholder::GetState;
let vault_state = match self.props.actors.key_holder.ask(GetState {}).await { let vault_state = match self.props.actors.key_holder.ask(GetState {}).await {
Ok(state) => state, Ok(state) => state,
Err(err) => { Err(err) => {
error!(?err, actor = "client", "keyholder.query.failed"); error!(?err, actor = "client", "keyholder.query.failed");
return Err(ClientSessionError::Internal); return Err(Error::Internal);
} }
}; };
@@ -77,7 +75,7 @@ impl ClientSession {
impl Actor for ClientSession { impl Actor for ClientSession {
type Args = Self; type Args = Self;
type Error = ClientSessionError; type Error = Error;
async fn on_start( async fn on_start(
args: Self::Args, args: Self::Args,
@@ -88,13 +86,13 @@ impl Actor for ClientSession {
.flow_coordinator .flow_coordinator
.ask(RegisterClient { actor: this }) .ask(RegisterClient { actor: this })
.await .await
.map_err(|_| ClientSessionError::ConnectionRegistrationFailed)?; .map_err(|_| Error::ConnectionRegistrationFailed)?;
Ok(args) Ok(args)
} }
} }
impl ClientSession { impl ClientSession {
pub const fn new_test(db: db::DatabasePool, actors: GlobalActors) -> Self { pub fn new_test(db: db::DatabasePool, actors: GlobalActors) -> Self {
let props = ClientConnection::new(db, actors); let props = ClientConnection::new(db, actors);
Self { Self {
props, props,
@@ -104,7 +102,7 @@ impl ClientSession {
} }
#[derive(Debug, thiserror::Error)] #[derive(Debug, thiserror::Error)]
pub enum ClientSessionError { pub enum Error {
#[error("Connection registration failed")] #[error("Connection registration failed")]
ConnectionRegistrationFailed, ConnectionRegistrationFailed,
#[error("Internal error")] #[error("Internal error")]
@@ -113,9 +111,9 @@ pub enum ClientSessionError {
#[derive(Debug, thiserror::Error)] #[derive(Debug, thiserror::Error)]
pub enum SignTransactionRpcError { pub enum SignTransactionRpcError {
#[error("Internal error")]
Internal,
#[error("Policy evaluation failed")] #[error("Policy evaluation failed")]
Vet(#[from] VetError), Vet(#[from] VetError),
#[error("Internal error")]
Internal,
} }

View File

@@ -1,6 +1,4 @@
use alloy::{ use alloy::{consensus::TxEip1559, primitives::Address, signers::Signature};
consensus::TxEip1559, network::TxSignerSync as _, primitives::Address, signers::Signature,
};
use diesel::{ use diesel::{
ExpressionMethods, OptionalExtension as _, QueryDsl, SelectableHelper as _, dsl::insert_into, ExpressionMethods, OptionalExtension as _, QueryDsl, SelectableHelper as _, dsl::insert_into,
}; };
@@ -37,7 +35,7 @@ pub enum SignTransactionError {
Database(#[from] DatabaseError), Database(#[from] DatabaseError),
#[error("Keyholder error: {0}")] #[error("Keyholder error: {0}")]
Keyholder(#[from] crate::actors::keyholder::KeyHolderError), Keyholder(#[from] crate::actors::keyholder::Error),
#[error("Keyholder mailbox error")] #[error("Keyholder mailbox error")]
KeyholderSend, KeyholderSend,
@@ -50,9 +48,9 @@ pub enum SignTransactionError {
} }
#[derive(Debug, thiserror::Error)] #[derive(Debug, thiserror::Error)]
pub enum EvmActorError { pub enum Error {
#[error("Keyholder error: {0}")] #[error("Keyholder error: {0}")]
Keyholder(#[from] crate::actors::keyholder::KeyHolderError), Keyholder(#[from] crate::actors::keyholder::Error),
#[error("Keyholder mailbox error")] #[error("Keyholder mailbox error")]
KeyholderSend, KeyholderSend,
@@ -61,7 +59,7 @@ pub enum EvmActorError {
Database(#[from] DatabaseError), Database(#[from] DatabaseError),
#[error("Integrity violation: {0}")] #[error("Integrity violation: {0}")]
Integrity(#[from] integrity::IntegrityError), Integrity(#[from] integrity::Error),
} }
#[derive(Actor)] #[derive(Actor)]
@@ -90,7 +88,7 @@ impl EvmActor {
#[messages] #[messages]
impl EvmActor { impl EvmActor {
#[message] #[message]
pub async fn generate(&mut self) -> Result<(i32, Address), EvmActorError> { pub async fn generate(&mut self) -> Result<(i32, Address), Error> {
let (mut key_cell, address) = safe_signer::generate(&mut self.rng); let (mut key_cell, address) = safe_signer::generate(&mut self.rng);
let plaintext = key_cell.read_inline(|reader| SafeCell::new(reader.to_vec())); let plaintext = key_cell.read_inline(|reader| SafeCell::new(reader.to_vec()));
@@ -99,7 +97,7 @@ impl EvmActor {
.keyholder .keyholder
.ask(CreateNew { plaintext }) .ask(CreateNew { plaintext })
.await .await
.map_err(|_| EvmActorError::KeyholderSend)?; .map_err(|_| Error::KeyholderSend)?;
let mut conn = self.db.get().await.map_err(DatabaseError::from)?; let mut conn = self.db.get().await.map_err(DatabaseError::from)?;
let wallet_id = insert_into(schema::evm_wallet::table) let wallet_id = insert_into(schema::evm_wallet::table)
@@ -116,7 +114,7 @@ impl EvmActor {
} }
#[message] #[message]
pub async fn list_wallets(&self) -> Result<Vec<(i32, Address)>, EvmActorError> { pub async fn list_wallets(&self) -> Result<Vec<(i32, Address)>, Error> {
let mut conn = self.db.get().await.map_err(DatabaseError::from)?; let mut conn = self.db.get().await.map_err(DatabaseError::from)?;
let rows: Vec<models::EvmWallet> = schema::evm_wallet::table let rows: Vec<models::EvmWallet> = schema::evm_wallet::table
.select(models::EvmWallet::as_select()) .select(models::EvmWallet::as_select())
@@ -138,7 +136,7 @@ impl EvmActor {
&mut self, &mut self,
basic: SharedGrantSettings, basic: SharedGrantSettings,
grant: SpecificGrant, grant: SpecificGrant,
) -> Result<i32, EvmActorError> { ) -> Result<i32, Error> {
match grant { match grant {
SpecificGrant::EtherTransfer(settings) => self SpecificGrant::EtherTransfer(settings) => self
.engine .engine
@@ -147,7 +145,7 @@ impl EvmActor {
specific: settings, specific: settings,
}) })
.await .await
.map_err(EvmActorError::from), .map_err(Error::from),
SpecificGrant::TokenTransfer(settings) => self SpecificGrant::TokenTransfer(settings) => self
.engine .engine
.create_grant::<TokenTransfer>(CombinedSettings { .create_grant::<TokenTransfer>(CombinedSettings {
@@ -155,13 +153,12 @@ impl EvmActor {
specific: settings, specific: settings,
}) })
.await .await
.map_err(EvmActorError::from), .map_err(Error::from),
} }
} }
#[message] #[message]
#[expect(clippy::unused_async, reason = "reserved for impl")] pub async fn useragent_delete_grant(&mut self, _grant_id: i32) -> Result<(), Error> {
pub async fn useragent_delete_grant(&mut self, _grant_id: i32) -> Result<(), EvmActorError> {
// let mut conn = self.db.get().await.map_err(DatabaseError::from)?; // let mut conn = self.db.get().await.map_err(DatabaseError::from)?;
// let keyholder = self.keyholder.clone(); // let keyholder = self.keyholder.clone();
@@ -186,15 +183,11 @@ impl EvmActor {
} }
#[message] #[message]
pub async fn useragent_list_grants( pub async fn useragent_list_grants(&mut self) -> Result<Vec<Grant<SpecificGrant>>, Error> {
&mut self,
) -> Result<Vec<Grant<SpecificGrant>>, EvmActorError> {
match self.engine.list_all_grants().await { match self.engine.list_all_grants().await {
Ok(grants) => Ok(grants), Ok(grants) => Ok(grants),
Err(ListError::Database(db_err)) => Err(EvmActorError::Database(db_err)), Err(ListError::Database(db_err)) => Err(Error::Database(db_err)),
Err(ListError::Integrity(integrity_err)) => { Err(ListError::Integrity(integrity_err)) => Err(Error::Integrity(integrity_err)),
Err(EvmActorError::Integrity(integrity_err))
}
} }
} }
@@ -274,6 +267,7 @@ impl EvmActor {
.evaluate_transaction(wallet_access, transaction.clone(), RunKind::Execution) .evaluate_transaction(wallet_access, transaction.clone(), RunKind::Execution)
.await?; .await?;
use alloy::network::TxSignerSync as _;
Ok(signer.sign_transaction_sync(&mut transaction)?) Ok(signer.sign_transaction_sync(&mut transaction)?)
} }
} }

View File

@@ -41,7 +41,7 @@ impl Actor for ClientApprovalController {
async fn on_start( async fn on_start(
Args { Args {
client, client,
user_agents, mut user_agents,
reply, reply,
}: Self::Args, }: Self::Args,
actor_ref: ActorRef<Self>, actor_ref: ActorRef<Self>,
@@ -52,9 +52,8 @@ impl Actor for ClientApprovalController {
reply: Some(reply), reply: Some(reply),
}; };
for user_agent in user_agents { for user_agent in user_agents.drain(..) {
actor_ref.link(&user_agent).await; actor_ref.link(&user_agent).await;
let _ = user_agent let _ = user_agent
.tell(BeginNewClientApproval { .tell(BeginNewClientApproval {
client: client.clone(), client: client.clone(),
@@ -86,7 +85,7 @@ impl Actor for ClientApprovalController {
#[messages] #[messages]
impl ClientApprovalController { impl ClientApprovalController {
#[message(ctx)] #[message(ctx)]
pub fn client_approval_answer(&mut self, approved: bool, ctx: &mut Context<Self, ()>) { pub async fn client_approval_answer(&mut self, approved: bool, ctx: &mut Context<Self, ()>) {
if !approved { if !approved {
// Denial wins immediately regardless of other pending responses. // Denial wins immediately regardless of other pending responses.
self.send_reply(Ok(false)); self.send_reply(Ok(false));

View File

@@ -92,7 +92,7 @@ impl FlowCoordinator {
} }
#[message(ctx)] #[message(ctx)]
pub fn request_client_approval( pub async fn request_client_approval(
&mut self, &mut self,
client: ClientProfile, client: ClientProfile,
ctx: &mut Context<Self, DelegatedReply<Result<bool, ApprovalError>>>, ctx: &mut Context<Self, DelegatedReply<Result<bool, ApprovalError>>>,

View File

@@ -36,12 +36,19 @@ enum State {
} }
#[derive(Debug, thiserror::Error)] #[derive(Debug, thiserror::Error)]
pub enum KeyHolderError { pub enum Error {
#[error("Keyholder is already bootstrapped")] #[error("Keyholder is already bootstrapped")]
AlreadyBootstrapped, AlreadyBootstrapped,
#[error("Keyholder is not bootstrapped")]
NotBootstrapped,
#[error("Invalid key provided")]
InvalidKey,
#[error("Broken database")] #[error("Requested aead entry not found")]
BrokenDatabase, NotFound,
#[error("Encryption error: {0}")]
Encryption(#[from] chacha20poly1305::aead::Error),
#[error("Database error: {0}")] #[error("Database error: {0}")]
DatabaseConnection(#[from] db::PoolError), DatabaseConnection(#[from] db::PoolError),
@@ -49,21 +56,11 @@ pub enum KeyHolderError {
#[error("Database transaction error: {0}")] #[error("Database transaction error: {0}")]
DatabaseTransaction(#[from] diesel::result::Error), DatabaseTransaction(#[from] diesel::result::Error),
#[error("Encryption error: {0}")] #[error("Broken database")]
Encryption(#[from] chacha20poly1305::aead::Error), BrokenDatabase,
#[error("Invalid key provided")]
InvalidKey,
#[error("Keyholder is not bootstrapped")]
NotBootstrapped,
#[error("Requested aead entry not found")]
NotFound,
} }
/// Manages vault root key and tracks current state of the vault (bootstrapped/unbootstrapped, sealed/unsealed). /// Manages vault root key and tracks current state of the vault (bootstrapped/unbootstrapped, sealed/unsealed).
///
/// Provides API for encrypting and decrypting data using the vault root key. /// Provides API for encrypting and decrypting data using the vault root key.
/// Abstraction over database to make sure nonces are never reused and encryption keys are never exposed in plaintext outside of this actor. /// Abstraction over database to make sure nonces are never reused and encryption keys are never exposed in plaintext outside of this actor.
#[derive(Actor)] #[derive(Actor)]
@@ -74,7 +71,7 @@ pub struct KeyHolder {
#[messages] #[messages]
impl KeyHolder { impl KeyHolder {
pub async fn new(db: db::DatabasePool) -> Result<Self, KeyHolderError> { pub async fn new(db: db::DatabasePool) -> Result<Self, Error> {
let state = { let state = {
let mut conn = db.get().await?; let mut conn = db.get().await?;
@@ -97,10 +94,7 @@ impl KeyHolder {
// Exclusive transaction to avoid race condtions if multiple keyholders write // Exclusive transaction to avoid race condtions if multiple keyholders write
// additional layer of protection against nonce-reuse // additional layer of protection against nonce-reuse
async fn get_new_nonce( async fn get_new_nonce(pool: &db::DatabasePool, root_key_id: i32) -> Result<Nonce, Error> {
pool: &db::DatabasePool,
root_key_id: i32,
) -> Result<Nonce, KeyHolderError> {
let mut conn = pool.get().await?; let mut conn = pool.get().await?;
let nonce = conn let nonce = conn
@@ -112,12 +106,12 @@ impl KeyHolder {
.first(conn) .first(conn)
.await?; .await?;
let mut nonce = Nonce::try_from(current_nonce.as_slice()).map_err(|()| { let mut nonce = Nonce::try_from(current_nonce.as_slice()).map_err(|_| {
error!( error!(
"Broken database: invalid nonce for root key history id={}", "Broken database: invalid nonce for root key history id={}",
root_key_id root_key_id
); );
KeyHolderError::BrokenDatabase Error::BrokenDatabase
})?; })?;
nonce.increment(); nonce.increment();
@@ -127,7 +121,7 @@ impl KeyHolder {
.execute(conn) .execute(conn)
.await?; .await?;
Result::<_, KeyHolderError>::Ok(nonce) Result::<_, Error>::Ok(nonce)
}) })
}) })
.await?; .await?;
@@ -136,12 +130,9 @@ impl KeyHolder {
} }
#[message] #[message]
pub async fn bootstrap( pub async fn bootstrap(&mut self, seal_key_raw: SafeCell<Vec<u8>>) -> Result<(), Error> {
&mut self,
seal_key_raw: SafeCell<Vec<u8>>,
) -> Result<(), KeyHolderError> {
if !matches!(self.state, State::Unbootstrapped) { if !matches!(self.state, State::Unbootstrapped) {
return Err(KeyHolderError::AlreadyBootstrapped); return Err(Error::AlreadyBootstrapped);
} }
let salt = v1::generate_salt(); let salt = v1::generate_salt();
let mut seal_key = derive_key(seal_key_raw, &salt); let mut seal_key = derive_key(seal_key_raw, &salt);
@@ -157,7 +148,7 @@ impl KeyHolder {
.encrypt(&root_key_nonce, v1::ROOT_KEY_TAG, root_key_reader) .encrypt(&root_key_nonce, v1::ROOT_KEY_TAG, root_key_reader)
.map_err(|err| { .map_err(|err| {
error!(?err, "Fatal bootstrap error"); error!(?err, "Fatal bootstrap error");
KeyHolderError::Encryption(err) Error::Encryption(err)
}) })
})?; })?;
@@ -201,15 +192,12 @@ impl KeyHolder {
} }
#[message] #[message]
pub async fn try_unseal( pub async fn try_unseal(&mut self, seal_key_raw: SafeCell<Vec<u8>>) -> Result<(), Error> {
&mut self,
seal_key_raw: SafeCell<Vec<u8>>,
) -> Result<(), KeyHolderError> {
let State::Sealed { let State::Sealed {
root_key_history_id, root_key_history_id,
} = &self.state } = &self.state
else { else {
return Err(KeyHolderError::NotBootstrapped); return Err(Error::NotBootstrapped);
}; };
// We don't want to hold connection while doing expensive KDF work // We don't want to hold connection while doing expensive KDF work
@@ -225,16 +213,16 @@ impl KeyHolder {
let salt = &current_key.salt; let salt = &current_key.salt;
let salt = v1::Salt::try_from(salt.as_slice()).map_err(|_| { let salt = v1::Salt::try_from(salt.as_slice()).map_err(|_| {
error!("Broken database: invalid salt for root key"); error!("Broken database: invalid salt for root key");
KeyHolderError::BrokenDatabase Error::BrokenDatabase
})?; })?;
let mut seal_key = derive_key(seal_key_raw, &salt); let mut seal_key = derive_key(seal_key_raw, &salt);
let mut root_key = SafeCell::new(current_key.ciphertext.clone()); let mut root_key = SafeCell::new(current_key.ciphertext.clone());
let nonce = Nonce::try_from(current_key.root_key_encryption_nonce.as_slice()).map_err( let nonce = v1::Nonce::try_from(current_key.root_key_encryption_nonce.as_slice()).map_err(
|()| { |_| {
error!("Broken database: invalid nonce for root key"); error!("Broken database: invalid nonce for root key");
KeyHolderError::BrokenDatabase Error::BrokenDatabase
}, },
)?; )?;
@@ -242,14 +230,14 @@ impl KeyHolder {
.decrypt_in_place(&nonce, v1::ROOT_KEY_TAG, &mut root_key) .decrypt_in_place(&nonce, v1::ROOT_KEY_TAG, &mut root_key)
.map_err(|err| { .map_err(|err| {
error!(?err, "Failed to unseal root key: invalid seal key"); error!(?err, "Failed to unseal root key: invalid seal key");
KeyHolderError::InvalidKey Error::InvalidKey
})?; })?;
self.state = State::Unsealed { self.state = State::Unsealed {
root_key_history_id: current_key.id, root_key_history_id: current_key.id,
root_key: KeyCell::try_from(root_key).map_err(|err| { root_key: KeyCell::try_from(root_key).map_err(|err| {
error!(?err, "Broken database: invalid encryption key size"); error!(?err, "Broken database: invalid encryption key size");
KeyHolderError::BrokenDatabase Error::BrokenDatabase
})?, })?,
}; };
@@ -259,9 +247,9 @@ impl KeyHolder {
} }
#[message] #[message]
pub async fn decrypt(&mut self, aead_id: i32) -> Result<SafeCell<Vec<u8>>, KeyHolderError> { pub async fn decrypt(&mut self, aead_id: i32) -> Result<SafeCell<Vec<u8>>, Error> {
let State::Unsealed { root_key, .. } = &mut self.state else { let State::Unsealed { root_key, .. } = &mut self.state else {
return Err(KeyHolderError::NotBootstrapped); return Err(Error::NotBootstrapped);
}; };
let row: models::AeadEncrypted = { let row: models::AeadEncrypted = {
@@ -272,15 +260,15 @@ impl KeyHolder {
.first(&mut conn) .first(&mut conn)
.await .await
.optional()? .optional()?
.ok_or(KeyHolderError::NotFound)? .ok_or(Error::NotFound)?
}; };
let nonce = Nonce::try_from(row.current_nonce.as_slice()).map_err(|()| { let nonce = v1::Nonce::try_from(row.current_nonce.as_slice()).map_err(|_| {
error!( error!(
"Broken database: invalid nonce for aead_encrypted id={}", "Broken database: invalid nonce for aead_encrypted id={}",
aead_id aead_id
); );
KeyHolderError::BrokenDatabase Error::BrokenDatabase
})?; })?;
let mut output = SafeCell::new(row.ciphertext); let mut output = SafeCell::new(row.ciphertext);
root_key.decrypt_in_place(&nonce, v1::TAG, &mut output)?; root_key.decrypt_in_place(&nonce, v1::TAG, &mut output)?;
@@ -289,17 +277,14 @@ impl KeyHolder {
// Creates new `aead_encrypted` entry in the database and returns it's ID // Creates new `aead_encrypted` entry in the database and returns it's ID
#[message] #[message]
pub async fn create_new( pub async fn create_new(&mut self, mut plaintext: SafeCell<Vec<u8>>) -> Result<i32, Error> {
&mut self,
mut plaintext: SafeCell<Vec<u8>>,
) -> Result<i32, KeyHolderError> {
let State::Unsealed { let State::Unsealed {
root_key, root_key,
root_key_history_id, root_key_history_id,
.. ..
} = &mut self.state } = &mut self.state
else { else {
return Err(KeyHolderError::NotBootstrapped); return Err(Error::NotBootstrapped);
}; };
// Order matters here - `get_new_nonce` acquires connection, so we need to call it before next acquire // Order matters here - `get_new_nonce` acquires connection, so we need to call it before next acquire
@@ -335,19 +320,21 @@ impl KeyHolder {
} }
#[message] #[message]
pub fn sign_integrity(&mut self, mac_input: Vec<u8>) -> Result<(i32, Vec<u8>), KeyHolderError> { pub fn sign_integrity(&mut self, mac_input: Vec<u8>) -> Result<(i32, Vec<u8>), Error> {
let State::Unsealed { let State::Unsealed {
root_key, root_key,
root_key_history_id, root_key_history_id,
} = &mut self.state } = &mut self.state
else { else {
return Err(KeyHolderError::NotBootstrapped); return Err(Error::NotBootstrapped);
}; };
let mut hmac = root_key.0.read_inline(|k| { let mut hmac = root_key
HmacSha256::new_from_slice(k) .0
.unwrap_or_else(|_| unreachable!("HMAC accepts keys of any size")) .read_inline(|k| match HmacSha256::new_from_slice(k) {
}); Ok(v) => v,
Err(_) => unreachable!("HMAC accepts keys of any size"),
});
hmac.update(&root_key_history_id.to_be_bytes()); hmac.update(&root_key_history_id.to_be_bytes());
hmac.update(&mac_input); hmac.update(&mac_input);
@@ -361,23 +348,25 @@ impl KeyHolder {
mac_input: Vec<u8>, mac_input: Vec<u8>,
expected_mac: Vec<u8>, expected_mac: Vec<u8>,
key_version: i32, key_version: i32,
) -> Result<bool, KeyHolderError> { ) -> Result<bool, Error> {
let State::Unsealed { let State::Unsealed {
root_key, root_key,
root_key_history_id, root_key_history_id,
} = &mut self.state } = &mut self.state
else { else {
return Err(KeyHolderError::NotBootstrapped); return Err(Error::NotBootstrapped);
}; };
if *root_key_history_id != key_version { if *root_key_history_id != key_version {
return Ok(false); return Ok(false);
} }
let mut hmac = root_key.0.read_inline(|k| { let mut hmac = root_key
HmacSha256::new_from_slice(k) .0
.unwrap_or_else(|_| unreachable!("HMAC accepts keys of any size")) .read_inline(|k| match HmacSha256::new_from_slice(k) {
}); Ok(v) => v,
Err(_) => unreachable!("HMAC accepts keys of any size"),
});
hmac.update(&key_version.to_be_bytes()); hmac.update(&key_version.to_be_bytes());
hmac.update(&mac_input); hmac.update(&mac_input);
@@ -385,13 +374,13 @@ impl KeyHolder {
} }
#[message] #[message]
pub fn seal(&mut self) -> Result<(), KeyHolderError> { pub fn seal(&mut self) -> Result<(), Error> {
let State::Unsealed { let State::Unsealed {
root_key_history_id, root_key_history_id,
.. ..
} = &self.state } = &self.state
else { else {
return Err(KeyHolderError::NotBootstrapped); return Err(Error::NotBootstrapped);
}; };
self.state = State::Sealed { self.state = State::Sealed {
root_key_history_id: *root_key_history_id, root_key_history_id: *root_key_history_id,
@@ -402,7 +391,12 @@ impl KeyHolder {
#[cfg(test)] #[cfg(test)]
mod tests { mod tests {
use arbiter_crypto::safecell::SafeCellHandle as _; use diesel::SelectableHelper;
use diesel_async::RunQueryDsl;
use crate::db::{self};
use arbiter_crypto::safecell::{SafeCell, SafeCellHandle as _};
use super::*; use super::*;
@@ -418,12 +412,12 @@ mod tests {
async fn nonce_monotonic_even_when_nonce_allocation_interleaves() { async fn nonce_monotonic_even_when_nonce_allocation_interleaves() {
let db = db::create_test_pool().await; let db = db::create_test_pool().await;
let mut actor = bootstrapped_actor(&db).await; let mut actor = bootstrapped_actor(&db).await;
let State::Unsealed { let root_key_history_id = match actor.state {
root_key_history_id, State::Unsealed {
.. root_key_history_id,
} = actor.state ..
else { } => root_key_history_id,
panic!("expected unsealed state"); _ => panic!("expected unsealed state"),
}; };
let n1 = KeyHolder::get_new_nonce(&db, root_key_history_id) let n1 = KeyHolder::get_new_nonce(&db, root_key_history_id)
@@ -435,8 +429,8 @@ mod tests {
assert!(n2.to_vec() > n1.to_vec(), "nonce must increase"); assert!(n2.to_vec() > n1.to_vec(), "nonce must increase");
let mut conn = db.get().await.unwrap(); let mut conn = db.get().await.unwrap();
let root_row = schema::root_key_history::table let root_row: models::RootKeyHistory = schema::root_key_history::table
.select(RootKeyHistory::as_select()) .select(models::RootKeyHistory::as_select())
.first(&mut conn) .first(&mut conn)
.await .await
.unwrap(); .unwrap();

View File

@@ -11,18 +11,18 @@ use crate::{
pub mod bootstrap; pub mod bootstrap;
pub mod client; pub mod client;
pub mod evm; mod evm;
pub mod flow_coordinator; pub mod flow_coordinator;
pub mod keyholder; pub mod keyholder;
pub mod user_agent; pub mod user_agent;
#[derive(Error, Debug)] #[derive(Error, Debug)]
pub enum GlobalActorsSpawnError { pub enum SpawnError {
#[error("Failed to spawn Bootstrapper actor")] #[error("Failed to spawn Bootstrapper actor")]
Bootstrapper(#[from] bootstrap::BootstrappError), Bootstrapper(#[from] bootstrap::Error),
#[error("Failed to spawn KeyHolder actor")] #[error("Failed to spawn KeyHolder actor")]
KeyHolder(#[from] keyholder::KeyHolderError), KeyHolder(#[from] keyholder::Error),
} }
/// Long-lived actors that are shared across all connections and handle global state and operations /// Long-lived actors that are shared across all connections and handle global state and operations
@@ -35,7 +35,7 @@ pub struct GlobalActors {
} }
impl GlobalActors { impl GlobalActors {
pub async fn spawn(db: db::DatabasePool) -> Result<Self, GlobalActorsSpawnError> { pub async fn spawn(db: db::DatabasePool) -> Result<Self, SpawnError> {
let key_holder = KeyHolder::spawn(KeyHolder::new(db.clone()).await?); let key_holder = KeyHolder::spawn(KeyHolder::new(db.clone()).await?);
Ok(Self { Ok(Self {
bootstrapper: Bootstrapper::spawn(Bootstrapper::new(&db).await?), bootstrapper: Bootstrapper::spawn(Bootstrapper::new(&db).await?),

View File

@@ -7,9 +7,7 @@ use crate::actors::user_agent::{
auth::state::{AuthContext, AuthStateMachine}, auth::state::{AuthContext, AuthStateMachine},
}; };
mod state; mod state;
use state::{ use state::*;
AuthError, AuthEvents, AuthStates, BootstrapAuthRequest, ChallengeRequest, ChallengeSolution,
};
#[derive(Debug, Clone)] #[derive(Debug, Clone)]
pub enum Inbound { pub enum Inbound {

View File

@@ -204,14 +204,14 @@ pub struct AuthContext<'a, T> {
} }
impl<'a, T> AuthContext<'a, T> { impl<'a, T> AuthContext<'a, T> {
pub const fn new(conn: &'a mut UserAgentConnection, transport: T) -> Self { pub fn new(conn: &'a mut UserAgentConnection, transport: T) -> Self {
Self { conn, transport } Self { conn, transport }
} }
} }
impl<T> AuthStateMachineContext for AuthContext<'_, T> impl<T> AuthStateMachineContext for AuthContext<'_, T>
where where
T: Bi<super::Inbound, Result<Outbound, Error>> + Send, T: Bi<super::Inbound, Result<super::Outbound, Error>> + Send,
{ {
type Error = Error; type Error = Error;
@@ -237,6 +237,8 @@ where
}) })
} }
#[allow(missing_docs)]
#[allow(clippy::result_unit_err)]
async fn verify_bootstrap_token( async fn verify_bootstrap_token(
&mut self, &mut self,
BootstrapAuthRequest { pubkey, token }: BootstrapAuthRequest, BootstrapAuthRequest { pubkey, token }: BootstrapAuthRequest,
@@ -259,23 +261,28 @@ where
return Err(Error::InvalidBootstrapToken); return Err(Error::InvalidBootstrapToken);
} }
if token_ok { match token_ok {
register_key(&self.conn.db, &self.conn.actors.key_holder, &pubkey).await?; true => {
self.transport register_key(&self.conn.db, &self.conn.actors.key_holder, &pubkey).await?;
.send(Ok(Outbound::AuthSuccess)) self.transport
.await .send(Ok(Outbound::AuthSuccess))
.map_err(|_| Error::Transport)?; .await
Ok(pubkey) .map_err(|_| Error::Transport)?;
} else { Ok(pubkey)
error!("Invalid bootstrap token provided"); }
self.transport false => {
.send(Err(Error::InvalidBootstrapToken)) error!("Invalid bootstrap token provided");
.await self.transport
.map_err(|_| Error::Transport)?; .send(Err(Error::InvalidBootstrapToken))
Err(Error::InvalidBootstrapToken) .await
.map_err(|_| Error::Transport)?;
Err(Error::InvalidBootstrapToken)
}
} }
} }
#[allow(missing_docs)]
#[allow(clippy::unused_unit)]
async fn verify_solution( async fn verify_solution(
&mut self, &mut self,
ChallengeContext { ChallengeContext {
@@ -284,25 +291,28 @@ where
}: &ChallengeContext, }: &ChallengeContext,
ChallengeSolution { solution }: ChallengeSolution, ChallengeSolution { solution }: ChallengeSolution,
) -> Result<authn::PublicKey, Self::Error> { ) -> Result<authn::PublicKey, Self::Error> {
let signature = authn::Signature::try_from(solution.as_slice()).map_err(|()| { let signature = authn::Signature::try_from(solution.as_slice()).map_err(|_| {
error!("Failed to decode signature in challenge solution"); error!("Failed to decode signature in challenge solution");
Error::InvalidChallengeSolution Error::InvalidChallengeSolution
})?; })?;
let valid = key.verify(*challenge_nonce, USERAGENT_CONTEXT, &signature); let valid = key.verify(*challenge_nonce, USERAGENT_CONTEXT, &signature);
if valid { match valid {
self.transport true => {
.send(Ok(Outbound::AuthSuccess)) self.transport
.await .send(Ok(Outbound::AuthSuccess))
.map_err(|_| Error::Transport)?; .await
Ok(key.clone()) .map_err(|_| Error::Transport)?;
} else { Ok(key.clone())
self.transport }
.send(Err(Error::InvalidChallengeSolution)) false => {
.await self.transport
.map_err(|_| Error::Transport)?; .send(Err(Error::InvalidChallengeSolution))
Err(Error::InvalidChallengeSolution) .await
.map_err(|_| Error::Transport)?;
Err(Error::InvalidChallengeSolution)
}
} }
} }
} }

View File

@@ -28,7 +28,7 @@ pub struct UserAgentConnection {
} }
impl UserAgentConnection { impl UserAgentConnection {
pub const fn new(db: db::DatabasePool, actors: GlobalActors) -> Self { pub fn new(db: db::DatabasePool, actors: GlobalActors) -> Self {
Self { db, actors } Self { db, actors }
} }
} }

View File

@@ -17,28 +17,28 @@ mod state;
use state::{DummyContext, UserAgentEvents, UserAgentStateMachine}; use state::{DummyContext, UserAgentEvents, UserAgentStateMachine};
#[derive(Debug, Error)] #[derive(Debug, Error)]
pub enum UserAgentSessionError { pub enum Error {
#[error("Internal error: {message}")]
Internal { message: Cow<'static, str> },
#[error("State transition failed")] #[error("State transition failed")]
State, State,
#[error("Internal error: {message}")]
Internal { message: Cow<'static, str> },
} }
impl From<crate::db::PoolError> for UserAgentSessionError { impl From<crate::db::PoolError> for Error {
fn from(err: crate::db::PoolError) -> Self { fn from(err: crate::db::PoolError) -> Self {
error!(?err, "Database pool error"); error!(?err, "Database pool error");
Self::internal("Database pool error") Self::internal("Database pool error")
} }
} }
impl From<diesel::result::Error> for UserAgentSessionError { impl From<diesel::result::Error> for Error {
fn from(err: diesel::result::Error) -> Self { fn from(err: diesel::result::Error) -> Self {
error!(?err, "Database error"); error!(?err, "Database error");
Self::internal("Database error") Self::internal("Database error")
} }
} }
impl UserAgentSessionError { impl Error {
pub fn internal(message: impl Into<Cow<'static, str>>) -> Self { pub fn internal(message: impl Into<Cow<'static, str>>) -> Self {
Self::Internal { Self::Internal {
message: message.into(), message: message.into(),
@@ -67,7 +67,7 @@ impl UserAgentSession {
props, props,
state: UserAgentStateMachine::new(DummyContext), state: UserAgentStateMachine::new(DummyContext),
sender, sender,
pending_client_approvals: HashMap::default(), pending_client_approvals: Default::default(),
} }
} }
@@ -87,10 +87,10 @@ impl UserAgentSession {
Self::new(UserAgentConnection::new(db, actors), Box::new(DummySender)) Self::new(UserAgentConnection::new(db, actors), Box::new(DummySender))
} }
fn transition(&mut self, event: UserAgentEvents) -> Result<(), UserAgentSessionError> { fn transition(&mut self, event: UserAgentEvents) -> Result<(), Error> {
self.state.process_event(event).map_err(|e| { self.state.process_event(event).map_err(|e| {
error!(?e, "State transition failed"); error!(?e, "State transition failed");
UserAgentSessionError::State Error::State
})?; })?;
Ok(()) Ok(())
} }
@@ -132,11 +132,11 @@ impl UserAgentSession {
impl Actor for UserAgentSession { impl Actor for UserAgentSession {
type Args = Self; type Args = Self;
type Error = UserAgentSessionError; type Error = Error;
async fn on_start( async fn on_start(
args: Self::Args, args: Self::Args,
this: ActorRef<Self>, this: kameo::prelude::ActorRef<Self>,
) -> Result<Self, Self::Error> { ) -> Result<Self, Self::Error> {
args.props args.props
.actors .actors
@@ -150,9 +150,7 @@ impl Actor for UserAgentSession {
?err, ?err,
"Failed to register user agent connection with flow coordinator" "Failed to register user agent connection with flow coordinator"
); );
UserAgentSessionError::internal( Error::internal("Failed to register user agent connection with flow coordinator")
"Failed to register user agent connection with flow coordinator",
)
})?; })?;
Ok(args) Ok(args)
} }

View File

@@ -11,13 +11,12 @@ use diesel_async::{AsyncConnection, RunQueryDsl};
use kameo::error::SendError; use kameo::error::SendError;
use kameo::messages; use kameo::messages;
use kameo::prelude::Context; use kameo::prelude::Context;
use thiserror::Error;
use tracing::{error, info}; use tracing::{error, info};
use x25519_dalek::{EphemeralSecret, PublicKey}; use x25519_dalek::{EphemeralSecret, PublicKey};
use crate::{actors::flow_coordinator::client_connect_approval::ClientApprovalAnswer, evm::policies::SharedGrantSettings}; use crate::actors::flow_coordinator::client_connect_approval::ClientApprovalAnswer;
use crate::actors::keyholder::KeyHolderState; use crate::actors::keyholder::KeyHolderState;
use crate::actors::user_agent::session::UserAgentSessionError; use crate::actors::user_agent::session::Error;
use crate::actors::{ use crate::actors::{
evm::{ evm::{
ClientSignTransaction, Generate, ListWallets, SignTransactionError as EvmSignError, ClientSignTransaction, Generate, ListWallets, SignTransactionError as EvmSignError,
@@ -35,12 +34,10 @@ use crate::db::models::{
use crate::evm::policies::{Grant, SpecificGrant}; use crate::evm::policies::{Grant, SpecificGrant};
impl UserAgentSession { impl UserAgentSession {
fn take_unseal_secret(&self) -> Result<(EphemeralSecret, PublicKey), UserAgentSessionError> { fn take_unseal_secret(&mut self) -> Result<(EphemeralSecret, PublicKey), Error> {
let UserAgentStates::WaitingForUnsealKey(unseal_context) = self.state.state() else { let UserAgentStates::WaitingForUnsealKey(unseal_context) = self.state.state() else {
error!("Received encrypted key in invalid state"); error!("Received encrypted key in invalid state");
return Err(UserAgentSessionError::internal( return Err(Error::internal("Invalid state for unseal encrypted key"));
"Invalid state for unseal encrypted key",
));
}; };
let ephemeral_secret = { let ephemeral_secret = {
@@ -50,14 +47,13 @@ impl UserAgentSession {
)] )]
let mut secret_lock = unseal_context.secret.lock().unwrap(); let mut secret_lock = unseal_context.secret.lock().unwrap();
let secret = secret_lock.take(); let secret = secret_lock.take();
if let Some(secret) = secret { match secret {
secret Some(secret) => secret,
} else { None => {
drop(secret_lock); drop(secret_lock);
error!("Ephemeral secret already taken"); error!("Ephemeral secret already taken");
return Err(UserAgentSessionError::internal( return Err(Error::internal("Ephemeral secret already taken"));
"Ephemeral secret already taken", }
));
} }
}; };
@@ -83,7 +79,7 @@ impl UserAgentSession {
}); });
match decryption_result { match decryption_result {
Ok(()) => Ok(key_buffer), Ok(_) => Ok(key_buffer),
Err(err) => { Err(err) => {
error!(?err, "Failed to decrypt encrypted key material"); error!(?err, "Failed to decrypt encrypted key material");
Err(()) Err(())
@@ -101,7 +97,7 @@ pub enum UnsealError {
#[error("Invalid key provided for unsealing")] #[error("Invalid key provided for unsealing")]
InvalidKey, InvalidKey,
#[error("Internal error during unsealing process")] #[error("Internal error during unsealing process")]
General(#[from] UserAgentSessionError), General(#[from] super::Error),
} }
#[derive(Debug, Error)] #[derive(Debug, Error)]
@@ -112,7 +108,7 @@ pub enum BootstrapError {
AlreadyBootstrapped, AlreadyBootstrapped,
#[error("Internal error during bootstrapping process")] #[error("Internal error during bootstrapping process")]
General(#[from] UserAgentSessionError), General(#[from] super::Error),
} }
#[derive(Debug, Error)] #[derive(Debug, Error)]
@@ -136,16 +132,16 @@ pub enum GrantMutationError {
#[messages] #[messages]
impl UserAgentSession { impl UserAgentSession {
#[message] #[message]
pub fn handle_unseal_request( pub async fn handle_unseal_request(
&mut self, &mut self,
client_pubkey: PublicKey, client_pubkey: x25519_dalek::PublicKey,
) -> Result<UnsealStartResponse, UserAgentSessionError> { ) -> Result<UnsealStartResponse, Error> {
let secret = EphemeralSecret::random(); let secret = EphemeralSecret::random();
let public_key = PublicKey::from(&secret); let public_key = PublicKey::from(&secret);
self.transition(UserAgentEvents::UnsealRequest(UnsealContext { self.transition(UserAgentEvents::UnsealRequest(UnsealContext {
client_public_key: client_pubkey,
secret: Mutex::new(Some(secret)), secret: Mutex::new(Some(secret)),
client_public_key: client_pubkey,
}))?; }))?;
Ok(UnsealStartResponse { Ok(UnsealStartResponse {
@@ -162,24 +158,27 @@ impl UserAgentSession {
) -> Result<(), UnsealError> { ) -> Result<(), UnsealError> {
let (ephemeral_secret, client_public_key) = match self.take_unseal_secret() { let (ephemeral_secret, client_public_key) = match self.take_unseal_secret() {
Ok(values) => values, Ok(values) => values,
Err(UserAgentSessionError::State) => { Err(Error::State) => {
self.transition(UserAgentEvents::ReceivedInvalidKey)?; self.transition(UserAgentEvents::ReceivedInvalidKey)?;
return Err(UnsealError::InvalidKey); return Err(UnsealError::InvalidKey);
} }
Err(_err) => { Err(_err) => {
return Err(UserAgentSessionError::internal("Failed to take unseal secret").into()); return Err(Error::internal("Failed to take unseal secret").into());
} }
}; };
let Ok(seal_key_buffer) = Self::decrypt_client_key_material( let seal_key_buffer = match Self::decrypt_client_key_material(
ephemeral_secret, ephemeral_secret,
client_public_key, client_public_key,
&nonce, &nonce,
&ciphertext, &ciphertext,
&associated_data, &associated_data,
) else { ) {
self.transition(UserAgentEvents::ReceivedInvalidKey)?; Ok(buffer) => buffer,
return Err(UnsealError::InvalidKey); Err(()) => {
self.transition(UserAgentEvents::ReceivedInvalidKey)?;
return Err(UnsealError::InvalidKey);
}
}; };
match self match self
@@ -191,12 +190,12 @@ impl UserAgentSession {
}) })
.await .await
{ {
Ok(()) => { Ok(_) => {
info!("Successfully unsealed key with client-provided key"); info!("Successfully unsealed key with client-provided key");
self.transition(UserAgentEvents::ReceivedValidKey)?; self.transition(UserAgentEvents::ReceivedValidKey)?;
Ok(()) Ok(())
} }
Err(SendError::HandlerError(keyholder::KeyHolderError::InvalidKey)) => { Err(SendError::HandlerError(keyholder::Error::InvalidKey)) => {
self.transition(UserAgentEvents::ReceivedInvalidKey)?; self.transition(UserAgentEvents::ReceivedInvalidKey)?;
Err(UnsealError::InvalidKey) Err(UnsealError::InvalidKey)
} }
@@ -208,7 +207,7 @@ impl UserAgentSession {
Err(err) => { Err(err) => {
error!(?err, "Failed to send unseal request to keyholder"); error!(?err, "Failed to send unseal request to keyholder");
self.transition(UserAgentEvents::ReceivedInvalidKey)?; self.transition(UserAgentEvents::ReceivedInvalidKey)?;
Err(UserAgentSessionError::internal("Vault actor error").into()) Err(Error::internal("Vault actor error").into())
} }
} }
} }
@@ -222,22 +221,25 @@ impl UserAgentSession {
) -> Result<(), BootstrapError> { ) -> Result<(), BootstrapError> {
let (ephemeral_secret, client_public_key) = match self.take_unseal_secret() { let (ephemeral_secret, client_public_key) = match self.take_unseal_secret() {
Ok(values) => values, Ok(values) => values,
Err(UserAgentSessionError::State) => { Err(Error::State) => {
self.transition(UserAgentEvents::ReceivedInvalidKey)?; self.transition(UserAgentEvents::ReceivedInvalidKey)?;
return Err(BootstrapError::InvalidKey); return Err(BootstrapError::InvalidKey);
} }
Err(err) => return Err(err.into()), Err(err) => return Err(err.into()),
}; };
let Ok(seal_key_buffer) = Self::decrypt_client_key_material( let seal_key_buffer = match Self::decrypt_client_key_material(
ephemeral_secret, ephemeral_secret,
client_public_key, client_public_key,
&nonce, &nonce,
&ciphertext, &ciphertext,
&associated_data, &associated_data,
) else { ) {
self.transition(UserAgentEvents::ReceivedInvalidKey)?; Ok(buffer) => buffer,
return Err(BootstrapError::InvalidKey); Err(()) => {
self.transition(UserAgentEvents::ReceivedInvalidKey)?;
return Err(BootstrapError::InvalidKey);
}
}; };
match self match self
@@ -249,12 +251,12 @@ impl UserAgentSession {
}) })
.await .await
{ {
Ok(()) => { Ok(_) => {
info!("Successfully bootstrapped vault with client-provided key"); info!("Successfully bootstrapped vault with client-provided key");
self.transition(UserAgentEvents::ReceivedValidKey)?; self.transition(UserAgentEvents::ReceivedValidKey)?;
Ok(()) Ok(())
} }
Err(SendError::HandlerError(keyholder::KeyHolderError::AlreadyBootstrapped)) => { Err(SendError::HandlerError(keyholder::Error::AlreadyBootstrapped)) => {
self.transition(UserAgentEvents::ReceivedInvalidKey)?; self.transition(UserAgentEvents::ReceivedInvalidKey)?;
Err(BootstrapError::AlreadyBootstrapped) Err(BootstrapError::AlreadyBootstrapped)
} }
@@ -266,7 +268,7 @@ impl UserAgentSession {
Err(err) => { Err(err) => {
error!(?err, "Failed to send bootstrap request to keyholder"); error!(?err, "Failed to send bootstrap request to keyholder");
self.transition(UserAgentEvents::ReceivedInvalidKey)?; self.transition(UserAgentEvents::ReceivedInvalidKey)?;
Err(BootstrapError::General(UserAgentSessionError::internal( Err(BootstrapError::General(Error::internal(
"Vault actor error", "Vault actor error",
))) )))
} }
@@ -277,16 +279,14 @@ impl UserAgentSession {
#[messages] #[messages]
impl UserAgentSession { impl UserAgentSession {
#[message] #[message]
pub(crate) async fn handle_query_vault_state( pub(crate) async fn handle_query_vault_state(&mut self) -> Result<KeyHolderState, Error> {
&mut self,
) -> Result<KeyHolderState, UserAgentSessionError> {
use crate::actors::keyholder::GetState; use crate::actors::keyholder::GetState;
let vault_state = match self.props.actors.key_holder.ask(GetState {}).await { let vault_state = match self.props.actors.key_holder.ask(GetState {}).await {
Ok(state) => state, Ok(state) => state,
Err(err) => { Err(err) => {
error!(?err, actor = "useragent", "keyholder.query.failed"); error!(?err, actor = "useragent", "keyholder.query.failed");
return Err(UserAgentSessionError::internal("Vault is in broken state")); return Err(Error::internal("Vault is in broken state"));
} }
}; };
@@ -297,32 +297,26 @@ impl UserAgentSession {
#[messages] #[messages]
impl UserAgentSession { impl UserAgentSession {
#[message] #[message]
pub(crate) async fn handle_evm_wallet_create( pub(crate) async fn handle_evm_wallet_create(&mut self) -> Result<(i32, Address), Error> {
&mut self,
) -> Result<(i32, Address), UserAgentSessionError> {
match self.props.actors.evm.ask(Generate {}).await { match self.props.actors.evm.ask(Generate {}).await {
Ok(address) => Ok(address), Ok(address) => Ok(address),
Err(SendError::HandlerError(err)) => Err(UserAgentSessionError::internal(format!( Err(SendError::HandlerError(err)) => Err(Error::internal(format!(
"EVM wallet generation failed: {err}" "EVM wallet generation failed: {err}"
))), ))),
Err(err) => { Err(err) => {
error!(?err, "EVM actor unreachable during wallet create"); error!(?err, "EVM actor unreachable during wallet create");
Err(UserAgentSessionError::internal("EVM actor unreachable")) Err(Error::internal("EVM actor unreachable"))
} }
} }
} }
#[message] #[message]
pub(crate) async fn handle_evm_wallet_list( pub(crate) async fn handle_evm_wallet_list(&mut self) -> Result<Vec<(i32, Address)>, Error> {
&mut self,
) -> Result<Vec<(i32, Address)>, UserAgentSessionError> {
match self.props.actors.evm.ask(ListWallets {}).await { match self.props.actors.evm.ask(ListWallets {}).await {
Ok(wallets) => Ok(wallets), Ok(wallets) => Ok(wallets),
Err(err) => { Err(err) => {
error!(?err, "EVM wallet list failed"); error!(?err, "EVM wallet list failed");
Err(UserAgentSessionError::internal( Err(Error::internal("Failed to list EVM wallets"))
"Failed to list EVM wallets",
))
} }
} }
} }
@@ -331,14 +325,12 @@ impl UserAgentSession {
#[messages] #[messages]
impl UserAgentSession { impl UserAgentSession {
#[message] #[message]
pub(crate) async fn handle_grant_list( pub(crate) async fn handle_grant_list(&mut self) -> Result<Vec<Grant<SpecificGrant>>, Error> {
&mut self,
) -> Result<Vec<Grant<SpecificGrant>>, UserAgentSessionError> {
match self.props.actors.evm.ask(UseragentListGrants {}).await { match self.props.actors.evm.ask(UseragentListGrants {}).await {
Ok(grants) => Ok(grants), Ok(grants) => Ok(grants),
Err(err) => { Err(err) => {
error!(?err, "EVM grant list failed"); error!(?err, "EVM grant list failed");
Err(UserAgentSessionError::internal("Failed to list EVM grants")) Err(Error::internal("Failed to list EVM grants"))
} }
} }
} }
@@ -346,8 +338,8 @@ impl UserAgentSession {
#[message] #[message]
pub(crate) async fn handle_grant_create( pub(crate) async fn handle_grant_create(
&mut self, &mut self,
basic: SharedGrantSettings, basic: crate::evm::policies::SharedGrantSettings,
grant: SpecificGrant, grant: crate::evm::policies::SpecificGrant,
) -> Result<i32, GrantMutationError> { ) -> Result<i32, GrantMutationError> {
match self match self
.props .props
@@ -365,7 +357,6 @@ impl UserAgentSession {
} }
#[message] #[message]
#[expect(clippy::unused_async, reason = "false positive")]
pub(crate) async fn handle_grant_delete( pub(crate) async fn handle_grant_delete(
&mut self, &mut self,
grant_id: i32, grant_id: i32,
@@ -383,7 +374,7 @@ impl UserAgentSession {
// Err(GrantMutationError::Internal) // Err(GrantMutationError::Internal)
// } // }
// } // }
let _ = grant_id; let _ = grant_id;
todo!() todo!()
} }
@@ -420,7 +411,7 @@ impl UserAgentSession {
pub(crate) async fn handle_grant_evm_wallet_access( pub(crate) async fn handle_grant_evm_wallet_access(
&mut self, &mut self,
entries: Vec<NewEvmWalletAccess>, entries: Vec<NewEvmWalletAccess>,
) -> Result<(), UserAgentSessionError> { ) -> Result<(), Error> {
let mut conn = self.props.db.get().await?; let mut conn = self.props.db.get().await?;
conn.transaction(|conn| { conn.transaction(|conn| {
Box::pin(async move { Box::pin(async move {
@@ -434,7 +425,7 @@ impl UserAgentSession {
.await?; .await?;
} }
Result::<_, UserAgentSessionError>::Ok(()) Result::<_, Error>::Ok(())
}) })
}) })
.await?; .await?;
@@ -445,7 +436,7 @@ impl UserAgentSession {
pub(crate) async fn handle_revoke_evm_wallet_access( pub(crate) async fn handle_revoke_evm_wallet_access(
&mut self, &mut self,
entries: Vec<i32>, entries: Vec<i32>,
) -> Result<(), UserAgentSessionError> { ) -> Result<(), Error> {
let mut conn = self.props.db.get().await?; let mut conn = self.props.db.get().await?;
conn.transaction(|conn| { conn.transaction(|conn| {
Box::pin(async move { Box::pin(async move {
@@ -457,7 +448,7 @@ impl UserAgentSession {
.await?; .await?;
} }
Result::<_, UserAgentSessionError>::Ok(()) Result::<_, Error>::Ok(())
}) })
}) })
.await?; .await?;
@@ -467,9 +458,10 @@ impl UserAgentSession {
#[message] #[message]
pub(crate) async fn handle_list_wallet_access( pub(crate) async fn handle_list_wallet_access(
&mut self, &mut self,
) -> Result<Vec<EvmWalletAccess>, UserAgentSessionError> { ) -> Result<Vec<EvmWalletAccess>, Error> {
let mut conn = self.props.db.get().await?; let mut conn = self.props.db.get().await?;
let access_entries = crate::db::schema::evm_wallet_access::table use crate::db::schema::evm_wallet_access;
let access_entries = evm_wallet_access::table
.select(EvmWalletAccess::as_select()) .select(EvmWalletAccess::as_select())
.load::<_>(&mut conn) .load::<_>(&mut conn)
.await?; .await?;
@@ -484,14 +476,14 @@ impl UserAgentSession {
&mut self, &mut self,
approved: bool, approved: bool,
pubkey: authn::PublicKey, pubkey: authn::PublicKey,
ctx: &Context<Self, Result<(), UserAgentSessionError>>, ctx: &mut Context<Self, Result<(), Error>>,
) -> Result<(), UserAgentSessionError> { ) -> Result<(), Error> {
let Some(pending_approval) = self.pending_client_approvals.remove(&pubkey.to_bytes()) let pending_approval = match self.pending_client_approvals.remove(&pubkey.to_bytes()) {
else { Some(approval) => approval,
error!("Received client connection response for unknown client"); None => {
return Err(UserAgentSessionError::internal( error!("Received client connection response for unknown client");
"Unknown client in connection response", return Err(Error::internal("Unknown client in connection response"));
)); }
}; };
pending_approval pending_approval
@@ -503,9 +495,7 @@ impl UserAgentSession {
?err, ?err,
"Failed to send client approval response to controller" "Failed to send client approval response to controller"
); );
UserAgentSessionError::internal( Error::internal("Failed to send client approval response to controller")
"Failed to send client approval response to controller",
)
})?; })?;
ctx.actor_ref().unlink(&pending_approval.controller).await; ctx.actor_ref().unlink(&pending_approval.controller).await;
@@ -516,7 +506,7 @@ impl UserAgentSession {
#[message] #[message]
pub(crate) async fn handle_sdk_client_list( pub(crate) async fn handle_sdk_client_list(
&mut self, &mut self,
) -> Result<Vec<(ProgramClient, ProgramClientMetadata)>, UserAgentSessionError> { ) -> Result<Vec<(ProgramClient, ProgramClientMetadata)>, Error> {
use crate::db::schema::{client_metadata, program_client}; use crate::db::schema::{client_metadata, program_client};
let mut conn = self.props.db.get().await?; let mut conn = self.props.db.get().await?;

View File

@@ -19,6 +19,8 @@ smlang::statemachine!(
pub struct DummyContext; pub struct DummyContext;
impl UserAgentStateMachineContext for DummyContext { impl UserAgentStateMachineContext for DummyContext {
#[allow(missing_docs)]
#[allow(clippy::unused_unit)]
fn generate_temp_keypair(&mut self, event_data: UnsealContext) -> Result<UnsealContext, ()> { fn generate_temp_keypair(&mut self, event_data: UnsealContext) -> Result<UnsealContext, ()> {
Ok(event_data) Ok(event_data)
} }

View File

@@ -25,22 +25,22 @@ pub enum InitError {
Tls(#[from] tls::InitError), Tls(#[from] tls::InitError),
#[error("Actor spawn failed: {0}")] #[error("Actor spawn failed: {0}")]
ActorSpawn(#[from] crate::actors::GlobalActorsSpawnError), ActorSpawn(#[from] crate::actors::SpawnError),
#[error("I/O Error: {0}")] #[error("I/O Error: {0}")]
Io(#[from] std::io::Error), Io(#[from] std::io::Error),
} }
pub struct __ServerContextInner { pub struct _ServerContextInner {
pub db: db::DatabasePool, pub db: db::DatabasePool,
pub tls: TlsManager, pub tls: TlsManager,
pub actors: GlobalActors, pub actors: GlobalActors,
} }
#[derive(Clone)] #[derive(Clone)]
pub struct ServerContext(Arc<__ServerContextInner>); pub struct ServerContext(Arc<_ServerContextInner>);
impl std::ops::Deref for ServerContext { impl std::ops::Deref for ServerContext {
type Target = __ServerContextInner; type Target = _ServerContextInner;
fn deref(&self) -> &Self::Target { fn deref(&self) -> &Self::Target {
&self.0 &self.0
@@ -49,7 +49,7 @@ impl std::ops::Deref for ServerContext {
impl ServerContext { impl ServerContext {
pub async fn new(db: db::DatabasePool) -> Result<Self, InitError> { pub async fn new(db: db::DatabasePool) -> Result<Self, InitError> {
Ok(Self(Arc::new(__ServerContextInner { Ok(Self(Arc::new(_ServerContextInner {
actors: GlobalActors::spawn(db.clone()).await?, actors: GlobalActors::spawn(db.clone()).await?,
tls: TlsManager::new(db.clone()).await?, tls: TlsManager::new(db.clone()).await?,
db, db,

View File

@@ -22,10 +22,9 @@ use crate::db::{
}; };
const ENCODE_CONFIG: pem::EncodeConfig = { const ENCODE_CONFIG: pem::EncodeConfig = {
let line_ending = if cfg!(target_family = "windows") { let line_ending = match cfg!(target_family = "windows") {
pem::LineEnding::CRLF true => pem::LineEnding::CRLF,
} else { false => pem::LineEnding::LF,
pem::LineEnding::LF
}; };
pem::EncodeConfig::new().set_line_ending(line_ending) pem::EncodeConfig::new().set_line_ending(line_ending)
}; };
@@ -53,14 +52,11 @@ pub enum InitError {
pub type PemCert = String; pub type PemCert = String;
pub fn encode_cert_to_pem(cert: &CertificateDer<'_>) -> PemCert { pub fn encode_cert_to_pem(cert: &CertificateDer) -> PemCert {
pem::encode_config(&Pem::new("CERTIFICATE", cert.to_vec()), ENCODE_CONFIG) pem::encode_config(&Pem::new("CERTIFICATE", cert.to_vec()), ENCODE_CONFIG)
} }
#[expect( #[allow(unused)]
unused,
reason = "may be needed for future cert rotation implementation"
)]
struct SerializedTls { struct SerializedTls {
cert_pem: PemCert, cert_pem: PemCert,
cert_key_pem: String, cert_key_pem: String,
@@ -89,7 +85,7 @@ impl TlsCa {
let cert_key_pem = certified_issuer.key().serialize_pem(); let cert_key_pem = certified_issuer.key().serialize_pem();
#[expect( #[allow(
clippy::unwrap_used, clippy::unwrap_used,
reason = "Broken cert couldn't bootstrap server anyway" reason = "Broken cert couldn't bootstrap server anyway"
)] )]
@@ -128,11 +124,7 @@ impl TlsCa {
}) })
} }
#[expect( #[allow(unused)]
unused,
clippy::unnecessary_wraps,
reason = "may be needed for future cert rotation implementation"
)]
fn serialize(&self) -> Result<SerializedTls, InitError> { fn serialize(&self) -> Result<SerializedTls, InitError> {
let cert_key_pem = self.issuer.key().serialize_pem(); let cert_key_pem = self.issuer.key().serialize_pem();
Ok(SerializedTls { Ok(SerializedTls {
@@ -141,10 +133,7 @@ impl TlsCa {
}) })
} }
#[expect( #[allow(unused)]
unused,
reason = "may be needed for future cert rotation implementation"
)]
fn try_deserialize(cert_pem: &str, cert_key_pem: &str) -> Result<Self, InitError> { fn try_deserialize(cert_pem: &str, cert_key_pem: &str) -> Result<Self, InitError> {
let keypair = let keypair =
KeyPair::from_pem(cert_key_pem).map_err(InitError::KeyDeserializationError)?; KeyPair::from_pem(cert_key_pem).map_err(InitError::KeyDeserializationError)?;
@@ -245,10 +234,10 @@ impl TlsManager {
} }
} }
pub const fn cert(&self) -> &CertificateDer<'static> { pub fn cert(&self) -> &CertificateDer<'static> {
&self.cert &self.cert
} }
pub const fn ca_cert(&self) -> &CertificateDer<'static> { pub fn ca_cert(&self) -> &CertificateDer<'static> {
&self.ca_cert &self.ca_cert
} }

View File

@@ -5,8 +5,8 @@ use rand::{
rngs::{StdRng, SysRng}, rngs::{StdRng, SysRng},
}; };
pub const ROOT_KEY_TAG: &[u8] = b"arbiter/seal/v1"; pub const ROOT_KEY_TAG: &[u8] = "arbiter/seal/v1".as_bytes();
pub const TAG: &[u8] = b"arbiter/private-key/v1"; pub const TAG: &[u8] = "arbiter/private-key/v1".as_bytes();
pub const NONCE_LENGTH: usize = 24; pub const NONCE_LENGTH: usize = 24;
@@ -15,13 +15,11 @@ pub struct Nonce(pub [u8; NONCE_LENGTH]);
impl Nonce { impl Nonce {
pub fn increment(&mut self) { pub fn increment(&mut self) {
for i in (0..self.0.len()).rev() { for i in (0..self.0.len()).rev() {
if let Some(byte) = self.0.get_mut(i) { if self.0[i] == 0xFF {
if *byte == 0xFF { self.0[i] = 0;
*byte = 0; } else {
} else { self.0[i] += 1;
*byte += 1; break;
break;
}
} }
} }
} }
@@ -47,14 +45,19 @@ pub type Salt = [u8; ArgonSalt::RECOMMENDED_LENGTH];
pub fn generate_salt() -> Salt { pub fn generate_salt() -> Salt {
let mut salt = Salt::default(); let mut salt = Salt::default();
let mut rng = #[allow(
StdRng::try_from_rng(&mut SysRng).expect("Rng failure is unrecoverable and should panic"); clippy::unwrap_used,
reason = "Rng failure is unrecoverable and should panic"
)]
let mut rng = StdRng::try_from_rng(&mut SysRng).unwrap();
rng.fill_bytes(&mut salt); rng.fill_bytes(&mut salt);
salt salt
} }
#[cfg(test)] #[cfg(test)]
mod tests { mod tests {
use std::ops::Deref as _;
use super::*; use super::*;
use crate::crypto::derive_key; use crate::crypto::derive_key;
use arbiter_crypto::safecell::{SafeCell, SafeCellHandle as _}; use arbiter_crypto::safecell::{SafeCell, SafeCellHandle as _};
@@ -72,7 +75,7 @@ mod tests {
let key1_reader = key1.0.read(); let key1_reader = key1.0.read();
let key2_reader = key2.0.read(); let key2_reader = key2.0.read();
assert_eq!(&*key1_reader, &*key2_reader); assert_eq!(key1_reader.deref(), key2_reader.deref());
} }
#[test] #[test]
@@ -83,13 +86,14 @@ mod tests {
let mut key = derive_key(password, &salt); let mut key = derive_key(password, &salt);
let key_reader = key.0.read(); let key_reader = key.0.read();
let key_ref = key_reader.deref();
assert_ne!(key_reader.as_slice(), &[0u8; 32][..]); assert_ne!(key_ref.as_slice(), &[0u8; 32][..]);
} }
#[test] #[test]
// We should fuzz this // We should fuzz this
pub fn nonce_increment() { pub fn test_nonce_increment() {
let mut nonce = Nonce([0u8; NONCE_LENGTH]); let mut nonce = Nonce([0u8; NONCE_LENGTH]);
nonce.increment(); nonce.increment();

View File

@@ -18,12 +18,12 @@ use crate::{
}; };
#[derive(Debug, thiserror::Error)] #[derive(Debug, thiserror::Error)]
pub enum IntegrityError { pub enum Error {
#[error("Database error: {0}")] #[error("Database error: {0}")]
Database(#[from] db::DatabaseError), Database(#[from] db::DatabaseError),
#[error("KeyHolder error: {0}")] #[error("KeyHolder error: {0}")]
Keyholder(#[from] keyholder::KeyHolderError), Keyholder(#[from] keyholder::Error),
#[error("KeyHolder mailbox error")] #[error("KeyHolder mailbox error")]
KeyholderSend, KeyholderSend,
@@ -67,11 +67,6 @@ fn payload_hash(payload: &impl Hashable) -> [u8; 32] {
} }
fn push_len_prefixed(out: &mut Vec<u8>, bytes: &[u8]) { fn push_len_prefixed(out: &mut Vec<u8>, bytes: &[u8]) {
#[expect(
clippy::cast_possible_truncation,
clippy::as_conversions,
reason = "fixme! #85"
)]
out.extend_from_slice(&(bytes.len() as u32).to_be_bytes()); out.extend_from_slice(&(bytes.len() as u32).to_be_bytes());
out.extend_from_slice(bytes); out.extend_from_slice(bytes);
} }
@@ -111,7 +106,7 @@ pub async fn sign_entity<E: Integrable>(
keyholder: &ActorRef<KeyHolder>, keyholder: &ActorRef<KeyHolder>,
entity: &E, entity: &E,
entity_id: impl IntoId, entity_id: impl IntoId,
) -> Result<(), IntegrityError> { ) -> Result<(), Error> {
let payload_hash = payload_hash(&entity); let payload_hash = payload_hash(&entity);
let entity_id = entity_id.into_id(); let entity_id = entity_id.into_id();
@@ -122,8 +117,8 @@ pub async fn sign_entity<E: Integrable>(
.ask(SignIntegrity { mac_input }) .ask(SignIntegrity { mac_input })
.await .await
.map_err(|err| match err { .map_err(|err| match err {
SendError::HandlerError(inner) => IntegrityError::Keyholder(inner), kameo::error::SendError::HandlerError(inner) => Error::Keyholder(inner),
_ => IntegrityError::KeyholderSend, _ => Error::KeyholderSend,
})?; })?;
insert_into(integrity_envelope::table) insert_into(integrity_envelope::table)
@@ -132,7 +127,7 @@ pub async fn sign_entity<E: Integrable>(
entity_id, entity_id,
payload_version: E::VERSION, payload_version: E::VERSION,
key_version, key_version,
mac: mac.clone(), mac: mac.to_vec(),
}) })
.on_conflict(( .on_conflict((
integrity_envelope::entity_id, integrity_envelope::entity_id,
@@ -156,7 +151,7 @@ pub async fn verify_entity<E: Integrable>(
keyholder: &ActorRef<KeyHolder>, keyholder: &ActorRef<KeyHolder>,
entity: &E, entity: &E,
entity_id: impl IntoId, entity_id: impl IntoId,
) -> Result<AttestationStatus, IntegrityError> { ) -> Result<AttestationStatus, Error> {
let entity_id = entity_id.into_id(); let entity_id = entity_id.into_id();
let envelope: IntegrityEnvelope = integrity_envelope::table let envelope: IntegrityEnvelope = integrity_envelope::table
.filter(integrity_envelope::entity_kind.eq(E::KIND)) .filter(integrity_envelope::entity_kind.eq(E::KIND))
@@ -164,14 +159,14 @@ pub async fn verify_entity<E: Integrable>(
.first(conn) .first(conn)
.await .await
.map_err(|err| match err { .map_err(|err| match err {
diesel::result::Error::NotFound => IntegrityError::MissingEnvelope { diesel::result::Error::NotFound => Error::MissingEnvelope {
entity_kind: E::KIND, entity_kind: E::KIND,
}, },
other => IntegrityError::Database(db::DatabaseError::from(other)), other => Error::Database(db::DatabaseError::from(other)),
})?; })?;
if envelope.payload_version != E::VERSION { if envelope.payload_version != E::VERSION {
return Err(IntegrityError::PayloadVersionMismatch { return Err(Error::PayloadVersionMismatch {
entity_kind: E::KIND, entity_kind: E::KIND,
expected: E::VERSION, expected: E::VERSION,
found: envelope.payload_version, found: envelope.payload_version,
@@ -191,13 +186,13 @@ pub async fn verify_entity<E: Integrable>(
match result { match result {
Ok(true) => Ok(AttestationStatus::Attested), Ok(true) => Ok(AttestationStatus::Attested),
Ok(false) => Err(IntegrityError::MacMismatch { Ok(false) => Err(Error::MacMismatch {
entity_kind: E::KIND, entity_kind: E::KIND,
}), }),
Err(SendError::HandlerError(keyholder::KeyHolderError::NotBootstrapped)) => { Err(SendError::HandlerError(keyholder::Error::NotBootstrapped)) => {
Ok(AttestationStatus::Unavailable) Ok(AttestationStatus::Unavailable)
} }
Err(_) => Err(IntegrityError::KeyholderSend), Err(_) => Err(Error::KeyholderSend),
} }
} }
@@ -213,7 +208,7 @@ mod tests {
}; };
use arbiter_crypto::safecell::{SafeCell, SafeCellHandle as _}; use arbiter_crypto::safecell::{SafeCell, SafeCellHandle as _};
use super::{Integrable, IntegrityError, sign_entity, verify_entity}; use super::{Error, Integrable, sign_entity, verify_entity};
#[derive(Clone, arbiter_macros::Hashable)] #[derive(Clone, arbiter_macros::Hashable)]
struct DummyEntity { struct DummyEntity {
payload_version: i32, payload_version: i32,
@@ -236,12 +231,12 @@ mod tests {
#[tokio::test] #[tokio::test]
async fn sign_writes_envelope_and_verify_passes() { async fn sign_writes_envelope_and_verify_passes() {
const ENTITY_ID: &[u8] = b"entity-id-7";
let db = db::create_test_pool().await; let db = db::create_test_pool().await;
let keyholder = bootstrapped_keyholder(&db).await; let keyholder = bootstrapped_keyholder(&db).await;
let mut conn = db.get().await.unwrap(); let mut conn = db.get().await.unwrap();
const ENTITY_ID: &[u8] = b"entity-id-7";
let entity = DummyEntity { let entity = DummyEntity {
payload_version: 1, payload_version: 1,
payload: b"payload-v1".to_vec(), payload: b"payload-v1".to_vec(),
@@ -267,12 +262,12 @@ mod tests {
#[tokio::test] #[tokio::test]
async fn tampered_mac_fails_verification() { async fn tampered_mac_fails_verification() {
const ENTITY_ID: &[u8] = b"entity-id-11";
let db = db::create_test_pool().await; let db = db::create_test_pool().await;
let keyholder = bootstrapped_keyholder(&db).await; let keyholder = bootstrapped_keyholder(&db).await;
let mut conn = db.get().await.unwrap(); let mut conn = db.get().await.unwrap();
const ENTITY_ID: &[u8] = b"entity-id-11";
let entity = DummyEntity { let entity = DummyEntity {
payload_version: 1, payload_version: 1,
payload: b"payload-v1".to_vec(), payload: b"payload-v1".to_vec(),
@@ -293,17 +288,17 @@ mod tests {
let err = verify_entity(&mut conn, &keyholder, &entity, ENTITY_ID) let err = verify_entity(&mut conn, &keyholder, &entity, ENTITY_ID)
.await .await
.unwrap_err(); .unwrap_err();
assert!(matches!(err, IntegrityError::MacMismatch { .. })); assert!(matches!(err, Error::MacMismatch { .. }));
} }
#[tokio::test] #[tokio::test]
async fn changed_payload_fails_verification() { async fn changed_payload_fails_verification() {
const ENTITY_ID: &[u8] = b"entity-id-21";
let db = db::create_test_pool().await; let db = db::create_test_pool().await;
let keyholder = bootstrapped_keyholder(&db).await; let keyholder = bootstrapped_keyholder(&db).await;
let mut conn = db.get().await.unwrap(); let mut conn = db.get().await.unwrap();
const ENTITY_ID: &[u8] = b"entity-id-21";
let entity = DummyEntity { let entity = DummyEntity {
payload_version: 1, payload_version: 1,
payload: b"payload-v1".to_vec(), payload: b"payload-v1".to_vec(),
@@ -321,6 +316,6 @@ mod tests {
let err = verify_entity(&mut conn, &keyholder, &tampered, ENTITY_ID) let err = verify_entity(&mut conn, &keyholder, &tampered, ENTITY_ID)
.await .await
.unwrap_err(); .unwrap_err();
assert!(matches!(err, IntegrityError::MacMismatch { .. })); assert!(matches!(err, Error::MacMismatch { .. }));
} }
} }

View File

@@ -1,3 +1,5 @@
use std::ops::Deref as _;
use argon2::{Algorithm, Argon2}; use argon2::{Algorithm, Argon2};
use chacha20poly1305::{ use chacha20poly1305::{
AeadInPlace, Key, KeyInit as _, XChaCha20Poly1305, XNonce, AeadInPlace, Key, KeyInit as _, XChaCha20Poly1305, XNonce,
@@ -39,8 +41,11 @@ impl TryFrom<SafeCell<Vec<u8>>> for KeyCell {
impl KeyCell { impl KeyCell {
pub fn new_secure_random() -> Self { pub fn new_secure_random() -> Self {
let key = SafeCell::new_inline(|key_buffer: &mut Key| { let key = SafeCell::new_inline(|key_buffer: &mut Key| {
let mut rng = StdRng::try_from_rng(&mut SysRng) #[allow(
.expect("Rng failure is unrecoverable and should panic"); clippy::unwrap_used,
reason = "Rng failure is unrecoverable and should panic"
)]
let mut rng = StdRng::try_from_rng(&mut SysRng).unwrap();
rng.fill_bytes(key_buffer); rng.fill_bytes(key_buffer);
}); });
@@ -54,7 +59,8 @@ impl KeyCell {
mut buffer: impl AsMut<Vec<u8>>, mut buffer: impl AsMut<Vec<u8>>,
) -> Result<(), Error> { ) -> Result<(), Error> {
let key_reader = self.0.read(); let key_reader = self.0.read();
let cipher = XChaCha20Poly1305::new(&key_reader); let key_ref = key_reader.deref();
let cipher = XChaCha20Poly1305::new(key_ref);
let nonce = XNonce::from_slice(nonce.0.as_ref()); let nonce = XNonce::from_slice(nonce.0.as_ref());
let buffer = buffer.as_mut(); let buffer = buffer.as_mut();
cipher.encrypt_in_place(nonce, associated_data, buffer) cipher.encrypt_in_place(nonce, associated_data, buffer)
@@ -66,7 +72,8 @@ impl KeyCell {
buffer: &mut SafeCell<Vec<u8>>, buffer: &mut SafeCell<Vec<u8>>,
) -> Result<(), Error> { ) -> Result<(), Error> {
let key_reader = self.0.read(); let key_reader = self.0.read();
let cipher = XChaCha20Poly1305::new(&key_reader); let key_ref = key_reader.deref();
let cipher = XChaCha20Poly1305::new(key_ref);
let nonce = XNonce::from_slice(nonce.0.as_ref()); let nonce = XNonce::from_slice(nonce.0.as_ref());
let mut buffer = buffer.write(); let mut buffer = buffer.write();
let buffer: &mut Vec<u8> = buffer.as_mut(); let buffer: &mut Vec<u8> = buffer.as_mut();
@@ -80,7 +87,8 @@ impl KeyCell {
plaintext: impl AsRef<[u8]>, plaintext: impl AsRef<[u8]>,
) -> Result<Vec<u8>, Error> { ) -> Result<Vec<u8>, Error> {
let key_reader = self.0.read(); let key_reader = self.0.read();
let mut cipher = XChaCha20Poly1305::new(&key_reader); let key_ref = key_reader.deref();
let mut cipher = XChaCha20Poly1305::new(key_ref);
let nonce = XNonce::from_slice(nonce.0.as_ref()); let nonce = XNonce::from_slice(nonce.0.as_ref());
let ciphertext = cipher.encrypt( let ciphertext = cipher.encrypt(
@@ -108,15 +116,20 @@ pub fn derive_key(mut password: SafeCell<Vec<u8>>, salt: &Salt) -> KeyCell {
} }
}; };
#[allow(clippy::unwrap_used)]
let hasher = Argon2::new(Algorithm::Argon2id, argon2::Version::V0x13, params); let hasher = Argon2::new(Algorithm::Argon2id, argon2::Version::V0x13, params);
let mut key = SafeCell::new(Key::default()); let mut key = SafeCell::new(Key::default());
password.read_inline(|password_source| { password.read_inline(|password_source| {
let mut key_buffer = key.write(); let mut key_buffer = key.write();
let key_buffer: &mut [u8] = key_buffer.as_mut(); let key_buffer: &mut [u8] = key_buffer.as_mut();
#[allow(
clippy::unwrap_used,
reason = "Better fail completely than return a weak key"
)]
hasher hasher
.hash_password_into(password_source, salt, key_buffer) .hash_password_into(password_source.deref(), salt, key_buffer)
.expect("Better fail completely than return a weak key"); .unwrap();
}); });
key.into() key.into()

View File

@@ -23,14 +23,14 @@ const MIGRATIONS: EmbeddedMigrations = embed_migrations!("migrations");
#[derive(Error, Debug)] #[derive(Error, Debug)]
pub enum DatabaseSetupError { pub enum DatabaseSetupError {
#[error(transparent)] #[error("Failed to determine home directory")]
ConcurrencySetup(diesel::result::Error), HomeDir(std::io::Error),
#[error(transparent)] #[error(transparent)]
Connection(diesel::ConnectionError), Connection(diesel::ConnectionError),
#[error("Failed to determine home directory")] #[error(transparent)]
HomeDir(std::io::Error), ConcurrencySetup(diesel::result::Error),
#[error(transparent)] #[error(transparent)]
Migration(Box<dyn std::error::Error + Send + Sync>), Migration(Box<dyn std::error::Error + Send + Sync>),
@@ -41,11 +41,10 @@ pub enum DatabaseSetupError {
#[derive(Error, Debug)] #[derive(Error, Debug)]
pub enum DatabaseError { pub enum DatabaseError {
#[error("Database query error")]
Connection(#[from] diesel::result::Error),
#[error("Database connection error")] #[error("Database connection error")]
Pool(#[from] PoolError), Pool(#[from] PoolError),
#[error("Database query error")]
Connection(#[from] diesel::result::Error),
} }
#[tracing::instrument(level = "info")] #[tracing::instrument(level = "info")]
@@ -94,16 +93,13 @@ fn initialize_database(url: &str) -> Result<(), DatabaseSetupError> {
} }
#[tracing::instrument(level = "info")] #[tracing::instrument(level = "info")]
/// Creates a connection pool for the `SQLite` database.
///
/// # Panics
/// Panics if the database path is not valid UTF-8.
pub async fn create_pool(url: Option<&str>) -> Result<DatabasePool, DatabaseSetupError> { pub async fn create_pool(url: Option<&str>) -> Result<DatabasePool, DatabaseSetupError> {
let database_url = url.map(String::from).unwrap_or( let database_url = url.map(String::from).unwrap_or(
#[allow(clippy::expect_used)]
database_path()? database_path()?
.to_str() .to_str()
.expect("database path is not valid UTF-8") .expect("database path is not valid UTF-8")
.to_owned(), .to_string(),
); );
initialize_database(&database_url)?; initialize_database(&database_url)?;
@@ -138,19 +134,19 @@ pub async fn create_pool(url: Option<&str>) -> Result<DatabasePool, DatabaseSetu
} }
#[mutants::skip] #[mutants::skip]
#[expect(clippy::missing_panics_doc, reason = "Tests oriented function")]
/// Creates a test database pool with a temporary `SQLite` database file.
pub async fn create_test_pool() -> DatabasePool { pub async fn create_test_pool() -> DatabasePool {
use rand::distr::{Alphanumeric, SampleString as _}; use rand::distr::{Alphanumeric, SampleString as _};
let tempfile_name = Alphanumeric.sample_string(&mut rand::rng(), 16); let tempfile_name = Alphanumeric.sample_string(&mut rand::rng(), 16);
let file = std::env::temp_dir().join(tempfile_name); let file = std::env::temp_dir().join(tempfile_name);
#[allow(clippy::expect_used)]
let url = file let url = file
.to_str() .to_str()
.expect("temp file path is not valid UTF-8") .expect("temp file path is not valid UTF-8")
.to_owned(); .to_string();
#[allow(clippy::expect_used)]
create_pool(Some(&url)) create_pool(Some(&url))
.await .await
.expect("Failed to create test database pool") .expect("Failed to create test database pool")

View File

@@ -1,7 +1,5 @@
#![allow( #![allow(unused)]
clippy::duplicated_attributes, #![allow(clippy::all)]
reason = "restructed's #[view] causes false positives"
)]
use crate::db::schema::{ use crate::db::schema::{
self, aead_encrypted, arbiter_settings, evm_basic_grant, evm_ether_transfer_grant, self, aead_encrypted, arbiter_settings, evm_basic_grant, evm_ether_transfer_grant,
@@ -9,6 +7,7 @@ use crate::db::schema::{
evm_token_transfer_log, evm_token_transfer_volume_limit, evm_transaction_log, evm_wallet, evm_token_transfer_log, evm_token_transfer_volume_limit, evm_transaction_log, evm_wallet,
integrity_envelope, root_key_history, tls_history, integrity_envelope, root_key_history, tls_history,
}; };
use chrono::{DateTime, Utc};
use diesel::{prelude::*, sqlite::Sqlite}; use diesel::{prelude::*, sqlite::Sqlite};
use restructed::Models; use restructed::Models;
@@ -28,16 +27,16 @@ pub mod types {
pub struct SqliteTimestamp(pub DateTime<Utc>); pub struct SqliteTimestamp(pub DateTime<Utc>);
impl SqliteTimestamp { impl SqliteTimestamp {
pub fn now() -> Self { pub fn now() -> Self {
Self(Utc::now()) SqliteTimestamp(Utc::now())
} }
} }
impl From<DateTime<Utc>> for SqliteTimestamp { impl From<chrono::DateTime<Utc>> for SqliteTimestamp {
fn from(dt: DateTime<Utc>) -> Self { fn from(dt: chrono::DateTime<Utc>) -> Self {
Self(dt) SqliteTimestamp(dt)
} }
} }
impl From<SqliteTimestamp> for DateTime<Utc> { impl From<SqliteTimestamp> for chrono::DateTime<Utc> {
fn from(ts: SqliteTimestamp) -> Self { fn from(ts: SqliteTimestamp) -> Self {
ts.0 ts.0
} }
@@ -48,11 +47,6 @@ pub mod types {
&'b self, &'b self,
out: &mut diesel::serialize::Output<'b, '_, Sqlite>, out: &mut diesel::serialize::Output<'b, '_, Sqlite>,
) -> diesel::serialize::Result { ) -> diesel::serialize::Result {
#[expect(
clippy::cast_possible_truncation,
clippy::as_conversions,
reason = "fixme! #84; this will break up in 2038 :3"
)]
let unix_timestamp = self.0.timestamp() as i32; let unix_timestamp = self.0.timestamp() as i32;
out.set_value(unix_timestamp); out.set_value(unix_timestamp);
Ok(IsNull::No) Ok(IsNull::No)
@@ -75,47 +69,7 @@ pub mod types {
let datetime = let datetime =
DateTime::from_timestamp(unix_timestamp, 0).ok_or("Timestamp is out of bounds")?; DateTime::from_timestamp(unix_timestamp, 0).ok_or("Timestamp is out of bounds")?;
Ok(Self(datetime)) Ok(SqliteTimestamp(datetime))
}
}
#[derive(Debug, FromSqlRow, AsExpression, Clone)]
#[diesel(sql_type = Integer)]
#[repr(transparent)] // hint compiler to optimize the wrapper struct away
pub struct ChainId(pub i32);
#[expect(
clippy::cast_sign_loss,
clippy::cast_possible_truncation,
clippy::as_conversions,
reason = "safe because chain_id is stored as i32 but is guaranteed to be a valid ChainId by the API when creating grants"
)]
const _: () = {
impl From<ChainId> for alloy::primitives::ChainId {
fn from(chain_id: ChainId) -> Self {
chain_id.0 as Self
}
}
impl From<alloy::primitives::ChainId> for ChainId {
fn from(chain_id: alloy::primitives::ChainId) -> Self {
Self(chain_id as _)
}
}
};
impl FromSql<Integer, Sqlite> for ChainId {
fn from_sql(
bytes: <Sqlite as diesel::backend::Backend>::RawValue<'_>,
) -> diesel::deserialize::Result<Self> {
FromSql::<Integer, Sqlite>::from_sql(bytes).map(Self)
}
}
impl ToSql<Integer, Sqlite> for ChainId {
fn to_sql<'b>(
&'b self,
out: &mut diesel::serialize::Output<'b, '_, Sqlite>,
) -> diesel::serialize::Result {
ToSql::<Integer, Sqlite>::to_sql(&self.0, out)
} }
} }
} }
@@ -283,7 +237,7 @@ pub struct EvmEtherTransferLimit {
pub struct EvmBasicGrant { pub struct EvmBasicGrant {
pub id: i32, pub id: i32,
pub wallet_access_id: i32, // references evm_wallet_access.id pub wallet_access_id: i32, // references evm_wallet_access.id
pub chain_id: ChainId, pub chain_id: i32,
pub valid_from: Option<SqliteTimestamp>, pub valid_from: Option<SqliteTimestamp>,
pub valid_until: Option<SqliteTimestamp>, pub valid_until: Option<SqliteTimestamp>,
pub max_gas_fee_per_gas: Option<Vec<u8>>, pub max_gas_fee_per_gas: Option<Vec<u8>>,
@@ -306,7 +260,7 @@ pub struct EvmTransactionLog {
pub id: i32, pub id: i32,
pub grant_id: i32, pub grant_id: i32,
pub wallet_access_id: i32, pub wallet_access_id: i32,
pub chain_id: ChainId, pub chain_id: i32,
pub eth_value: Vec<u8>, pub eth_value: Vec<u8>,
pub signed_at: SqliteTimestamp, pub signed_at: SqliteTimestamp,
} }
@@ -381,7 +335,7 @@ pub struct EvmTokenTransferLog {
pub id: i32, pub id: i32,
pub grant_id: i32, pub grant_id: i32,
pub log_id: i32, pub log_id: i32,
pub chain_id: ChainId, pub chain_id: i32,
pub token_contract: Vec<u8>, pub token_contract: Vec<u8>,
pub recipient_address: Vec<u8>, pub recipient_address: Vec<u8>,
pub value: Vec<u8>, pub value: Vec<u8>,

View File

@@ -45,7 +45,7 @@ sol! {
sol! { sol! {
/// Permit2 — Uniswap's canonical token approval manager. /// Permit2 — Uniswap's canonical token approval manager.
/// Replaces per-contract ERC-20 `approve()` with a single approval hub. /// Replaces per-contract ERC-20 approve() with a single approval hub.
#[derive(Debug)] #[derive(Debug)]
interface IPermit2 { interface IPermit2 {
struct TokenPermissions { struct TokenPermissions {

View File

@@ -34,14 +34,14 @@ mod utils;
#[derive(Debug, thiserror::Error)] #[derive(Debug, thiserror::Error)]
pub enum PolicyError { pub enum PolicyError {
#[error("Database error")] #[error("Database error")]
Database(#[from] DatabaseError), Database(#[from] crate::db::DatabaseError),
#[error("Transaction violates policy: {0:?}")] #[error("Transaction violates policy: {0:?}")]
Violations(Vec<EvalViolation>), Violations(Vec<EvalViolation>),
#[error("No matching grant found")] #[error("No matching grant found")]
NoMatchingGrant, NoMatchingGrant,
#[error("Integrity error: {0}")] #[error("Integrity error: {0}")]
Integrity(#[from] integrity::IntegrityError), Integrity(#[from] integrity::Error),
} }
#[derive(Debug, thiserror::Error)] #[derive(Debug, thiserror::Error)]
@@ -66,10 +66,10 @@ pub enum AnalyzeError {
#[derive(Debug, thiserror::Error)] #[derive(Debug, thiserror::Error)]
pub enum ListError { pub enum ListError {
#[error("Database error")] #[error("Database error")]
Database(#[from] DatabaseError), Database(#[from] crate::db::DatabaseError),
#[error("Integrity verification failed for grant")] #[error("Integrity verification failed for grant")]
Integrity(#[from] integrity::IntegrityError), Integrity(#[from] integrity::Error),
} }
/// Controls whether a transaction should be executed or only validated /// Controls whether a transaction should be executed or only validated
@@ -127,7 +127,7 @@ async fn check_shared_constraints(
.get_result(conn) .get_result(conn)
.await?; .await?;
if count >= rate_limit.count.into() { if count >= rate_limit.count as i64 {
violations.push(EvalViolation::RateLimitExceeded); violations.push(EvalViolation::RateLimitExceeded);
} }
} }
@@ -185,7 +185,7 @@ impl Engine {
.values(&NewEvmTransactionLog { .values(&NewEvmTransactionLog {
grant_id: grant.common_settings_id, grant_id: grant.common_settings_id,
wallet_access_id: context.target.id, wallet_access_id: context.target.id,
chain_id: context.chain.into(), chain_id: context.chain as i32,
eth_value: utils::u256_to_bytes(context.value).to_vec(), eth_value: utils::u256_to_bytes(context.value).to_vec(),
signed_at: Utc::now().into(), signed_at: Utc::now().into(),
}) })
@@ -207,7 +207,7 @@ impl Engine {
} }
impl Engine { impl Engine {
pub const fn new(db: db::DatabasePool, keyholder: ActorRef<KeyHolder>) -> Self { pub fn new(db: db::DatabasePool, keyholder: ActorRef<KeyHolder>) -> Self {
Self { db, keyholder } Self { db, keyholder }
} }
@@ -226,15 +226,9 @@ impl Engine {
Box::pin(async move { Box::pin(async move {
use schema::evm_basic_grant; use schema::evm_basic_grant;
#[expect(
clippy::cast_possible_truncation,
clippy::cast_possible_wrap,
clippy::as_conversions,
reason = "fixme! #86"
)]
let basic_grant: EvmBasicGrant = insert_into(evm_basic_grant::table) let basic_grant: EvmBasicGrant = insert_into(evm_basic_grant::table)
.values(&NewEvmBasicGrant { .values(&NewEvmBasicGrant {
chain_id: full_grant.shared.chain.into(), chain_id: full_grant.shared.chain as i32,
wallet_access_id: full_grant.shared.wallet_access_id, wallet_access_id: full_grant.shared.wallet_access_id,
valid_from: full_grant.shared.valid_from.map(SqliteTimestamp), valid_from: full_grant.shared.valid_from.map(SqliteTimestamp),
valid_until: full_grant.shared.valid_until.map(SqliteTimestamp), valid_until: full_grant.shared.valid_until.map(SqliteTimestamp),
@@ -319,7 +313,7 @@ impl Engine {
let TxKind::Call(to) = transaction.to else { let TxKind::Call(to) = transaction.to else {
return Err(VetError::ContractCreationNotSupported); return Err(VetError::ContractCreationNotSupported);
}; };
let context = EvalContext { let context = policies::EvalContext {
target, target,
chain: transaction.chain_id, chain: transaction.chain_id,
to, to,
@@ -400,6 +394,7 @@ mod tests {
chain: CHAIN_ID, chain: CHAIN_ID,
valid_from: None, valid_from: None,
valid_until: None, valid_until: None,
revoked_at: None,
max_gas_fee_per_gas: None, max_gas_fee_per_gas: None,
max_priority_fee_per_gas: None, max_priority_fee_per_gas: None,
rate_limit: None, rate_limit: None,
@@ -410,16 +405,10 @@ mod tests {
conn: &mut DatabaseConnection, conn: &mut DatabaseConnection,
shared: &SharedGrantSettings, shared: &SharedGrantSettings,
) -> EvmBasicGrant { ) -> EvmBasicGrant {
#[expect(
clippy::cast_possible_truncation,
clippy::cast_possible_wrap,
clippy::as_conversions,
reason = "fixme! #86"
)]
insert_into(evm_basic_grant::table) insert_into(evm_basic_grant::table)
.values(NewEvmBasicGrant { .values(NewEvmBasicGrant {
wallet_access_id: shared.wallet_access_id, wallet_access_id: shared.wallet_access_id,
chain_id: shared.chain.into(), chain_id: shared.chain as i32,
valid_from: shared.valid_from.map(SqliteTimestamp), valid_from: shared.valid_from.map(SqliteTimestamp),
valid_until: shared.valid_until.map(SqliteTimestamp), valid_until: shared.valid_until.map(SqliteTimestamp),
max_gas_fee_per_gas: shared max_gas_fee_per_gas: shared
@@ -583,7 +572,7 @@ mod tests {
.values(NewEvmTransactionLog { .values(NewEvmTransactionLog {
grant_id: basic_grant.id, grant_id: basic_grant.id,
wallet_access_id: WALLET_ACCESS_ID, wallet_access_id: WALLET_ACCESS_ID,
chain_id: CHAIN_ID.into(), chain_id: CHAIN_ID as i32,
eth_value: super::utils::u256_to_bytes(U256::ZERO).to_vec(), eth_value: super::utils::u256_to_bytes(U256::ZERO).to_vec(),
signed_at: SqliteTimestamp(Utc::now()), signed_at: SqliteTimestamp(Utc::now()),
}) })
@@ -608,4 +597,24 @@ mod tests {
assert!(violations.is_empty()); assert!(violations.is_empty());
} }
} }
#[test]
fn shared_settings_hash_changes_when_revoked_at_changes() {
use arbiter_crypto::hashing::Hashable;
use sha2::Digest;
let active = shared_settings();
let revoked = SharedGrantSettings {
revoked_at: Some(Utc::now()),
..shared_settings()
};
let mut active_hash = sha2::Sha256::new();
active.hash(&mut active_hash);
let mut revoked_hash = sha2::Sha256::new();
revoked.hash(&mut revoked_hash);
assert_ne!(active_hash.finalize(), revoked_hash.finalize());
}
} }

View File

@@ -11,7 +11,7 @@ use thiserror::Error;
use crate::{ use crate::{
crypto::integrity::v1::Integrable, crypto::integrity::v1::Integrable,
db::models::{EvmBasicGrant, EvmWalletAccess}, db::models::{self, EvmBasicGrant, EvmWalletAccess},
evm::utils, evm::utils,
}; };
@@ -87,10 +87,10 @@ pub trait Policy: Sized {
// Create a new grant in the database based on the provided grant details, and return its ID // Create a new grant in the database based on the provided grant details, and return its ID
fn create_grant( fn create_grant(
basic: &EvmBasicGrant, basic: &models::EvmBasicGrant,
grant: &Self::Settings, grant: &Self::Settings,
conn: &mut impl AsyncConnection<Backend = Sqlite>, conn: &mut impl AsyncConnection<Backend = Sqlite>,
) -> impl Future<Output = QueryResult<DatabaseID>> + Send; ) -> impl std::future::Future<Output = QueryResult<DatabaseID>> + Send;
// Try to find an existing grant that matches the transaction context, and return its details if found // Try to find an existing grant that matches the transaction context, and return its details if found
// Additionally, return ID of basic grant for shared-logic checks like rate limits and validity periods // Additionally, return ID of basic grant for shared-logic checks like rate limits and validity periods
@@ -146,6 +146,7 @@ pub struct SharedGrantSettings {
pub valid_from: Option<DateTime<Utc>>, pub valid_from: Option<DateTime<Utc>>,
pub valid_until: Option<DateTime<Utc>>, pub valid_until: Option<DateTime<Utc>>,
pub revoked_at: Option<DateTime<Utc>>,
pub max_gas_fee_per_gas: Option<U256>, pub max_gas_fee_per_gas: Option<U256>,
pub max_priority_fee_per_gas: Option<U256>, pub max_priority_fee_per_gas: Option<U256>,
@@ -157,9 +158,10 @@ impl SharedGrantSettings {
pub(crate) fn try_from_model(model: EvmBasicGrant) -> QueryResult<Self> { pub(crate) fn try_from_model(model: EvmBasicGrant) -> QueryResult<Self> {
Ok(Self { Ok(Self {
wallet_access_id: model.wallet_access_id, wallet_access_id: model.wallet_access_id,
chain: model.chain_id.into(), chain: model.chain_id as u64, // safe because chain_id is stored as i32 but is guaranteed to be a valid ChainId by the API when creating grants
valid_from: model.valid_from.map(Into::into), valid_from: model.valid_from.map(Into::into),
valid_until: model.valid_until.map(Into::into), valid_until: model.valid_until.map(Into::into),
revoked_at: model.revoked_at.map(Into::into),
max_gas_fee_per_gas: model max_gas_fee_per_gas: model
.max_gas_fee_per_gas .max_gas_fee_per_gas
.map(|b| utils::try_bytes_to_u256(&b)) .map(|b| utils::try_bytes_to_u256(&b))
@@ -168,11 +170,10 @@ impl SharedGrantSettings {
.max_priority_fee_per_gas .max_priority_fee_per_gas
.map(|b| utils::try_bytes_to_u256(&b)) .map(|b| utils::try_bytes_to_u256(&b))
.transpose()?, .transpose()?,
#[expect(clippy::cast_sign_loss, clippy::as_conversions, reason = "fixme! #86")]
rate_limit: match (model.rate_limit_count, model.rate_limit_window_secs) { rate_limit: match (model.rate_limit_count, model.rate_limit_window_secs) {
(Some(count), Some(window_secs)) => Some(TransactionRateLimit { (Some(count), Some(window_secs)) => Some(TransactionRateLimit {
count: count as u32, count: count as u32,
window: Duration::seconds(window_secs.into()), window: Duration::seconds(window_secs as i64),
}), }),
_ => None, _ => None,
}, },
@@ -182,7 +183,7 @@ impl SharedGrantSettings {
pub async fn query_by_id( pub async fn query_by_id(
conn: &mut impl AsyncConnection<Backend = Sqlite>, conn: &mut impl AsyncConnection<Backend = Sqlite>,
id: i32, id: i32,
) -> QueryResult<Self> { ) -> diesel::result::QueryResult<Self> {
use crate::db::schema::evm_basic_grant; use crate::db::schema::evm_basic_grant;
let basic_grant: EvmBasicGrant = evm_basic_grant::table let basic_grant: EvmBasicGrant = evm_basic_grant::table

View File

@@ -4,8 +4,8 @@ use std::fmt::Display;
use alloy::primitives::{Address, U256}; use alloy::primitives::{Address, U256};
use chrono::{DateTime, Duration, Utc}; use chrono::{DateTime, Duration, Utc};
use diesel::dsl::{auto_type, insert_into}; use diesel::dsl::{auto_type, insert_into};
use diesel::prelude::*;
use diesel::sqlite::Sqlite; use diesel::sqlite::Sqlite;
use diesel::{ExpressionMethods, JoinOnDsl, prelude::*};
use diesel_async::{AsyncConnection, RunQueryDsl}; use diesel_async::{AsyncConnection, RunQueryDsl};
use crate::crypto::integrity::v1::Integrable; use crate::crypto::integrity::v1::Integrable;
@@ -19,7 +19,7 @@ use crate::evm::policies::{
}; };
use crate::{ use crate::{
db::{ db::{
models::{NewEvmEtherTransferGrant, NewEvmEtherTransferGrantTarget}, models::{self, NewEvmEtherTransferGrant, NewEvmEtherTransferGrantTarget},
schema::{evm_ether_transfer_grant, evm_ether_transfer_grant_target}, schema::{evm_ether_transfer_grant, evm_ether_transfer_grant_target},
}, },
evm::{policies::Policy, utils}, evm::{policies::Policy, utils},
@@ -46,8 +46,8 @@ impl Display for Meaning {
} }
} }
impl From<Meaning> for SpecificMeaning { impl From<Meaning> for SpecificMeaning {
fn from(val: Meaning) -> Self { fn from(val: Meaning) -> SpecificMeaning {
Self::EtherTransfer(val) SpecificMeaning::EtherTransfer(val)
} }
} }
@@ -62,8 +62,8 @@ impl Integrable for Settings {
} }
impl From<Settings> for SpecificGrant { impl From<Settings> for SpecificGrant {
fn from(val: Settings) -> Self { fn from(val: Settings) -> SpecificGrant {
Self::EtherTransfer(val) SpecificGrant::EtherTransfer(val)
} }
} }
@@ -74,7 +74,9 @@ async fn query_relevant_past_transaction(
) -> QueryResult<Vec<(U256, DateTime<Utc>)>> { ) -> QueryResult<Vec<(U256, DateTime<Utc>)>> {
let past_transactions: Vec<(Vec<u8>, SqliteTimestamp)> = evm_transaction_log::table let past_transactions: Vec<(Vec<u8>, SqliteTimestamp)> = evm_transaction_log::table
.filter(evm_transaction_log::grant_id.eq(grant_id)) .filter(evm_transaction_log::grant_id.eq(grant_id))
.filter(evm_transaction_log::signed_at.ge(SqliteTimestamp(Utc::now() - longest_window))) .filter(
evm_transaction_log::signed_at.ge(SqliteTimestamp(chrono::Utc::now() - longest_window)),
)
.select(( .select((
evm_transaction_log::eth_value, evm_transaction_log::eth_value,
evm_transaction_log::signed_at, evm_transaction_log::signed_at,
@@ -101,7 +103,7 @@ async fn check_rate_limits(
let past_transaction = query_relevant_past_transaction(grant.id, window, db).await?; let past_transaction = query_relevant_past_transaction(grant.id, window, db).await?;
let window_start = Utc::now() - grant.settings.specific.limit.window; let window_start = chrono::Utc::now() - grant.settings.specific.limit.window;
let prospective_cumulative_volume: U256 = past_transaction let prospective_cumulative_volume: U256 = past_transaction
.iter() .iter()
.filter(|(_, timestamp)| timestamp >= &window_start) .filter(|(_, timestamp)| timestamp >= &window_start)
@@ -151,15 +153,10 @@ impl Policy for EtherTransfer {
} }
async fn create_grant( async fn create_grant(
basic: &EvmBasicGrant, basic: &models::EvmBasicGrant,
grant: &Self::Settings, grant: &Self::Settings,
conn: &mut impl AsyncConnection<Backend = Sqlite>, conn: &mut impl AsyncConnection<Backend = Sqlite>,
) -> QueryResult<DatabaseID> { ) -> diesel::result::QueryResult<DatabaseID> {
#[expect(
clippy::cast_possible_truncation,
clippy::as_conversions,
reason = "fixme! #86"
)]
let limit_id: i32 = insert_into(evm_ether_transfer_limit::table) let limit_id: i32 = insert_into(evm_ether_transfer_limit::table)
.values(NewEvmEtherTransferLimit { .values(NewEvmEtherTransferLimit {
window_secs: grant.limit.window.num_seconds() as i32, window_secs: grant.limit.window.num_seconds() as i32,
@@ -194,7 +191,7 @@ impl Policy for EtherTransfer {
async fn try_find_grant( async fn try_find_grant(
context: &EvalContext, context: &EvalContext,
conn: &mut impl AsyncConnection<Backend = Sqlite>, conn: &mut impl AsyncConnection<Backend = Sqlite>,
) -> QueryResult<Option<Grant<Self::Settings>>> { ) -> diesel::result::QueryResult<Option<Grant<Self::Settings>>> {
let target_bytes = context.to.to_vec(); let target_bytes = context.to.to_vec();
// Find a grant where: // Find a grant where:
@@ -248,7 +245,7 @@ impl Policy for EtherTransfer {
limit: VolumeRateLimit { limit: VolumeRateLimit {
max_volume: utils::try_bytes_to_u256(&limit.max_volume) max_volume: utils::try_bytes_to_u256(&limit.max_volume)
.map_err(|err| diesel::result::Error::DeserializationError(Box::new(err)))?, .map_err(|err| diesel::result::Error::DeserializationError(Box::new(err)))?,
window: Duration::seconds(limit.window_secs.into()), window: chrono::Duration::seconds(limit.window_secs as i64),
}, },
}; };
@@ -268,7 +265,7 @@ impl Policy for EtherTransfer {
_log_id: i32, _log_id: i32,
_grant: &Grant<Self::Settings>, _grant: &Grant<Self::Settings>,
_conn: &mut impl AsyncConnection<Backend = Sqlite>, _conn: &mut impl AsyncConnection<Backend = Sqlite>,
) -> QueryResult<()> { ) -> diesel::result::QueryResult<()> {
// Basic log is sufficient // Basic log is sufficient
Ok(()) Ok(())
@@ -321,7 +318,7 @@ impl Policy for EtherTransfer {
.map(|(basic, specific)| { .map(|(basic, specific)| {
let targets: Vec<Address> = targets_by_grant let targets: Vec<Address> = targets_by_grant
.get(&specific.id) .get(&specific.id)
.map(Vec::as_slice) .map(|v| v.as_slice())
.unwrap_or_default() .unwrap_or_default()
.iter() .iter()
.filter_map(|t| { .filter_map(|t| {
@@ -345,7 +342,7 @@ impl Policy for EtherTransfer {
max_volume: utils::try_bytes_to_u256(&limit.max_volume).map_err( max_volume: utils::try_bytes_to_u256(&limit.max_volume).map_err(
|e| diesel::result::Error::DeserializationError(Box::new(e)), |e| diesel::result::Error::DeserializationError(Box::new(e)),
)?, )?,
window: Duration::seconds(limit.window_secs.into()), window: Duration::seconds(limit.window_secs as i64),
}, },
}, },
}, },

View File

@@ -21,7 +21,7 @@ use crate::evm::{
use super::{EtherTransfer, Settings}; use super::{EtherTransfer, Settings};
const WALLET_ACCESS_ID: i32 = 1; const WALLET_ACCESS_ID: i32 = 1;
const CHAIN_ID: alloy::primitives::ChainId = 1; const CHAIN_ID: u64 = 1;
const ALLOWED: Address = address!("1111111111111111111111111111111111111111"); const ALLOWED: Address = address!("1111111111111111111111111111111111111111");
const OTHER: Address = address!("2222222222222222222222222222222222222222"); const OTHER: Address = address!("2222222222222222222222222222222222222222");
@@ -47,7 +47,7 @@ async fn insert_basic(conn: &mut DatabaseConnection, revoked: bool) -> EvmBasicG
insert_into(evm_basic_grant::table) insert_into(evm_basic_grant::table)
.values(NewEvmBasicGrant { .values(NewEvmBasicGrant {
wallet_access_id: WALLET_ACCESS_ID, wallet_access_id: WALLET_ACCESS_ID,
chain_id: CHAIN_ID.into(), chain_id: CHAIN_ID as i32,
valid_from: None, valid_from: None,
valid_until: None, valid_until: None,
max_gas_fee_per_gas: None, max_gas_fee_per_gas: None,
@@ -78,6 +78,7 @@ fn shared() -> SharedGrantSettings {
chain: CHAIN_ID, chain: CHAIN_ID,
valid_from: None, valid_from: None,
valid_until: None, valid_until: None,
revoked_at: None,
max_gas_fee_per_gas: None, max_gas_fee_per_gas: None,
max_priority_fee_per_gas: None, max_priority_fee_per_gas: None,
rate_limit: None, rate_limit: None,
@@ -160,7 +161,7 @@ async fn evaluate_passes_when_volume_within_limit() {
.values(NewEvmTransactionLog { .values(NewEvmTransactionLog {
grant_id, grant_id,
wallet_access_id: WALLET_ACCESS_ID, wallet_access_id: WALLET_ACCESS_ID,
chain_id: CHAIN_ID.into(), chain_id: CHAIN_ID as i32,
eth_value: utils::u256_to_bytes(U256::from(500u64)).to_vec(), eth_value: utils::u256_to_bytes(U256::from(500u64)).to_vec(),
signed_at: SqliteTimestamp(Utc::now()), signed_at: SqliteTimestamp(Utc::now()),
}) })
@@ -202,7 +203,7 @@ async fn evaluate_rejects_volume_over_limit() {
.values(NewEvmTransactionLog { .values(NewEvmTransactionLog {
grant_id, grant_id,
wallet_access_id: WALLET_ACCESS_ID, wallet_access_id: WALLET_ACCESS_ID,
chain_id: CHAIN_ID.into(), chain_id: CHAIN_ID as i32,
eth_value: utils::u256_to_bytes(U256::from(1_000u64)).to_vec(), eth_value: utils::u256_to_bytes(U256::from(1_000u64)).to_vec(),
signed_at: SqliteTimestamp(Utc::now()), signed_at: SqliteTimestamp(Utc::now()),
}) })
@@ -245,7 +246,7 @@ async fn evaluate_passes_at_exactly_volume_limit() {
.values(NewEvmTransactionLog { .values(NewEvmTransactionLog {
grant_id, grant_id,
wallet_access_id: WALLET_ACCESS_ID, wallet_access_id: WALLET_ACCESS_ID,
chain_id: CHAIN_ID.into(), chain_id: CHAIN_ID as i32,
eth_value: utils::u256_to_bytes(U256::from(900u64)).to_vec(), eth_value: utils::u256_to_bytes(U256::from(900u64)).to_vec(),
signed_at: SqliteTimestamp(Utc::now()), signed_at: SqliteTimestamp(Utc::now()),
}) })

View File

@@ -27,8 +27,8 @@ use alloy::{
use arbiter_tokens_registry::evm::nonfungible::{self, TokenInfo}; use arbiter_tokens_registry::evm::nonfungible::{self, TokenInfo};
use chrono::{DateTime, Duration, Utc}; use chrono::{DateTime, Duration, Utc};
use diesel::dsl::{auto_type, insert_into}; use diesel::dsl::{auto_type, insert_into};
use diesel::prelude::*;
use diesel::sqlite::Sqlite; use diesel::sqlite::Sqlite;
use diesel::{ExpressionMethods, prelude::*};
use diesel_async::{AsyncConnection, RunQueryDsl}; use diesel_async::{AsyncConnection, RunQueryDsl};
use super::{DatabaseID, EvalContext, EvalViolation}; use super::{DatabaseID, EvalContext, EvalViolation};
@@ -56,8 +56,8 @@ impl std::fmt::Display for Meaning {
} }
} }
impl From<Meaning> for SpecificMeaning { impl From<Meaning> for SpecificMeaning {
fn from(val: Meaning) -> Self { fn from(val: Meaning) -> SpecificMeaning {
Self::TokenTransfer(val) SpecificMeaning::TokenTransfer(val)
} }
} }
@@ -73,8 +73,8 @@ impl Integrable for Settings {
} }
impl From<Settings> for SpecificGrant { impl From<Settings> for SpecificGrant {
fn from(val: Settings) -> Self { fn from(val: Settings) -> SpecificGrant {
Self::TokenTransfer(val) SpecificGrant::TokenTransfer(val)
} }
} }
@@ -85,7 +85,10 @@ async fn query_relevant_past_transfers(
) -> QueryResult<Vec<(U256, DateTime<Utc>)>> { ) -> QueryResult<Vec<(U256, DateTime<Utc>)>> {
let past_logs: Vec<(Vec<u8>, SqliteTimestamp)> = evm_token_transfer_log::table let past_logs: Vec<(Vec<u8>, SqliteTimestamp)> = evm_token_transfer_log::table
.filter(evm_token_transfer_log::grant_id.eq(grant_id)) .filter(evm_token_transfer_log::grant_id.eq(grant_id))
.filter(evm_token_transfer_log::created_at.ge(SqliteTimestamp(Utc::now() - longest_window))) .filter(
evm_token_transfer_log::created_at
.ge(SqliteTimestamp(chrono::Utc::now() - longest_window)),
)
.select(( .select((
evm_token_transfer_log::value, evm_token_transfer_log::value,
evm_token_transfer_log::created_at, evm_token_transfer_log::created_at,
@@ -125,7 +128,7 @@ async fn check_volume_rate_limits(
let past_transfers = query_relevant_past_transfers(grant.id, longest_window, db).await?; let past_transfers = query_relevant_past_transfers(grant.id, longest_window, db).await?;
for limit in &grant.settings.specific.volume_limits { for limit in &grant.settings.specific.volume_limits {
let window_start = Utc::now() - limit.window; let window_start = chrono::Utc::now() - limit.window;
let prospective_cumulative_volume: U256 = past_transfers let prospective_cumulative_volume: U256 = past_transfers
.iter() .iter()
.filter(|(_, timestamp)| timestamp >= &window_start) .filter(|(_, timestamp)| timestamp >= &window_start)
@@ -201,11 +204,6 @@ impl Policy for TokenTransfer {
.await?; .await?;
for limit in &grant.volume_limits { for limit in &grant.volume_limits {
#[expect(
clippy::cast_possible_truncation,
clippy::as_conversions,
reason = "fixme! #86"
)]
insert_into(evm_token_transfer_volume_limit::table) insert_into(evm_token_transfer_volume_limit::table)
.values(NewEvmTokenTransferVolumeLimit { .values(NewEvmTokenTransferVolumeLimit {
grant_id, grant_id,
@@ -255,7 +253,7 @@ impl Policy for TokenTransfer {
max_volume: utils::try_bytes_to_u256(&row.max_volume).map_err(|err| { max_volume: utils::try_bytes_to_u256(&row.max_volume).map_err(|err| {
diesel::result::Error::DeserializationError(Box::new(err)) diesel::result::Error::DeserializationError(Box::new(err))
})?, })?,
window: Duration::seconds(row.window_secs.into()), window: Duration::seconds(row.window_secs as i64),
}) })
}) })
.collect::<QueryResult<Vec<_>>>()?; .collect::<QueryResult<Vec<_>>>()?;
@@ -305,7 +303,7 @@ impl Policy for TokenTransfer {
.values(NewEvmTokenTransferLog { .values(NewEvmTokenTransferLog {
grant_id: grant.id, grant_id: grant.id,
log_id, log_id,
chain_id: context.chain.into(), chain_id: context.chain as i32,
token_contract: context.to.to_vec(), token_contract: context.to.to_vec(),
recipient_address: meaning.to.to_vec(), recipient_address: meaning.to.to_vec(),
value: utils::u256_to_bytes(meaning.value).to_vec(), value: utils::u256_to_bytes(meaning.value).to_vec(),
@@ -354,7 +352,7 @@ impl Policy for TokenTransfer {
.map(|(basic, specific)| { .map(|(basic, specific)| {
let volume_limits: Vec<VolumeRateLimit> = limits_by_grant let volume_limits: Vec<VolumeRateLimit> = limits_by_grant
.get(&specific.id) .get(&specific.id)
.map(Vec::as_slice) .map(|v| v.as_slice())
.unwrap_or_default() .unwrap_or_default()
.iter() .iter()
.map(|row| { .map(|row| {
@@ -362,7 +360,7 @@ impl Policy for TokenTransfer {
max_volume: utils::try_bytes_to_u256(&row.max_volume).map_err(|e| { max_volume: utils::try_bytes_to_u256(&row.max_volume).map_err(|e| {
diesel::result::Error::DeserializationError(Box::new(e)) diesel::result::Error::DeserializationError(Box::new(e))
})?, })?,
window: Duration::seconds(row.window_secs.into()), window: Duration::seconds(row.window_secs as i64),
}) })
}) })
.collect::<QueryResult<Vec<_>>>()?; .collect::<QueryResult<Vec<_>>>()?;

View File

@@ -59,7 +59,7 @@ async fn insert_basic(conn: &mut DatabaseConnection, revoked: bool) -> EvmBasicG
insert_into(evm_basic_grant::table) insert_into(evm_basic_grant::table)
.values(NewEvmBasicGrant { .values(NewEvmBasicGrant {
wallet_access_id: WALLET_ACCESS_ID, wallet_access_id: WALLET_ACCESS_ID,
chain_id: CHAIN_ID.into(), chain_id: CHAIN_ID as i32,
valid_from: None, valid_from: None,
valid_until: None, valid_until: None,
max_gas_fee_per_gas: None, max_gas_fee_per_gas: None,
@@ -95,6 +95,7 @@ fn shared() -> SharedGrantSettings {
chain: CHAIN_ID, chain: CHAIN_ID,
valid_from: None, valid_from: None,
valid_until: None, valid_until: None,
revoked_at: None,
max_gas_fee_per_gas: None, max_gas_fee_per_gas: None,
max_priority_fee_per_gas: None, max_priority_fee_per_gas: None,
rate_limit: None, rate_limit: None,
@@ -238,11 +239,12 @@ async fn evaluate_passes_volume_at_exact_limit() {
.unwrap(); .unwrap();
// Record a past transfer of 900, with current transfer 100 => exactly 1000 limit // Record a past transfer of 900, with current transfer 100 => exactly 1000 limit
insert_into(db::schema::evm_token_transfer_log::table) use crate::db::{models::NewEvmTokenTransferLog, schema::evm_token_transfer_log};
.values(db::models::NewEvmTokenTransferLog { insert_into(evm_token_transfer_log::table)
.values(NewEvmTokenTransferLog {
grant_id, grant_id,
log_id: 0, log_id: 0,
chain_id: CHAIN_ID.into(), chain_id: CHAIN_ID as i32,
token_contract: DAI.to_vec(), token_contract: DAI.to_vec(),
recipient_address: RECIPIENT.to_vec(), recipient_address: RECIPIENT.to_vec(),
value: utils::u256_to_bytes(U256::from(900u64)).to_vec(), value: utils::u256_to_bytes(U256::from(900u64)).to_vec(),
@@ -282,11 +284,12 @@ async fn evaluate_rejects_volume_over_limit() {
.await .await
.unwrap(); .unwrap();
insert_into(db::schema::evm_token_transfer_log::table) use crate::db::{models::NewEvmTokenTransferLog, schema::evm_token_transfer_log};
.values(db::models::NewEvmTokenTransferLog { insert_into(evm_token_transfer_log::table)
.values(NewEvmTokenTransferLog {
grant_id, grant_id,
log_id: 0, log_id: 0,
chain_id: CHAIN_ID.into(), chain_id: CHAIN_ID as i32,
token_contract: DAI.to_vec(), token_contract: DAI.to_vec(),
recipient_address: RECIPIENT.to_vec(), recipient_address: RECIPIENT.to_vec(),
value: utils::u256_to_bytes(U256::from(1_000u64)).to_vec(), value: utils::u256_to_bytes(U256::from(1_000u64)).to_vec(),

View File

@@ -82,8 +82,8 @@ impl SafeSigner {
}) })
} }
#[expect(clippy::significant_drop_tightening, reason = "false positive")]
fn sign_hash_inner(&self, hash: &B256) -> Result<Signature> { fn sign_hash_inner(&self, hash: &B256) -> Result<Signature> {
#[allow(clippy::expect_used)]
let mut cell = self.key.lock().expect("SafeSigner mutex poisoned"); let mut cell = self.key.lock().expect("SafeSigner mutex poisoned");
let reader = cell.read(); let reader = cell.read();
let sig: (ecdsa::Signature, RecoveryId) = reader.sign_prehash(hash.as_ref())?; let sig: (ecdsa::Signature, RecoveryId) = reader.sign_prehash(hash.as_ref())?;
@@ -96,6 +96,7 @@ impl SafeSigner {
{ {
return Err(Error::TransactionChainIdMismatch { return Err(Error::TransactionChainIdMismatch {
signer: chain_id, signer: chain_id,
#[allow(clippy::expect_used)]
tx: tx.chain_id().expect("Chain ID is guaranteed to be set"), tx: tx.chain_id().expect("Chain ID is guaranteed to be set"),
}); });
} }

View File

@@ -7,7 +7,7 @@ pub struct LengthError {
pub actual: usize, pub actual: usize,
} }
pub const fn u256_to_bytes(value: U256) -> [u8; 32] { pub fn u256_to_bytes(value: U256) -> [u8; 32] {
value.to_le_bytes() value.to_le_bytes()
} }
pub fn bytes_to_u256(bytes: &[u8]) -> Option<U256> { pub fn bytes_to_u256(bytes: &[u8]) -> Option<U256> {

View File

@@ -98,7 +98,8 @@ pub async fn start(mut conn: ClientConnection, mut bi: GrpcBi<ClientRequest, Cli
Err(err) => { Err(err) => {
let _ = bi let _ = bi
.send(Err(Status::unauthenticated(format!( .send(Err(Status::unauthenticated(format!(
"Authentication failed: {err}", "Authentication failed: {}",
err
)))) ))))
.await; .await;
warn!(error = ?err, "Client authentication failed"); warn!(error = ?err, "Client authentication failed");

View File

@@ -5,7 +5,8 @@ use arbiter_proto::{
client::{ client::{
ClientRequest, ClientResponse, ClientRequest, ClientResponse,
auth::{ auth::{
self as proto_auth, AuthChallengeRequest as ProtoAuthChallengeRequest, self as proto_auth, AuthChallenge as ProtoAuthChallenge,
AuthChallengeRequest as ProtoAuthChallengeRequest,
AuthChallengeSolution as ProtoAuthChallengeSolution, AuthResult as ProtoAuthResult, AuthChallengeSolution as ProtoAuthChallengeSolution, AuthResult as ProtoAuthResult,
request::Payload as AuthRequestPayload, response::Payload as AuthResponsePayload, request::Payload as AuthRequestPayload, response::Payload as AuthResponsePayload,
}, },
@@ -21,7 +22,7 @@ use tonic::Status;
use tracing::warn; use tracing::warn;
use crate::{ use crate::{
actors::client::{ClientConnection, auth}, actors::client::{self, ClientConnection, auth},
grpc::request_tracker::RequestTracker, grpc::request_tracker::RequestTracker,
}; };
@@ -31,7 +32,7 @@ pub struct AuthTransportAdapter<'a> {
} }
impl<'a> AuthTransportAdapter<'a> { impl<'a> AuthTransportAdapter<'a> {
pub const fn new( pub fn new(
bi: &'a mut GrpcBi<ClientRequest, ClientResponse>, bi: &'a mut GrpcBi<ClientRequest, ClientResponse>,
request_tracker: &'a mut RequestTracker, request_tracker: &'a mut RequestTracker,
) -> Self { ) -> Self {
@@ -41,6 +42,40 @@ impl<'a> AuthTransportAdapter<'a> {
} }
} }
fn response_to_proto(response: auth::Outbound) -> AuthResponsePayload {
match response {
auth::Outbound::AuthChallenge { pubkey, nonce } => {
AuthResponsePayload::Challenge(ProtoAuthChallenge {
pubkey: pubkey.to_bytes(),
nonce,
})
}
auth::Outbound::AuthSuccess => {
AuthResponsePayload::Result(ProtoAuthResult::Success.into())
}
}
}
fn error_to_proto(error: auth::Error) -> AuthResponsePayload {
AuthResponsePayload::Result(
match error {
auth::Error::InvalidChallengeSolution => ProtoAuthResult::InvalidSignature,
auth::Error::ApproveError(auth::ApproveError::Denied) => {
ProtoAuthResult::ApprovalDenied
}
auth::Error::ApproveError(auth::ApproveError::Upstream(
crate::actors::flow_coordinator::ApprovalError::NoUserAgentsConnected,
)) => ProtoAuthResult::NoUserAgentsOnline,
auth::Error::ApproveError(auth::ApproveError::Internal)
| auth::Error::DatabasePoolUnavailable
| auth::Error::DatabaseOperationFailed
| auth::Error::IntegrityCheckFailed
| auth::Error::Transport => ProtoAuthResult::Internal,
}
.into(),
)
}
async fn send_client_response( async fn send_client_response(
&mut self, &mut self,
payload: AuthResponsePayload, payload: AuthResponsePayload,
@@ -62,14 +97,14 @@ impl<'a> AuthTransportAdapter<'a> {
} }
#[async_trait] #[async_trait]
impl Sender<Result<auth::Outbound, auth::ClientAuthError>> for AuthTransportAdapter<'_> { impl Sender<Result<auth::Outbound, auth::Error>> for AuthTransportAdapter<'_> {
async fn send( async fn send(
&mut self, &mut self,
item: Result<auth::Outbound, auth::ClientAuthError>, item: Result<auth::Outbound, auth::Error>,
) -> Result<(), TransportError> { ) -> Result<(), TransportError> {
let payload = match item { let payload = match item {
Ok(message) => message.into(), Ok(message) => AuthTransportAdapter::response_to_proto(message),
Err(err) => AuthResponsePayload::Result(ProtoAuthResult::from(err).into()), Err(err) => AuthTransportAdapter::error_to_proto(err),
}; };
self.send_client_response(payload).await self.send_client_response(payload).await
@@ -148,7 +183,7 @@ impl Receiver<auth::Inbound> for AuthTransportAdapter<'_> {
} }
} }
impl Bi<auth::Inbound, Result<auth::Outbound, auth::ClientAuthError>> for AuthTransportAdapter<'_> {} impl Bi<auth::Inbound, Result<auth::Outbound, auth::Error>> for AuthTransportAdapter<'_> {}
fn client_metadata_from_proto(metadata: ProtoClientInfo) -> ClientMetadata { fn client_metadata_from_proto(metadata: ProtoClientInfo) -> ClientMetadata {
ClientMetadata { ClientMetadata {
@@ -162,7 +197,7 @@ pub async fn start(
conn: &mut ClientConnection, conn: &mut ClientConnection,
bi: &mut GrpcBi<ClientRequest, ClientResponse>, bi: &mut GrpcBi<ClientRequest, ClientResponse>,
request_tracker: &mut RequestTracker, request_tracker: &mut RequestTracker,
) -> Result<i32, auth::ClientAuthError> { ) -> Result<i32, auth::Error> {
let mut transport = AuthTransportAdapter::new(bi, request_tracker); let mut transport = AuthTransportAdapter::new(bi, request_tracker);
auth::authenticate(conn, &mut transport).await client::auth::authenticate(conn, &mut transport).await
} }

View File

@@ -23,7 +23,7 @@ use crate::{
}, },
}; };
const fn wrap_response(payload: EvmResponsePayload) -> ClientResponsePayload { fn wrap_response(payload: EvmResponsePayload) -> ClientResponsePayload {
ClientResponsePayload::Evm(proto_evm::Response { ClientResponsePayload::Evm(proto_evm::Response {
payload: Some(payload), payload: Some(payload),
}) })

View File

@@ -13,7 +13,7 @@ use tonic::Status;
use tracing::warn; use tracing::warn;
use crate::actors::{ use crate::actors::{
client::session::{ClientSession, ClientSessionError, HandleQueryVaultState}, client::session::{ClientSession, Error, HandleQueryVaultState},
keyholder::KeyHolderState, keyholder::KeyHolderState,
}; };
@@ -28,14 +28,12 @@ pub(super) async fn dispatch(
}; };
match payload { match payload {
VaultRequestPayload::QueryState(()) => { VaultRequestPayload::QueryState(_) => {
let state = match actor.ask(HandleQueryVaultState {}).await { let state = match actor.ask(HandleQueryVaultState {}).await {
Ok(KeyHolderState::Unbootstrapped) => ProtoVaultState::Unbootstrapped, Ok(KeyHolderState::Unbootstrapped) => ProtoVaultState::Unbootstrapped,
Ok(KeyHolderState::Sealed) => ProtoVaultState::Sealed, Ok(KeyHolderState::Sealed) => ProtoVaultState::Sealed,
Ok(KeyHolderState::Unsealed) => ProtoVaultState::Unsealed, Ok(KeyHolderState::Unsealed) => ProtoVaultState::Unsealed,
Err(SendError::HandlerError(ClientSessionError::Internal)) => { Err(SendError::HandlerError(Error::Internal)) => ProtoVaultState::Error,
ProtoVaultState::Error
}
Err(err) => { Err(err) => {
warn!(error = ?err, "Failed to query vault state"); warn!(error = ?err, "Failed to query vault state");
ProtoVaultState::Error ProtoVaultState::Error

View File

@@ -31,16 +31,16 @@ impl Convert for SpecificMeaning {
fn convert(self) -> Self::Output { fn convert(self) -> Self::Output {
let kind = match self { let kind = match self {
Self::EtherTransfer(meaning) => ProtoSpecificMeaningKind::EtherTransfer( SpecificMeaning::EtherTransfer(meaning) => ProtoSpecificMeaningKind::EtherTransfer(
arbiter_proto::proto::shared::evm::EtherTransferMeaning { arbiter_proto::proto::shared::evm::EtherTransferMeaning {
to: meaning.to.to_vec(), to: meaning.to.to_vec(),
value: u256_to_proto_bytes(meaning.value), value: u256_to_proto_bytes(meaning.value),
}, },
), ),
Self::TokenTransfer(meaning) => ProtoSpecificMeaningKind::TokenTransfer( SpecificMeaning::TokenTransfer(meaning) => ProtoSpecificMeaningKind::TokenTransfer(
arbiter_proto::proto::shared::evm::TokenTransferMeaning { arbiter_proto::proto::shared::evm::TokenTransferMeaning {
token: Some(ProtoTokenInfo { token: Some(ProtoTokenInfo {
symbol: meaning.token.symbol.to_owned(), symbol: meaning.token.symbol.to_string(),
address: meaning.token.contract.to_vec(), address: meaning.token.contract.to_vec(),
chain_id: meaning.token.chain, chain_id: meaning.token.chain,
}), }),
@@ -61,21 +61,25 @@ impl Convert for EvalViolation {
fn convert(self) -> Self::Output { fn convert(self) -> Self::Output {
let kind = match self { let kind = match self {
Self::InvalidTarget { target } => { EvalViolation::InvalidTarget { target } => {
ProtoEvalViolationKind::InvalidTarget(target.to_vec()) ProtoEvalViolationKind::InvalidTarget(target.to_vec())
} }
Self::GasLimitExceeded { EvalViolation::GasLimitExceeded {
max_gas_fee_per_gas, max_gas_fee_per_gas,
max_priority_fee_per_gas, max_priority_fee_per_gas,
} => ProtoEvalViolationKind::GasLimitExceeded(GasLimitExceededViolation { } => ProtoEvalViolationKind::GasLimitExceeded(GasLimitExceededViolation {
max_gas_fee_per_gas: max_gas_fee_per_gas.map(u256_to_proto_bytes), max_gas_fee_per_gas: max_gas_fee_per_gas.map(u256_to_proto_bytes),
max_priority_fee_per_gas: max_priority_fee_per_gas.map(u256_to_proto_bytes), max_priority_fee_per_gas: max_priority_fee_per_gas.map(u256_to_proto_bytes),
}), }),
Self::RateLimitExceeded => ProtoEvalViolationKind::RateLimitExceeded(()), EvalViolation::RateLimitExceeded => ProtoEvalViolationKind::RateLimitExceeded(()),
Self::VolumetricLimitExceeded => ProtoEvalViolationKind::VolumetricLimitExceeded(()), EvalViolation::VolumetricLimitExceeded => {
Self::InvalidTime => ProtoEvalViolationKind::InvalidTime(()), ProtoEvalViolationKind::VolumetricLimitExceeded(())
Self::InvalidTransactionType => ProtoEvalViolationKind::InvalidTransactionType(()), }
Self::MismatchingChainId { expected, actual } => { EvalViolation::InvalidTime => ProtoEvalViolationKind::InvalidTime(()),
EvalViolation::InvalidTransactionType => {
ProtoEvalViolationKind::InvalidTransactionType(())
}
EvalViolation::MismatchingChainId { expected, actual } => {
ProtoEvalViolationKind::ChainIdMismatch(proto_eval_violation::ChainIdMismatch { ProtoEvalViolationKind::ChainIdMismatch(proto_eval_violation::ChainIdMismatch {
expected, expected,
actual, actual,
@@ -92,13 +96,13 @@ impl Convert for VetError {
fn convert(self) -> Self::Output { fn convert(self) -> Self::Output {
let kind = match self { let kind = match self {
Self::ContractCreationNotSupported => { VetError::ContractCreationNotSupported => {
ProtoTransactionEvalErrorKind::ContractCreationNotSupported(()) ProtoTransactionEvalErrorKind::ContractCreationNotSupported(())
} }
Self::UnsupportedTransactionType => { VetError::UnsupportedTransactionType => {
ProtoTransactionEvalErrorKind::UnsupportedTransactionType(()) ProtoTransactionEvalErrorKind::UnsupportedTransactionType(())
} }
Self::Evaluated(meaning, policy_error) => match policy_error { VetError::Evaluated(meaning, policy_error) => match policy_error {
PolicyError::NoMatchingGrant => { PolicyError::NoMatchingGrant => {
ProtoTransactionEvalErrorKind::NoMatchingGrant(NoMatchingGrantError { ProtoTransactionEvalErrorKind::NoMatchingGrant(NoMatchingGrantError {
meaning: Some(meaning.convert()), meaning: Some(meaning.convert()),

View File

@@ -20,7 +20,7 @@ impl RequestTracker {
// This is used to set the response id for auth responses, which need to match the request id of the auth challenge request. // This is used to set the response id for auth responses, which need to match the request id of the auth challenge request.
// -1 offset is needed because request() increments the next_request_id after returning the current request id. // -1 offset is needed because request() increments the next_request_id after returning the current request id.
pub const fn current_request_id(&self) -> i32 { pub fn current_request_id(&self) -> i32 {
self.next_request_id - 1 self.next_request_id - 1
} }
} }

View File

@@ -28,7 +28,7 @@ pub struct AuthTransportAdapter<'a> {
} }
impl<'a> AuthTransportAdapter<'a> { impl<'a> AuthTransportAdapter<'a> {
pub const fn new( pub fn new(
bi: &'a mut GrpcBi<UserAgentRequest, UserAgentResponse>, bi: &'a mut GrpcBi<UserAgentRequest, UserAgentResponse>,
request_tracker: &'a mut RequestTracker, request_tracker: &'a mut RequestTracker,
) -> Self { ) -> Self {
@@ -140,7 +140,7 @@ impl Receiver<auth::Inbound> for AuthTransportAdapter<'_> {
AuthRequestPayload::ChallengeRequest(ProtoAuthChallengeRequest { AuthRequestPayload::ChallengeRequest(ProtoAuthChallengeRequest {
pubkey, pubkey,
bootstrap_token, bootstrap_token,
.. key_type: _,
}) => { }) => {
let Ok(pubkey) = authn::PublicKey::try_from(pubkey.as_slice()) else { let Ok(pubkey) = authn::PublicKey::try_from(pubkey.as_slice()) else {
warn!( warn!(

View File

@@ -37,7 +37,7 @@ use crate::{
}, },
}; };
const fn wrap_evm_response(payload: EvmResponsePayload) -> UserAgentResponsePayload { fn wrap_evm_response(payload: EvmResponsePayload) -> UserAgentResponsePayload {
UserAgentResponsePayload::Evm(proto_evm::Response { UserAgentResponsePayload::Evm(proto_evm::Response {
payload: Some(payload), payload: Some(payload),
}) })
@@ -52,8 +52,8 @@ pub(super) async fn dispatch(
}; };
match payload { match payload {
EvmRequestPayload::WalletCreate(()) => handle_wallet_create(actor).await, EvmRequestPayload::WalletCreate(_) => handle_wallet_create(actor).await,
EvmRequestPayload::WalletList(()) => handle_wallet_list(actor).await, EvmRequestPayload::WalletList(_) => handle_wallet_list(actor).await,
EvmRequestPayload::GrantCreate(req) => handle_grant_create(actor, req).await, EvmRequestPayload::GrantCreate(req) => handle_grant_create(actor, req).await,
EvmRequestPayload::GrantDelete(req) => handle_grant_delete(actor, req).await, EvmRequestPayload::GrantDelete(req) => handle_grant_delete(actor, req).await,
EvmRequestPayload::GrantList(_) => handle_grant_list(actor).await, EvmRequestPayload::GrantList(_) => handle_grant_list(actor).await,

View File

@@ -22,11 +22,11 @@ use crate::{
grpc::TryConvert, grpc::TryConvert,
}; };
fn address_from_bytes(bytes: &[u8]) -> Result<Address, Status> { fn address_from_bytes(bytes: Vec<u8>) -> Result<Address, Status> {
if bytes.len() != 20 { if bytes.len() != 20 {
return Err(Status::invalid_argument("Invalid EVM address")); return Err(Status::invalid_argument("Invalid EVM address"));
} }
Ok(Address::from_slice(bytes)) Ok(Address::from_slice(&bytes))
} }
fn u256_from_proto_bytes(bytes: &[u8]) -> Result<U256, Status> { fn u256_from_proto_bytes(bytes: &[u8]) -> Result<U256, Status> {
@@ -41,7 +41,7 @@ impl TryConvert for ProtoTimestamp {
type Error = Status; type Error = Status;
fn try_convert(self) -> Result<DateTime<Utc>, Status> { fn try_convert(self) -> Result<DateTime<Utc>, Status> {
Utc.timestamp_opt(self.seconds, self.nanos.try_into().unwrap_or_default()) Utc.timestamp_opt(self.seconds, self.nanos as u32)
.single() .single()
.ok_or_else(|| Status::invalid_argument("Invalid timestamp")) .ok_or_else(|| Status::invalid_argument("Invalid timestamp"))
} }
@@ -87,6 +87,7 @@ impl TryConvert for ProtoSharedSettings {
.valid_until .valid_until
.map(ProtoTimestamp::try_convert) .map(ProtoTimestamp::try_convert)
.transpose()?, .transpose()?,
revoked_at: None,
max_gas_fee_per_gas: self max_gas_fee_per_gas: self
.max_gas_fee_per_gas .max_gas_fee_per_gas
.as_deref() .as_deref()
@@ -116,8 +117,7 @@ impl TryConvert for ProtoSpecificGrant {
limit, limit,
})) => Ok(SpecificGrant::EtherTransfer(ether_transfer::Settings { })) => Ok(SpecificGrant::EtherTransfer(ether_transfer::Settings {
target: targets target: targets
.iter() .into_iter()
.map(Vec::as_slice)
.map(address_from_bytes) .map(address_from_bytes)
.collect::<Result<_, _>>()?, .collect::<Result<_, _>>()?,
limit: limit limit: limit
@@ -131,10 +131,8 @@ impl TryConvert for ProtoSpecificGrant {
target, target,
volume_limits, volume_limits,
})) => Ok(SpecificGrant::TokenTransfer(token_transfers::Settings { })) => Ok(SpecificGrant::TokenTransfer(token_transfers::Settings {
token_contract: address_from_bytes(&token_contract)?, token_contract: address_from_bytes(token_contract)?,
target: target target: target.map(address_from_bytes).transpose()?,
.map(|target| address_from_bytes(&target))
.transpose()?,
volume_limits: volume_limits volume_limits: volume_limits
.into_iter() .into_iter()
.map(ProtoVolumeRateLimit::try_convert) .map(ProtoVolumeRateLimit::try_convert)

View File

@@ -22,7 +22,7 @@ impl Convert for DateTime<Utc> {
fn convert(self) -> ProtoTimestamp { fn convert(self) -> ProtoTimestamp {
ProtoTimestamp { ProtoTimestamp {
seconds: self.timestamp(), seconds: self.timestamp(),
nanos: self.timestamp_subsec_nanos().try_into().unwrap_or(i32::MAX), nanos: self.timestamp_subsec_nanos() as i32,
} }
} }
} }
@@ -74,13 +74,13 @@ impl Convert for SpecificGrant {
fn convert(self) -> ProtoSpecificGrant { fn convert(self) -> ProtoSpecificGrant {
let grant = match self { let grant = match self {
Self::EtherTransfer(s) => { SpecificGrant::EtherTransfer(s) => {
ProtoSpecificGrantType::EtherTransfer(ProtoEtherTransferSettings { ProtoSpecificGrantType::EtherTransfer(ProtoEtherTransferSettings {
targets: s.target.into_iter().map(|a| a.to_vec()).collect(), targets: s.target.into_iter().map(|a| a.to_vec()).collect(),
limit: Some(s.limit.convert()), limit: Some(s.limit.convert()),
}) })
} }
Self::TokenTransfer(s) => { SpecificGrant::TokenTransfer(s) => {
ProtoSpecificGrantType::TokenTransfer(ProtoTokenTransferSettings { ProtoSpecificGrantType::TokenTransfer(ProtoTokenTransferSettings {
token_contract: s.token_contract.to_vec(), token_contract: s.token_contract.to_vec(),
target: s.target.map(|a| a.to_vec()), target: s.target.map(|a| a.to_vec()),

View File

@@ -32,7 +32,7 @@ use crate::{
grpc::Convert, grpc::Convert,
}; };
const fn wrap_sdk_client_response(payload: SdkClientResponsePayload) -> UserAgentResponsePayload { fn wrap_sdk_client_response(payload: SdkClientResponsePayload) -> UserAgentResponsePayload {
UserAgentResponsePayload::SdkClient(proto_sdk_client::Response { UserAgentResponsePayload::SdkClient(proto_sdk_client::Response {
payload: Some(payload), payload: Some(payload),
}) })
@@ -75,14 +75,14 @@ pub(super) async fn dispatch(
SdkClientRequestPayload::Revoke(_) => Err(Status::unimplemented( SdkClientRequestPayload::Revoke(_) => Err(Status::unimplemented(
"SdkClientRevoke is not yet implemented", "SdkClientRevoke is not yet implemented",
)), )),
SdkClientRequestPayload::List(()) => handle_list(actor).await, SdkClientRequestPayload::List(_) => handle_list(actor).await,
SdkClientRequestPayload::GrantWalletAccess(req) => { SdkClientRequestPayload::GrantWalletAccess(req) => {
handle_grant_wallet_access(actor, req).await handle_grant_wallet_access(actor, req).await
} }
SdkClientRequestPayload::RevokeWalletAccess(req) => { SdkClientRequestPayload::RevokeWalletAccess(req) => {
handle_revoke_wallet_access(actor, req).await handle_revoke_wallet_access(actor, req).await
} }
SdkClientRequestPayload::ListWalletAccess(()) => handle_list_wallet_access(actor).await, SdkClientRequestPayload::ListWalletAccess(_) => handle_list_wallet_access(actor).await,
} }
} }
@@ -91,7 +91,7 @@ async fn handle_connection_response(
resp: ProtoSdkClientConnectionResponse, resp: ProtoSdkClientConnectionResponse,
) -> Result<Option<UserAgentResponsePayload>, Status> { ) -> Result<Option<UserAgentResponsePayload>, Status> {
let pubkey = authn::PublicKey::try_from(resp.pubkey.as_slice()) let pubkey = authn::PublicKey::try_from(resp.pubkey.as_slice())
.map_err(|()| Status::invalid_argument("Invalid ML-DSA public key"))?; .map_err(|_| Status::invalid_argument("Invalid ML-DSA public key"))?;
actor actor
.ask(HandleNewClientApprove { .ask(HandleNewClientApprove {
@@ -116,17 +116,12 @@ async fn handle_list(
.into_iter() .into_iter()
.map(|(client, metadata)| ProtoSdkClientEntry { .map(|(client, metadata)| ProtoSdkClientEntry {
id: client.id, id: client.id,
pubkey: client.public_key.clone(), pubkey: client.public_key.to_vec(),
info: Some(ProtoClientMetadata { info: Some(ProtoClientMetadata {
name: metadata.name, name: metadata.name,
description: metadata.description, description: metadata.description,
version: metadata.version, version: metadata.version,
}), }),
#[expect(
clippy::cast_possible_truncation,
clippy::as_conversions,
reason = "fixme! #84"
)]
created_at: client.created_at.0.timestamp() as i32, created_at: client.created_at.0.timestamp() as i32,
}) })
.collect(), .collect(),
@@ -147,7 +142,7 @@ async fn handle_grant_wallet_access(
actor: &ActorRef<UserAgentSession>, actor: &ActorRef<UserAgentSession>,
req: ProtoSdkClientGrantWalletAccess, req: ProtoSdkClientGrantWalletAccess,
) -> Result<Option<UserAgentResponsePayload>, Status> { ) -> Result<Option<UserAgentResponsePayload>, Status> {
let entries: Vec<NewEvmWalletAccess> = req.accesses.into_iter().map(Convert::convert).collect(); let entries: Vec<NewEvmWalletAccess> = req.accesses.into_iter().map(|a| a.convert()).collect();
match actor.ask(HandleGrantEvmWalletAccess { entries }).await { match actor.ask(HandleGrantEvmWalletAccess { entries }).await {
Ok(()) => { Ok(()) => {
info!("Successfully granted wallet access"); info!("Successfully granted wallet access");
@@ -187,7 +182,7 @@ async fn handle_list_wallet_access(
match actor.ask(HandleListWalletAccess {}).await { match actor.ask(HandleListWalletAccess {}).await {
Ok(accesses) => Ok(Some(wrap_sdk_client_response( Ok(accesses) => Ok(Some(wrap_sdk_client_response(
SdkClientResponsePayload::ListWalletAccess(ListWalletAccessResponse { SdkClientResponsePayload::ListWalletAccess(ListWalletAccessResponse {
accesses: accesses.into_iter().map(Convert::convert).collect(), accesses: accesses.into_iter().map(|a| a.convert()).collect(),
}), }),
))), ))),
Err(err) => { Err(err) => {

View File

@@ -31,13 +31,13 @@ use crate::actors::{
}, },
}; };
const fn wrap_vault_response(payload: VaultResponsePayload) -> UserAgentResponsePayload { fn wrap_vault_response(payload: VaultResponsePayload) -> UserAgentResponsePayload {
UserAgentResponsePayload::Vault(proto_vault::Response { UserAgentResponsePayload::Vault(proto_vault::Response {
payload: Some(payload), payload: Some(payload),
}) })
} }
const fn wrap_unseal_response(payload: UnsealResponsePayload) -> UserAgentResponsePayload { fn wrap_unseal_response(payload: UnsealResponsePayload) -> UserAgentResponsePayload {
wrap_vault_response(VaultResponsePayload::Unseal(proto_unseal::Response { wrap_vault_response(VaultResponsePayload::Unseal(proto_unseal::Response {
payload: Some(payload), payload: Some(payload),
})) }))
@@ -58,7 +58,7 @@ pub(super) async fn dispatch(
}; };
match payload { match payload {
VaultRequestPayload::QueryState(()) => handle_query_vault_state(actor).await, VaultRequestPayload::QueryState(_) => handle_query_vault_state(actor).await,
VaultRequestPayload::Unseal(req) => dispatch_unseal_request(actor, req).await, VaultRequestPayload::Unseal(req) => dispatch_unseal_request(actor, req).await,
VaultRequestPayload::Bootstrap(req) => handle_bootstrap_request(actor, req).await, VaultRequestPayload::Bootstrap(req) => handle_bootstrap_request(actor, req).await,
} }

View File

@@ -1,3 +1,4 @@
#![forbid(unsafe_code)]
use crate::context::ServerContext; use crate::context::ServerContext;
pub mod actors; pub mod actors;
@@ -13,7 +14,7 @@ pub struct Server {
} }
impl Server { impl Server {
pub const fn new(context: ServerContext) -> Self { pub fn new(context: ServerContext) -> Self {
Self { context } Self { context }
} }
} }

View File

@@ -15,7 +15,7 @@ use arbiter_server::{
}; };
use diesel::{ExpressionMethods as _, NullableExpressionMethods as _, QueryDsl as _, insert_into}; use diesel::{ExpressionMethods as _, NullableExpressionMethods as _, QueryDsl as _, insert_into};
use diesel_async::RunQueryDsl; use diesel_async::RunQueryDsl;
use ml_dsa::{KeyGen, MlDsa87, SigningKey, VerifyingKey, signature::Keypair}; use ml_dsa::{KeyGen, MlDsa87, SigningKey, VerifyingKey, signature::Keypair as _};
use super::common::ChannelTransport; use super::common::ChannelTransport;
@@ -27,10 +27,6 @@ fn metadata(name: &str, description: Option<&str>, version: Option<&str>) -> Cli
} }
} }
fn verifying_key(key: &SigningKey<MlDsa87>) -> VerifyingKey<MlDsa87> {
<SigningKey<MlDsa87> as Keypair>::verifying_key(key)
}
async fn insert_registered_client( async fn insert_registered_client(
db: &db::DatabasePool, db: &db::DatabasePool,
actors: &GlobalActors, actors: &GlobalActors,
@@ -52,7 +48,7 @@ async fn insert_registered_client(
.unwrap(); .unwrap();
let client_id: i32 = insert_into(program_client::table) let client_id: i32 = insert_into(program_client::table)
.values(( .values((
program_client::public_key.eq(pubkey.encode().0.to_vec()), program_client::public_key.eq(pubkey.encode().to_vec()),
program_client::metadata_id.eq(metadata_id), program_client::metadata_id.eq(metadata_id),
)) ))
.returning(program_client::id) .returning(program_client::id)
@@ -87,9 +83,9 @@ fn sign_client_challenge(
async fn insert_bootstrap_sentinel_useragent(db: &db::DatabasePool) { async fn insert_bootstrap_sentinel_useragent(db: &db::DatabasePool) {
let mut conn = db.get().await.unwrap(); let mut conn = db.get().await.unwrap();
let sentinel_key = verifying_key(&MlDsa87::key_gen(&mut rand::rng())) let sentinel_key = MlDsa87::key_gen(&mut rand::rng())
.verifying_key()
.encode() .encode()
.0
.to_vec(); .to_vec();
insert_into(schema::useragent_client::table) insert_into(schema::useragent_client::table)
@@ -118,7 +114,7 @@ async fn spawn_test_actors(db: &db::DatabasePool) -> GlobalActors {
#[tokio::test] #[tokio::test]
#[test_log::test] #[test_log::test]
pub async fn unregistered_pubkey_rejected() { pub async fn test_unregistered_pubkey_rejected() {
let db = db::create_test_pool().await; let db = db::create_test_pool().await;
let (server_transport, mut test_transport) = ChannelTransport::new(); let (server_transport, mut test_transport) = ChannelTransport::new();
@@ -133,7 +129,7 @@ pub async fn unregistered_pubkey_rejected() {
test_transport test_transport
.send(auth::Inbound::AuthChallengeRequest { .send(auth::Inbound::AuthChallengeRequest {
pubkey: verifying_key(&new_key).into(), pubkey: new_key.verifying_key().into(),
metadata: metadata("client", Some("desc"), Some("1.0.0")), metadata: metadata("client", Some("desc"), Some("1.0.0")),
}) })
.await .await
@@ -145,18 +141,18 @@ pub async fn unregistered_pubkey_rejected() {
#[tokio::test] #[tokio::test]
#[test_log::test] #[test_log::test]
pub async fn challenge_auth() { pub async fn test_challenge_auth() {
let db = db::create_test_pool().await; let db = db::create_test_pool().await;
let actors = spawn_test_actors(&db).await; let actors = spawn_test_actors(&db).await;
let new_key = MlDsa87::key_gen(&mut rand::rng()); let new_key = MlDsa87::key_gen(&mut rand::rng());
Box::pin(insert_registered_client( insert_registered_client(
&db, &db,
&actors, &actors,
verifying_key(&new_key), new_key.verifying_key(),
&metadata("client", Some("desc"), Some("1.0.0")), &metadata("client", Some("desc"), Some("1.0.0")),
)) )
.await; .await;
let (server_transport, mut test_transport) = ChannelTransport::new(); let (server_transport, mut test_transport) = ChannelTransport::new();
@@ -169,7 +165,7 @@ pub async fn challenge_auth() {
// Send challenge request // Send challenge request
test_transport test_transport
.send(auth::Inbound::AuthChallengeRequest { .send(auth::Inbound::AuthChallengeRequest {
pubkey: verifying_key(&new_key).into(), pubkey: new_key.verifying_key().into(),
metadata: metadata("client", Some("desc"), Some("1.0.0")), metadata: metadata("client", Some("desc"), Some("1.0.0")),
}) })
.await .await
@@ -183,7 +179,7 @@ pub async fn challenge_auth() {
let challenge = match response { let challenge = match response {
Ok(resp) => match resp { Ok(resp) => match resp {
auth::Outbound::AuthChallenge { pubkey, nonce } => (pubkey, nonce), auth::Outbound::AuthChallenge { pubkey, nonce } => (pubkey, nonce),
other @ auth::Outbound::AuthSuccess => panic!("Expected AuthChallenge, got {other:?}"), other => panic!("Expected AuthChallenge, got {other:?}"),
}, },
Err(err) => panic!("Expected Ok response, got Err({err:?})"), Err(err) => panic!("Expected Ok response, got Err({err:?})"),
}; };
@@ -212,19 +208,13 @@ pub async fn challenge_auth() {
#[tokio::test] #[tokio::test]
#[test_log::test] #[test_log::test]
pub async fn metadata_unchanged_does_not_append_history() { pub async fn test_metadata_unchanged_does_not_append_history() {
let db = db::create_test_pool().await; let db = db::create_test_pool().await;
let actors = spawn_test_actors(&db).await; let actors = spawn_test_actors(&db).await;
let new_key = MlDsa87::key_gen(&mut rand::rng()); let new_key = MlDsa87::key_gen(&mut rand::rng());
let requested = metadata("client", Some("desc"), Some("1.0.0")); let requested = metadata("client", Some("desc"), Some("1.0.0"));
Box::pin(insert_registered_client( insert_registered_client(&db, &actors, new_key.verifying_key(), &requested).await;
&db,
&actors,
verifying_key(&new_key),
&requested,
))
.await;
let props = ClientConnection::new(db.clone(), actors); let props = ClientConnection::new(db.clone(), actors);
@@ -236,7 +226,7 @@ pub async fn metadata_unchanged_does_not_append_history() {
test_transport test_transport
.send(auth::Inbound::AuthChallengeRequest { .send(auth::Inbound::AuthChallengeRequest {
pubkey: verifying_key(&new_key).into(), pubkey: new_key.verifying_key().into(),
metadata: requested, metadata: requested,
}) })
.await .await
@@ -245,7 +235,7 @@ pub async fn metadata_unchanged_does_not_append_history() {
let response = test_transport.recv().await.unwrap().unwrap(); let response = test_transport.recv().await.unwrap().unwrap();
let (pubkey, nonce) = match response { let (pubkey, nonce) = match response {
auth::Outbound::AuthChallenge { pubkey, nonce } => (pubkey, nonce), auth::Outbound::AuthChallenge { pubkey, nonce } => (pubkey, nonce),
auth::Outbound::AuthSuccess => panic!("Expected AuthChallenge, got AuthSuccess"), other => panic!("Expected AuthChallenge, got {other:?}"),
}; };
let signature = sign_client_challenge(&new_key, nonce, &pubkey); let signature = sign_client_challenge(&new_key, nonce, &pubkey);
test_transport test_transport
@@ -275,17 +265,17 @@ pub async fn metadata_unchanged_does_not_append_history() {
#[tokio::test] #[tokio::test]
#[test_log::test] #[test_log::test]
pub async fn metadata_change_appends_history_and_repoints_binding() { pub async fn test_metadata_change_appends_history_and_repoints_binding() {
let db = db::create_test_pool().await; let db = db::create_test_pool().await;
let actors = spawn_test_actors(&db).await; let actors = spawn_test_actors(&db).await;
let new_key = MlDsa87::key_gen(&mut rand::rng()); let new_key = MlDsa87::key_gen(&mut rand::rng());
Box::pin(insert_registered_client( insert_registered_client(
&db, &db,
&actors, &actors,
verifying_key(&new_key), new_key.verifying_key(),
&metadata("client", Some("old"), Some("1.0.0")), &metadata("client", Some("old"), Some("1.0.0")),
)) )
.await; .await;
let props = ClientConnection::new(db.clone(), actors); let props = ClientConnection::new(db.clone(), actors);
@@ -298,7 +288,7 @@ pub async fn metadata_change_appends_history_and_repoints_binding() {
test_transport test_transport
.send(auth::Inbound::AuthChallengeRequest { .send(auth::Inbound::AuthChallengeRequest {
pubkey: verifying_key(&new_key).into(), pubkey: new_key.verifying_key().into(),
metadata: metadata("client", Some("new"), Some("2.0.0")), metadata: metadata("client", Some("new"), Some("2.0.0")),
}) })
.await .await
@@ -307,14 +297,14 @@ pub async fn metadata_change_appends_history_and_repoints_binding() {
let response = test_transport.recv().await.unwrap().unwrap(); let response = test_transport.recv().await.unwrap().unwrap();
let (pubkey, nonce) = match response { let (pubkey, nonce) = match response {
auth::Outbound::AuthChallenge { pubkey, nonce } => (pubkey, nonce), auth::Outbound::AuthChallenge { pubkey, nonce } => (pubkey, nonce),
auth::Outbound::AuthSuccess => panic!("Expected AuthChallenge, got AuthSuccess"), other => panic!("Expected AuthChallenge, got {other:?}"),
}; };
let signature = sign_client_challenge(&new_key, nonce, &pubkey); let signature = sign_client_challenge(&new_key, nonce, &pubkey);
test_transport test_transport
.send(auth::Inbound::AuthChallengeSolution { signature }) .send(auth::Inbound::AuthChallengeSolution { signature })
.await .await
.unwrap(); .unwrap();
drop(test_transport.recv().await.unwrap()); let _ = test_transport.recv().await.unwrap();
task.await.unwrap(); task.await.unwrap();
{ {
@@ -362,7 +352,7 @@ pub async fn metadata_change_appends_history_and_repoints_binding() {
#[tokio::test] #[tokio::test]
#[test_log::test] #[test_log::test]
pub async fn challenge_auth_rejects_integrity_tag_mismatch() { pub async fn test_challenge_auth_rejects_integrity_tag_mismatch() {
let db = db::create_test_pool().await; let db = db::create_test_pool().await;
let actors = spawn_test_actors(&db).await; let actors = spawn_test_actors(&db).await;
@@ -384,7 +374,7 @@ pub async fn challenge_auth_rejects_integrity_tag_mismatch() {
.unwrap(); .unwrap();
insert_into(program_client::table) insert_into(program_client::table)
.values(( .values((
program_client::public_key.eq(verifying_key(&new_key).encode().0.to_vec()), program_client::public_key.eq(new_key.verifying_key().encode().to_vec()),
program_client::metadata_id.eq(metadata_id), program_client::metadata_id.eq(metadata_id),
)) ))
.execute(&mut conn) .execute(&mut conn)
@@ -401,7 +391,7 @@ pub async fn challenge_auth_rejects_integrity_tag_mismatch() {
test_transport test_transport
.send(auth::Inbound::AuthChallengeRequest { .send(auth::Inbound::AuthChallengeRequest {
pubkey: verifying_key(&new_key).into(), pubkey: new_key.verifying_key().into(),
metadata: requested, metadata: requested,
}) })
.await .await
@@ -411,10 +401,7 @@ pub async fn challenge_auth_rejects_integrity_tag_mismatch() {
.recv() .recv()
.await .await
.expect("should receive auth rejection"); .expect("should receive auth rejection");
assert!(matches!( assert!(matches!(response, Err(auth::Error::IntegrityCheckFailed)));
response,
Err(auth::ClientAuthError::IntegrityCheckFailed)
));
task.await.unwrap(); task.await.unwrap();
} }

View File

@@ -1,4 +1,3 @@
#![allow(dead_code, reason = "Common test utilities that may not be used in every test")]
use arbiter_crypto::safecell::{SafeCell, SafeCellHandle as _}; use arbiter_crypto::safecell::{SafeCell, SafeCellHandle as _};
use arbiter_proto::transport::{Bi, Error, Receiver, Sender}; use arbiter_proto::transport::{Bi, Error, Receiver, Sender};
use arbiter_server::{ use arbiter_server::{
@@ -11,6 +10,7 @@ use diesel::QueryDsl;
use diesel_async::RunQueryDsl; use diesel_async::RunQueryDsl;
use tokio::sync::mpsc; use tokio::sync::mpsc;
#[allow(dead_code)]
pub async fn bootstrapped_keyholder(db: &db::DatabasePool) -> KeyHolder { pub async fn bootstrapped_keyholder(db: &db::DatabasePool) -> KeyHolder {
let mut actor = KeyHolder::new(db.clone()).await.unwrap(); let mut actor = KeyHolder::new(db.clone()).await.unwrap();
actor actor
@@ -20,6 +20,7 @@ pub async fn bootstrapped_keyholder(db: &db::DatabasePool) -> KeyHolder {
actor actor
} }
#[allow(dead_code)]
pub async fn root_key_history_id(db: &db::DatabasePool) -> i32 { pub async fn root_key_history_id(db: &db::DatabasePool) -> i32 {
let mut conn = db.get().await.unwrap(); let mut conn = db.get().await.unwrap();
let id = schema::arbiter_settings::table let id = schema::arbiter_settings::table
@@ -30,12 +31,14 @@ pub async fn root_key_history_id(db: &db::DatabasePool) -> i32 {
id.expect("root_key_id should be set after bootstrap") id.expect("root_key_id should be set after bootstrap")
} }
#[allow(dead_code)]
pub struct ChannelTransport<T, Y> { pub struct ChannelTransport<T, Y> {
receiver: mpsc::Receiver<T>, receiver: mpsc::Receiver<T>,
sender: mpsc::Sender<Y>, sender: mpsc::Sender<Y>,
} }
impl<T, Y> ChannelTransport<T, Y> { impl<T, Y> ChannelTransport<T, Y> {
#[allow(dead_code)]
pub fn new() -> (Self, ChannelTransport<Y, T>) { pub fn new() -> (Self, ChannelTransport<Y, T>) {
let (tx1, rx1) = mpsc::channel(10); let (tx1, rx1) = mpsc::channel(10);
let (tx2, rx2) = mpsc::channel(10); let (tx2, rx2) = mpsc::channel(10);

View File

@@ -2,7 +2,7 @@ use std::collections::{HashMap, HashSet};
use arbiter_crypto::safecell::{SafeCell, SafeCellHandle as _}; use arbiter_crypto::safecell::{SafeCell, SafeCellHandle as _};
use arbiter_server::{ use arbiter_server::{
actors::keyholder::{CreateNew, KeyHolder, KeyHolderError}, actors::keyholder::{CreateNew, Error, KeyHolder},
db::{self, models, schema}, db::{self, models, schema},
}; };
@@ -122,7 +122,7 @@ async fn insert_failure_does_not_create_partial_row() {
.create_new(SafeCell::new(b"should fail".to_vec())) .create_new(SafeCell::new(b"should fail".to_vec()))
.await .await
.unwrap_err(); .unwrap_err();
assert!(matches!(err, KeyHolderError::DatabaseTransaction(_))); assert!(matches!(err, Error::DatabaseTransaction(_)));
let mut conn = db.get().await.unwrap(); let mut conn = db.get().await.unwrap();
sql_query("DROP TRIGGER fail_aead_insert;") sql_query("DROP TRIGGER fail_aead_insert;")

View File

@@ -1,6 +1,6 @@
use arbiter_crypto::safecell::{SafeCell, SafeCellHandle as _}; use arbiter_crypto::safecell::{SafeCell, SafeCellHandle as _};
use arbiter_server::{ use arbiter_server::{
actors::keyholder::{KeyHolder, KeyHolderError}, actors::keyholder::{Error, KeyHolder},
crypto::encryption::v1::{Nonce, ROOT_KEY_TAG}, crypto::encryption::v1::{Nonce, ROOT_KEY_TAG},
db::{self, models, schema}, db::{self, models, schema},
}; };
@@ -12,7 +12,7 @@ use crate::common;
#[tokio::test] #[tokio::test]
#[test_log::test] #[test_log::test]
async fn bootstrap() { async fn test_bootstrap() {
let db = db::create_test_pool().await; let db = db::create_test_pool().await;
let mut actor = KeyHolder::new(db.clone()).await.unwrap(); let mut actor = KeyHolder::new(db.clone()).await.unwrap();
@@ -35,18 +35,18 @@ async fn bootstrap() {
#[tokio::test] #[tokio::test]
#[test_log::test] #[test_log::test]
async fn bootstrap_rejects_double() { async fn test_bootstrap_rejects_double() {
let db = db::create_test_pool().await; let db = db::create_test_pool().await;
let mut actor = common::bootstrapped_keyholder(&db).await; let mut actor = common::bootstrapped_keyholder(&db).await;
let seal_key2 = SafeCell::new(b"test-seal-key".to_vec()); let seal_key2 = SafeCell::new(b"test-seal-key".to_vec());
let err = actor.bootstrap(seal_key2).await.unwrap_err(); let err = actor.bootstrap(seal_key2).await.unwrap_err();
assert!(matches!(err, KeyHolderError::AlreadyBootstrapped)); assert!(matches!(err, Error::AlreadyBootstrapped));
} }
#[tokio::test] #[tokio::test]
#[test_log::test] #[test_log::test]
async fn create_new_before_bootstrap_fails() { async fn test_create_new_before_bootstrap_fails() {
let db = db::create_test_pool().await; let db = db::create_test_pool().await;
let mut actor = KeyHolder::new(db).await.unwrap(); let mut actor = KeyHolder::new(db).await.unwrap();
@@ -54,34 +54,34 @@ async fn create_new_before_bootstrap_fails() {
.create_new(SafeCell::new(b"data".to_vec())) .create_new(SafeCell::new(b"data".to_vec()))
.await .await
.unwrap_err(); .unwrap_err();
assert!(matches!(err, KeyHolderError::NotBootstrapped)); assert!(matches!(err, Error::NotBootstrapped));
} }
#[tokio::test] #[tokio::test]
#[test_log::test] #[test_log::test]
async fn decrypt_before_bootstrap_fails() { async fn test_decrypt_before_bootstrap_fails() {
let db = db::create_test_pool().await; let db = db::create_test_pool().await;
let mut actor = KeyHolder::new(db).await.unwrap(); let mut actor = KeyHolder::new(db).await.unwrap();
let err = actor.decrypt(1).await.unwrap_err(); let err = actor.decrypt(1).await.unwrap_err();
assert!(matches!(err, KeyHolderError::NotBootstrapped)); assert!(matches!(err, Error::NotBootstrapped));
} }
#[tokio::test] #[tokio::test]
#[test_log::test] #[test_log::test]
async fn new_restores_sealed_state() { async fn test_new_restores_sealed_state() {
let db = db::create_test_pool().await; let db = db::create_test_pool().await;
let actor = common::bootstrapped_keyholder(&db).await; let actor = common::bootstrapped_keyholder(&db).await;
drop(actor); drop(actor);
let mut actor2 = KeyHolder::new(db).await.unwrap(); let mut actor2 = KeyHolder::new(db).await.unwrap();
let err = actor2.decrypt(1).await.unwrap_err(); let err = actor2.decrypt(1).await.unwrap_err();
assert!(matches!(err, KeyHolderError::NotBootstrapped)); assert!(matches!(err, Error::NotBootstrapped));
} }
#[tokio::test] #[tokio::test]
#[test_log::test] #[test_log::test]
async fn unseal_correct_password() { async fn test_unseal_correct_password() {
let db = db::create_test_pool().await; let db = db::create_test_pool().await;
let mut actor = common::bootstrapped_keyholder(&db).await; let mut actor = common::bootstrapped_keyholder(&db).await;
@@ -102,7 +102,7 @@ async fn unseal_correct_password() {
#[tokio::test] #[tokio::test]
#[test_log::test] #[test_log::test]
async fn unseal_wrong_then_correct_password() { async fn test_unseal_wrong_then_correct_password() {
let db = db::create_test_pool().await; let db = db::create_test_pool().await;
let mut actor = common::bootstrapped_keyholder(&db).await; let mut actor = common::bootstrapped_keyholder(&db).await;
@@ -117,7 +117,7 @@ async fn unseal_wrong_then_correct_password() {
let bad_key = SafeCell::new(b"wrong-password".to_vec()); let bad_key = SafeCell::new(b"wrong-password".to_vec());
let err = actor.try_unseal(bad_key).await.unwrap_err(); let err = actor.try_unseal(bad_key).await.unwrap_err();
assert!(matches!(err, KeyHolderError::InvalidKey)); assert!(matches!(err, Error::InvalidKey));
let good_key = SafeCell::new(b"test-seal-key".to_vec()); let good_key = SafeCell::new(b"test-seal-key".to_vec());
actor.try_unseal(good_key).await.unwrap(); actor.try_unseal(good_key).await.unwrap();

View File

@@ -2,7 +2,7 @@ use std::collections::HashSet;
use arbiter_crypto::safecell::{SafeCell, SafeCellHandle as _}; use arbiter_crypto::safecell::{SafeCell, SafeCellHandle as _};
use arbiter_server::{ use arbiter_server::{
actors::keyholder::KeyHolderError, actors::keyholder::Error,
crypto::encryption::v1::Nonce, crypto::encryption::v1::Nonce,
db::{self, models, schema}, db::{self, models, schema},
}; };
@@ -14,7 +14,7 @@ use crate::common;
#[tokio::test] #[tokio::test]
#[test_log::test] #[test_log::test]
async fn create_decrypt_roundtrip() { async fn test_create_decrypt_roundtrip() {
let db = db::create_test_pool().await; let db = db::create_test_pool().await;
let mut actor = common::bootstrapped_keyholder(&db).await; let mut actor = common::bootstrapped_keyholder(&db).await;
@@ -30,17 +30,17 @@ async fn create_decrypt_roundtrip() {
#[tokio::test] #[tokio::test]
#[test_log::test] #[test_log::test]
async fn decrypt_nonexistent_returns_not_found() { async fn test_decrypt_nonexistent_returns_not_found() {
let db = db::create_test_pool().await; let db = db::create_test_pool().await;
let mut actor = common::bootstrapped_keyholder(&db).await; let mut actor = common::bootstrapped_keyholder(&db).await;
let err = actor.decrypt(9999).await.unwrap_err(); let err = actor.decrypt(9999).await.unwrap_err();
assert!(matches!(err, KeyHolderError::NotFound)); assert!(matches!(err, Error::NotFound));
} }
#[tokio::test] #[tokio::test]
#[test_log::test] #[test_log::test]
async fn ciphertext_differs_across_entries() { async fn test_ciphertext_differs_across_entries() {
let db = db::create_test_pool().await; let db = db::create_test_pool().await;
let mut actor = common::bootstrapped_keyholder(&db).await; let mut actor = common::bootstrapped_keyholder(&db).await;
@@ -78,7 +78,7 @@ async fn ciphertext_differs_across_entries() {
#[tokio::test] #[tokio::test]
#[test_log::test] #[test_log::test]
async fn nonce_never_reused() { async fn test_nonce_never_reused() {
let db = db::create_test_pool().await; let db = db::create_test_pool().await;
let mut actor = common::bootstrapped_keyholder(&db).await; let mut actor = common::bootstrapped_keyholder(&db).await;
@@ -142,7 +142,7 @@ async fn broken_db_nonce_format_fails_closed() {
.create_new(SafeCell::new(b"must fail".to_vec())) .create_new(SafeCell::new(b"must fail".to_vec()))
.await .await
.unwrap_err(); .unwrap_err();
assert!(matches!(err, KeyHolderError::BrokenDatabase)); assert!(matches!(err, Error::BrokenDatabase));
let db = db::create_test_pool().await; let db = db::create_test_pool().await;
let mut actor = common::bootstrapped_keyholder(&db).await; let mut actor = common::bootstrapped_keyholder(&db).await;
@@ -159,5 +159,5 @@ async fn broken_db_nonce_format_fails_closed() {
drop(conn); drop(conn);
let err = actor.decrypt(id).await.unwrap_err(); let err = actor.decrypt(id).await.unwrap_err();
assert!(matches!(err, KeyHolderError::BrokenDatabase)); assert!(matches!(err, Error::BrokenDatabase));
} }

View File

@@ -16,14 +16,10 @@ use arbiter_server::{
}; };
use diesel::{ExpressionMethods as _, QueryDsl, insert_into}; use diesel::{ExpressionMethods as _, QueryDsl, insert_into};
use diesel_async::RunQueryDsl; use diesel_async::RunQueryDsl;
use ml_dsa::{KeyGen, MlDsa87, SigningKey, VerifyingKey, signature::Keypair}; use ml_dsa::{KeyGen, MlDsa87, SigningKey, signature::Keypair as _};
use super::common::ChannelTransport; use super::common::ChannelTransport;
fn verifying_key(key: &SigningKey<MlDsa87>) -> VerifyingKey<MlDsa87> {
<SigningKey<MlDsa87> as Keypair>::verifying_key(key)
}
fn sign_useragent_challenge( fn sign_useragent_challenge(
key: &SigningKey<MlDsa87>, key: &SigningKey<MlDsa87>,
nonce: i32, nonce: i32,
@@ -38,7 +34,7 @@ fn sign_useragent_challenge(
#[tokio::test] #[tokio::test]
#[test_log::test] #[test_log::test]
pub async fn bootstrap_token_auth() { pub async fn test_bootstrap_token_auth() {
let db = db::create_test_pool().await; let db = db::create_test_pool().await;
let actors = GlobalActors::spawn(db.clone()).await.unwrap(); let actors = GlobalActors::spawn(db.clone()).await.unwrap();
actors actors
@@ -60,7 +56,7 @@ pub async fn bootstrap_token_auth() {
let new_key = MlDsa87::key_gen(&mut rand::rng()); let new_key = MlDsa87::key_gen(&mut rand::rng());
test_transport test_transport
.send(auth::Inbound::AuthChallengeRequest { .send(auth::Inbound::AuthChallengeRequest {
pubkey: verifying_key(&new_key).into(), pubkey: new_key.verifying_key().into(),
bootstrap_token: Some(token), bootstrap_token: Some(token),
}) })
.await .await
@@ -83,12 +79,12 @@ pub async fn bootstrap_token_auth() {
.first::<Vec<u8>>(&mut conn) .first::<Vec<u8>>(&mut conn)
.await .await
.unwrap(); .unwrap();
assert_eq!(stored_pubkey, verifying_key(&new_key).encode().0.to_vec()); assert_eq!(stored_pubkey, new_key.verifying_key().encode().to_vec());
} }
#[tokio::test] #[tokio::test]
#[test_log::test] #[test_log::test]
pub async fn bootstrap_invalid_token_auth() { pub async fn test_bootstrap_invalid_token_auth() {
let db = db::create_test_pool().await; let db = db::create_test_pool().await;
let actors = GlobalActors::spawn(db.clone()).await.unwrap(); let actors = GlobalActors::spawn(db.clone()).await.unwrap();
@@ -102,8 +98,8 @@ pub async fn bootstrap_invalid_token_auth() {
let new_key = MlDsa87::key_gen(&mut rand::rng()); let new_key = MlDsa87::key_gen(&mut rand::rng());
test_transport test_transport
.send(auth::Inbound::AuthChallengeRequest { .send(auth::Inbound::AuthChallengeRequest {
pubkey: verifying_key(&new_key).into(), pubkey: new_key.verifying_key().into(),
bootstrap_token: Some("invalid_token".to_owned()), bootstrap_token: Some("invalid_token".to_string()),
}) })
.await .await
.unwrap(); .unwrap();
@@ -124,7 +120,7 @@ pub async fn bootstrap_invalid_token_auth() {
#[tokio::test] #[tokio::test]
#[test_log::test] #[test_log::test]
pub async fn challenge_auth() { pub async fn test_challenge_auth() {
let db = db::create_test_pool().await; let db = db::create_test_pool().await;
let actors = GlobalActors::spawn(db.clone()).await.unwrap(); let actors = GlobalActors::spawn(db.clone()).await.unwrap();
actors actors
@@ -136,7 +132,7 @@ pub async fn challenge_auth() {
.unwrap(); .unwrap();
let new_key = MlDsa87::key_gen(&mut rand::rng()); let new_key = MlDsa87::key_gen(&mut rand::rng());
let pubkey_bytes = authn::PublicKey::from(verifying_key(&new_key)).to_bytes(); let pubkey_bytes = new_key.verifying_key().encode().to_vec();
{ {
let mut conn = db.get().await.unwrap(); let mut conn = db.get().await.unwrap();
@@ -153,7 +149,7 @@ pub async fn challenge_auth() {
&mut conn, &mut conn,
&actors.key_holder, &actors.key_holder,
&UserAgentCredentials { &UserAgentCredentials {
pubkey: verifying_key(&new_key).into(), pubkey: new_key.verifying_key().into(),
nonce: 1, nonce: 1,
}, },
id, id,
@@ -171,7 +167,7 @@ pub async fn challenge_auth() {
test_transport test_transport
.send(auth::Inbound::AuthChallengeRequest { .send(auth::Inbound::AuthChallengeRequest {
pubkey: verifying_key(&new_key).into(), pubkey: new_key.verifying_key().into(),
bootstrap_token: None, bootstrap_token: None,
}) })
.await .await
@@ -184,7 +180,7 @@ pub async fn challenge_auth() {
let challenge = match response { let challenge = match response {
Ok(resp) => match resp { Ok(resp) => match resp {
auth::Outbound::AuthChallenge { nonce } => nonce, auth::Outbound::AuthChallenge { nonce } => nonce,
auth::Outbound::AuthSuccess => panic!("Expected AuthChallenge, got AuthSuccess"), other => panic!("Expected AuthChallenge, got {other:?}"),
}, },
Err(err) => panic!("Expected Ok response, got Err({err:?})"), Err(err) => panic!("Expected Ok response, got Err({err:?})"),
}; };
@@ -212,7 +208,7 @@ pub async fn challenge_auth() {
#[tokio::test] #[tokio::test]
#[test_log::test] #[test_log::test]
pub async fn challenge_auth_rejects_integrity_tag_mismatch_when_unsealed() { pub async fn test_challenge_auth_rejects_integrity_tag_mismatch_when_unsealed() {
let db = db::create_test_pool().await; let db = db::create_test_pool().await;
let actors = GlobalActors::spawn(db.clone()).await.unwrap(); let actors = GlobalActors::spawn(db.clone()).await.unwrap();
@@ -225,7 +221,7 @@ pub async fn challenge_auth_rejects_integrity_tag_mismatch_when_unsealed() {
.unwrap(); .unwrap();
let new_key = MlDsa87::key_gen(&mut rand::rng()); let new_key = MlDsa87::key_gen(&mut rand::rng());
let pubkey_bytes = authn::PublicKey::from(verifying_key(&new_key)).to_bytes(); let pubkey_bytes = new_key.verifying_key().encode().to_vec();
{ {
let mut conn = db.get().await.unwrap(); let mut conn = db.get().await.unwrap();
@@ -248,7 +244,7 @@ pub async fn challenge_auth_rejects_integrity_tag_mismatch_when_unsealed() {
test_transport test_transport
.send(auth::Inbound::AuthChallengeRequest { .send(auth::Inbound::AuthChallengeRequest {
pubkey: verifying_key(&new_key).into(), pubkey: new_key.verifying_key().into(),
bootstrap_token: None, bootstrap_token: None,
}) })
.await .await
@@ -262,7 +258,7 @@ pub async fn challenge_auth_rejects_integrity_tag_mismatch_when_unsealed() {
#[tokio::test] #[tokio::test]
#[test_log::test] #[test_log::test]
pub async fn challenge_auth_rejects_invalid_signature() { pub async fn test_challenge_auth_rejects_invalid_signature() {
let db = db::create_test_pool().await; let db = db::create_test_pool().await;
let actors = GlobalActors::spawn(db.clone()).await.unwrap(); let actors = GlobalActors::spawn(db.clone()).await.unwrap();
actors actors
@@ -274,7 +270,7 @@ pub async fn challenge_auth_rejects_invalid_signature() {
.unwrap(); .unwrap();
let new_key = MlDsa87::key_gen(&mut rand::rng()); let new_key = MlDsa87::key_gen(&mut rand::rng());
let pubkey_bytes = authn::PublicKey::from(verifying_key(&new_key)).to_bytes(); let pubkey_bytes = new_key.verifying_key().encode().to_vec();
{ {
let mut conn = db.get().await.unwrap(); let mut conn = db.get().await.unwrap();
@@ -291,7 +287,7 @@ pub async fn challenge_auth_rejects_invalid_signature() {
&mut conn, &mut conn,
&actors.key_holder, &actors.key_holder,
&UserAgentCredentials { &UserAgentCredentials {
pubkey: verifying_key(&new_key).into(), pubkey: new_key.verifying_key().into(),
nonce: 1, nonce: 1,
}, },
id, id,
@@ -309,7 +305,7 @@ pub async fn challenge_auth_rejects_invalid_signature() {
test_transport test_transport
.send(auth::Inbound::AuthChallengeRequest { .send(auth::Inbound::AuthChallengeRequest {
pubkey: verifying_key(&new_key).into(), pubkey: new_key.verifying_key().into(),
bootstrap_token: None, bootstrap_token: None,
}) })
.await .await
@@ -322,7 +318,7 @@ pub async fn challenge_auth_rejects_invalid_signature() {
let challenge = match response { let challenge = match response {
Ok(resp) => match resp { Ok(resp) => match resp {
auth::Outbound::AuthChallenge { nonce } => nonce, auth::Outbound::AuthChallenge { nonce } => nonce,
auth::Outbound::AuthSuccess => panic!("Expected AuthChallenge, got AuthSuccess"), other => panic!("Expected AuthChallenge, got {other:?}"),
}, },
Err(err) => panic!("Expected Ok response, got Err({err:?})"), Err(err) => panic!("Expected Ok response, got Err({err:?})"),
}; };

View File

@@ -69,7 +69,7 @@ async fn client_dh_encrypt(
#[tokio::test] #[tokio::test]
#[test_log::test] #[test_log::test]
pub async fn unseal_success() { pub async fn test_unseal_success() {
let seal_key = b"test-seal-key"; let seal_key = b"test-seal-key";
let (_db, user_agent) = setup_sealed_user_agent(seal_key).await; let (_db, user_agent) = setup_sealed_user_agent(seal_key).await;
@@ -81,7 +81,7 @@ pub async fn unseal_success() {
#[tokio::test] #[tokio::test]
#[test_log::test] #[test_log::test]
pub async fn unseal_wrong_seal_key() { pub async fn test_unseal_wrong_seal_key() {
let (_db, user_agent) = setup_sealed_user_agent(b"correct-key").await; let (_db, user_agent) = setup_sealed_user_agent(b"correct-key").await;
let encrypted_key = client_dh_encrypt(&user_agent, b"wrong-key").await; let encrypted_key = client_dh_encrypt(&user_agent, b"wrong-key").await;
@@ -97,7 +97,7 @@ pub async fn unseal_wrong_seal_key() {
#[tokio::test] #[tokio::test]
#[test_log::test] #[test_log::test]
pub async fn unseal_corrupted_ciphertext() { pub async fn test_unseal_corrupted_ciphertext() {
let (_db, user_agent) = setup_sealed_user_agent(b"test-key").await; let (_db, user_agent) = setup_sealed_user_agent(b"test-key").await;
let client_secret = EphemeralSecret::random(); let client_secret = EphemeralSecret::random();
@@ -128,7 +128,7 @@ pub async fn unseal_corrupted_ciphertext() {
#[tokio::test] #[tokio::test]
#[test_log::test] #[test_log::test]
pub async fn unseal_retry_after_invalid_key() { pub async fn test_unseal_retry_after_invalid_key() {
let seal_key = b"real-seal-key"; let seal_key = b"real-seal-key";
let (_db, user_agent) = setup_sealed_user_agent(seal_key).await; let (_db, user_agent) = setup_sealed_user_agent(seal_key).await;