Compare commits
12 Commits
7a999b7b6e
...
developmen
Author | SHA1 | Date | |
---|---|---|---|
|
53e9a2d4f0 | ||
cb1fb0f0ec | |||
97c24d146b | |||
6bff52e8b7 | |||
|
a7c978efd4 | ||
|
0b0d546b4e | ||
|
2f5e18df98 | ||
|
77169c073c | ||
|
ce12f26a91 | ||
|
130822b69b | ||
|
7439980b33 | ||
|
453e86edd2 |
114
Cargo.lock
generated
114
Cargo.lock
generated
@@ -65,13 +65,13 @@ checksum = "7c02d123df017efcdfbd739ef81735b36c5ba83ec3c59c80a9d7ecc718f92e50"
|
||||
|
||||
[[package]]
|
||||
name = "async-trait"
|
||||
version = "0.1.88"
|
||||
version = "0.1.89"
|
||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||
checksum = "e539d3fca749fcee5236ab05e93a52867dd549cc157c8cb7f99595f3cedffdb5"
|
||||
checksum = "9035ad2d096bed7955a320ee7e2230574d28fd3c3a0f186cbea1ff3c7eed5dbb"
|
||||
dependencies = [
|
||||
"proc-macro2",
|
||||
"quote",
|
||||
"syn 2.0.104",
|
||||
"syn 2.0.106",
|
||||
]
|
||||
|
||||
[[package]]
|
||||
@@ -123,9 +123,9 @@ dependencies = [
|
||||
|
||||
[[package]]
|
||||
name = "bitflags"
|
||||
version = "2.9.1"
|
||||
version = "2.9.2"
|
||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||
checksum = "1b8e56985ec62d17e9c1001dc89c88ecd7dc08e47eba5ec7c29c7b5eeecde967"
|
||||
checksum = "6a65b545ab31d687cff52899d4890855fec459eb6afe0da6417b8a18da87aa29"
|
||||
|
||||
[[package]]
|
||||
name = "bitvec"
|
||||
@@ -168,7 +168,7 @@ dependencies = [
|
||||
"proc-macro-crate",
|
||||
"proc-macro2",
|
||||
"quote",
|
||||
"syn 2.0.104",
|
||||
"syn 2.0.106",
|
||||
]
|
||||
|
||||
[[package]]
|
||||
@@ -213,18 +213,18 @@ checksum = "d71b6127be86fdcfddb610f7182ac57211d4b18a3e9c82eb2d17662f2227ad6a"
|
||||
|
||||
[[package]]
|
||||
name = "cc"
|
||||
version = "1.2.31"
|
||||
version = "1.2.33"
|
||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||
checksum = "c3a42d84bb6b69d3a8b3eaacf0d88f179e1929695e1ad012b6cf64d9caaa5fd2"
|
||||
checksum = "3ee0f8803222ba5a7e2777dd72ca451868909b1ac410621b676adf07280e9b5f"
|
||||
dependencies = [
|
||||
"shlex",
|
||||
]
|
||||
|
||||
[[package]]
|
||||
name = "cfg-if"
|
||||
version = "1.0.1"
|
||||
version = "1.0.3"
|
||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||
checksum = "9555578bc9e57714c812a1f84e4fc5b4d21fcb063490c624de019f7464c91268"
|
||||
checksum = "2fd1289c04a9ea8cb22300a459a72a385d7c73d3259e2ed7dcb2af674838cfa9"
|
||||
|
||||
[[package]]
|
||||
name = "cfg_aliases"
|
||||
@@ -372,7 +372,7 @@ checksum = "162ee34ebcb7c64a8abebc059ce0fee27c2262618d7b60ed8faf72fef13c3650"
|
||||
dependencies = [
|
||||
"proc-macro2",
|
||||
"quote",
|
||||
"syn 2.0.104",
|
||||
"syn 2.0.106",
|
||||
]
|
||||
|
||||
[[package]]
|
||||
@@ -452,9 +452,9 @@ dependencies = [
|
||||
|
||||
[[package]]
|
||||
name = "hashbrown"
|
||||
version = "0.15.4"
|
||||
version = "0.15.5"
|
||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||
checksum = "5971ac85611da7067dbfcabef3c70ebb5606018acd9e2a3903a0da507521e0d5"
|
||||
checksum = "9229cfe53dfd69f0609a49f65461bd93001ea1ef889cd5529dd176593f5338a1"
|
||||
|
||||
[[package]]
|
||||
name = "heck"
|
||||
@@ -491,10 +491,12 @@ dependencies = [
|
||||
name = "heromodels-derive"
|
||||
version = "0.1.0"
|
||||
dependencies = [
|
||||
"heromodels_core",
|
||||
"proc-macro2",
|
||||
"quote",
|
||||
"serde",
|
||||
"syn 2.0.104",
|
||||
"serde_json",
|
||||
"syn 2.0.106",
|
||||
]
|
||||
|
||||
[[package]]
|
||||
@@ -545,7 +547,7 @@ source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||
checksum = "fe4cd85333e22411419a0bcae1297d25e58c9443848b11dc6a86fefe8c78a661"
|
||||
dependencies = [
|
||||
"equivalent",
|
||||
"hashbrown 0.15.4",
|
||||
"hashbrown 0.15.5",
|
||||
]
|
||||
|
||||
[[package]]
|
||||
@@ -597,7 +599,7 @@ checksum = "03343451ff899767262ec32146f6d559dd759fdadf42ff0e227c7c48f72594b4"
|
||||
dependencies = [
|
||||
"proc-macro2",
|
||||
"quote",
|
||||
"syn 2.0.104",
|
||||
"syn 2.0.106",
|
||||
]
|
||||
|
||||
[[package]]
|
||||
@@ -627,9 +629,9 @@ dependencies = [
|
||||
|
||||
[[package]]
|
||||
name = "jsonb"
|
||||
version = "0.5.3"
|
||||
version = "0.5.4"
|
||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||
checksum = "96cbb4fba292867a2d86ed83dbe5f9d036f423bf6a491b7d884058b2fde42fcd"
|
||||
checksum = "a452366d21e8d3cbca680c41388e01d6a88739afef7877961946a6da409f9ccd"
|
||||
dependencies = [
|
||||
"byteorder",
|
||||
"ethnum",
|
||||
@@ -647,9 +649,20 @@ dependencies = [
|
||||
|
||||
[[package]]
|
||||
name = "libc"
|
||||
version = "0.2.174"
|
||||
version = "0.2.175"
|
||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||
checksum = "1171693293099992e19cddea4e8b849964e9846f4acee11b3948bcc337be8776"
|
||||
checksum = "6a82ae493e598baaea5209805c49bbf2ea7de956d50d7da0da1164f9c6d28543"
|
||||
|
||||
[[package]]
|
||||
name = "libredox"
|
||||
version = "0.1.9"
|
||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||
checksum = "391290121bad3d37fbddad76d8f5d1c1c314cfc646d143d7e07a3086ddff0ce3"
|
||||
dependencies = [
|
||||
"bitflags",
|
||||
"libc",
|
||||
"redox_syscall",
|
||||
]
|
||||
|
||||
[[package]]
|
||||
name = "lock_api"
|
||||
@@ -760,6 +773,7 @@ dependencies = [
|
||||
[[package]]
|
||||
name = "ourdb"
|
||||
version = "0.1.0"
|
||||
source = "git+https://git.ourworld.tf/herocode/herolib_rust#aa0248ef17cb0117bb69f1d9f278f995bb417f16"
|
||||
dependencies = [
|
||||
"crc32fast",
|
||||
"log",
|
||||
@@ -792,9 +806,9 @@ dependencies = [
|
||||
|
||||
[[package]]
|
||||
name = "percent-encoding"
|
||||
version = "2.3.1"
|
||||
version = "2.3.2"
|
||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||
checksum = "e3148f5046208a5d56bcfc03053e3ca6334e51da8dfb19b6cdc8b306fae3283e"
|
||||
checksum = "9b4f627cb1b25917193a259e49bdad08f671f8d9708acfd5fe0a8c1455d87220"
|
||||
|
||||
[[package]]
|
||||
name = "phf"
|
||||
@@ -906,9 +920,9 @@ dependencies = [
|
||||
|
||||
[[package]]
|
||||
name = "proc-macro2"
|
||||
version = "1.0.95"
|
||||
version = "1.0.101"
|
||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||
checksum = "02b3e5e68a3a1a02aad3ec490a98007cbc13c37cbe84a3cd7b8e406d76e7f778"
|
||||
checksum = "89ae43fd86e4158d6db51ad8e2b80f313af9cc74f5c0e03ccb87de09998732de"
|
||||
dependencies = [
|
||||
"unicode-ident",
|
||||
]
|
||||
@@ -1079,12 +1093,13 @@ checksum = "a5a11a05ee1ce44058fa3d5961d05194fdbe3ad6b40f904af764d81b86450e6b"
|
||||
dependencies = [
|
||||
"proc-macro2",
|
||||
"quote",
|
||||
"syn 2.0.104",
|
||||
"syn 2.0.106",
|
||||
]
|
||||
|
||||
[[package]]
|
||||
name = "rhailib-macros"
|
||||
version = "0.1.0"
|
||||
source = "git+https://git.ourworld.tf/herocode/herolib_rust#aa0248ef17cb0117bb69f1d9f278f995bb417f16"
|
||||
dependencies = [
|
||||
"rhai",
|
||||
"serde",
|
||||
@@ -1143,9 +1158,9 @@ checksum = "56f7d92ca342cea22a06f2121d944b4fd82af56988c270852495420f961d4ace"
|
||||
|
||||
[[package]]
|
||||
name = "rustversion"
|
||||
version = "1.0.21"
|
||||
version = "1.0.22"
|
||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||
checksum = "8a0d197bd2c9dc6e53b84da9556a69ba4cdfab8619eb41a8bd1cc2027a0f6b1d"
|
||||
checksum = "b39cdef0fa800fc44525c84ccb54a029961a8215f9619753635a9c0d2538d46d"
|
||||
|
||||
[[package]]
|
||||
name = "ryu"
|
||||
@@ -1191,14 +1206,14 @@ checksum = "5b0276cf7f2c73365f7157c8123c21cd9a50fbbd844757af28ca1f5925fc2a00"
|
||||
dependencies = [
|
||||
"proc-macro2",
|
||||
"quote",
|
||||
"syn 2.0.104",
|
||||
"syn 2.0.106",
|
||||
]
|
||||
|
||||
[[package]]
|
||||
name = "serde_json"
|
||||
version = "1.0.142"
|
||||
version = "1.0.143"
|
||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||
checksum = "030fedb782600dcbd6f02d479bf0d817ac3bb40d644745b769d6a96bc3afc5a7"
|
||||
checksum = "d401abef1d108fbd9cbaebc3e46611f4b1021f714a0597a71f41ee463f5f4a5a"
|
||||
dependencies = [
|
||||
"indexmap",
|
||||
"itoa",
|
||||
@@ -1247,9 +1262,9 @@ checksum = "56199f7ddabf13fe5074ce809e7d3f42b42ae711800501b5b16ea82ad029c39d"
|
||||
|
||||
[[package]]
|
||||
name = "slab"
|
||||
version = "0.4.10"
|
||||
version = "0.4.11"
|
||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||
checksum = "04dc19736151f35336d325007ac991178d504a119863a2fcb3758cdb5e52c50d"
|
||||
checksum = "7a2ae44ef20feb57a68b23d846850f861394c2e02dc425a50098ae8c90267589"
|
||||
|
||||
[[package]]
|
||||
name = "smallvec"
|
||||
@@ -1327,7 +1342,7 @@ dependencies = [
|
||||
"proc-macro2",
|
||||
"quote",
|
||||
"rustversion",
|
||||
"syn 2.0.104",
|
||||
"syn 2.0.106",
|
||||
]
|
||||
|
||||
[[package]]
|
||||
@@ -1349,9 +1364,9 @@ dependencies = [
|
||||
|
||||
[[package]]
|
||||
name = "syn"
|
||||
version = "2.0.104"
|
||||
version = "2.0.106"
|
||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||
checksum = "17b6f705963418cdb9927482fa304bc562ece2fdd4f616084c50b7023b435a40"
|
||||
checksum = "ede7c438028d4436d71104916910f5bb611972c5cfd7f89b8300a8186e6fada6"
|
||||
dependencies = [
|
||||
"proc-macro2",
|
||||
"quote",
|
||||
@@ -1387,7 +1402,7 @@ checksum = "4fee6c4efc90059e10f81e6d42c60a18f76588c3d74cb83a0b242a2b6c7504c1"
|
||||
dependencies = [
|
||||
"proc-macro2",
|
||||
"quote",
|
||||
"syn 2.0.104",
|
||||
"syn 2.0.106",
|
||||
]
|
||||
|
||||
[[package]]
|
||||
@@ -1401,9 +1416,9 @@ dependencies = [
|
||||
|
||||
[[package]]
|
||||
name = "tinyvec"
|
||||
version = "1.9.0"
|
||||
version = "1.10.0"
|
||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||
checksum = "09b3661f17e86524eccd4371ab0429194e0d7c008abb45f7a7495b1719463c71"
|
||||
checksum = "bfa5fdc3bce6191a1dbc8c02d5c8bffcf557bafa17c124c5264a458f1b0613fa"
|
||||
dependencies = [
|
||||
"tinyvec_macros",
|
||||
]
|
||||
@@ -1442,7 +1457,7 @@ checksum = "6e06d43f1345a3bcd39f6a56dbb7dcab2ba47e68e8ac134855e7e2bdbaf8cab8"
|
||||
dependencies = [
|
||||
"proc-macro2",
|
||||
"quote",
|
||||
"syn 2.0.104",
|
||||
"syn 2.0.106",
|
||||
]
|
||||
|
||||
[[package]]
|
||||
@@ -1504,6 +1519,7 @@ dependencies = [
|
||||
[[package]]
|
||||
name = "tst"
|
||||
version = "0.1.0"
|
||||
source = "git+https://git.ourworld.tf/herocode/herolib_rust#aa0248ef17cb0117bb69f1d9f278f995bb417f16"
|
||||
dependencies = [
|
||||
"ourdb",
|
||||
"thiserror",
|
||||
@@ -1550,9 +1566,9 @@ checksum = "6d49784317cd0d1ee7ec5c716dd598ec5b4483ea832a2dced265471cc0f690ae"
|
||||
|
||||
[[package]]
|
||||
name = "uuid"
|
||||
version = "1.17.0"
|
||||
version = "1.18.0"
|
||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||
checksum = "3cf4199d1e5d15ddd86a694e4d0dffa9c323ce759fea589f00fef9d81cc1931d"
|
||||
checksum = "f33196643e165781c20a5ead5582283a7dacbb87855d867fbc2df3f81eddc1be"
|
||||
dependencies = [
|
||||
"getrandom 0.3.3",
|
||||
"js-sys",
|
||||
@@ -1614,7 +1630,7 @@ dependencies = [
|
||||
"log",
|
||||
"proc-macro2",
|
||||
"quote",
|
||||
"syn 2.0.104",
|
||||
"syn 2.0.106",
|
||||
"wasm-bindgen-shared",
|
||||
]
|
||||
|
||||
@@ -1636,7 +1652,7 @@ checksum = "8ae87ea40c9f689fc23f209965b6fb8a99ad69aeeb0231408be24920604395de"
|
||||
dependencies = [
|
||||
"proc-macro2",
|
||||
"quote",
|
||||
"syn 2.0.104",
|
||||
"syn 2.0.106",
|
||||
"wasm-bindgen-backend",
|
||||
"wasm-bindgen-shared",
|
||||
]
|
||||
@@ -1662,11 +1678,11 @@ dependencies = [
|
||||
|
||||
[[package]]
|
||||
name = "whoami"
|
||||
version = "1.6.0"
|
||||
version = "1.6.1"
|
||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||
checksum = "6994d13118ab492c3c80c1f81928718159254c53c472bf9ce36f8dae4add02a7"
|
||||
checksum = "5d4a4db5077702ca3015d3d02d74974948aba2ad9e12ab7df718ee64ccd7e97d"
|
||||
dependencies = [
|
||||
"redox_syscall",
|
||||
"libredox",
|
||||
"wasite",
|
||||
"web-sys",
|
||||
]
|
||||
@@ -1692,7 +1708,7 @@ checksum = "a47fddd13af08290e67f4acabf4b459f647552718f683a7b415d290ac744a836"
|
||||
dependencies = [
|
||||
"proc-macro2",
|
||||
"quote",
|
||||
"syn 2.0.104",
|
||||
"syn 2.0.106",
|
||||
]
|
||||
|
||||
[[package]]
|
||||
@@ -1703,7 +1719,7 @@ checksum = "bd9211b69f8dcdfa817bfd14bf1c97c9188afa36f4750130fcdf3f400eca9fa8"
|
||||
dependencies = [
|
||||
"proc-macro2",
|
||||
"quote",
|
||||
"syn 2.0.104",
|
||||
"syn 2.0.106",
|
||||
]
|
||||
|
||||
[[package]]
|
||||
@@ -1856,5 +1872,5 @@ checksum = "9ecf5b4cc5364572d7f4c329661bcc82724222973f2cab6f050a4e5c22f75181"
|
||||
dependencies = [
|
||||
"proc-macro2",
|
||||
"quote",
|
||||
"syn 2.0.104",
|
||||
"syn 2.0.106",
|
||||
]
|
||||
|
@@ -14,4 +14,6 @@ quote = "1.0"
|
||||
proc-macro2 = "1.0"
|
||||
|
||||
[dev-dependencies]
|
||||
serde = { version = "1.0", features = ["derive"] }
|
||||
serde = { version = "1.0", features = ["derive"] }
|
||||
serde_json = "1.0"
|
||||
heromodels_core = { path = "../heromodels_core" }
|
@@ -1,6 +1,6 @@
|
||||
use proc_macro::TokenStream;
|
||||
use quote::{format_ident, quote};
|
||||
use syn::{Data, DeriveInput, Fields, parse_macro_input};
|
||||
use syn::{parse_macro_input, Data, DeriveInput, Fields, Lit, Meta, MetaList, MetaNameValue};
|
||||
|
||||
/// Convert a string to snake_case
|
||||
fn to_snake_case(s: &str) -> String {
|
||||
@@ -47,86 +47,165 @@ pub fn model(_attr: TokenStream, item: TokenStream) -> TokenStream {
|
||||
let db_prefix = to_snake_case(&name_str);
|
||||
|
||||
// Extract fields with #[index] attribute
|
||||
let mut indexed_fields = Vec::new();
|
||||
let mut custom_index_names = std::collections::HashMap::new();
|
||||
// Supports both top-level (no args) and nested path-based indexes declared on a field
|
||||
#[derive(Clone)]
|
||||
enum IndexDecl {
|
||||
TopLevel {
|
||||
field_ident: syn::Ident,
|
||||
field_ty: syn::Type,
|
||||
},
|
||||
NestedPath {
|
||||
on_field_ident: syn::Ident,
|
||||
path: String, // dotted path relative to the field
|
||||
},
|
||||
}
|
||||
|
||||
let mut index_decls: Vec<IndexDecl> = Vec::new();
|
||||
|
||||
if let Data::Struct(ref mut data_struct) = input.data {
|
||||
if let Fields::Named(ref mut fields_named) = data_struct.fields {
|
||||
for field in &mut fields_named.named {
|
||||
let mut attr_idx = None;
|
||||
let mut to_remove: Vec<usize> = Vec::new();
|
||||
for (i, attr) in field.attrs.iter().enumerate() {
|
||||
if attr.path().is_ident("index") {
|
||||
attr_idx = Some(i);
|
||||
if let Some(ref field_name) = field.ident {
|
||||
// Check if the attribute has parameters
|
||||
let mut custom_name = None;
|
||||
if !attr.path().is_ident("index") {
|
||||
continue;
|
||||
}
|
||||
to_remove.push(i);
|
||||
|
||||
// Parse attribute arguments if any
|
||||
let meta = attr.meta.clone();
|
||||
if let syn::Meta::List(list) = meta {
|
||||
if let Ok(nested) = list.parse_args_with(syn::punctuated::Punctuated::<syn::Meta, syn::Token![,]>::parse_terminated) {
|
||||
for meta in nested {
|
||||
if let syn::Meta::NameValue(name_value) = meta {
|
||||
if name_value.path.is_ident("name") {
|
||||
if let syn::Expr::Lit(syn::ExprLit { lit: syn::Lit::Str(lit_str), .. }) = name_value.value {
|
||||
custom_name = Some(lit_str.value());
|
||||
}
|
||||
if let Some(ref field_name) = field.ident {
|
||||
match &attr.meta {
|
||||
Meta::Path(_) => {
|
||||
// Simple top-level index on this field
|
||||
index_decls.push(IndexDecl::TopLevel {
|
||||
field_ident: field_name.clone(),
|
||||
field_ty: field.ty.clone(),
|
||||
});
|
||||
}
|
||||
Meta::List(MetaList { .. }) => {
|
||||
// Parse for path = "..."; name is assumed equal to path
|
||||
// We support syntax: #[index(path = "a.b.c")]
|
||||
if let Ok(nested) = attr.parse_args_with(
|
||||
syn::punctuated::Punctuated::<Meta, syn::Token![,]>::parse_terminated,
|
||||
) {
|
||||
for meta in nested {
|
||||
if let Meta::NameValue(MetaNameValue { path, value, .. }) = meta {
|
||||
if path.is_ident("path") {
|
||||
if let syn::Expr::Lit(syn::ExprLit { lit: Lit::Str(lit_str), .. }) = value {
|
||||
let p = lit_str.value();
|
||||
index_decls.push(IndexDecl::NestedPath {
|
||||
on_field_ident: field_name.clone(),
|
||||
path: p,
|
||||
});
|
||||
}
|
||||
}
|
||||
}
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
indexed_fields.push((field_name.clone(), field.ty.clone()));
|
||||
|
||||
if let Some(name) = custom_name {
|
||||
custom_index_names.insert(field_name.to_string(), name);
|
||||
}
|
||||
_ => {}
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
if let Some(idx) = attr_idx {
|
||||
// remove all #[index] attributes we processed
|
||||
// remove from the back to keep indices valid
|
||||
to_remove.sort_unstable();
|
||||
to_remove.drain(..).rev().for_each(|idx| {
|
||||
field.attrs.remove(idx);
|
||||
}
|
||||
});
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
// Generate Model trait implementation
|
||||
let db_keys_impl = if indexed_fields.is_empty() {
|
||||
let db_keys_impl = if index_decls.is_empty() {
|
||||
quote! {
|
||||
fn db_keys(&self) -> Vec<heromodels_core::IndexKey> {
|
||||
Vec::new()
|
||||
}
|
||||
}
|
||||
} else {
|
||||
let field_keys = indexed_fields.iter().map(|(field_name, _)| {
|
||||
let name_str = custom_index_names
|
||||
.get(&field_name.to_string())
|
||||
.cloned()
|
||||
.unwrap_or(field_name.to_string());
|
||||
quote! {
|
||||
heromodels_core::IndexKey {
|
||||
name: #name_str,
|
||||
value: self.#field_name.to_string(),
|
||||
// Build code for keys from each index declaration
|
||||
let mut key_snippets: Vec<proc_macro2::TokenStream> = Vec::new();
|
||||
|
||||
for decl in &index_decls {
|
||||
match decl.clone() {
|
||||
IndexDecl::TopLevel { field_ident, .. } => {
|
||||
let name_str = field_ident.to_string();
|
||||
key_snippets.push(quote! {
|
||||
keys.push(heromodels_core::IndexKey {
|
||||
name: #name_str,
|
||||
value: self.#field_ident.to_string(),
|
||||
});
|
||||
});
|
||||
}
|
||||
IndexDecl::NestedPath { on_field_ident, path } => {
|
||||
// Name is equal to provided path
|
||||
let name_str = path.clone();
|
||||
// Generate traversal code using serde_json to support arrays and objects generically
|
||||
// Split the path into static segs for iteration
|
||||
let segs: Vec<String> = path.split('.').map(|s| s.to_string()).collect();
|
||||
let segs_iter = segs.iter().map(|s| s.as_str());
|
||||
let segs_array = quote! { [ #( #segs_iter ),* ] };
|
||||
|
||||
key_snippets.push(quote! {
|
||||
// Serialize the target field to JSON for generic traversal
|
||||
let __hm_json_val = ::serde_json::to_value(&self.#on_field_ident).unwrap_or(::serde_json::Value::Null);
|
||||
let mut __hm_stack: Vec<&::serde_json::Value> = vec![&__hm_json_val];
|
||||
for __hm_seg in #segs_array.iter() {
|
||||
let mut __hm_next: Vec<&::serde_json::Value> = Vec::new();
|
||||
for __hm_v in &__hm_stack {
|
||||
match __hm_v {
|
||||
::serde_json::Value::Array(arr) => {
|
||||
for __hm_e in arr {
|
||||
if let ::serde_json::Value::Object(map) = __hm_e {
|
||||
if let Some(x) = map.get(*__hm_seg) { __hm_next.push(x); }
|
||||
}
|
||||
}
|
||||
}
|
||||
::serde_json::Value::Object(map) => {
|
||||
if let Some(x) = map.get(*__hm_seg) { __hm_next.push(x); }
|
||||
}
|
||||
_ => {}
|
||||
}
|
||||
}
|
||||
__hm_stack = __hm_next;
|
||||
if __hm_stack.is_empty() { break; }
|
||||
}
|
||||
for __hm_leaf in __hm_stack {
|
||||
match __hm_leaf {
|
||||
::serde_json::Value::Null => {},
|
||||
::serde_json::Value::Array(_) => {},
|
||||
::serde_json::Value::Object(_) => {},
|
||||
other => {
|
||||
// Convert primitives to string without surrounding quotes for strings
|
||||
let mut s = other.to_string();
|
||||
if let ::serde_json::Value::String(_) = other { s = s.trim_matches('"').to_string(); }
|
||||
keys.push(heromodels_core::IndexKey { name: #name_str, value: s });
|
||||
}
|
||||
}
|
||||
}
|
||||
});
|
||||
}
|
||||
}
|
||||
});
|
||||
}
|
||||
|
||||
quote! {
|
||||
fn db_keys(&self) -> Vec<heromodels_core::IndexKey> {
|
||||
vec![
|
||||
#(#field_keys),*
|
||||
]
|
||||
let mut keys: Vec<heromodels_core::IndexKey> = Vec::new();
|
||||
#(#key_snippets)*
|
||||
keys
|
||||
}
|
||||
}
|
||||
};
|
||||
|
||||
let indexed_field_names = indexed_fields
|
||||
let indexed_field_names: Vec<String> = index_decls
|
||||
.iter()
|
||||
.map(|f| f.0.to_string())
|
||||
.collect::<Vec<_>>();
|
||||
.map(|d| match d {
|
||||
IndexDecl::TopLevel { field_ident, .. } => field_ident.to_string(),
|
||||
IndexDecl::NestedPath { path, .. } => path.clone(),
|
||||
})
|
||||
.collect();
|
||||
|
||||
let model_impl = quote! {
|
||||
impl heromodels_core::Model for #struct_name {
|
||||
@@ -152,51 +231,33 @@ pub fn model(_attr: TokenStream, item: TokenStream) -> TokenStream {
|
||||
}
|
||||
};
|
||||
|
||||
// Generate Index trait implementations
|
||||
// Generate Index trait implementations only for top-level fields, keep existing behavior
|
||||
let mut index_impls = proc_macro2::TokenStream::new();
|
||||
for decl in &index_decls {
|
||||
if let IndexDecl::TopLevel { field_ident, field_ty } = decl {
|
||||
let name_str = field_ident.to_string();
|
||||
let index_struct_name = format_ident!("{}", &name_str);
|
||||
let field_type = field_ty.clone();
|
||||
|
||||
for (field_name, field_type) in &indexed_fields {
|
||||
let name_str = field_name.to_string();
|
||||
let index_impl = quote! {
|
||||
pub struct #index_struct_name;
|
||||
|
||||
// Get custom index name if specified, otherwise use field name
|
||||
let index_key = match custom_index_names.get(&name_str) {
|
||||
Some(custom_name) => custom_name.clone(),
|
||||
None => name_str.clone(),
|
||||
};
|
||||
impl heromodels_core::Index for #index_struct_name {
|
||||
type Model = super::#struct_name;
|
||||
type Key = #field_type;
|
||||
|
||||
// Convert field name to PascalCase for struct name
|
||||
// let struct_name_str = to_pascal_case(&name_str);
|
||||
// let index_struct_name = format_ident!("{}", struct_name_str);
|
||||
let index_struct_name = format_ident!("{}", &name_str);
|
||||
fn key() -> &'static str { #name_str }
|
||||
|
||||
// Default to str for key type
|
||||
let index_impl = quote! {
|
||||
pub struct #index_struct_name;
|
||||
|
||||
impl heromodels_core::Index for #index_struct_name {
|
||||
type Model = super::#struct_name;
|
||||
type Key = #field_type;
|
||||
|
||||
fn key() -> &'static str {
|
||||
#index_key
|
||||
fn field_name() -> &'static str { #name_str }
|
||||
}
|
||||
|
||||
fn field_name() -> &'static str {
|
||||
#name_str
|
||||
}
|
||||
}
|
||||
};
|
||||
|
||||
index_impls.extend(index_impl);
|
||||
};
|
||||
index_impls.extend(index_impl);
|
||||
}
|
||||
}
|
||||
|
||||
if !index_impls.is_empty() {
|
||||
let index_mod_name = format_ident!("{}_index", db_prefix);
|
||||
index_impls = quote! {
|
||||
pub mod #index_mod_name {
|
||||
#index_impls
|
||||
}
|
||||
}
|
||||
index_impls = quote! { pub mod #index_mod_name { #index_impls } };
|
||||
}
|
||||
|
||||
// Combine the original struct with the generated implementations
|
||||
|
@@ -1,7 +1,38 @@
|
||||
use heromodels_derive::model;
|
||||
use serde::{Deserialize, Serialize};
|
||||
|
||||
// Define the necessary structs and traits for testing
|
||||
// Make the current crate visible as an extern crate named `heromodels_core`
|
||||
extern crate self as heromodels_core;
|
||||
extern crate serde_json; // ensure ::serde_json path resolves
|
||||
|
||||
// Mock the heromodels_core API at crate root (visible via the alias above)
|
||||
#[derive(Debug, Clone, PartialEq, Eq)]
|
||||
pub struct IndexKey {
|
||||
pub name: &'static str,
|
||||
pub value: String,
|
||||
}
|
||||
|
||||
pub trait Model: std::fmt::Debug + Clone + Serialize + for<'de> Deserialize<'de> + Send + Sync + 'static {
|
||||
fn db_prefix() -> &'static str
|
||||
where
|
||||
Self: Sized;
|
||||
fn get_id(&self) -> u32;
|
||||
fn base_data_mut(&mut self) -> &mut BaseModelData;
|
||||
fn db_keys(&self) -> Vec<IndexKey> {
|
||||
Vec::new()
|
||||
}
|
||||
fn indexed_fields() -> Vec<&'static str> {
|
||||
Vec::new()
|
||||
}
|
||||
}
|
||||
|
||||
pub trait Index {
|
||||
type Model: Model;
|
||||
type Key: ToString + ?Sized;
|
||||
fn key() -> &'static str;
|
||||
fn field_name() -> &'static str;
|
||||
}
|
||||
|
||||
#[derive(Debug, Clone, Serialize, Deserialize)]
|
||||
pub struct BaseModelData {
|
||||
pub id: u32,
|
||||
@@ -11,41 +42,18 @@ pub struct BaseModelData {
|
||||
}
|
||||
|
||||
impl BaseModelData {
|
||||
pub fn new(id: u32) -> Self {
|
||||
let now = 1000; // Mock timestamp
|
||||
Self {
|
||||
id,
|
||||
created_at: now,
|
||||
modified_at: now,
|
||||
comments: Vec::new(),
|
||||
}
|
||||
pub fn new() -> Self {
|
||||
let now = 1000;
|
||||
Self { id: 0, created_at: now, modified_at: now, comments: Vec::new() }
|
||||
}
|
||||
pub fn update_modified(&mut self) { self.modified_at += 1; }
|
||||
}
|
||||
|
||||
#[derive(Debug, Clone, PartialEq, Eq)]
|
||||
pub struct IndexKey {
|
||||
pub name: &'static str,
|
||||
pub value: String,
|
||||
}
|
||||
|
||||
pub trait Model: std::fmt::Debug + Clone {
|
||||
fn db_prefix() -> &'static str;
|
||||
fn get_id(&self) -> u32;
|
||||
fn base_data_mut(&mut self) -> &mut BaseModelData;
|
||||
fn db_keys(&self) -> Vec<IndexKey>;
|
||||
}
|
||||
|
||||
pub trait Index {
|
||||
type Model: Model;
|
||||
type Key: ?Sized;
|
||||
fn key() -> &'static str;
|
||||
}
|
||||
|
||||
// Test struct using the model macro
|
||||
// Top-level field index tests
|
||||
#[derive(Debug, Clone, Serialize, Deserialize)]
|
||||
#[model]
|
||||
struct TestUser {
|
||||
base_data: BaseModelData,
|
||||
pub struct TestUser {
|
||||
base_data: heromodels_core::BaseModelData,
|
||||
|
||||
#[index]
|
||||
username: String,
|
||||
@@ -54,25 +62,12 @@ struct TestUser {
|
||||
is_active: bool,
|
||||
}
|
||||
|
||||
// Test struct with custom index name
|
||||
#[derive(Debug, Clone, Serialize, Deserialize)]
|
||||
#[model]
|
||||
struct TestUserWithCustomIndex {
|
||||
base_data: BaseModelData,
|
||||
|
||||
#[index(name = "custom_username")]
|
||||
username: String,
|
||||
|
||||
#[index]
|
||||
is_active: bool,
|
||||
}
|
||||
|
||||
#[test]
|
||||
fn test_basic_model() {
|
||||
assert_eq!(TestUser::db_prefix(), "test_user");
|
||||
|
||||
let user = TestUser {
|
||||
base_data: BaseModelData::new(1),
|
||||
base_data: heromodels_core::BaseModelData::new(),
|
||||
username: "test".to_string(),
|
||||
is_active: true,
|
||||
};
|
||||
@@ -85,22 +80,47 @@ fn test_basic_model() {
|
||||
assert_eq!(keys[1].value, "true");
|
||||
}
|
||||
|
||||
// Nested path index tests (including vector traversal)
|
||||
#[derive(Debug, Clone, Serialize, Deserialize, Default)]
|
||||
struct GPU { gpu_brand: String }
|
||||
|
||||
#[derive(Debug, Clone, Serialize, Deserialize, Default)]
|
||||
struct CPU { cpu_brand: String }
|
||||
|
||||
#[derive(Debug, Clone, Serialize, Deserialize, Default)]
|
||||
struct DeviceInfo { vendor: String, cpu: Vec<CPU>, gpu: Vec<GPU> }
|
||||
|
||||
#[derive(Debug, Clone, Serialize, Deserialize)]
|
||||
#[model]
|
||||
pub struct NodeLike {
|
||||
base_data: heromodels_core::BaseModelData,
|
||||
|
||||
#[index(path = "vendor")]
|
||||
#[index(path = "cpu.cpu_brand")]
|
||||
#[index(path = "gpu.gpu_brand")]
|
||||
devices: DeviceInfo,
|
||||
}
|
||||
|
||||
#[test]
|
||||
fn test_custom_index_name() {
|
||||
let user = TestUserWithCustomIndex {
|
||||
base_data: BaseModelData::new(1),
|
||||
username: "test".to_string(),
|
||||
is_active: true,
|
||||
fn test_nested_indexes() {
|
||||
let n = NodeLike {
|
||||
base_data: heromodels_core::BaseModelData::new(),
|
||||
devices: DeviceInfo {
|
||||
vendor: "SuperVendor".to_string(),
|
||||
cpu: vec![CPU { cpu_brand: "Intel".into() }, CPU { cpu_brand: "AMD".into() }],
|
||||
gpu: vec![GPU { gpu_brand: "NVIDIA".into() }, GPU { gpu_brand: "AMD".into() }],
|
||||
},
|
||||
};
|
||||
|
||||
// Check that the Username struct uses the custom index name
|
||||
assert_eq!(Username::key(), "custom_username");
|
||||
let mut keys = n.db_keys();
|
||||
// Sort for deterministic assertions
|
||||
keys.sort_by(|a,b| a.name.cmp(b.name).then(a.value.cmp(&b.value)));
|
||||
|
||||
// Check that the db_keys method returns the correct keys
|
||||
let keys = user.db_keys();
|
||||
assert_eq!(keys.len(), 2);
|
||||
assert_eq!(keys[0].name, "custom_username");
|
||||
assert_eq!(keys[0].value, "test");
|
||||
assert_eq!(keys[1].name, "is_active");
|
||||
assert_eq!(keys[1].value, "true");
|
||||
// Expect 1 (vendor) + 2 (cpu brands) + 2 (gpu brands) = 5 keys
|
||||
assert_eq!(keys.len(), 5);
|
||||
assert!(keys.iter().any(|k| k.name == "vendor" && k.value == "SuperVendor"));
|
||||
assert!(keys.iter().any(|k| k.name == "cpu.cpu_brand" && k.value == "Intel"));
|
||||
assert!(keys.iter().any(|k| k.name == "cpu.cpu_brand" && k.value == "AMD"));
|
||||
assert!(keys.iter().any(|k| k.name == "gpu.gpu_brand" && k.value == "NVIDIA"));
|
||||
assert!(keys.iter().any(|k| k.name == "gpu.gpu_brand" && k.value == "AMD"));
|
||||
}
|
||||
|
1637
heromodels/Cargo.lock
generated
1637
heromodels/Cargo.lock
generated
File diff suppressed because it is too large
Load Diff
@@ -10,11 +10,11 @@ serde = { version = "1.0", features = ["derive"] }
|
||||
serde_json = "1.0"
|
||||
bincode = { version = "2", features = ["serde"] }
|
||||
chrono = { version = "0.4", features = ["serde"] }
|
||||
ourdb = { path = "../../herolib_rust/packages/data/ourdb" }
|
||||
tst = { path = "../../herolib_rust/packages/data/tst" }
|
||||
ourdb = { git = "https://git.ourworld.tf/herocode/herolib_rust", package = "ourdb" }
|
||||
tst = { git = "https://git.ourworld.tf/herocode/herolib_rust", package = "tst" }
|
||||
heromodels-derive = { path = "../heromodels-derive" }
|
||||
heromodels_core = { path = "../heromodels_core" }
|
||||
rhailib-macros = { path = "../../herolib_rust/rhailib/src/macros" }
|
||||
rhailib-macros = { git = "https://git.ourworld.tf/herocode/herolib_rust", package = "rhailib-macros" }
|
||||
rhai = { version = "1.21.0", features = [
|
||||
"std",
|
||||
"sync",
|
||||
@@ -53,11 +53,19 @@ path = "examples/finance_example/main.rs"
|
||||
name = "flow_example"
|
||||
path = "examples/flow_example.rs"
|
||||
|
||||
[[example]]
|
||||
name = "biz_rhai"
|
||||
path = "examples/biz_rhai/example.rs"
|
||||
required-features = ["rhai"]
|
||||
# [[example]]
|
||||
# name = "biz_rhai"
|
||||
# path = "examples/biz_rhai/example.rs"
|
||||
# required-features = ["rhai"]
|
||||
|
||||
[[example]]
|
||||
name = "postgres_model_example"
|
||||
path = "examples/postgres_example/example.rs"
|
||||
|
||||
[[example]]
|
||||
name = "heroledger_example"
|
||||
path = "examples/heroledger_example/example.rs"
|
||||
|
||||
[[example]]
|
||||
name = "grid4_example"
|
||||
path = "examples/grid4_example/example.rs"
|
||||
|
@@ -1,10 +1,25 @@
|
||||
use chrono::{Duration, Utc};
|
||||
use chrono::{Duration, Utc, NaiveDateTime};
|
||||
use heromodels::db::{Collection, Db};
|
||||
use heromodels::models::User;
|
||||
use heromodels::models::calendar::{AttendanceStatus, Attendee, Calendar, Event, EventStatus};
|
||||
use heromodels_core::Model;
|
||||
|
||||
fn main() {
|
||||
// Helper to format i64 timestamps
|
||||
let fmt_time = |ts: i64| -> String {
|
||||
let ndt = NaiveDateTime::from_timestamp_opt(ts, 0)
|
||||
.unwrap_or(NaiveDateTime::from_timestamp_opt(0, 0).unwrap());
|
||||
chrono::DateTime::<Utc>::from_utc(ndt, Utc)
|
||||
.format("%Y-%m-%d %H:%M")
|
||||
.to_string()
|
||||
};
|
||||
let fmt_date = |ts: i64| -> String {
|
||||
let ndt = NaiveDateTime::from_timestamp_opt(ts, 0)
|
||||
.unwrap_or(NaiveDateTime::from_timestamp_opt(0, 0).unwrap());
|
||||
chrono::DateTime::<Utc>::from_utc(ndt, Utc)
|
||||
.format("%Y-%m-%d")
|
||||
.to_string()
|
||||
};
|
||||
// Create a new DB instance, reset before every run
|
||||
let db_path = "/tmp/ourdb_calendar_example";
|
||||
let db = heromodels::db::hero::OurDB::new(db_path, true).expect("Can create DB");
|
||||
@@ -47,50 +62,21 @@ fn main() {
|
||||
println!("- User 2 (ID: {}): {}", user2_id, stored_user2.full_name);
|
||||
println!("- User 3 (ID: {}): {}", user3_id, stored_user3.full_name);
|
||||
|
||||
// --- Create Attendees ---
|
||||
// --- Create Attendees (embedded in events, not stored separately) ---
|
||||
println!("\n--- Creating Attendees ---");
|
||||
let attendee1 = Attendee::new(user1_id).status(AttendanceStatus::Accepted);
|
||||
let attendee2 = Attendee::new(user2_id).status(AttendanceStatus::Tentative);
|
||||
let attendee3 = Attendee::new(user3_id); // Default NoResponse
|
||||
|
||||
// Store attendees in database and get their IDs
|
||||
let attendee_collection = db
|
||||
.collection::<Attendee>()
|
||||
.expect("can open attendee collection");
|
||||
|
||||
let (attendee1_id, stored_attendee1) = attendee_collection
|
||||
.set(&attendee1)
|
||||
.expect("can set attendee1");
|
||||
let (attendee2_id, stored_attendee2) = attendee_collection
|
||||
.set(&attendee2)
|
||||
.expect("can set attendee2");
|
||||
let (attendee3_id, stored_attendee3) = attendee_collection
|
||||
.set(&attendee3)
|
||||
.expect("can set attendee3");
|
||||
|
||||
println!("Created attendees:");
|
||||
println!(
|
||||
"- Attendee 1 (ID: {}): Contact ID {}, Status: {:?}",
|
||||
attendee1_id, stored_attendee1.contact_id, stored_attendee1.status
|
||||
);
|
||||
println!(
|
||||
"- Attendee 2 (ID: {}): Contact ID {}, Status: {:?}",
|
||||
attendee2_id, stored_attendee2.contact_id, stored_attendee2.status
|
||||
);
|
||||
println!(
|
||||
"- Attendee 3 (ID: {}): Contact ID {}, Status: {:?}",
|
||||
attendee3_id, stored_attendee3.contact_id, stored_attendee3.status
|
||||
);
|
||||
|
||||
// --- Create Events with Attendees ---
|
||||
println!("\n--- Creating Events with Enhanced Features ---");
|
||||
let now = Utc::now();
|
||||
let event1_start = (now + Duration::hours(1)).timestamp();
|
||||
let event1_end = (now + Duration::hours(2)).timestamp();
|
||||
|
||||
let event1 = Event::new(
|
||||
"Team Meeting",
|
||||
now + Duration::hours(1),
|
||||
now + Duration::hours(2),
|
||||
)
|
||||
let event1 = Event::new()
|
||||
.title("Team Meeting")
|
||||
.reschedule(event1_start, event1_end)
|
||||
.description("Weekly sync-up meeting to discuss project progress.")
|
||||
.location("Conference Room A")
|
||||
.color("#FF5722") // Red-orange color
|
||||
@@ -99,14 +85,14 @@ fn main() {
|
||||
.category("Work")
|
||||
.reminder_minutes(15)
|
||||
.timezone("UTC")
|
||||
.add_attendee(attendee1_id)
|
||||
.add_attendee(attendee2_id);
|
||||
.add_attendee(attendee1.clone())
|
||||
.add_attendee(attendee2.clone());
|
||||
|
||||
let event2 = Event::new(
|
||||
"Project Brainstorm",
|
||||
now + Duration::days(1),
|
||||
now + Duration::days(1) + Duration::minutes(90),
|
||||
)
|
||||
let event2_start = (now + Duration::days(1)).timestamp();
|
||||
let event2_end = (now + Duration::days(1) + Duration::minutes(90)).timestamp();
|
||||
let event2 = Event::new()
|
||||
.title("Project Brainstorm")
|
||||
.reschedule(event2_start, event2_end)
|
||||
.description("Brainstorming session for new project features.")
|
||||
.location("Innovation Lab")
|
||||
.color("#4CAF50") // Green color
|
||||
@@ -115,28 +101,28 @@ fn main() {
|
||||
.category("Planning")
|
||||
.reminder_minutes(30)
|
||||
.is_recurring(true)
|
||||
.add_attendee(attendee1_id)
|
||||
.add_attendee(attendee3_id);
|
||||
.add_attendee(attendee1.clone())
|
||||
.add_attendee(attendee3.clone());
|
||||
|
||||
let event3 = Event::new(
|
||||
"Client Call",
|
||||
now + Duration::days(2),
|
||||
now + Duration::days(2) + Duration::hours(1),
|
||||
)
|
||||
let event3_start = (now + Duration::days(2)).timestamp();
|
||||
let event3_end = (now + Duration::days(2) + Duration::hours(1)).timestamp();
|
||||
let event3 = Event::new()
|
||||
.title("Client Call")
|
||||
.reschedule(event3_start, event3_end)
|
||||
.description("Quarterly review with key client.")
|
||||
.color("#9C27B0") // Purple color
|
||||
.created_by(user3_id)
|
||||
.status(EventStatus::Published)
|
||||
.category("Client")
|
||||
.reminder_minutes(60)
|
||||
.add_attendee(attendee2_id);
|
||||
.add_attendee(attendee2.clone());
|
||||
|
||||
// Create an all-day event
|
||||
let event4 = Event::new(
|
||||
"Company Holiday",
|
||||
now + Duration::days(7),
|
||||
now + Duration::days(7) + Duration::hours(24),
|
||||
)
|
||||
let event4_start = (now + Duration::days(7)).timestamp();
|
||||
let event4_end = (now + Duration::days(7) + Duration::hours(24)).timestamp();
|
||||
let event4 = Event::new()
|
||||
.title("Company Holiday")
|
||||
.reschedule(event4_start, event4_end)
|
||||
.description("National holiday - office closed.")
|
||||
.color("#FFC107") // Amber color
|
||||
.all_day(true)
|
||||
@@ -148,7 +134,7 @@ fn main() {
|
||||
println!(
|
||||
"- Event 1: '{}' at {} with {} attendees",
|
||||
event1.title,
|
||||
event1.start_time.format("%Y-%m-%d %H:%M"),
|
||||
fmt_time(event1.start_time),
|
||||
event1.attendees.len()
|
||||
);
|
||||
println!(
|
||||
@@ -174,12 +160,19 @@ fn main() {
|
||||
);
|
||||
println!(" All-day: {}", event1.all_day);
|
||||
println!(" Recurring: {}", event1.is_recurring);
|
||||
println!(" Attendee IDs: {:?}", event1.attendees);
|
||||
println!(
|
||||
" Attendee IDs: {:?}",
|
||||
event1
|
||||
.attendees
|
||||
.iter()
|
||||
.map(|a| a.contact_id)
|
||||
.collect::<Vec<u32>>()
|
||||
);
|
||||
|
||||
println!(
|
||||
"- Event 2: '{}' at {} with {} attendees",
|
||||
event2.title,
|
||||
event2.start_time.format("%Y-%m-%d %H:%M"),
|
||||
fmt_time(event2.start_time),
|
||||
event2.attendees.len()
|
||||
);
|
||||
println!(
|
||||
@@ -205,12 +198,19 @@ fn main() {
|
||||
);
|
||||
println!(" All-day: {}", event2.all_day);
|
||||
println!(" Recurring: {}", event2.is_recurring);
|
||||
println!(" Attendee IDs: {:?}", event2.attendees);
|
||||
println!(
|
||||
" Attendee IDs: {:?}",
|
||||
event2
|
||||
.attendees
|
||||
.iter()
|
||||
.map(|a| a.contact_id)
|
||||
.collect::<Vec<u32>>()
|
||||
);
|
||||
|
||||
println!(
|
||||
"- Event 3: '{}' at {} with {} attendees",
|
||||
event3.title,
|
||||
event3.start_time.format("%Y-%m-%d %H:%M"),
|
||||
fmt_time(event3.start_time),
|
||||
event3.attendees.len()
|
||||
);
|
||||
println!(
|
||||
@@ -236,12 +236,19 @@ fn main() {
|
||||
);
|
||||
println!(" All-day: {}", event3.all_day);
|
||||
println!(" Recurring: {}", event3.is_recurring);
|
||||
println!(" Attendee IDs: {:?}", event3.attendees);
|
||||
println!(
|
||||
" Attendee IDs: {:?}",
|
||||
event3
|
||||
.attendees
|
||||
.iter()
|
||||
.map(|a| a.contact_id)
|
||||
.collect::<Vec<u32>>()
|
||||
);
|
||||
|
||||
println!(
|
||||
"- Event 4: '{}' at {} (All-day: {})",
|
||||
event4.title,
|
||||
event4.start_time.format("%Y-%m-%d"),
|
||||
fmt_date(event4.start_time),
|
||||
event4.all_day
|
||||
);
|
||||
println!(
|
||||
@@ -262,25 +269,37 @@ fn main() {
|
||||
let new_start = now + Duration::hours(2);
|
||||
let new_end = now + Duration::hours(3);
|
||||
let mut updated_event1 = event1.clone();
|
||||
updated_event1 = updated_event1.reschedule(new_start, new_end);
|
||||
updated_event1 = updated_event1.reschedule(new_start.timestamp(), new_end.timestamp());
|
||||
println!(
|
||||
"Rescheduled '{}' to {}",
|
||||
updated_event1.title,
|
||||
new_start.format("%Y-%m-%d %H:%M")
|
||||
fmt_time(new_start.timestamp())
|
||||
);
|
||||
|
||||
// Remove an attendee
|
||||
updated_event1 = updated_event1.remove_attendee(attendee1_id);
|
||||
updated_event1 = updated_event1.remove_attendee(user1_id);
|
||||
println!(
|
||||
"Removed attendee {} from '{}'. Remaining attendee IDs: {:?}",
|
||||
attendee1_id, updated_event1.title, updated_event1.attendees
|
||||
user1_id,
|
||||
updated_event1.title,
|
||||
updated_event1
|
||||
.attendees
|
||||
.iter()
|
||||
.map(|a| a.contact_id)
|
||||
.collect::<Vec<u32>>()
|
||||
);
|
||||
|
||||
// Add a new attendee
|
||||
updated_event1 = updated_event1.add_attendee(attendee3_id);
|
||||
updated_event1 = updated_event1.add_attendee(attendee3.clone());
|
||||
println!(
|
||||
"Added attendee {} to '{}'. Current attendee IDs: {:?}",
|
||||
attendee3_id, updated_event1.title, updated_event1.attendees
|
||||
user3_id,
|
||||
updated_event1.title,
|
||||
updated_event1
|
||||
.attendees
|
||||
.iter()
|
||||
.map(|a| a.contact_id)
|
||||
.collect::<Vec<u32>>()
|
||||
);
|
||||
|
||||
// --- Demonstrate Event Status Changes ---
|
||||
@@ -300,11 +319,11 @@ fn main() {
|
||||
println!("Cancelled event: '{}'", cancelled_event.title);
|
||||
|
||||
// Update event with new features
|
||||
let enhanced_event = Event::new(
|
||||
"Enhanced Meeting",
|
||||
now + Duration::days(5),
|
||||
now + Duration::days(5) + Duration::hours(2),
|
||||
)
|
||||
let enhanced_start = (now + Duration::days(5)).timestamp();
|
||||
let enhanced_end = (now + Duration::days(5) + Duration::hours(2)).timestamp();
|
||||
let enhanced_event = Event::new()
|
||||
.title("Enhanced Meeting")
|
||||
.reschedule(enhanced_start, enhanced_end)
|
||||
.description("Meeting with all new features demonstrated.")
|
||||
.location("Virtual - Zoom")
|
||||
.color("#673AB7") // Deep purple
|
||||
@@ -314,9 +333,9 @@ fn main() {
|
||||
.reminder_minutes(45)
|
||||
.timezone("America/New_York")
|
||||
.is_recurring(true)
|
||||
.add_attendee(attendee1_id)
|
||||
.add_attendee(attendee2_id)
|
||||
.add_attendee(attendee3_id);
|
||||
.add_attendee(attendee1)
|
||||
.add_attendee(attendee2)
|
||||
.add_attendee(attendee3);
|
||||
|
||||
println!("Created enhanced event with all features:");
|
||||
println!(" Title: {}", enhanced_event.title);
|
||||
@@ -485,13 +504,13 @@ fn main() {
|
||||
println!("\n--- Modifying Calendar ---");
|
||||
|
||||
// Create and store a new event
|
||||
let new_event = Event::new(
|
||||
"1-on-1 Meeting",
|
||||
now + Duration::days(3),
|
||||
now + Duration::days(3) + Duration::minutes(30),
|
||||
)
|
||||
.description("One-on-one meeting with team member.")
|
||||
.location("Office");
|
||||
let ne_start = (now + Duration::days(3)).timestamp();
|
||||
let ne_end = (now + Duration::days(3) + Duration::minutes(30)).timestamp();
|
||||
let new_event = Event::new()
|
||||
.title("1-on-1 Meeting")
|
||||
.reschedule(ne_start, ne_end)
|
||||
.description("One-on-one meeting with team member.")
|
||||
.location("Office");
|
||||
|
||||
let (new_event_id, _stored_new_event) =
|
||||
event_collection.set(&new_event).expect("can set new event");
|
||||
@@ -565,7 +584,7 @@ fn main() {
|
||||
"- Event ID: {}, Title: '{}', Start: {}, Attendees: {}",
|
||||
event.get_id(),
|
||||
event.title,
|
||||
event.start_time.format("%Y-%m-%d %H:%M"),
|
||||
fmt_time(event.start_time),
|
||||
event.attendees.len()
|
||||
);
|
||||
}
|
||||
@@ -583,22 +602,16 @@ fn main() {
|
||||
retrieved_event1.attendees.len()
|
||||
);
|
||||
|
||||
// Look up attendee details for each attendee ID
|
||||
for &attendee_id in &retrieved_event1.attendees {
|
||||
if let Some(attendee) = attendee_collection
|
||||
.get_by_id(attendee_id)
|
||||
.expect("can try to get attendee")
|
||||
// Look up attendee details directly from embedded attendees
|
||||
for attendee in &retrieved_event1.attendees {
|
||||
if let Some(user) = user_collection
|
||||
.get_by_id(attendee.contact_id)
|
||||
.expect("can try to get user")
|
||||
{
|
||||
// Look up user details for the attendee's contact_id
|
||||
if let Some(user) = user_collection
|
||||
.get_by_id(attendee.contact_id)
|
||||
.expect("can try to get user")
|
||||
{
|
||||
println!(
|
||||
" - Attendee ID {}: {} (User: {}, Status: {:?})",
|
||||
attendee_id, user.full_name, attendee.contact_id, attendee.status
|
||||
);
|
||||
}
|
||||
println!(
|
||||
" - User {}: {} (Status: {:?})",
|
||||
attendee.contact_id, user.full_name, attendee.status
|
||||
);
|
||||
}
|
||||
}
|
||||
}
|
||||
|
@@ -1,26 +1,26 @@
|
||||
use circles_launcher::{new_launcher};
|
||||
use heromodels::models::circle::circle::{new_circle};
|
||||
use secp256k1::{Secp256k1, SecretKey, PublicKey};
|
||||
use circles_launcher::new_launcher;
|
||||
use heromodels::models::circle::circle::new_circle;
|
||||
use rand::rngs::OsRng;
|
||||
use secp256k1::{PublicKey, Secp256k1, SecretKey};
|
||||
|
||||
#[tokio::main]
|
||||
async fn main() -> Result<(), Box<dyn std::error::Error>> {
|
||||
async fn main() -> Result<(), Box<dyn std::error::Error>> {
|
||||
// Generate valid secp256k1 keypairs for testing
|
||||
let secp = Secp256k1::new();
|
||||
let mut rng = OsRng;
|
||||
|
||||
|
||||
let secret_key1 = SecretKey::new(&mut rng);
|
||||
let public_key1 = PublicKey::from_secret_key(&secp, &secret_key1);
|
||||
let pk1_hex = hex::encode(public_key1.serialize());
|
||||
|
||||
|
||||
let secret_key2 = SecretKey::new(&mut rng);
|
||||
let public_key2 = PublicKey::from_secret_key(&secp, &secret_key2);
|
||||
let pk2_hex = hex::encode(public_key2.serialize());
|
||||
|
||||
|
||||
let secret_key3 = SecretKey::new(&mut rng);
|
||||
let public_key3 = PublicKey::from_secret_key(&secp, &secret_key3);
|
||||
let pk3_hex = hex::encode(public_key3.serialize());
|
||||
|
||||
|
||||
println!("Generated test public keys:");
|
||||
println!(" PK1: {}", pk1_hex);
|
||||
println!(" PK2: {}", pk2_hex);
|
||||
@@ -36,4 +36,4 @@ async fn main() -> Result<(), Box<dyn std::error::Error>> {
|
||||
.save();
|
||||
|
||||
Ok(())
|
||||
}
|
||||
}
|
||||
|
199
heromodels/examples/grid4_bid_example.rs
Normal file
199
heromodels/examples/grid4_bid_example.rs
Normal file
@@ -0,0 +1,199 @@
|
||||
use heromodels::db::{Collection, Db};
|
||||
use heromodels::models::grid4::{Bid, BidStatus, BillingPeriod};
|
||||
use heromodels::models::grid4::bid::bid_index::customer_id;
|
||||
use heromodels_core::Model;
|
||||
|
||||
// Helper function to print bid details
|
||||
fn print_bid_details(bid: &Bid) {
|
||||
println!("\n--- Bid Details ---");
|
||||
println!("ID: {}", bid.get_id());
|
||||
println!("Customer ID: {}", bid.customer_id);
|
||||
println!("Compute Slices: {}", bid.compute_slices_nr);
|
||||
println!("Compute Slice Price: ${:.2}", bid.compute_slice_price);
|
||||
println!("Storage Slices: {}", bid.storage_slices_nr);
|
||||
println!("Storage Slice Price: ${:.2}", bid.storage_slice_price);
|
||||
println!("Status: {:?}", bid.status);
|
||||
println!("Obligation: {}", bid.obligation);
|
||||
println!("Start Date: {}", bid.start_date);
|
||||
println!("End Date: {}", bid.end_date);
|
||||
println!("Billing Period: {:?}", bid.billing_period);
|
||||
println!("Signature User: {}", bid.signature_user);
|
||||
println!("Created At: {}", bid.base_data.created_at);
|
||||
println!("Modified At: {}", bid.base_data.modified_at);
|
||||
}
|
||||
|
||||
fn main() {
|
||||
// Create a new DB instance in /tmp/grid4_db, and reset before every run
|
||||
let db = heromodels::db::hero::OurDB::new("/tmp/grid4_db", true).expect("Can create DB");
|
||||
|
||||
println!("Grid4 Bid Models - Basic Usage Example");
|
||||
println!("=====================================");
|
||||
|
||||
// Create bids with different configurations
|
||||
|
||||
// Bid 1 - Small compute request
|
||||
let bid1 = Bid::new()
|
||||
.customer_id(101)
|
||||
.compute_slices_nr(4)
|
||||
.compute_slice_price(0.05)
|
||||
.storage_slices_nr(10)
|
||||
.storage_slice_price(0.02)
|
||||
.status(BidStatus::Pending)
|
||||
.obligation(false)
|
||||
.start_date(1640995200) // 2022-01-01
|
||||
.end_date(1672531200) // 2023-01-01
|
||||
.billing_period(BillingPeriod::Monthly)
|
||||
.signature_user("sig_user_101_abc123".to_string());
|
||||
|
||||
// Bid 2 - Large compute request with obligation
|
||||
let bid2 = Bid::new()
|
||||
.customer_id(102)
|
||||
.compute_slices_nr(16)
|
||||
.compute_slice_price(0.04)
|
||||
.storage_slices_nr(50)
|
||||
.storage_slice_price(0.015)
|
||||
.status(BidStatus::Confirmed)
|
||||
.obligation(true)
|
||||
.start_date(1640995200)
|
||||
.end_date(1704067200) // 2024-01-01
|
||||
.billing_period(BillingPeriod::Yearly)
|
||||
.signature_user("sig_user_102_def456".to_string());
|
||||
|
||||
// Bid 3 - Storage-heavy request
|
||||
let bid3 = Bid::new()
|
||||
.customer_id(103)
|
||||
.compute_slices_nr(2)
|
||||
.compute_slice_price(0.06)
|
||||
.storage_slices_nr(100)
|
||||
.storage_slice_price(0.01)
|
||||
.status(BidStatus::Assigned)
|
||||
.obligation(true)
|
||||
.start_date(1640995200)
|
||||
.end_date(1672531200)
|
||||
.billing_period(BillingPeriod::Hourly)
|
||||
.signature_user("sig_user_103_ghi789".to_string());
|
||||
|
||||
// Bid 4 - Cancelled bid
|
||||
let bid4 = Bid::new()
|
||||
.customer_id(104)
|
||||
.compute_slices_nr(8)
|
||||
.compute_slice_price(0.055)
|
||||
.storage_slices_nr(25)
|
||||
.storage_slice_price(0.018)
|
||||
.status(BidStatus::Cancelled)
|
||||
.obligation(false)
|
||||
.start_date(1640995200)
|
||||
.end_date(1672531200)
|
||||
.billing_period(BillingPeriod::Monthly)
|
||||
.signature_user("sig_user_104_jkl012".to_string());
|
||||
|
||||
// Save all bids to database and get their assigned IDs and updated models
|
||||
let (bid1_id, db_bid1) = db
|
||||
.collection()
|
||||
.expect("can open bid collection")
|
||||
.set(&bid1)
|
||||
.expect("can set bid");
|
||||
let (bid2_id, db_bid2) = db
|
||||
.collection()
|
||||
.expect("can open bid collection")
|
||||
.set(&bid2)
|
||||
.expect("can set bid");
|
||||
let (bid3_id, db_bid3) = db
|
||||
.collection()
|
||||
.expect("can open bid collection")
|
||||
.set(&bid3)
|
||||
.expect("can set bid");
|
||||
let (bid4_id, db_bid4) = db
|
||||
.collection()
|
||||
.expect("can open bid collection")
|
||||
.set(&bid4)
|
||||
.expect("can set bid");
|
||||
|
||||
println!("Bid 1 assigned ID: {}", bid1_id);
|
||||
println!("Bid 2 assigned ID: {}", bid2_id);
|
||||
println!("Bid 3 assigned ID: {}", bid3_id);
|
||||
println!("Bid 4 assigned ID: {}", bid4_id);
|
||||
|
||||
// Print all bids retrieved from database
|
||||
println!("\n--- Bids Retrieved from Database ---");
|
||||
println!("\n1. Small compute bid:");
|
||||
print_bid_details(&db_bid1);
|
||||
|
||||
println!("\n2. Large compute bid with obligation:");
|
||||
print_bid_details(&db_bid2);
|
||||
|
||||
println!("\n3. Storage-heavy bid:");
|
||||
print_bid_details(&db_bid3);
|
||||
|
||||
println!("\n4. Cancelled bid:");
|
||||
print_bid_details(&db_bid4);
|
||||
|
||||
// Demonstrate different ways to retrieve bids from the database
|
||||
println!("\n--- Retrieving Bids by Different Methods ---");
|
||||
println!("\n1. By Customer ID Index (Customer 102):");
|
||||
|
||||
let customer_bids = db
|
||||
.collection::<Bid>()
|
||||
.expect("can open bid collection")
|
||||
.get::<customer_id, _>(&102u32)
|
||||
.expect("can load bids by customer");
|
||||
|
||||
assert_eq!(customer_bids.len(), 1);
|
||||
print_bid_details(&customer_bids[0]);
|
||||
|
||||
println!("\n2. Updating Bid Status:");
|
||||
let mut updated_bid = db_bid1.clone();
|
||||
updated_bid.status = BidStatus::Confirmed;
|
||||
|
||||
let (_, confirmed_bid) = db
|
||||
.collection::<Bid>()
|
||||
.expect("can open bid collection")
|
||||
.set(&updated_bid)
|
||||
.expect("can update bid");
|
||||
|
||||
println!("Updated bid status to Confirmed:");
|
||||
print_bid_details(&confirmed_bid);
|
||||
|
||||
// 3. Delete a bid and show the updated results
|
||||
println!("\n3. After Deleting a Bid:");
|
||||
println!("Deleting bid with ID: {}", bid4_id);
|
||||
db.collection::<Bid>()
|
||||
.expect("can open bid collection")
|
||||
.delete_by_id(bid4_id)
|
||||
.expect("can delete existing bid");
|
||||
|
||||
// Show remaining bids
|
||||
let all_bids = db
|
||||
.collection::<Bid>()
|
||||
.expect("can open bid collection")
|
||||
.get_all()
|
||||
.expect("can load all bids");
|
||||
|
||||
println!("Remaining bids count: {}", all_bids.len());
|
||||
assert_eq!(all_bids.len(), 3);
|
||||
|
||||
// Calculate total compute and storage requested
|
||||
println!("\n--- Bid Analytics ---");
|
||||
let total_compute_slices: i32 = all_bids.iter().map(|b| b.compute_slices_nr).sum();
|
||||
let total_storage_slices: i32 = all_bids.iter().map(|b| b.storage_slices_nr).sum();
|
||||
let avg_compute_price: f64 = all_bids.iter().map(|b| b.compute_slice_price).sum::<f64>() / all_bids.len() as f64;
|
||||
let avg_storage_price: f64 = all_bids.iter().map(|b| b.storage_slice_price).sum::<f64>() / all_bids.len() as f64;
|
||||
|
||||
println!("Total Compute Slices Requested: {}", total_compute_slices);
|
||||
println!("Total Storage Slices Requested: {}", total_storage_slices);
|
||||
println!("Average Compute Price: ${:.3}", avg_compute_price);
|
||||
println!("Average Storage Price: ${:.3}", avg_storage_price);
|
||||
|
||||
// Count bids by status
|
||||
let confirmed_count = all_bids.iter().filter(|b| matches!(b.status, BidStatus::Confirmed)).count();
|
||||
let assigned_count = all_bids.iter().filter(|b| matches!(b.status, BidStatus::Assigned)).count();
|
||||
let pending_count = all_bids.iter().filter(|b| matches!(b.status, BidStatus::Pending)).count();
|
||||
|
||||
println!("\nBids by Status:");
|
||||
println!(" Confirmed: {}", confirmed_count);
|
||||
println!(" Assigned: {}", assigned_count);
|
||||
println!(" Pending: {}", pending_count);
|
||||
|
||||
println!("\n--- Model Information ---");
|
||||
println!("Bid DB Prefix: {}", Bid::db_prefix());
|
||||
}
|
301
heromodels/examples/grid4_contract_example.rs
Normal file
301
heromodels/examples/grid4_contract_example.rs
Normal file
@@ -0,0 +1,301 @@
|
||||
use heromodels::db::{Collection, Db};
|
||||
use heromodels::models::grid4::{Contract, ContractStatus};
|
||||
use heromodels::models::grid4::contract::contract_index::customer_id;
|
||||
use heromodels_core::Model;
|
||||
|
||||
// Helper function to print contract details
|
||||
fn print_contract_details(contract: &Contract) {
|
||||
println!("\n--- Contract Details ---");
|
||||
println!("ID: {}", contract.get_id());
|
||||
println!("Customer ID: {}", contract.customer_id);
|
||||
println!("Compute Slices: {}", contract.compute_slices.len());
|
||||
println!("Storage Slices: {}", contract.storage_slices.len());
|
||||
println!("Compute Slice Price: ${:.2}", contract.compute_slice_price);
|
||||
println!("Storage Slice Price: ${:.2}", contract.storage_slice_price);
|
||||
println!("Network Slice Price: ${:.2}", contract.network_slice_price);
|
||||
println!("Status: {:?}", contract.status);
|
||||
println!("Start Date: {}", contract.start_date);
|
||||
println!("End Date: {}", contract.end_date);
|
||||
println!("Billing Period: {:?}", contract.billing_period);
|
||||
println!("Signature User: {}", contract.signature_user);
|
||||
println!("Signature Hoster: {}", contract.signature_hoster);
|
||||
println!("Created At: {}", contract.base_data.created_at);
|
||||
println!("Modified At: {}", contract.base_data.modified_at);
|
||||
|
||||
// Print compute slices details
|
||||
if !contract.compute_slices.is_empty() {
|
||||
println!(" Compute Slices:");
|
||||
for (i, slice) in contract.compute_slices.iter().enumerate() {
|
||||
println!(" {}. Node: {}, ID: {}, Memory: {:.1}GB, Storage: {:.1}GB, Passmark: {}, vCores: {}",
|
||||
i + 1, slice.node_id, slice.id, slice.mem_gb, slice.storage_gb, slice.passmark, slice.vcores);
|
||||
}
|
||||
}
|
||||
|
||||
// Print storage slices details
|
||||
if !contract.storage_slices.is_empty() {
|
||||
println!(" Storage Slices:");
|
||||
for (i, slice) in contract.storage_slices.iter().enumerate() {
|
||||
println!(" {}. Node: {}, ID: {}, Size: {}GB",
|
||||
i + 1, slice.node_id, slice.id, slice.storage_size_gb);
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
fn main() {
|
||||
// Create a new DB instance in /tmp/grid4_contracts_db, and reset before every run
|
||||
let db = heromodels::db::hero::OurDB::new("/tmp/grid4_contracts_db", true).expect("Can create DB");
|
||||
|
||||
println!("Grid4 Contract Models - Basic Usage Example");
|
||||
println!("==========================================");
|
||||
|
||||
// Create compute slices for contracts
|
||||
let compute_slice1 = ComputeSliceProvisioned::new()
|
||||
.node_id(1001)
|
||||
.id(1)
|
||||
.mem_gb(2.0)
|
||||
.storage_gb(20.0)
|
||||
.passmark(2500)
|
||||
.vcores(2)
|
||||
.cpu_oversubscription(150)
|
||||
.tags("web-server,production".to_string());
|
||||
|
||||
let compute_slice2 = ComputeSliceProvisioned::new()
|
||||
.node_id(1002)
|
||||
.id(2)
|
||||
.mem_gb(4.0)
|
||||
.storage_gb(40.0)
|
||||
.passmark(5000)
|
||||
.vcores(4)
|
||||
.cpu_oversubscription(120)
|
||||
.tags("database,high-performance".to_string());
|
||||
|
||||
let compute_slice3 = ComputeSliceProvisioned::new()
|
||||
.node_id(1003)
|
||||
.id(1)
|
||||
.mem_gb(8.0)
|
||||
.storage_gb(80.0)
|
||||
.passmark(10000)
|
||||
.vcores(8)
|
||||
.cpu_oversubscription(100)
|
||||
.tags("ml-training,gpu-enabled".to_string());
|
||||
|
||||
// Create storage slices for contracts
|
||||
let storage_slice1 = StorageSliceProvisioned::new()
|
||||
.node_id(2001)
|
||||
.id(1)
|
||||
.storage_size_gb(100)
|
||||
.tags("backup,cold-storage".to_string());
|
||||
|
||||
let storage_slice2 = StorageSliceProvisioned::new()
|
||||
.node_id(2002)
|
||||
.id(2)
|
||||
.storage_size_gb(500)
|
||||
.tags("data-lake,analytics".to_string());
|
||||
|
||||
let storage_slice3 = StorageSliceProvisioned::new()
|
||||
.node_id(2003)
|
||||
.id(1)
|
||||
.storage_size_gb(1000)
|
||||
.tags("archive,long-term".to_string());
|
||||
|
||||
// Create contracts with different configurations
|
||||
|
||||
// Contract 1 - Small web hosting contract
|
||||
let contract1 = Contract::new()
|
||||
.customer_id(201)
|
||||
.add_compute_slice(compute_slice1.clone())
|
||||
.add_storage_slice(storage_slice1.clone())
|
||||
.compute_slice_price(0.05)
|
||||
.storage_slice_price(0.02)
|
||||
.network_slice_price(0.01)
|
||||
.status(ContractStatus::Active)
|
||||
.start_date(1640995200) // 2022-01-01
|
||||
.end_date(1672531200) // 2023-01-01
|
||||
.billing_period(BillingPeriod::Monthly)
|
||||
.signature_user("contract_user_201_abc123".to_string())
|
||||
.signature_hoster("hoster_node1001_xyz789".to_string());
|
||||
|
||||
// Contract 2 - Database hosting contract
|
||||
let contract2 = Contract::new()
|
||||
.customer_id(202)
|
||||
.add_compute_slice(compute_slice2.clone())
|
||||
.add_storage_slice(storage_slice2.clone())
|
||||
.compute_slice_price(0.04)
|
||||
.storage_slice_price(0.015)
|
||||
.network_slice_price(0.008)
|
||||
.status(ContractStatus::Active)
|
||||
.start_date(1640995200)
|
||||
.end_date(1704067200) // 2024-01-01
|
||||
.billing_period(BillingPeriod::Yearly)
|
||||
.signature_user("contract_user_202_def456".to_string())
|
||||
.signature_hoster("hoster_node1002_uvw123".to_string());
|
||||
|
||||
// Contract 3 - ML training contract (paused)
|
||||
let contract3 = Contract::new()
|
||||
.customer_id(203)
|
||||
.add_compute_slice(compute_slice3.clone())
|
||||
.add_storage_slice(storage_slice3.clone())
|
||||
.compute_slice_price(0.08)
|
||||
.storage_slice_price(0.01)
|
||||
.network_slice_price(0.015)
|
||||
.status(ContractStatus::Paused)
|
||||
.start_date(1640995200)
|
||||
.end_date(1672531200)
|
||||
.billing_period(BillingPeriod::Hourly)
|
||||
.signature_user("contract_user_203_ghi789".to_string())
|
||||
.signature_hoster("hoster_node1003_rst456".to_string());
|
||||
|
||||
// Contract 4 - Multi-slice enterprise contract
|
||||
let contract4 = Contract::new()
|
||||
.customer_id(204)
|
||||
.add_compute_slice(compute_slice1.clone())
|
||||
.add_compute_slice(compute_slice2.clone())
|
||||
.add_storage_slice(storage_slice1.clone())
|
||||
.add_storage_slice(storage_slice2.clone())
|
||||
.compute_slice_price(0.045)
|
||||
.storage_slice_price(0.018)
|
||||
.network_slice_price(0.012)
|
||||
.status(ContractStatus::Active)
|
||||
.start_date(1640995200)
|
||||
.end_date(1735689600) // 2025-01-01
|
||||
.billing_period(BillingPeriod::Monthly)
|
||||
.signature_user("contract_user_204_jkl012".to_string())
|
||||
.signature_hoster("hoster_enterprise_mno345".to_string());
|
||||
|
||||
// Save all contracts to database and get their assigned IDs and updated models
|
||||
let (contract1_id, db_contract1) = db
|
||||
.collection()
|
||||
.expect("can open contract collection")
|
||||
.set(&contract1)
|
||||
.expect("can set contract");
|
||||
let (contract2_id, db_contract2) = db
|
||||
.collection()
|
||||
.expect("can open contract collection")
|
||||
.set(&contract2)
|
||||
.expect("can set contract");
|
||||
let (contract3_id, db_contract3) = db
|
||||
.collection()
|
||||
.expect("can open contract collection")
|
||||
.set(&contract3)
|
||||
.expect("can set contract");
|
||||
let (contract4_id, db_contract4) = db
|
||||
.collection()
|
||||
.expect("can open contract collection")
|
||||
.set(&contract4)
|
||||
.expect("can set contract");
|
||||
|
||||
println!("Contract 1 assigned ID: {}", contract1_id);
|
||||
println!("Contract 2 assigned ID: {}", contract2_id);
|
||||
println!("Contract 3 assigned ID: {}", contract3_id);
|
||||
println!("Contract 4 assigned ID: {}", contract4_id);
|
||||
|
||||
// Print all contracts retrieved from database
|
||||
println!("\n--- Contracts Retrieved from Database ---");
|
||||
println!("\n1. Web hosting contract:");
|
||||
print_contract_details(&db_contract1);
|
||||
|
||||
println!("\n2. Database hosting contract:");
|
||||
print_contract_details(&db_contract2);
|
||||
|
||||
println!("\n3. ML training contract (paused):");
|
||||
print_contract_details(&db_contract3);
|
||||
|
||||
println!("\n4. Enterprise multi-slice contract:");
|
||||
print_contract_details(&db_contract4);
|
||||
|
||||
// Demonstrate different ways to retrieve contracts from the database
|
||||
|
||||
// 1. Retrieve by customer ID index
|
||||
println!("\n--- Retrieving Contracts by Different Methods ---");
|
||||
println!("\n1. By Customer ID Index (Customer 202):");
|
||||
let customer_contracts = db
|
||||
.collection::<Contract>()
|
||||
.expect("can open contract collection")
|
||||
.get::<customer_id, _>(&202u32)
|
||||
.expect("can load contracts by customer");
|
||||
|
||||
assert_eq!(customer_contracts.len(), 1);
|
||||
print_contract_details(&customer_contracts[0]);
|
||||
|
||||
// 2. Update contract status
|
||||
println!("\n2. Resuming Paused Contract:");
|
||||
let mut updated_contract = db_contract3.clone();
|
||||
updated_contract.status = ContractStatus::Active;
|
||||
|
||||
let (_, resumed_contract) = db
|
||||
.collection::<Contract>()
|
||||
.expect("can open contract collection")
|
||||
.set(&updated_contract)
|
||||
.expect("can update contract");
|
||||
|
||||
println!("Updated contract status to Active:");
|
||||
print_contract_details(&resumed_contract);
|
||||
|
||||
// 3. Cancel a contract
|
||||
println!("\n3. Cancelling a Contract:");
|
||||
let mut cancelled_contract = db_contract1.clone();
|
||||
cancelled_contract.status = ContractStatus::Cancelled;
|
||||
|
||||
let (_, final_contract) = db
|
||||
.collection::<Contract>()
|
||||
.expect("can open contract collection")
|
||||
.set(&cancelled_contract)
|
||||
.expect("can update contract");
|
||||
|
||||
println!("Cancelled contract:");
|
||||
print_contract_details(&final_contract);
|
||||
|
||||
// Show remaining active contracts
|
||||
let all_contracts = db
|
||||
.collection::<Contract>()
|
||||
.expect("can open contract collection")
|
||||
.get_all()
|
||||
.expect("can load all contracts");
|
||||
|
||||
println!("\n--- Contract Analytics ---");
|
||||
let active_contracts: Vec<_> = all_contracts.iter()
|
||||
.filter(|c| matches!(c.status, ContractStatus::Active))
|
||||
.collect();
|
||||
let paused_contracts: Vec<_> = all_contracts.iter()
|
||||
.filter(|c| matches!(c.status, ContractStatus::Paused))
|
||||
.collect();
|
||||
let cancelled_contracts: Vec<_> = all_contracts.iter()
|
||||
.filter(|c| matches!(c.status, ContractStatus::Cancelled))
|
||||
.collect();
|
||||
|
||||
println!("Total Contracts: {}", all_contracts.len());
|
||||
println!("Active Contracts: {}", active_contracts.len());
|
||||
println!("Paused Contracts: {}", paused_contracts.len());
|
||||
println!("Cancelled Contracts: {}", cancelled_contracts.len());
|
||||
|
||||
// Calculate total provisioned resources
|
||||
let total_compute_slices: usize = all_contracts.iter().map(|c| c.compute_slices.len()).sum();
|
||||
let total_storage_slices: usize = all_contracts.iter().map(|c| c.storage_slices.len()).sum();
|
||||
let total_memory_gb: f64 = all_contracts.iter()
|
||||
.flat_map(|c| &c.compute_slices)
|
||||
.map(|s| s.mem_gb)
|
||||
.sum();
|
||||
let total_storage_gb: i32 = all_contracts.iter()
|
||||
.flat_map(|c| &c.storage_slices)
|
||||
.map(|s| s.storage_size_gb)
|
||||
.sum();
|
||||
|
||||
println!("\nProvisioned Resources:");
|
||||
println!(" Total Compute Slices: {}", total_compute_slices);
|
||||
println!(" Total Storage Slices: {}", total_storage_slices);
|
||||
println!(" Total Memory: {:.1} GB", total_memory_gb);
|
||||
println!(" Total Storage: {} GB", total_storage_gb);
|
||||
|
||||
// Calculate average pricing
|
||||
let avg_compute_price: f64 = all_contracts.iter().map(|c| c.compute_slice_price).sum::<f64>() / all_contracts.len() as f64;
|
||||
let avg_storage_price: f64 = all_contracts.iter().map(|c| c.storage_slice_price).sum::<f64>() / all_contracts.len() as f64;
|
||||
let avg_network_price: f64 = all_contracts.iter().map(|c| c.network_slice_price).sum::<f64>() / all_contracts.len() as f64;
|
||||
|
||||
println!("\nAverage Pricing:");
|
||||
println!(" Compute: ${:.3} per slice", avg_compute_price);
|
||||
println!(" Storage: ${:.3} per slice", avg_storage_price);
|
||||
println!(" Network: ${:.3} per slice", avg_network_price);
|
||||
|
||||
println!("\n--- Model Information ---");
|
||||
println!("Contract DB Prefix: {}", Contract::db_prefix());
|
||||
}
|
12
heromodels/examples/grid4_example/README.md
Normal file
12
heromodels/examples/grid4_example/README.md
Normal file
@@ -0,0 +1,12 @@
|
||||
# Grid4 Node Example (OurDB)
|
||||
|
||||
This example demonstrates how to use the Grid4 `Node` model against the embedded OurDB backend.
|
||||
|
||||
- Creates an in-memory/on-disk OurDB under `/tmp`.
|
||||
- Demonstrates CRUD and simple index lookups on `country`, `nodegroupid`, and `pubkey`.
|
||||
|
||||
Run it:
|
||||
|
||||
```bash
|
||||
cargo run -p heromodels --example grid4_example
|
||||
```
|
66
heromodels/examples/grid4_example/example.rs
Normal file
66
heromodels/examples/grid4_example/example.rs
Normal file
@@ -0,0 +1,66 @@
|
||||
use heromodels::db::hero::OurDB;
|
||||
use heromodels::db::{Collection, Db};
|
||||
use heromodels::models::grid4::node::node_index::{country, nodegroupid, pubkey};
|
||||
use heromodels::models::grid4::node::{ComputeSlice, DeviceInfo, Node};
|
||||
use std::sync::Arc;
|
||||
|
||||
fn main() {
|
||||
// Create a temp OurDB
|
||||
let ts = std::time::SystemTime::now()
|
||||
.duration_since(std::time::UNIX_EPOCH)
|
||||
.unwrap()
|
||||
.as_nanos();
|
||||
let path = format!("/tmp/grid4_example_{}", ts);
|
||||
let _ = std::fs::remove_dir_all(&path);
|
||||
let db = Arc::new(OurDB::new(&path, true).expect("create OurDB"));
|
||||
|
||||
let nodes = db.collection::<Node>().expect("open node collection");
|
||||
|
||||
// Build a node
|
||||
let cs = ComputeSlice::new()
|
||||
.nodeid(1)
|
||||
.slice_id(1)
|
||||
.mem_gb(64.0)
|
||||
.storage_gb(1024.0)
|
||||
.passmark(8000)
|
||||
.vcores(24)
|
||||
.gpus(2)
|
||||
.price_cc(0.5);
|
||||
|
||||
let dev = DeviceInfo {
|
||||
vendor: "ACME".into(),
|
||||
..Default::default()
|
||||
};
|
||||
|
||||
let n = Node::new()
|
||||
.nodegroupid(7)
|
||||
.uptime(98)
|
||||
.add_compute_slice(cs)
|
||||
.devices(dev)
|
||||
.country("BE")
|
||||
.pubkey("PUB_NODE_X")
|
||||
.build();
|
||||
|
||||
// Store
|
||||
let (id, stored) = nodes.set(&n).expect("store node");
|
||||
println!("Stored node id={id} pubkey={} country={}", stored.pubkey, stored.country);
|
||||
|
||||
// Query by indexes
|
||||
let by_country = nodes.get::<country, _>("BE").expect("query country");
|
||||
println!("Found {} nodes in country=BE", by_country.len());
|
||||
|
||||
let by_group = nodes.get::<nodegroupid, _>(&7).expect("query group");
|
||||
println!("Found {} nodes in group=7", by_group.len());
|
||||
|
||||
let by_key = nodes.get::<pubkey, _>("PUB_NODE_X").expect("query pubkey");
|
||||
println!("Found {} with pubkey PUB_NODE_X", by_key.len());
|
||||
|
||||
// Update
|
||||
let updated = stored.clone().country("NL");
|
||||
let (_, back) = nodes.set(&updated).expect("update node");
|
||||
println!("Updated node country={}", back.country);
|
||||
|
||||
// Delete
|
||||
nodes.delete_by_id(id).expect("delete node");
|
||||
println!("Deleted node id={id}");
|
||||
}
|
390
heromodels/examples/grid4_node_example.rs
Normal file
390
heromodels/examples/grid4_node_example.rs
Normal file
@@ -0,0 +1,390 @@
|
||||
use heromodels::db::{Collection, Db};
|
||||
use heromodels::models::grid4::{Node, NodeDevice, ComputeSlice, StorageSlice};
|
||||
use heromodels::models::grid4::node::node_index::{nodegroupid, country};
|
||||
use heromodels_core::Model;
|
||||
|
||||
// Helper function to print node details
|
||||
fn print_node_details(node: &Node) {
|
||||
println!("\n--- Node Details ---");
|
||||
println!("ID: {}", node.get_id());
|
||||
println!("NodeGroup ID: {}", node.nodegroupid);
|
||||
println!("Uptime: {}%", node.uptime);
|
||||
println!("Country: {}", node.country);
|
||||
println!("Birth Time: {}", node.birthtime);
|
||||
println!("Public Key: {}", node.pubkey);
|
||||
println!("Compute Slices: {}", node.computeslices.len());
|
||||
println!("Storage Slices: {}", node.storageslices.len());
|
||||
println!("Created At: {}", node.base_data.created_at);
|
||||
println!("Modified At: {}", node.base_data.modified_at);
|
||||
|
||||
// Print capacity details
|
||||
println!(" Capacity:");
|
||||
println!(" Storage: {:.1} GB", node.capacity.storage_gb);
|
||||
println!(" Memory: {:.1} GB", node.capacity.mem_gb);
|
||||
println!(" GPU Memory: {:.1} GB", node.capacity.mem_gb_gpu);
|
||||
println!(" Passmark: {}", node.capacity.passmark);
|
||||
println!(" vCores: {}", node.capacity.vcores);
|
||||
|
||||
// Print device info
|
||||
println!(" Devices:");
|
||||
println!(" Vendor: {}", node.devices.vendor);
|
||||
println!(" CPUs: {}", node.devices.cpu.len());
|
||||
println!(" GPUs: {}", node.devices.gpu.len());
|
||||
println!(" Memory: {}", node.devices.memory.len());
|
||||
println!(" Storage: {}", node.devices.storage.len());
|
||||
println!(" Network: {}", node.devices.network.len());
|
||||
|
||||
// Print compute slices
|
||||
if !node.computeslices.is_empty() {
|
||||
println!(" Compute Slices:");
|
||||
for (i, slice) in node.computeslices.iter().enumerate() {
|
||||
println!(" {}. ID: {}, Memory: {:.1}GB, Storage: {:.1}GB, vCores: {}, GPUs: {}",
|
||||
i + 1, slice.id, slice.mem_gb, slice.storage_gb, slice.vcores, slice.gpus);
|
||||
}
|
||||
}
|
||||
|
||||
// Print storage slices
|
||||
if !node.storageslices.is_empty() {
|
||||
println!(" Storage Slices:");
|
||||
for (i, slice) in node.storageslices.iter().enumerate() {
|
||||
println!(" {}. ID: {}", i + 1, slice.id);
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
fn main() {
|
||||
// Create a new DB instance in /tmp/grid4_nodes_db, and reset before every run
|
||||
let db = heromodels::db::hero::OurDB::new("/tmp/grid4_nodes_db", true).expect("Can create DB");
|
||||
|
||||
println!("Grid4 Node Models - Basic Usage Example");
|
||||
println!("======================================");
|
||||
|
||||
// Create device components for nodes
|
||||
|
||||
// CPU devices
|
||||
let cpu1 = CPUDevice {
|
||||
id: "cpu_intel_i7_12700k".to_string(),
|
||||
cores: 12,
|
||||
passmark: 28500,
|
||||
description: "Intel Core i7-12700K".to_string(),
|
||||
cpu_brand: "Intel".to_string(),
|
||||
cpu_version: "12th Gen".to_string(),
|
||||
};
|
||||
|
||||
let cpu2 = CPUDevice {
|
||||
id: "cpu_amd_ryzen_9_5900x".to_string(),
|
||||
cores: 12,
|
||||
passmark: 32000,
|
||||
description: "AMD Ryzen 9 5900X".to_string(),
|
||||
cpu_brand: "AMD".to_string(),
|
||||
cpu_version: "Zen 3".to_string(),
|
||||
};
|
||||
|
||||
// GPU devices
|
||||
let gpu1 = GPUDevice {
|
||||
id: "gpu_rtx_3080".to_string(),
|
||||
cores: 8704,
|
||||
memory_gb: 10.0,
|
||||
description: "NVIDIA GeForce RTX 3080".to_string(),
|
||||
gpu_brand: "NVIDIA".to_string(),
|
||||
gpu_version: "RTX 30 Series".to_string(),
|
||||
};
|
||||
|
||||
let gpu2 = GPUDevice {
|
||||
id: "gpu_rtx_4090".to_string(),
|
||||
cores: 16384,
|
||||
memory_gb: 24.0,
|
||||
description: "NVIDIA GeForce RTX 4090".to_string(),
|
||||
gpu_brand: "NVIDIA".to_string(),
|
||||
gpu_version: "RTX 40 Series".to_string(),
|
||||
};
|
||||
|
||||
// Memory devices
|
||||
let memory1 = MemoryDevice {
|
||||
id: "mem_ddr4_32gb".to_string(),
|
||||
size_gb: 32.0,
|
||||
description: "DDR4-3200 32GB Kit".to_string(),
|
||||
};
|
||||
|
||||
let memory2 = MemoryDevice {
|
||||
id: "mem_ddr5_64gb".to_string(),
|
||||
size_gb: 64.0,
|
||||
description: "DDR5-5600 64GB Kit".to_string(),
|
||||
};
|
||||
|
||||
// Storage devices
|
||||
let storage1 = StorageDevice {
|
||||
id: "ssd_nvme_1tb".to_string(),
|
||||
size_gb: 1000.0,
|
||||
description: "NVMe SSD 1TB".to_string(),
|
||||
};
|
||||
|
||||
let storage2 = StorageDevice {
|
||||
id: "hdd_sata_4tb".to_string(),
|
||||
size_gb: 4000.0,
|
||||
description: "SATA HDD 4TB".to_string(),
|
||||
};
|
||||
|
||||
// Network devices
|
||||
let network1 = NetworkDevice {
|
||||
id: "eth_1gbit".to_string(),
|
||||
speed_mbps: 1000,
|
||||
description: "Gigabit Ethernet".to_string(),
|
||||
};
|
||||
|
||||
let network2 = NetworkDevice {
|
||||
id: "eth_10gbit".to_string(),
|
||||
speed_mbps: 10000,
|
||||
description: "10 Gigabit Ethernet".to_string(),
|
||||
};
|
||||
|
||||
// Create device info configurations
|
||||
let devices1 = DeviceInfo {
|
||||
vendor: "Dell".to_string(),
|
||||
cpu: vec![cpu1.clone()],
|
||||
gpu: vec![gpu1.clone()],
|
||||
memory: vec![memory1.clone()],
|
||||
storage: vec![storage1.clone(), storage2.clone()],
|
||||
network: vec![network1.clone()],
|
||||
};
|
||||
|
||||
let devices2 = DeviceInfo {
|
||||
vendor: "HP".to_string(),
|
||||
cpu: vec![cpu2.clone()],
|
||||
gpu: vec![gpu2.clone()],
|
||||
memory: vec![memory2.clone()],
|
||||
storage: vec![storage1.clone()],
|
||||
network: vec![network2.clone()],
|
||||
};
|
||||
|
||||
// Create node capacities
|
||||
let capacity1 = NodeCapacity {
|
||||
storage_gb: 5000.0,
|
||||
mem_gb: 32.0,
|
||||
mem_gb_gpu: 10.0,
|
||||
passmark: 28500,
|
||||
vcores: 24,
|
||||
};
|
||||
|
||||
let capacity2 = NodeCapacity {
|
||||
storage_gb: 1000.0,
|
||||
mem_gb: 64.0,
|
||||
mem_gb_gpu: 24.0,
|
||||
passmark: 32000,
|
||||
vcores: 24,
|
||||
};
|
||||
|
||||
// Create compute slices
|
||||
let compute_slice1 = ComputeSlice::new()
|
||||
.id(1)
|
||||
.mem_gb(4.0)
|
||||
.storage_gb(100.0)
|
||||
.passmark(3000)
|
||||
.vcores(2)
|
||||
.cpu_oversubscription(150)
|
||||
.storage_oversubscription(120)
|
||||
.gpus(0);
|
||||
|
||||
let compute_slice2 = ComputeSlice::new()
|
||||
.id(2)
|
||||
.mem_gb(8.0)
|
||||
.storage_gb(200.0)
|
||||
.passmark(6000)
|
||||
.vcores(4)
|
||||
.cpu_oversubscription(130)
|
||||
.storage_oversubscription(110)
|
||||
.gpus(1);
|
||||
|
||||
let compute_slice3 = ComputeSlice::new()
|
||||
.id(1)
|
||||
.mem_gb(16.0)
|
||||
.storage_gb(400.0)
|
||||
.passmark(12000)
|
||||
.vcores(8)
|
||||
.cpu_oversubscription(110)
|
||||
.storage_oversubscription(100)
|
||||
.gpus(1);
|
||||
|
||||
// Create storage slices
|
||||
let storage_slice1 = StorageSlice::new().id(1);
|
||||
let storage_slice2 = StorageSlice::new().id(2);
|
||||
let storage_slice3 = StorageSlice::new().id(3);
|
||||
|
||||
// Create nodes with different configurations
|
||||
|
||||
// Node 1 - Web hosting node
|
||||
let node1 = Node::new()
|
||||
.nodegroupid(1001)
|
||||
.uptime(98)
|
||||
.add_compute_slice(compute_slice1.clone())
|
||||
.add_compute_slice(compute_slice2.clone())
|
||||
.add_storage_slice(storage_slice1.clone())
|
||||
.add_storage_slice(storage_slice2.clone())
|
||||
.devices(devices1.clone())
|
||||
.country("US".to_string())
|
||||
.capacity(capacity1.clone())
|
||||
.birthtime(1640995200) // 2022-01-01
|
||||
.pubkey("node1_pubkey_abc123xyz789".to_string())
|
||||
.signature_node("node1_signature_def456".to_string())
|
||||
.signature_farmer("farmer1_signature_ghi789".to_string());
|
||||
|
||||
// Node 2 - High-performance computing node
|
||||
let node2 = Node::new()
|
||||
.nodegroupid(1002)
|
||||
.uptime(99)
|
||||
.add_compute_slice(compute_slice3.clone())
|
||||
.add_storage_slice(storage_slice3.clone())
|
||||
.devices(devices2.clone())
|
||||
.country("DE".to_string())
|
||||
.capacity(capacity2.clone())
|
||||
.birthtime(1672531200) // 2023-01-01
|
||||
.pubkey("node2_pubkey_jkl012mno345".to_string())
|
||||
.signature_node("node2_signature_pqr678".to_string())
|
||||
.signature_farmer("farmer2_signature_stu901".to_string());
|
||||
|
||||
// Node 3 - Storage-focused node
|
||||
let node3 = Node::new()
|
||||
.nodegroupid(1001)
|
||||
.uptime(95)
|
||||
.add_storage_slice(storage_slice1.clone())
|
||||
.add_storage_slice(storage_slice2.clone())
|
||||
.add_storage_slice(storage_slice3.clone())
|
||||
.devices(devices1.clone())
|
||||
.country("NL".to_string())
|
||||
.capacity(capacity1.clone())
|
||||
.birthtime(1704067200) // 2024-01-01
|
||||
.pubkey("node3_pubkey_vwx234yzab567".to_string())
|
||||
.signature_node("node3_signature_cde890".to_string())
|
||||
.signature_farmer("farmer1_signature_fgh123".to_string());
|
||||
|
||||
// Save all nodes to database and get their assigned IDs and updated models
|
||||
let (node1_id, db_node1) = db
|
||||
.collection()
|
||||
.expect("can open node collection")
|
||||
.set(&node1)
|
||||
.expect("can set node");
|
||||
let (node2_id, db_node2) = db
|
||||
.collection()
|
||||
.expect("can open node collection")
|
||||
.set(&node2)
|
||||
.expect("can set node");
|
||||
let (node3_id, db_node3) = db
|
||||
.collection()
|
||||
.expect("can open node collection")
|
||||
.set(&node3)
|
||||
.expect("can set node");
|
||||
|
||||
println!("Node 1 assigned ID: {}", node1_id);
|
||||
println!("Node 2 assigned ID: {}", node2_id);
|
||||
println!("Node 3 assigned ID: {}", node3_id);
|
||||
|
||||
// Print all nodes retrieved from database
|
||||
println!("\n--- Nodes Retrieved from Database ---");
|
||||
println!("\n1. Web hosting node:");
|
||||
print_node_details(&db_node1);
|
||||
|
||||
println!("\n2. High-performance computing node:");
|
||||
print_node_details(&db_node2);
|
||||
|
||||
println!("\n3. Storage-focused node:");
|
||||
print_node_details(&db_node3);
|
||||
|
||||
// Demonstrate different ways to retrieve nodes from the database
|
||||
|
||||
// 1. Retrieve by nodegroup ID index
|
||||
println!("\n--- Retrieving Nodes by Different Methods ---");
|
||||
println!("\n1. By NodeGroup ID Index (NodeGroup 1001):");
|
||||
let nodegroup_nodes = db
|
||||
.collection::<Node>()
|
||||
.expect("can open node collection")
|
||||
.get::<nodegroupid, _>(&1001i32)
|
||||
.expect("can load nodes by nodegroup");
|
||||
|
||||
assert_eq!(nodegroup_nodes.len(), 2);
|
||||
for (i, node) in nodegroup_nodes.iter().enumerate() {
|
||||
println!(" Node {}: ID {}, Country: {}, Uptime: {}%",
|
||||
i + 1, node.get_id(), node.country, node.uptime);
|
||||
}
|
||||
|
||||
// 2. Retrieve by country index
|
||||
println!("\n2. By Country Index (Germany - DE):");
|
||||
let country_nodes = db
|
||||
.collection::<Node>()
|
||||
.expect("can open node collection")
|
||||
.get::<country, _>("DE")
|
||||
.expect("can load nodes by country");
|
||||
|
||||
assert_eq!(country_nodes.len(), 1);
|
||||
print_node_details(&country_nodes[0]);
|
||||
|
||||
// 3. Update node uptime
|
||||
println!("\n3. Updating Node Uptime:");
|
||||
let mut updated_node = db_node1.clone();
|
||||
updated_node.uptime = 99;
|
||||
|
||||
let (_, uptime_updated_node) = db
|
||||
.collection::<Node>()
|
||||
.expect("can open node collection")
|
||||
.set(&updated_node)
|
||||
.expect("can update node");
|
||||
|
||||
println!("Updated node uptime to 99%:");
|
||||
println!(" Node ID: {}, New Uptime: {}%", uptime_updated_node.get_id(), uptime_updated_node.uptime);
|
||||
|
||||
// Show all nodes and calculate analytics
|
||||
let all_nodes = db
|
||||
.collection::<Node>()
|
||||
.expect("can open node collection")
|
||||
.get_all()
|
||||
.expect("can load all nodes");
|
||||
|
||||
println!("\n--- Node Analytics ---");
|
||||
println!("Total Nodes: {}", all_nodes.len());
|
||||
|
||||
// Calculate total capacity
|
||||
let total_storage_gb: f64 = all_nodes.iter().map(|n| n.capacity.storage_gb).sum();
|
||||
let total_memory_gb: f64 = all_nodes.iter().map(|n| n.capacity.mem_gb).sum();
|
||||
let total_gpu_memory_gb: f64 = all_nodes.iter().map(|n| n.capacity.mem_gb_gpu).sum();
|
||||
let total_vcores: i32 = all_nodes.iter().map(|n| n.capacity.vcores).sum();
|
||||
let avg_uptime: f64 = all_nodes.iter().map(|n| n.uptime as f64).sum::<f64>() / all_nodes.len() as f64;
|
||||
|
||||
println!("Total Capacity:");
|
||||
println!(" Storage: {:.1} GB", total_storage_gb);
|
||||
println!(" Memory: {:.1} GB", total_memory_gb);
|
||||
println!(" GPU Memory: {:.1} GB", total_gpu_memory_gb);
|
||||
println!(" vCores: {}", total_vcores);
|
||||
println!(" Average Uptime: {:.1}%", avg_uptime);
|
||||
|
||||
// Count nodes by country
|
||||
let mut country_counts = std::collections::HashMap::new();
|
||||
for node in &all_nodes {
|
||||
*country_counts.entry(&node.country).or_insert(0) += 1;
|
||||
}
|
||||
|
||||
println!("\nNodes by Country:");
|
||||
for (country, count) in country_counts {
|
||||
println!(" {}: {}", country, count);
|
||||
}
|
||||
|
||||
// Count total slices
|
||||
let total_compute_slices: usize = all_nodes.iter().map(|n| n.computeslices.len()).sum();
|
||||
let total_storage_slices: usize = all_nodes.iter().map(|n| n.storageslices.len()).sum();
|
||||
|
||||
println!("\nTotal Slices:");
|
||||
println!(" Compute Slices: {}", total_compute_slices);
|
||||
println!(" Storage Slices: {}", total_storage_slices);
|
||||
|
||||
// Vendor distribution
|
||||
let mut vendor_counts = std::collections::HashMap::new();
|
||||
for node in &all_nodes {
|
||||
*vendor_counts.entry(&node.devices.vendor).or_insert(0) += 1;
|
||||
}
|
||||
|
||||
println!("\nNodes by Vendor:");
|
||||
for (vendor, count) in vendor_counts {
|
||||
println!(" {}: {}", vendor, count);
|
||||
}
|
||||
|
||||
println!("\n--- Model Information ---");
|
||||
println!("Node DB Prefix: {}", Node::db_prefix());
|
||||
}
|
284
heromodels/examples/grid4_nodegroup_example.rs
Normal file
284
heromodels/examples/grid4_nodegroup_example.rs
Normal file
@@ -0,0 +1,284 @@
|
||||
use heromodels::db::{Collection, Db};
|
||||
use heromodels::models::grid4::{NodeGroup, PricingPolicy, SLAPolicy};
|
||||
use heromodels_core::Model;
|
||||
|
||||
// Helper function to print nodegroup details
|
||||
fn print_nodegroup_details(nodegroup: &NodeGroup) {
|
||||
println!("\n--- NodeGroup Details ---");
|
||||
println!("ID: {}", nodegroup.get_id());
|
||||
println!("Farmer ID: {}", nodegroup.farmerid);
|
||||
println!("Description: {}", nodegroup.description);
|
||||
println!("Secret: {}", nodegroup.secret);
|
||||
println!("Compute Slice Pricing (CC): {:.4}", nodegroup.compute_slice_normalized_pricing_cc);
|
||||
println!("Storage Slice Pricing (CC): {:.4}", nodegroup.storage_slice_normalized_pricing_cc);
|
||||
println!("Signature Farmer: {}", nodegroup.signature_farmer);
|
||||
println!("Created At: {}", nodegroup.base_data.created_at);
|
||||
println!("Modified At: {}", nodegroup.base_data.modified_at);
|
||||
|
||||
// Print SLA Policy details
|
||||
println!(" SLA Policy:");
|
||||
println!(" Uptime: {}%", nodegroup.slapolicy.sla_uptime);
|
||||
println!(" Bandwidth: {} Mbit/s", nodegroup.slapolicy.sla_bandwidth_mbit);
|
||||
println!(" Penalty: {}%", nodegroup.slapolicy.sla_penalty);
|
||||
|
||||
// Print Pricing Policy details
|
||||
println!(" Pricing Policy:");
|
||||
println!(" Marketplace Year Discounts: {:?}%", nodegroup.pricingpolicy.marketplace_year_discounts);
|
||||
}
|
||||
|
||||
fn main() {
|
||||
// Create a new DB instance in /tmp/grid4_nodegroups_db, and reset before every run
|
||||
let db = heromodels::db::hero::OurDB::new("/tmp/grid4_nodegroups_db", true).expect("Can create DB");
|
||||
|
||||
println!("Grid4 NodeGroup Models - Basic Usage Example");
|
||||
println!("===========================================");
|
||||
|
||||
// Create SLA policies
|
||||
let sla_policy_premium = SLAPolicy {
|
||||
sla_uptime: 99,
|
||||
sla_bandwidth_mbit: 1000,
|
||||
sla_penalty: 200,
|
||||
};
|
||||
|
||||
let sla_policy_standard = SLAPolicy {
|
||||
sla_uptime: 95,
|
||||
sla_bandwidth_mbit: 100,
|
||||
sla_penalty: 100,
|
||||
};
|
||||
|
||||
let sla_policy_basic = SLAPolicy {
|
||||
sla_uptime: 90,
|
||||
sla_bandwidth_mbit: 50,
|
||||
sla_penalty: 50,
|
||||
};
|
||||
|
||||
// Create pricing policies
|
||||
let pricing_policy_aggressive = PricingPolicy {
|
||||
marketplace_year_discounts: vec![40, 50, 60],
|
||||
};
|
||||
|
||||
let pricing_policy_standard = PricingPolicy {
|
||||
marketplace_year_discounts: vec![30, 40, 50],
|
||||
};
|
||||
|
||||
let pricing_policy_conservative = PricingPolicy {
|
||||
marketplace_year_discounts: vec![20, 30, 40],
|
||||
};
|
||||
|
||||
// Create nodegroups with different configurations
|
||||
|
||||
// NodeGroup 1 - Premium hosting provider
|
||||
let nodegroup1 = NodeGroup::new()
|
||||
.farmerid(501)
|
||||
.secret("encrypted_boot_secret_premium_abc123".to_string())
|
||||
.description("Premium hosting with 99% uptime SLA and high-speed connectivity".to_string())
|
||||
.slapolicy(sla_policy_premium.clone())
|
||||
.pricingpolicy(pricing_policy_aggressive.clone())
|
||||
.compute_slice_normalized_pricing_cc(0.0450)
|
||||
.storage_slice_normalized_pricing_cc(0.0180)
|
||||
.signature_farmer("farmer_501_premium_signature_xyz789".to_string());
|
||||
|
||||
// NodeGroup 2 - Standard business provider
|
||||
let nodegroup2 = NodeGroup::new()
|
||||
.farmerid(502)
|
||||
.secret("encrypted_boot_secret_standard_def456".to_string())
|
||||
.description("Standard business hosting with reliable performance".to_string())
|
||||
.slapolicy(sla_policy_standard.clone())
|
||||
.pricingpolicy(pricing_policy_standard.clone())
|
||||
.compute_slice_normalized_pricing_cc(0.0350)
|
||||
.storage_slice_normalized_pricing_cc(0.0150)
|
||||
.signature_farmer("farmer_502_standard_signature_uvw012".to_string());
|
||||
|
||||
// NodeGroup 3 - Budget-friendly provider
|
||||
let nodegroup3 = NodeGroup::new()
|
||||
.farmerid(503)
|
||||
.secret("encrypted_boot_secret_budget_ghi789".to_string())
|
||||
.description("Cost-effective hosting for development and testing".to_string())
|
||||
.slapolicy(sla_policy_basic.clone())
|
||||
.pricingpolicy(pricing_policy_conservative.clone())
|
||||
.compute_slice_normalized_pricing_cc(0.0250)
|
||||
.storage_slice_normalized_pricing_cc(0.0120)
|
||||
.signature_farmer("farmer_503_budget_signature_rst345".to_string());
|
||||
|
||||
// NodeGroup 4 - Enterprise provider
|
||||
let nodegroup4 = NodeGroup::new()
|
||||
.farmerid(504)
|
||||
.secret("encrypted_boot_secret_enterprise_jkl012".to_string())
|
||||
.description("Enterprise-grade infrastructure with maximum reliability".to_string())
|
||||
.slapolicy(sla_policy_premium.clone())
|
||||
.pricingpolicy(pricing_policy_standard.clone())
|
||||
.compute_slice_normalized_pricing_cc(0.0500)
|
||||
.storage_slice_normalized_pricing_cc(0.0200)
|
||||
.signature_farmer("farmer_504_enterprise_signature_mno678".to_string());
|
||||
|
||||
// Save all nodegroups to database and get their assigned IDs and updated models
|
||||
let (nodegroup1_id, db_nodegroup1) = db
|
||||
.collection()
|
||||
.expect("can open nodegroup collection")
|
||||
.set(&nodegroup1)
|
||||
.expect("can set nodegroup");
|
||||
let (nodegroup2_id, db_nodegroup2) = db
|
||||
.collection()
|
||||
.expect("can open nodegroup collection")
|
||||
.set(&nodegroup2)
|
||||
.expect("can set nodegroup");
|
||||
let (nodegroup3_id, db_nodegroup3) = db
|
||||
.collection()
|
||||
.expect("can open nodegroup collection")
|
||||
.set(&nodegroup3)
|
||||
.expect("can set nodegroup");
|
||||
let (nodegroup4_id, db_nodegroup4) = db
|
||||
.collection()
|
||||
.expect("can open nodegroup collection")
|
||||
.set(&nodegroup4)
|
||||
.expect("can set nodegroup");
|
||||
|
||||
println!("NodeGroup 1 assigned ID: {}", nodegroup1_id);
|
||||
println!("NodeGroup 2 assigned ID: {}", nodegroup2_id);
|
||||
println!("NodeGroup 3 assigned ID: {}", nodegroup3_id);
|
||||
println!("NodeGroup 4 assigned ID: {}", nodegroup4_id);
|
||||
|
||||
// Print all nodegroups retrieved from database
|
||||
println!("\n--- NodeGroups Retrieved from Database ---");
|
||||
println!("\n1. Premium hosting provider:");
|
||||
print_nodegroup_details(&db_nodegroup1);
|
||||
|
||||
println!("\n2. Standard business provider:");
|
||||
print_nodegroup_details(&db_nodegroup2);
|
||||
|
||||
println!("\n3. Budget-friendly provider:");
|
||||
print_nodegroup_details(&db_nodegroup3);
|
||||
|
||||
println!("\n4. Enterprise provider:");
|
||||
print_nodegroup_details(&db_nodegroup4);
|
||||
|
||||
// Demonstrate different ways to retrieve nodegroups from the database
|
||||
|
||||
// 1. Retrieve by farmer ID index
|
||||
println!("\n--- Retrieving NodeGroups by Different Methods ---");
|
||||
println!("\n1. By Farmer ID Index (Farmer 502):");
|
||||
let farmer_nodegroups = db
|
||||
.collection::<NodeGroup>()
|
||||
.expect("can open nodegroup collection")
|
||||
.get_by_index("farmerid", &502u32)
|
||||
.expect("can load nodegroups by farmer");
|
||||
|
||||
assert_eq!(farmer_nodegroups.len(), 1);
|
||||
print_nodegroup_details(&farmer_nodegroups[0]);
|
||||
|
||||
// 2. Update nodegroup pricing
|
||||
println!("\n2. Updating NodeGroup Pricing:");
|
||||
let mut updated_nodegroup = db_nodegroup3.clone();
|
||||
updated_nodegroup.compute_slice_normalized_pricing_cc = 0.0280;
|
||||
updated_nodegroup.storage_slice_normalized_pricing_cc = 0.0130;
|
||||
|
||||
let (_, price_updated_nodegroup) = db
|
||||
.collection::<NodeGroup>()
|
||||
.expect("can open nodegroup collection")
|
||||
.set(&updated_nodegroup)
|
||||
.expect("can update nodegroup");
|
||||
|
||||
println!("Updated pricing for budget provider:");
|
||||
println!(" Compute: {:.4} CC", price_updated_nodegroup.compute_slice_normalized_pricing_cc);
|
||||
println!(" Storage: {:.4} CC", price_updated_nodegroup.storage_slice_normalized_pricing_cc);
|
||||
|
||||
// 3. Update SLA policy
|
||||
println!("\n3. Updating SLA Policy:");
|
||||
let mut sla_updated_nodegroup = db_nodegroup2.clone();
|
||||
sla_updated_nodegroup.slapolicy.sla_uptime = 98;
|
||||
sla_updated_nodegroup.slapolicy.sla_bandwidth_mbit = 500;
|
||||
|
||||
let (_, sla_updated_nodegroup) = db
|
||||
.collection::<NodeGroup>()
|
||||
.expect("can open nodegroup collection")
|
||||
.set(&sla_updated_nodegroup)
|
||||
.expect("can update nodegroup");
|
||||
|
||||
println!("Updated SLA policy for standard provider:");
|
||||
println!(" Uptime: {}%", sla_updated_nodegroup.slapolicy.sla_uptime);
|
||||
println!(" Bandwidth: {} Mbit/s", sla_updated_nodegroup.slapolicy.sla_bandwidth_mbit);
|
||||
|
||||
// Show all nodegroups and calculate analytics
|
||||
let all_nodegroups = db
|
||||
.collection::<NodeGroup>()
|
||||
.expect("can open nodegroup collection")
|
||||
.get_all()
|
||||
.expect("can load all nodegroups");
|
||||
|
||||
println!("\n--- NodeGroup Analytics ---");
|
||||
println!("Total NodeGroups: {}", all_nodegroups.len());
|
||||
|
||||
// Calculate pricing statistics
|
||||
let avg_compute_price: f64 = all_nodegroups.iter()
|
||||
.map(|ng| ng.compute_slice_normalized_pricing_cc)
|
||||
.sum::<f64>() / all_nodegroups.len() as f64;
|
||||
let avg_storage_price: f64 = all_nodegroups.iter()
|
||||
.map(|ng| ng.storage_slice_normalized_pricing_cc)
|
||||
.sum::<f64>() / all_nodegroups.len() as f64;
|
||||
|
||||
let min_compute_price = all_nodegroups.iter()
|
||||
.map(|ng| ng.compute_slice_normalized_pricing_cc)
|
||||
.fold(f64::INFINITY, f64::min);
|
||||
let max_compute_price = all_nodegroups.iter()
|
||||
.map(|ng| ng.compute_slice_normalized_pricing_cc)
|
||||
.fold(f64::NEG_INFINITY, f64::max);
|
||||
|
||||
println!("Pricing Statistics:");
|
||||
println!(" Average Compute Price: {:.4} CC", avg_compute_price);
|
||||
println!(" Average Storage Price: {:.4} CC", avg_storage_price);
|
||||
println!(" Compute Price Range: {:.4} - {:.4} CC", min_compute_price, max_compute_price);
|
||||
|
||||
// Calculate SLA statistics
|
||||
let avg_uptime: f64 = all_nodegroups.iter()
|
||||
.map(|ng| ng.slapolicy.sla_uptime as f64)
|
||||
.sum::<f64>() / all_nodegroups.len() as f64;
|
||||
let avg_bandwidth: f64 = all_nodegroups.iter()
|
||||
.map(|ng| ng.slapolicy.sla_bandwidth_mbit as f64)
|
||||
.sum::<f64>() / all_nodegroups.len() as f64;
|
||||
let avg_penalty: f64 = all_nodegroups.iter()
|
||||
.map(|ng| ng.slapolicy.sla_penalty as f64)
|
||||
.sum::<f64>() / all_nodegroups.len() as f64;
|
||||
|
||||
println!("\nSLA Statistics:");
|
||||
println!(" Average Uptime Guarantee: {:.1}%", avg_uptime);
|
||||
println!(" Average Bandwidth Guarantee: {:.0} Mbit/s", avg_bandwidth);
|
||||
println!(" Average Penalty Rate: {:.0}%", avg_penalty);
|
||||
|
||||
// Count farmers
|
||||
let unique_farmers: std::collections::HashSet<_> = all_nodegroups.iter()
|
||||
.map(|ng| ng.farmerid)
|
||||
.collect();
|
||||
|
||||
println!("\nFarmer Statistics:");
|
||||
println!(" Unique Farmers: {}", unique_farmers.len());
|
||||
println!(" NodeGroups per Farmer: {:.1}", all_nodegroups.len() as f64 / unique_farmers.len() as f64);
|
||||
|
||||
// Analyze discount policies
|
||||
let total_discount_tiers: usize = all_nodegroups.iter()
|
||||
.map(|ng| ng.pricingpolicy.marketplace_year_discounts.len())
|
||||
.sum();
|
||||
let avg_discount_tiers: f64 = total_discount_tiers as f64 / all_nodegroups.len() as f64;
|
||||
|
||||
println!("\nDiscount Policy Statistics:");
|
||||
println!(" Average Discount Tiers: {:.1}", avg_discount_tiers);
|
||||
|
||||
// Find best value providers (high SLA, low price)
|
||||
println!("\n--- Provider Rankings ---");
|
||||
let mut providers_with_scores: Vec<_> = all_nodegroups.iter()
|
||||
.map(|ng| {
|
||||
let value_score = (ng.slapolicy.sla_uptime as f64) / ng.compute_slice_normalized_pricing_cc;
|
||||
(ng, value_score)
|
||||
})
|
||||
.collect();
|
||||
|
||||
providers_with_scores.sort_by(|a, b| b.1.partial_cmp(&a.1).unwrap());
|
||||
|
||||
println!("Best Value Providers (Uptime/Price ratio):");
|
||||
for (i, (ng, score)) in providers_with_scores.iter().enumerate() {
|
||||
println!(" {}. Farmer {}: {:.0} ({}% uptime, {:.4} CC)",
|
||||
i + 1, ng.farmerid, score, ng.slapolicy.sla_uptime, ng.compute_slice_normalized_pricing_cc);
|
||||
}
|
||||
|
||||
println!("\n--- Model Information ---");
|
||||
println!("NodeGroup DB Prefix: {}", NodeGroup::db_prefix());
|
||||
}
|
311
heromodels/examples/grid4_reputation_example.rs
Normal file
311
heromodels/examples/grid4_reputation_example.rs
Normal file
@@ -0,0 +1,311 @@
|
||||
use heromodels::db::{Collection, Db};
|
||||
use heromodels::models::grid4::{NodeGroupReputation, NodeReputation};
|
||||
use heromodels_core::Model;
|
||||
|
||||
// Helper function to print nodegroup reputation details
|
||||
fn print_nodegroup_reputation_details(reputation: &NodeGroupReputation) {
|
||||
println!("\n--- NodeGroup Reputation Details ---");
|
||||
println!("ID: {}", reputation.get_id());
|
||||
println!("NodeGroup ID: {}", reputation.nodegroup_id);
|
||||
println!("Reputation Score: {}/100", reputation.reputation);
|
||||
println!("Uptime: {}%", reputation.uptime);
|
||||
println!("Node Count: {}", reputation.nodes.len());
|
||||
println!("Created At: {}", reputation.base_data.created_at);
|
||||
println!("Modified At: {}", reputation.base_data.modified_at);
|
||||
|
||||
// Print individual node reputations
|
||||
if !reputation.nodes.is_empty() {
|
||||
println!(" Individual Node Reputations:");
|
||||
for (i, node_rep) in reputation.nodes.iter().enumerate() {
|
||||
println!(" {}. Node {}: Reputation {}/100, Uptime {}%",
|
||||
i + 1, node_rep.node_id, node_rep.reputation, node_rep.uptime);
|
||||
}
|
||||
|
||||
// Calculate average node reputation and uptime
|
||||
let avg_node_reputation: f64 = reputation.nodes.iter()
|
||||
.map(|n| n.reputation as f64)
|
||||
.sum::<f64>() / reputation.nodes.len() as f64;
|
||||
let avg_node_uptime: f64 = reputation.nodes.iter()
|
||||
.map(|n| n.uptime as f64)
|
||||
.sum::<f64>() / reputation.nodes.len() as f64;
|
||||
|
||||
println!(" Average Node Reputation: {:.1}/100", avg_node_reputation);
|
||||
println!(" Average Node Uptime: {:.1}%", avg_node_uptime);
|
||||
}
|
||||
}
|
||||
|
||||
fn main() {
|
||||
// Create a new DB instance in /tmp/grid4_reputation_db, and reset before every run
|
||||
let db = heromodels::db::hero::OurDB::new("/tmp/grid4_reputation_db", true).expect("Can create DB");
|
||||
|
||||
println!("Grid4 Reputation Models - Basic Usage Example");
|
||||
println!("============================================");
|
||||
|
||||
// Create individual node reputations
|
||||
|
||||
// High-performing nodes
|
||||
let node_rep1 = NodeReputation::new()
|
||||
.node_id(1001)
|
||||
.reputation(85)
|
||||
.uptime(99);
|
||||
|
||||
let node_rep2 = NodeReputation::new()
|
||||
.node_id(1002)
|
||||
.reputation(92)
|
||||
.uptime(98);
|
||||
|
||||
let node_rep3 = NodeReputation::new()
|
||||
.node_id(1003)
|
||||
.reputation(78)
|
||||
.uptime(97);
|
||||
|
||||
// Medium-performing nodes
|
||||
let node_rep4 = NodeReputation::new()
|
||||
.node_id(2001)
|
||||
.reputation(65)
|
||||
.uptime(94);
|
||||
|
||||
let node_rep5 = NodeReputation::new()
|
||||
.node_id(2002)
|
||||
.reputation(72)
|
||||
.uptime(96);
|
||||
|
||||
// Lower-performing nodes
|
||||
let node_rep6 = NodeReputation::new()
|
||||
.node_id(3001)
|
||||
.reputation(45)
|
||||
.uptime(88);
|
||||
|
||||
let node_rep7 = NodeReputation::new()
|
||||
.node_id(3002)
|
||||
.reputation(38)
|
||||
.uptime(85);
|
||||
|
||||
// New nodes with default reputation
|
||||
let node_rep8 = NodeReputation::new()
|
||||
.node_id(4001)
|
||||
.reputation(50) // default
|
||||
.uptime(0); // just started
|
||||
|
||||
let node_rep9 = NodeReputation::new()
|
||||
.node_id(4002)
|
||||
.reputation(50) // default
|
||||
.uptime(0); // just started
|
||||
|
||||
// Create nodegroup reputations with different performance profiles
|
||||
|
||||
// NodeGroup 1 - High-performance provider
|
||||
let nodegroup_rep1 = NodeGroupReputation::new()
|
||||
.nodegroup_id(1001)
|
||||
.reputation(85) // high reputation earned over time
|
||||
.uptime(98) // excellent uptime
|
||||
.add_node_reputation(node_rep1.clone())
|
||||
.add_node_reputation(node_rep2.clone())
|
||||
.add_node_reputation(node_rep3.clone());
|
||||
|
||||
// NodeGroup 2 - Medium-performance provider
|
||||
let nodegroup_rep2 = NodeGroupReputation::new()
|
||||
.nodegroup_id(1002)
|
||||
.reputation(68) // decent reputation
|
||||
.uptime(95) // good uptime
|
||||
.add_node_reputation(node_rep4.clone())
|
||||
.add_node_reputation(node_rep5.clone());
|
||||
|
||||
// NodeGroup 3 - Struggling provider
|
||||
let nodegroup_rep3 = NodeGroupReputation::new()
|
||||
.nodegroup_id(1003)
|
||||
.reputation(42) // below average reputation
|
||||
.uptime(87) // poor uptime
|
||||
.add_node_reputation(node_rep6.clone())
|
||||
.add_node_reputation(node_rep7.clone());
|
||||
|
||||
// NodeGroup 4 - New provider (default reputation)
|
||||
let nodegroup_rep4 = NodeGroupReputation::new()
|
||||
.nodegroup_id(1004)
|
||||
.reputation(50) // default starting reputation
|
||||
.uptime(0) // no history yet
|
||||
.add_node_reputation(node_rep8.clone())
|
||||
.add_node_reputation(node_rep9.clone());
|
||||
|
||||
// Save all nodegroup reputations to database and get their assigned IDs and updated models
|
||||
let (rep1_id, db_rep1) = db
|
||||
.collection()
|
||||
.expect("can open reputation collection")
|
||||
.set(&nodegroup_rep1)
|
||||
.expect("can set reputation");
|
||||
let (rep2_id, db_rep2) = db
|
||||
.collection()
|
||||
.expect("can open reputation collection")
|
||||
.set(&nodegroup_rep2)
|
||||
.expect("can set reputation");
|
||||
let (rep3_id, db_rep3) = db
|
||||
.collection()
|
||||
.expect("can open reputation collection")
|
||||
.set(&nodegroup_rep3)
|
||||
.expect("can set reputation");
|
||||
let (rep4_id, db_rep4) = db
|
||||
.collection()
|
||||
.expect("can open reputation collection")
|
||||
.set(&nodegroup_rep4)
|
||||
.expect("can set reputation");
|
||||
|
||||
println!("NodeGroup Reputation 1 assigned ID: {}", rep1_id);
|
||||
println!("NodeGroup Reputation 2 assigned ID: {}", rep2_id);
|
||||
println!("NodeGroup Reputation 3 assigned ID: {}", rep3_id);
|
||||
println!("NodeGroup Reputation 4 assigned ID: {}", rep4_id);
|
||||
|
||||
// Print all reputation records retrieved from database
|
||||
println!("\n--- Reputation Records Retrieved from Database ---");
|
||||
println!("\n1. High-performance provider:");
|
||||
print_nodegroup_reputation_details(&db_rep1);
|
||||
|
||||
println!("\n2. Medium-performance provider:");
|
||||
print_nodegroup_reputation_details(&db_rep2);
|
||||
|
||||
println!("\n3. Struggling provider:");
|
||||
print_nodegroup_reputation_details(&db_rep3);
|
||||
|
||||
println!("\n4. New provider:");
|
||||
print_nodegroup_reputation_details(&db_rep4);
|
||||
|
||||
// Demonstrate different ways to retrieve reputation records from the database
|
||||
|
||||
// 1. Retrieve by nodegroup ID index
|
||||
println!("\n--- Retrieving Reputation by Different Methods ---");
|
||||
println!("\n1. By NodeGroup ID Index (NodeGroup 1002):");
|
||||
let nodegroup_reps = db
|
||||
.collection::<NodeGroupReputation>()
|
||||
.expect("can open reputation collection")
|
||||
.get_by_index("nodegroup_id", &1002u32)
|
||||
.expect("can load reputation by nodegroup");
|
||||
|
||||
assert_eq!(nodegroup_reps.len(), 1);
|
||||
print_nodegroup_reputation_details(&nodegroup_reps[0]);
|
||||
|
||||
// 2. Update reputation scores (simulate performance improvement)
|
||||
println!("\n2. Updating Reputation Scores (Performance Improvement):");
|
||||
let mut improved_rep = db_rep3.clone();
|
||||
improved_rep.reputation = 55; // improved from 42
|
||||
improved_rep.uptime = 92; // improved from 87
|
||||
|
||||
// Also improve individual node reputations
|
||||
for node_rep in &mut improved_rep.nodes {
|
||||
node_rep.reputation += 10; // boost each node's reputation
|
||||
node_rep.uptime += 5; // improve uptime
|
||||
}
|
||||
|
||||
let (_, updated_rep) = db
|
||||
.collection::<NodeGroupReputation>()
|
||||
.expect("can open reputation collection")
|
||||
.set(&improved_rep)
|
||||
.expect("can update reputation");
|
||||
|
||||
println!("Updated reputation for struggling provider:");
|
||||
print_nodegroup_reputation_details(&updated_rep);
|
||||
|
||||
// 3. Add new node to existing nodegroup reputation
|
||||
println!("\n3. Adding New Node to Existing NodeGroup:");
|
||||
let new_node_rep = NodeReputation::new()
|
||||
.node_id(1004)
|
||||
.reputation(88)
|
||||
.uptime(99);
|
||||
|
||||
let mut expanded_rep = db_rep1.clone();
|
||||
expanded_rep.add_node_reputation(new_node_rep);
|
||||
// Recalculate nodegroup reputation based on node average
|
||||
let total_node_rep: i32 = expanded_rep.nodes.iter().map(|n| n.reputation).sum();
|
||||
expanded_rep.reputation = total_node_rep / expanded_rep.nodes.len() as i32;
|
||||
|
||||
let (_, expanded_rep) = db
|
||||
.collection::<NodeGroupReputation>()
|
||||
.expect("can open reputation collection")
|
||||
.set(&expanded_rep)
|
||||
.expect("can update reputation");
|
||||
|
||||
println!("Added new high-performing node to top provider:");
|
||||
print_nodegroup_reputation_details(&expanded_rep);
|
||||
|
||||
// Show all reputation records and calculate analytics
|
||||
let all_reps = db
|
||||
.collection::<NodeGroupReputation>()
|
||||
.expect("can open reputation collection")
|
||||
.get_all()
|
||||
.expect("can load all reputations");
|
||||
|
||||
println!("\n--- Reputation Analytics ---");
|
||||
println!("Total NodeGroup Reputations: {}", all_reps.len());
|
||||
|
||||
// Calculate overall statistics
|
||||
let avg_nodegroup_reputation: f64 = all_reps.iter()
|
||||
.map(|r| r.reputation as f64)
|
||||
.sum::<f64>() / all_reps.len() as f64;
|
||||
let avg_nodegroup_uptime: f64 = all_reps.iter()
|
||||
.filter(|r| r.uptime > 0) // exclude new providers with 0 uptime
|
||||
.map(|r| r.uptime as f64)
|
||||
.sum::<f64>() / all_reps.iter().filter(|r| r.uptime > 0).count() as f64;
|
||||
|
||||
println!("Overall Statistics:");
|
||||
println!(" Average NodeGroup Reputation: {:.1}/100", avg_nodegroup_reputation);
|
||||
println!(" Average NodeGroup Uptime: {:.1}%", avg_nodegroup_uptime);
|
||||
|
||||
// Count reputation tiers
|
||||
let excellent_reps = all_reps.iter().filter(|r| r.reputation >= 80).count();
|
||||
let good_reps = all_reps.iter().filter(|r| r.reputation >= 60 && r.reputation < 80).count();
|
||||
let average_reps = all_reps.iter().filter(|r| r.reputation >= 40 && r.reputation < 60).count();
|
||||
let poor_reps = all_reps.iter().filter(|r| r.reputation < 40).count();
|
||||
|
||||
println!("\nReputation Distribution:");
|
||||
println!(" Excellent (80-100): {}", excellent_reps);
|
||||
println!(" Good (60-79): {}", good_reps);
|
||||
println!(" Average (40-59): {}", average_reps);
|
||||
println!(" Poor (0-39): {}", poor_reps);
|
||||
|
||||
// Calculate total nodes and their statistics
|
||||
let total_nodes: usize = all_reps.iter().map(|r| r.nodes.len()).sum();
|
||||
let all_node_reps: Vec<i32> = all_reps.iter()
|
||||
.flat_map(|r| &r.nodes)
|
||||
.map(|n| n.reputation)
|
||||
.collect();
|
||||
let all_node_uptimes: Vec<i32> = all_reps.iter()
|
||||
.flat_map(|r| &r.nodes)
|
||||
.filter(|n| n.uptime > 0)
|
||||
.map(|n| n.uptime)
|
||||
.collect();
|
||||
|
||||
let avg_node_reputation: f64 = all_node_reps.iter().sum::<i32>() as f64 / all_node_reps.len() as f64;
|
||||
let avg_node_uptime: f64 = all_node_uptimes.iter().sum::<i32>() as f64 / all_node_uptimes.len() as f64;
|
||||
|
||||
println!("\nNode-Level Statistics:");
|
||||
println!(" Total Nodes: {}", total_nodes);
|
||||
println!(" Average Node Reputation: {:.1}/100", avg_node_reputation);
|
||||
println!(" Average Node Uptime: {:.1}%", avg_node_uptime);
|
||||
|
||||
// Find best and worst performing nodegroups
|
||||
let best_nodegroup = all_reps.iter().max_by_key(|r| r.reputation).unwrap();
|
||||
let worst_nodegroup = all_reps.iter().min_by_key(|r| r.reputation).unwrap();
|
||||
|
||||
println!("\nPerformance Leaders:");
|
||||
println!(" Best NodeGroup: {} (Reputation: {}, Uptime: {}%)",
|
||||
best_nodegroup.nodegroup_id, best_nodegroup.reputation, best_nodegroup.uptime);
|
||||
println!(" Worst NodeGroup: {} (Reputation: {}, Uptime: {}%)",
|
||||
worst_nodegroup.nodegroup_id, worst_nodegroup.reputation, worst_nodegroup.uptime);
|
||||
|
||||
// Rank nodegroups by reputation
|
||||
let mut ranked_nodegroups: Vec<_> = all_reps.iter().collect();
|
||||
ranked_nodegroups.sort_by(|a, b| b.reputation.cmp(&a.reputation));
|
||||
|
||||
println!("\nNodeGroup Rankings (by Reputation):");
|
||||
for (i, rep) in ranked_nodegroups.iter().enumerate() {
|
||||
let status = match rep.reputation {
|
||||
80..=100 => "Excellent",
|
||||
60..=79 => "Good",
|
||||
40..=59 => "Average",
|
||||
_ => "Poor",
|
||||
};
|
||||
println!(" {}. NodeGroup {}: {} ({}/100, {}% uptime)",
|
||||
i + 1, rep.nodegroup_id, status, rep.reputation, rep.uptime);
|
||||
}
|
||||
|
||||
println!("\n--- Model Information ---");
|
||||
println!("NodeGroupReputation DB Prefix: {}", NodeGroupReputation::db_prefix());
|
||||
}
|
@@ -1,6 +1,6 @@
|
||||
use heromodels::models::heroledger::rhai::register_heroledger_rhai_modules;
|
||||
use heromodels_core::db::hero::OurDB;
|
||||
use rhai::{Dynamic, Engine};
|
||||
use heromodels::models::heroledger::rhai::register_heroledger_rhai_modules;
|
||||
use std::sync::Arc;
|
||||
use std::{fs, path::Path};
|
||||
|
||||
|
15
heromodels/examples/heroledger_example/README.md
Normal file
15
heromodels/examples/heroledger_example/README.md
Normal file
@@ -0,0 +1,15 @@
|
||||
# Heroledger Postgres Example
|
||||
|
||||
This example demonstrates how to use the Heroledger `User` model against Postgres using the `heromodels::db::postgres` backend.
|
||||
|
||||
- Connects to Postgres with user `postgres` and password `test123` on `localhost:5432`.
|
||||
- Creates the table and indexes automatically on first use.
|
||||
- Shows basic CRUD and an index lookup on `username`.
|
||||
|
||||
Run it:
|
||||
|
||||
```bash
|
||||
cargo run -p heromodels --example heroledger_example
|
||||
```
|
||||
|
||||
Make sure Postgres is running locally and accessible with the credentials above.
|
54
heromodels/examples/heroledger_example/example.rs
Normal file
54
heromodels/examples/heroledger_example/example.rs
Normal file
@@ -0,0 +1,54 @@
|
||||
use heromodels::db::postgres::{Config, Postgres};
|
||||
use heromodels::db::{Collection, Db};
|
||||
use heromodels::models::heroledger::user::user_index::username;
|
||||
use heromodels::models::heroledger::user::{SecretBox, User};
|
||||
|
||||
fn main() {
|
||||
let db = Postgres::new(
|
||||
Config::new()
|
||||
.user(Some("postgres".into()))
|
||||
.password(Some("test123".into()))
|
||||
.host(Some("localhost".into()))
|
||||
.port(Some(5432)),
|
||||
)
|
||||
.expect("Can connect to Postgres");
|
||||
|
||||
println!("Heroledger User - Postgres Example");
|
||||
println!("==================================");
|
||||
|
||||
let users = db.collection::<User>().expect("open user collection");
|
||||
|
||||
// Clean
|
||||
if let Ok(existing) = users.get_all() {
|
||||
for u in existing {
|
||||
let _ = users.delete_by_id(u.get_id());
|
||||
}
|
||||
}
|
||||
|
||||
let sb = SecretBox::new().data(vec![1, 2, 3]).nonce(vec![9, 9, 9]).build();
|
||||
|
||||
let u = User::new(0)
|
||||
.username("alice")
|
||||
.pubkey("PUBKEY_A")
|
||||
.add_email("alice@example.com")
|
||||
.add_userprofile(sb)
|
||||
.build();
|
||||
|
||||
let (id, stored) = users.set(&u).expect("store user");
|
||||
println!("Stored user id={id} username={} pubkey={}", stored.username, stored.pubkey);
|
||||
|
||||
let by_idx = users.get::<username, _>("alice").expect("by username");
|
||||
println!("Found {} user(s) with username=alice", by_idx.len());
|
||||
|
||||
let fetched = users.get_by_id(id).expect("get by id").expect("exists");
|
||||
println!("Fetched by id={} username={} emails={:?}", id, fetched.username, fetched.email);
|
||||
|
||||
// Update
|
||||
let updated = fetched.clone().add_email("work@alice.example");
|
||||
let (_, back) = users.set(&updated).expect("update user");
|
||||
println!("Updated emails = {:?}", back.email);
|
||||
|
||||
// Delete
|
||||
users.delete_by_id(id).expect("delete user");
|
||||
println!("Deleted user id={id}");
|
||||
}
|
@@ -1,8 +1,11 @@
|
||||
use heromodels::db::postgres::Config;
|
||||
use heromodels::db::{Collection, Db};
|
||||
use heromodels::models::userexample::user::user_index::{is_active, username};
|
||||
use heromodels::models::userexample::user::user_index::{email, username};
|
||||
use heromodels::models::{Comment, User};
|
||||
use heromodels_core::Model;
|
||||
// For demonstrating embedded/nested indexes
|
||||
use heromodels::models::grid4::node::{ComputeSlice, DeviceInfo, Node};
|
||||
use heromodels::models::grid4::node::node_index::{country as node_country, pubkey as node_pubkey};
|
||||
|
||||
// Helper function to print user details
|
||||
fn print_user_details(user: &User) {
|
||||
@@ -37,14 +40,31 @@ fn main() {
|
||||
)
|
||||
.expect("Can connect to postgress");
|
||||
|
||||
// Unique suffix to avoid collisions with legacy rows from prior runs
|
||||
use std::time::{SystemTime, UNIX_EPOCH};
|
||||
let ts = SystemTime::now()
|
||||
.duration_since(UNIX_EPOCH)
|
||||
.unwrap()
|
||||
.as_secs();
|
||||
let user1_name = format!("johndoe_{}", ts);
|
||||
let user2_name = format!("janesmith_{}", ts);
|
||||
let user3_name = format!("willism_{}", ts);
|
||||
let user4_name = format!("carrols_{}", ts);
|
||||
let user1_email = format!("john.doe+{}@example.com", ts);
|
||||
let user2_email = format!("jane.smith+{}@example.com", ts);
|
||||
let user3_email = format!("willis.masters+{}@example.com", ts);
|
||||
let user4_email = format!("carrol.smith+{}@example.com", ts);
|
||||
|
||||
println!("Hero Models - Basic Usage Example");
|
||||
println!("================================");
|
||||
|
||||
// Clean up any existing data to ensure consistent results
|
||||
println!("Cleaning up existing data...");
|
||||
let user_collection = db.collection::<User>().expect("can open user collection");
|
||||
let comment_collection = db.collection::<Comment>().expect("can open comment collection");
|
||||
|
||||
let comment_collection = db
|
||||
.collection::<Comment>()
|
||||
.expect("can open comment collection");
|
||||
|
||||
// Clear all existing users and comments
|
||||
if let Ok(existing_users) = user_collection.get_all() {
|
||||
for user in existing_users {
|
||||
@@ -62,32 +82,32 @@ fn main() {
|
||||
|
||||
// User 1
|
||||
let user1 = User::new()
|
||||
.username("johndoe")
|
||||
.email("john.doe@example.com")
|
||||
.username(&user1_name)
|
||||
.email(&user1_email)
|
||||
.full_name("John Doe")
|
||||
.is_active(false)
|
||||
.build();
|
||||
|
||||
// User 2
|
||||
let user2 = User::new()
|
||||
.username("janesmith")
|
||||
.email("jane.smith@example.com")
|
||||
.username(&user2_name)
|
||||
.email(&user2_email)
|
||||
.full_name("Jane Smith")
|
||||
.is_active(true)
|
||||
.build();
|
||||
|
||||
// User 3
|
||||
let user3 = User::new()
|
||||
.username("willism")
|
||||
.email("willis.masters@example.com")
|
||||
.username(&user3_name)
|
||||
.email(&user3_email)
|
||||
.full_name("Willis Masters")
|
||||
.is_active(true)
|
||||
.build();
|
||||
|
||||
// User 4
|
||||
let user4 = User::new()
|
||||
.username("carrols")
|
||||
.email("carrol.smith@example.com")
|
||||
.username(&user4_name)
|
||||
.email(&user4_email)
|
||||
.full_name("Carrol Smith")
|
||||
.is_active(false)
|
||||
.build();
|
||||
@@ -143,66 +163,95 @@ fn main() {
|
||||
let stored_users = db
|
||||
.collection::<User>()
|
||||
.expect("can open user collection")
|
||||
.get::<username, _>("johndoe")
|
||||
.get::<username, _>(&user1_name)
|
||||
.expect("can load stored user");
|
||||
|
||||
assert_eq!(stored_users.len(), 1);
|
||||
print_user_details(&stored_users[0]);
|
||||
|
||||
// 2. Retrieve by active status
|
||||
println!("\n2. By Active Status (Active = true):");
|
||||
let active_users = db
|
||||
// 2. Retrieve by email index
|
||||
println!("\n2. By Email Index:");
|
||||
let by_email = db
|
||||
.collection::<User>()
|
||||
.expect("can open user collection")
|
||||
.get::<is_active, _>(&true)
|
||||
.expect("can load stored users");
|
||||
|
||||
assert_eq!(active_users.len(), 2);
|
||||
for active_user in active_users.iter() {
|
||||
print_user_details(active_user);
|
||||
}
|
||||
.get::<email, _>(&user2_email)
|
||||
.expect("can load stored user by email");
|
||||
assert_eq!(by_email.len(), 1);
|
||||
print_user_details(&by_email[0]);
|
||||
|
||||
// 3. Delete a user and show the updated results
|
||||
println!("\n3. After Deleting a User:");
|
||||
let user_to_delete_id = active_users[0].get_id();
|
||||
let user_to_delete_id = stored_users[0].get_id();
|
||||
println!("Deleting user with ID: {user_to_delete_id}");
|
||||
db.collection::<User>()
|
||||
.expect("can open user collection")
|
||||
.delete_by_id(user_to_delete_id)
|
||||
.expect("can delete existing user");
|
||||
|
||||
// Show remaining active users
|
||||
let active_users = db
|
||||
// Verify deletion by querying the same username again
|
||||
let should_be_empty = db
|
||||
.collection::<User>()
|
||||
.expect("can open user collection")
|
||||
.get::<is_active, _>(&true)
|
||||
.expect("can load stored users");
|
||||
|
||||
println!(" a. Remaining Active Users:");
|
||||
assert_eq!(active_users.len(), 1);
|
||||
for active_user in active_users.iter() {
|
||||
print_user_details(active_user);
|
||||
}
|
||||
|
||||
// Show inactive users
|
||||
let inactive_users = db
|
||||
.collection::<User>()
|
||||
.expect("can open user collection")
|
||||
.get::<is_active, _>(&false)
|
||||
.expect("can load stored users");
|
||||
|
||||
println!(" b. Inactive Users:");
|
||||
assert_eq!(inactive_users.len(), 2);
|
||||
for inactive_user in inactive_users.iter() {
|
||||
print_user_details(inactive_user);
|
||||
}
|
||||
.get::<username, _>(&user1_name)
|
||||
.expect("can query by username after delete");
|
||||
println!(" a. Query by username '{}' after delete -> {} results", user1_name, should_be_empty.len());
|
||||
assert_eq!(should_be_empty.len(), 0);
|
||||
|
||||
// Delete a user based on an index for good measure
|
||||
db.collection::<User>()
|
||||
.expect("can open user collection")
|
||||
.delete::<username, _>("janesmith")
|
||||
.delete::<username, _>(&user4_name)
|
||||
.expect("can delete existing user");
|
||||
|
||||
// Demonstrate embedded/nested indexes with Grid4 Node
|
||||
println!("\n--- Demonstrating Embedded/Nested Indexes (Grid4::Node) ---");
|
||||
println!("Node indexed fields: {:?}", Node::indexed_fields());
|
||||
|
||||
// Build a minimal node with nested data and persist it
|
||||
let cs = ComputeSlice::new()
|
||||
.nodeid(42)
|
||||
.slice_id(1)
|
||||
.mem_gb(32.0)
|
||||
.storage_gb(512.0)
|
||||
.passmark(6000)
|
||||
.vcores(16)
|
||||
.gpus(1)
|
||||
.price_cc(0.33);
|
||||
let dev = DeviceInfo { vendor: "ACME".into(), ..Default::default() };
|
||||
let node = Node::new()
|
||||
.nodegroupid(101)
|
||||
.uptime(99)
|
||||
.add_compute_slice(cs)
|
||||
.devices(dev)
|
||||
.country("BE")
|
||||
.pubkey("EX_NODE_PK_1")
|
||||
.build();
|
||||
let (node_id, _stored_node) = db
|
||||
.collection::<Node>()
|
||||
.expect("can open node collection")
|
||||
.set(&node)
|
||||
.expect("can set node");
|
||||
println!("Stored node id: {}", node_id);
|
||||
|
||||
// Query by top-level indexes
|
||||
let be_nodes = db
|
||||
.collection::<Node>()
|
||||
.expect("can open node collection")
|
||||
.get::<node_country, _>("BE")
|
||||
.expect("can query nodes by country");
|
||||
println!("Nodes in BE (count may include legacy rows): {}", be_nodes.len());
|
||||
|
||||
let by_pk = db
|
||||
.collection::<Node>()
|
||||
.expect("can open node collection")
|
||||
.get::<node_pubkey, _>("EX_NODE_PK_1")
|
||||
.expect("can query node by pubkey");
|
||||
assert!(by_pk.iter().any(|n| n.get_id() == node_id));
|
||||
|
||||
// Note: Nested path indexes (e.g., devices.vendor, computeslices.passmark) are created and used
|
||||
// for DB-side indexing, but are not yet exposed as typed Index keys in the API. They appear in
|
||||
// Node::indexed_fields() and contribute to Model::db_keys(), enabling performant JSONB GIN indexes.
|
||||
|
||||
println!("\n--- User Model Information ---");
|
||||
println!("User DB Prefix: {}", User::db_prefix());
|
||||
|
||||
@@ -212,7 +261,7 @@ fn main() {
|
||||
// 1. Create and save a comment
|
||||
println!("\n1. Creating a Comment:");
|
||||
let comment = Comment::new()
|
||||
.user_id(db_user1.get_id()) // commenter's user ID
|
||||
.user_id(db_user2.get_id()) // commenter's user ID (use an existing user)
|
||||
.content("This is a comment on the user")
|
||||
.build();
|
||||
|
||||
@@ -230,7 +279,7 @@ fn main() {
|
||||
|
||||
// 3. Associate the comment with a user
|
||||
println!("\n2. Associating Comment with User:");
|
||||
let mut updated_user = db_user1.clone();
|
||||
let mut updated_user = db_user2.clone();
|
||||
updated_user.base_data.add_comment(db_comment.get_id());
|
||||
|
||||
// Save the updated user and get the new version
|
||||
|
@@ -8,8 +8,8 @@ use std::{
|
||||
collections::HashSet,
|
||||
path::PathBuf,
|
||||
sync::{
|
||||
atomic::{AtomicU32, Ordering},
|
||||
Arc, Mutex,
|
||||
atomic::{AtomicU32, Ordering},
|
||||
},
|
||||
};
|
||||
|
||||
|
@@ -119,4 +119,4 @@ impl Circle {
|
||||
/// Creates a new circle builder
|
||||
pub fn new_circle() -> Circle {
|
||||
Circle::new()
|
||||
}
|
||||
}
|
||||
|
@@ -1,16 +1,17 @@
|
||||
use crate::db::Db;
|
||||
use rhailib_macros::{
|
||||
register_authorized_create_by_id_fn, register_authorized_delete_by_id_fn, register_authorized_get_by_id_fn,
|
||||
};
|
||||
use rhai::plugin::*;
|
||||
use rhai::{Array, Dynamic, Engine, EvalAltResult, Map, Module};
|
||||
use rhailib_macros::{
|
||||
register_authorized_create_by_id_fn, register_authorized_delete_by_id_fn,
|
||||
register_authorized_get_by_id_fn,
|
||||
};
|
||||
use std::collections::HashMap;
|
||||
use std::sync::Arc;
|
||||
|
||||
use crate::models::circle::Circle;
|
||||
type RhaiCircle = Circle;
|
||||
use crate::db::hero::OurDB;
|
||||
use crate::db::Collection;
|
||||
use crate::db::hero::OurDB;
|
||||
use crate::models::circle::ThemeData;
|
||||
|
||||
#[export_module]
|
||||
|
128
heromodels/src/models/grid4/bid.rs
Normal file
128
heromodels/src/models/grid4/bid.rs
Normal file
@@ -0,0 +1,128 @@
|
||||
use heromodels_core::BaseModelData;
|
||||
use heromodels_derive::model;
|
||||
use rhai::{CustomType, TypeBuilder};
|
||||
use serde::{Deserialize, Serialize};
|
||||
|
||||
/// Bid status enumeration
|
||||
#[derive(Debug, Clone, Serialize, Deserialize, PartialEq, Default)]
|
||||
pub enum BidStatus {
|
||||
#[default]
|
||||
Pending,
|
||||
Confirmed,
|
||||
Assigned,
|
||||
Cancelled,
|
||||
Done,
|
||||
}
|
||||
|
||||
/// Billing period enumeration
|
||||
#[derive(Debug, Clone, Serialize, Deserialize, PartialEq, Default)]
|
||||
pub enum BillingPeriod {
|
||||
#[default]
|
||||
Hourly,
|
||||
Monthly,
|
||||
Yearly,
|
||||
Biannually,
|
||||
Triannually,
|
||||
}
|
||||
|
||||
/// I can bid for infra, and optionally get accepted
|
||||
#[model]
|
||||
#[derive(Debug, Clone, Serialize, Deserialize, PartialEq, Default, CustomType)]
|
||||
pub struct Bid {
|
||||
pub base_data: BaseModelData,
|
||||
/// links back to customer for this capacity (user on ledger)
|
||||
#[index]
|
||||
pub customer_id: u32,
|
||||
/// nr of slices I need in 1 machine
|
||||
pub compute_slices_nr: i32,
|
||||
/// price per 1 GB slice I want to accept
|
||||
pub compute_slice_price: f64,
|
||||
/// nr of storage slices needed
|
||||
pub storage_slices_nr: i32,
|
||||
/// price per 1 GB storage slice I want to accept
|
||||
pub storage_slice_price: f64,
|
||||
pub status: BidStatus,
|
||||
/// if obligation then will be charged and money needs to be in escrow, otherwise its an intent
|
||||
pub obligation: bool,
|
||||
/// epoch timestamp
|
||||
pub start_date: u32,
|
||||
/// epoch timestamp
|
||||
pub end_date: u32,
|
||||
/// signature as done by a user/consumer to validate their identity and intent
|
||||
pub signature_user: String,
|
||||
pub billing_period: BillingPeriod,
|
||||
}
|
||||
|
||||
impl Bid {
|
||||
pub fn new() -> Self {
|
||||
Self {
|
||||
base_data: BaseModelData::new(),
|
||||
customer_id: 0,
|
||||
compute_slices_nr: 0,
|
||||
compute_slice_price: 0.0,
|
||||
storage_slices_nr: 0,
|
||||
storage_slice_price: 0.0,
|
||||
status: BidStatus::default(),
|
||||
obligation: false,
|
||||
start_date: 0,
|
||||
end_date: 0,
|
||||
signature_user: String::new(),
|
||||
billing_period: BillingPeriod::default(),
|
||||
}
|
||||
}
|
||||
|
||||
pub fn customer_id(mut self, v: u32) -> Self {
|
||||
self.customer_id = v;
|
||||
self
|
||||
}
|
||||
|
||||
pub fn compute_slices_nr(mut self, v: i32) -> Self {
|
||||
self.compute_slices_nr = v;
|
||||
self
|
||||
}
|
||||
|
||||
pub fn compute_slice_price(mut self, v: f64) -> Self {
|
||||
self.compute_slice_price = v;
|
||||
self
|
||||
}
|
||||
|
||||
pub fn storage_slices_nr(mut self, v: i32) -> Self {
|
||||
self.storage_slices_nr = v;
|
||||
self
|
||||
}
|
||||
|
||||
pub fn storage_slice_price(mut self, v: f64) -> Self {
|
||||
self.storage_slice_price = v;
|
||||
self
|
||||
}
|
||||
|
||||
pub fn status(mut self, v: BidStatus) -> Self {
|
||||
self.status = v;
|
||||
self
|
||||
}
|
||||
|
||||
pub fn obligation(mut self, v: bool) -> Self {
|
||||
self.obligation = v;
|
||||
self
|
||||
}
|
||||
|
||||
pub fn start_date(mut self, v: u32) -> Self {
|
||||
self.start_date = v;
|
||||
self
|
||||
}
|
||||
|
||||
pub fn end_date(mut self, v: u32) -> Self {
|
||||
self.end_date = v;
|
||||
self
|
||||
}
|
||||
|
||||
pub fn signature_user(mut self, v: impl ToString) -> Self {
|
||||
self.signature_user = v.to_string();
|
||||
self
|
||||
}
|
||||
|
||||
pub fn billing_period(mut self, v: BillingPeriod) -> Self {
|
||||
self.billing_period = v;
|
||||
self
|
||||
}
|
||||
}
|
39
heromodels/src/models/grid4/common.rs
Normal file
39
heromodels/src/models/grid4/common.rs
Normal file
@@ -0,0 +1,39 @@
|
||||
use rhai::{CustomType, TypeBuilder};
|
||||
use serde::{Deserialize, Serialize};
|
||||
|
||||
/// SLA policy matching the V spec `SLAPolicy`
|
||||
#[derive(Debug, Clone, Serialize, Deserialize, PartialEq, Default, CustomType)]
|
||||
pub struct SLAPolicy {
|
||||
/// should +90
|
||||
pub sla_uptime: i32,
|
||||
/// minimal mbits we can expect avg over 1h per node, 0 means we don't guarantee
|
||||
pub sla_bandwidth_mbit: i32,
|
||||
/// 0-100, percent of money given back in relation to month if sla breached,
|
||||
/// e.g. 200 means we return 2 months worth of rev if sla missed
|
||||
pub sla_penalty: i32,
|
||||
}
|
||||
|
||||
impl SLAPolicy {
|
||||
pub fn new() -> Self { Self::default() }
|
||||
pub fn sla_uptime(mut self, v: i32) -> Self { self.sla_uptime = v; self }
|
||||
pub fn sla_bandwidth_mbit(mut self, v: i32) -> Self { self.sla_bandwidth_mbit = v; self }
|
||||
pub fn sla_penalty(mut self, v: i32) -> Self { self.sla_penalty = v; self }
|
||||
pub fn build(self) -> Self { self }
|
||||
}
|
||||
|
||||
/// Pricing policy matching the V spec `PricingPolicy`
|
||||
#[derive(Debug, Clone, Serialize, Deserialize, PartialEq, Default, CustomType)]
|
||||
pub struct PricingPolicy {
|
||||
/// e.g. 30,40,50 means if user has more CC in wallet than 1 year utilization
|
||||
/// then this provider gives 30%, 2Y 40%, ...
|
||||
pub marketplace_year_discounts: Vec<i32>,
|
||||
/// e.g. 10,20,30
|
||||
pub volume_discounts: Vec<i32>,
|
||||
}
|
||||
|
||||
impl PricingPolicy {
|
||||
pub fn new() -> Self { Self { marketplace_year_discounts: vec![30, 40, 50], volume_discounts: vec![10, 20, 30] } }
|
||||
pub fn marketplace_year_discounts(mut self, v: Vec<i32>) -> Self { self.marketplace_year_discounts = v; self }
|
||||
pub fn volume_discounts(mut self, v: Vec<i32>) -> Self { self.volume_discounts = v; self }
|
||||
pub fn build(self) -> Self { self }
|
||||
}
|
219
heromodels/src/models/grid4/contract.rs
Normal file
219
heromodels/src/models/grid4/contract.rs
Normal file
@@ -0,0 +1,219 @@
|
||||
use heromodels_core::BaseModelData;
|
||||
use heromodels_derive::model;
|
||||
use rhai::{CustomType, TypeBuilder};
|
||||
use serde::{Deserialize, Serialize};
|
||||
use super::bid::BillingPeriod;
|
||||
|
||||
/// Contract status enumeration
|
||||
#[derive(Debug, Clone, Serialize, Deserialize, PartialEq, Default)]
|
||||
pub enum ContractStatus {
|
||||
#[default]
|
||||
Active,
|
||||
Cancelled,
|
||||
Error,
|
||||
Paused,
|
||||
}
|
||||
|
||||
/// Compute slice provisioned for a contract
|
||||
#[derive(Debug, Clone, Serialize, Deserialize, PartialEq, Default, CustomType)]
|
||||
pub struct ComputeSliceProvisioned {
|
||||
pub node_id: u32,
|
||||
/// the id of the slice in the node
|
||||
pub id: u16,
|
||||
pub mem_gb: f64,
|
||||
pub storage_gb: f64,
|
||||
pub passmark: i32,
|
||||
pub vcores: i32,
|
||||
pub cpu_oversubscription: i32,
|
||||
pub tags: String,
|
||||
}
|
||||
|
||||
/// Storage slice provisioned for a contract
|
||||
#[derive(Debug, Clone, Serialize, Deserialize, PartialEq, Default, CustomType)]
|
||||
pub struct StorageSliceProvisioned {
|
||||
pub node_id: u32,
|
||||
/// the id of the slice in the node, are tracked in the node itself
|
||||
pub id: u16,
|
||||
pub storage_size_gb: i32,
|
||||
pub tags: String,
|
||||
}
|
||||
|
||||
/// Contract for provisioned infrastructure
|
||||
#[model]
|
||||
#[derive(Debug, Clone, Serialize, Deserialize, PartialEq, Default, CustomType)]
|
||||
pub struct Contract {
|
||||
pub base_data: BaseModelData,
|
||||
/// links back to customer for this capacity (user on ledger)
|
||||
#[index]
|
||||
pub customer_id: u32,
|
||||
pub compute_slices: Vec<ComputeSliceProvisioned>,
|
||||
pub storage_slices: Vec<StorageSliceProvisioned>,
|
||||
/// price per 1 GB agreed upon
|
||||
pub compute_slice_price: f64,
|
||||
/// price per 1 GB agreed upon
|
||||
pub storage_slice_price: f64,
|
||||
/// price per 1 GB agreed upon (transfer)
|
||||
pub network_slice_price: f64,
|
||||
pub status: ContractStatus,
|
||||
/// epoch timestamp
|
||||
pub start_date: u32,
|
||||
/// epoch timestamp
|
||||
pub end_date: u32,
|
||||
/// signature as done by a user/consumer to validate their identity and intent
|
||||
pub signature_user: String,
|
||||
/// signature as done by the hoster
|
||||
pub signature_hoster: String,
|
||||
pub billing_period: BillingPeriod,
|
||||
}
|
||||
|
||||
impl Contract {
|
||||
pub fn new() -> Self {
|
||||
Self {
|
||||
base_data: BaseModelData::new(),
|
||||
customer_id: 0,
|
||||
compute_slices: Vec::new(),
|
||||
storage_slices: Vec::new(),
|
||||
compute_slice_price: 0.0,
|
||||
storage_slice_price: 0.0,
|
||||
network_slice_price: 0.0,
|
||||
status: ContractStatus::default(),
|
||||
start_date: 0,
|
||||
end_date: 0,
|
||||
signature_user: String::new(),
|
||||
signature_hoster: String::new(),
|
||||
billing_period: BillingPeriod::default(),
|
||||
}
|
||||
}
|
||||
|
||||
pub fn customer_id(mut self, v: u32) -> Self {
|
||||
self.customer_id = v;
|
||||
self
|
||||
}
|
||||
|
||||
pub fn add_compute_slice(mut self, slice: ComputeSliceProvisioned) -> Self {
|
||||
self.compute_slices.push(slice);
|
||||
self
|
||||
}
|
||||
|
||||
pub fn add_storage_slice(mut self, slice: StorageSliceProvisioned) -> Self {
|
||||
self.storage_slices.push(slice);
|
||||
self
|
||||
}
|
||||
|
||||
pub fn compute_slice_price(mut self, v: f64) -> Self {
|
||||
self.compute_slice_price = v;
|
||||
self
|
||||
}
|
||||
|
||||
pub fn storage_slice_price(mut self, v: f64) -> Self {
|
||||
self.storage_slice_price = v;
|
||||
self
|
||||
}
|
||||
|
||||
pub fn network_slice_price(mut self, v: f64) -> Self {
|
||||
self.network_slice_price = v;
|
||||
self
|
||||
}
|
||||
|
||||
pub fn status(mut self, v: ContractStatus) -> Self {
|
||||
self.status = v;
|
||||
self
|
||||
}
|
||||
|
||||
pub fn start_date(mut self, v: u32) -> Self {
|
||||
self.start_date = v;
|
||||
self
|
||||
}
|
||||
|
||||
pub fn end_date(mut self, v: u32) -> Self {
|
||||
self.end_date = v;
|
||||
self
|
||||
}
|
||||
|
||||
pub fn signature_user(mut self, v: impl ToString) -> Self {
|
||||
self.signature_user = v.to_string();
|
||||
self
|
||||
}
|
||||
|
||||
pub fn signature_hoster(mut self, v: impl ToString) -> Self {
|
||||
self.signature_hoster = v.to_string();
|
||||
self
|
||||
}
|
||||
|
||||
pub fn billing_period(mut self, v: BillingPeriod) -> Self {
|
||||
self.billing_period = v;
|
||||
self
|
||||
}
|
||||
}
|
||||
|
||||
impl ComputeSliceProvisioned {
|
||||
pub fn new() -> Self {
|
||||
Self::default()
|
||||
}
|
||||
|
||||
pub fn node_id(mut self, v: u32) -> Self {
|
||||
self.node_id = v;
|
||||
self
|
||||
}
|
||||
|
||||
pub fn id(mut self, v: u16) -> Self {
|
||||
self.id = v;
|
||||
self
|
||||
}
|
||||
|
||||
pub fn mem_gb(mut self, v: f64) -> Self {
|
||||
self.mem_gb = v;
|
||||
self
|
||||
}
|
||||
|
||||
pub fn storage_gb(mut self, v: f64) -> Self {
|
||||
self.storage_gb = v;
|
||||
self
|
||||
}
|
||||
|
||||
pub fn passmark(mut self, v: i32) -> Self {
|
||||
self.passmark = v;
|
||||
self
|
||||
}
|
||||
|
||||
pub fn vcores(mut self, v: i32) -> Self {
|
||||
self.vcores = v;
|
||||
self
|
||||
}
|
||||
|
||||
pub fn cpu_oversubscription(mut self, v: i32) -> Self {
|
||||
self.cpu_oversubscription = v;
|
||||
self
|
||||
}
|
||||
|
||||
pub fn tags(mut self, v: impl ToString) -> Self {
|
||||
self.tags = v.to_string();
|
||||
self
|
||||
}
|
||||
}
|
||||
|
||||
impl StorageSliceProvisioned {
|
||||
pub fn new() -> Self {
|
||||
Self::default()
|
||||
}
|
||||
|
||||
pub fn node_id(mut self, v: u32) -> Self {
|
||||
self.node_id = v;
|
||||
self
|
||||
}
|
||||
|
||||
pub fn id(mut self, v: u16) -> Self {
|
||||
self.id = v;
|
||||
self
|
||||
}
|
||||
|
||||
pub fn storage_size_gb(mut self, v: i32) -> Self {
|
||||
self.storage_size_gb = v;
|
||||
self
|
||||
}
|
||||
|
||||
pub fn tags(mut self, v: impl ToString) -> Self {
|
||||
self.tags = v.to_string();
|
||||
self
|
||||
}
|
||||
}
|
@@ -1,16 +1,18 @@
|
||||
pub mod bid;
|
||||
pub mod common;
|
||||
pub mod contract;
|
||||
pub mod node;
|
||||
pub mod nodegroup;
|
||||
pub mod reputation;
|
||||
pub mod reservation;
|
||||
|
||||
pub use bid::{Bid, BidStatus, BillingPeriod};
|
||||
pub use common::{PricingPolicy, SLAPolicy};
|
||||
pub use contract::{Contract, ContractStatus, ComputeSliceProvisioned, StorageSliceProvisioned};
|
||||
pub use node::{
|
||||
Node,
|
||||
DeviceInfo,
|
||||
StorageDevice,
|
||||
MemoryDevice,
|
||||
CPUDevice,
|
||||
GPUDevice,
|
||||
NetworkDevice,
|
||||
NodeCapacity,
|
||||
ComputeSlice,
|
||||
StorageSlice,
|
||||
PricingPolicy,
|
||||
SLAPolicy,
|
||||
};
|
||||
CPUDevice, ComputeSlice, DeviceInfo, GPUDevice, MemoryDevice, NetworkDevice, Node,
|
||||
NodeCapacity, StorageDevice, StorageSlice,
|
||||
};
|
||||
pub use nodegroup::NodeGroup;
|
||||
pub use reputation::{NodeGroupReputation, NodeReputation};
|
||||
pub use reservation::{Reservation, ReservationStatus};
|
||||
|
@@ -1,7 +1,8 @@
|
||||
use heromodels_core::BaseModelData;
|
||||
use heromodels_derive::model;
|
||||
use rhai::CustomType;
|
||||
use rhai::{CustomType, TypeBuilder};
|
||||
use serde::{Deserialize, Serialize};
|
||||
use super::common::{PricingPolicy, SLAPolicy};
|
||||
|
||||
/// Storage device information
|
||||
#[derive(Debug, Clone, Serialize, Deserialize, PartialEq, Default, CustomType)]
|
||||
@@ -94,57 +95,26 @@ pub struct NodeCapacity {
|
||||
pub vcores: i32,
|
||||
}
|
||||
|
||||
/// Pricing policy for slices (minimal version until full spec available)
|
||||
#[derive(Debug, Clone, Serialize, Deserialize, PartialEq, Default, CustomType)]
|
||||
pub struct PricingPolicy {
|
||||
/// Human friendly policy name (e.g. "fixed", "market")
|
||||
pub name: String,
|
||||
/// Optional free-form details as JSON-encoded string
|
||||
pub details: Option<String>,
|
||||
}
|
||||
|
||||
/// SLA policy for slices (minimal version until full spec available)
|
||||
#[derive(Debug, Clone, Serialize, Deserialize, PartialEq, Default, CustomType)]
|
||||
pub struct SLAPolicy {
|
||||
/// Uptime in percentage (0..100)
|
||||
pub uptime: f32,
|
||||
/// Max response time in ms
|
||||
pub max_response_time_ms: u32,
|
||||
}
|
||||
// PricingPolicy and SLAPolicy moved to `common.rs` to be shared across models.
|
||||
|
||||
/// Compute slice (typically represents a base unit of compute)
|
||||
#[model]
|
||||
#[derive(Debug, Clone, Serialize, Deserialize, PartialEq, Default, CustomType)]
|
||||
pub struct ComputeSlice {
|
||||
pub base_data: BaseModelData,
|
||||
/// the node in the grid, there is an object describing the node
|
||||
#[index]
|
||||
pub nodeid: u32,
|
||||
/// the id of the slice in the node
|
||||
#[index]
|
||||
pub id: i32,
|
||||
pub id: u16,
|
||||
pub mem_gb: f64,
|
||||
pub storage_gb: f64,
|
||||
pub passmark: i32,
|
||||
pub vcores: i32,
|
||||
pub cpu_oversubscription: i32,
|
||||
pub storage_oversubscription: i32,
|
||||
/// Min/max allowed price range for validation
|
||||
#[serde(default)]
|
||||
pub price_range: Vec<f64>,
|
||||
/// nr of GPU's see node to know what GPU's are
|
||||
pub gpus: u8,
|
||||
/// price per slice (even if the grouped one)
|
||||
pub price_cc: f64,
|
||||
pub pricing_policy: PricingPolicy,
|
||||
pub sla_policy: SLAPolicy,
|
||||
}
|
||||
|
||||
impl ComputeSlice {
|
||||
pub fn new() -> Self {
|
||||
Self {
|
||||
base_data: BaseModelData::new(),
|
||||
nodeid: 0,
|
||||
id: 0,
|
||||
mem_gb: 0.0,
|
||||
storage_gb: 0.0,
|
||||
@@ -152,63 +122,62 @@ impl ComputeSlice {
|
||||
vcores: 0,
|
||||
cpu_oversubscription: 0,
|
||||
storage_oversubscription: 0,
|
||||
price_range: vec![0.0, 0.0],
|
||||
gpus: 0,
|
||||
price_cc: 0.0,
|
||||
pricing_policy: PricingPolicy::default(),
|
||||
sla_policy: SLAPolicy::default(),
|
||||
}
|
||||
}
|
||||
|
||||
pub fn nodeid(mut self, nodeid: u32) -> Self { self.nodeid = nodeid; self }
|
||||
pub fn slice_id(mut self, id: i32) -> Self { self.id = id; self }
|
||||
pub fn mem_gb(mut self, v: f64) -> Self { self.mem_gb = v; self }
|
||||
pub fn storage_gb(mut self, v: f64) -> Self { self.storage_gb = v; self }
|
||||
pub fn passmark(mut self, v: i32) -> Self { self.passmark = v; self }
|
||||
pub fn vcores(mut self, v: i32) -> Self { self.vcores = v; self }
|
||||
pub fn cpu_oversubscription(mut self, v: i32) -> Self { self.cpu_oversubscription = v; self }
|
||||
pub fn storage_oversubscription(mut self, v: i32) -> Self { self.storage_oversubscription = v; self }
|
||||
pub fn price_range(mut self, min_max: Vec<f64>) -> Self { self.price_range = min_max; self }
|
||||
pub fn gpus(mut self, v: u8) -> Self { self.gpus = v; self }
|
||||
pub fn price_cc(mut self, v: f64) -> Self { self.price_cc = v; self }
|
||||
pub fn pricing_policy(mut self, p: PricingPolicy) -> Self { self.pricing_policy = p; self }
|
||||
pub fn sla_policy(mut self, p: SLAPolicy) -> Self { self.sla_policy = p; self }
|
||||
pub fn id(mut self, id: u16) -> Self {
|
||||
self.id = id;
|
||||
self
|
||||
}
|
||||
pub fn mem_gb(mut self, v: f64) -> Self {
|
||||
self.mem_gb = v;
|
||||
self
|
||||
}
|
||||
pub fn storage_gb(mut self, v: f64) -> Self {
|
||||
self.storage_gb = v;
|
||||
self
|
||||
}
|
||||
pub fn passmark(mut self, v: i32) -> Self {
|
||||
self.passmark = v;
|
||||
self
|
||||
}
|
||||
pub fn vcores(mut self, v: i32) -> Self {
|
||||
self.vcores = v;
|
||||
self
|
||||
}
|
||||
pub fn cpu_oversubscription(mut self, v: i32) -> Self {
|
||||
self.cpu_oversubscription = v;
|
||||
self
|
||||
}
|
||||
pub fn storage_oversubscription(mut self, v: i32) -> Self {
|
||||
self.storage_oversubscription = v;
|
||||
self
|
||||
}
|
||||
pub fn gpus(mut self, v: u8) -> Self {
|
||||
self.gpus = v;
|
||||
self
|
||||
}
|
||||
}
|
||||
|
||||
/// Storage slice (typically 1GB of storage)
|
||||
#[model]
|
||||
#[derive(Debug, Clone, Serialize, Deserialize, PartialEq, Default, CustomType)]
|
||||
pub struct StorageSlice {
|
||||
pub base_data: BaseModelData,
|
||||
/// the node in the grid
|
||||
#[index]
|
||||
pub nodeid: u32,
|
||||
/// the id of the slice in the node, are tracked in the node itself
|
||||
#[index]
|
||||
pub id: i32,
|
||||
/// price per slice (even if the grouped one)
|
||||
pub price_cc: f64,
|
||||
pub pricing_policy: PricingPolicy,
|
||||
pub sla_policy: SLAPolicy,
|
||||
pub id: u16,
|
||||
}
|
||||
|
||||
impl StorageSlice {
|
||||
pub fn new() -> Self {
|
||||
Self {
|
||||
base_data: BaseModelData::new(),
|
||||
nodeid: 0,
|
||||
id: 0,
|
||||
price_cc: 0.0,
|
||||
pricing_policy: PricingPolicy::default(),
|
||||
sla_policy: SLAPolicy::default(),
|
||||
}
|
||||
}
|
||||
|
||||
pub fn nodeid(mut self, nodeid: u32) -> Self { self.nodeid = nodeid; self }
|
||||
pub fn slice_id(mut self, id: i32) -> Self { self.id = id; self }
|
||||
pub fn price_cc(mut self, v: f64) -> Self { self.price_cc = v; self }
|
||||
pub fn pricing_policy(mut self, p: PricingPolicy) -> Self { self.pricing_policy = p; self }
|
||||
pub fn sla_policy(mut self, p: SLAPolicy) -> Self { self.sla_policy = p; self }
|
||||
pub fn id(mut self, id: u16) -> Self {
|
||||
self.id = id;
|
||||
self
|
||||
}
|
||||
}
|
||||
|
||||
/// Grid4 Node model
|
||||
@@ -224,13 +193,20 @@ pub struct Node {
|
||||
pub computeslices: Vec<ComputeSlice>,
|
||||
pub storageslices: Vec<StorageSlice>,
|
||||
pub devices: DeviceInfo,
|
||||
/// 2 letter code
|
||||
/// 2 letter code as specified in lib/data/countries/data/countryInfo.txt
|
||||
#[index]
|
||||
pub country: String,
|
||||
/// Hardware capacity details
|
||||
pub capacity: NodeCapacity,
|
||||
/// lets keep it simple and compatible
|
||||
pub provisiontime: u32,
|
||||
/// first time node was active
|
||||
pub birthtime: u32,
|
||||
/// node public key
|
||||
#[index]
|
||||
pub pubkey: String,
|
||||
/// signature done on node to validate pubkey with privkey
|
||||
pub signature_node: String,
|
||||
/// signature as done by farmers to validate their identity
|
||||
pub signature_farmer: String,
|
||||
}
|
||||
|
||||
impl Node {
|
||||
@@ -244,21 +220,61 @@ impl Node {
|
||||
devices: DeviceInfo::default(),
|
||||
country: String::new(),
|
||||
capacity: NodeCapacity::default(),
|
||||
provisiontime: 0,
|
||||
birthtime: 0,
|
||||
pubkey: String::new(),
|
||||
signature_node: String::new(),
|
||||
signature_farmer: String::new(),
|
||||
}
|
||||
}
|
||||
|
||||
pub fn nodegroupid(mut self, v: i32) -> Self { self.nodegroupid = v; self }
|
||||
pub fn uptime(mut self, v: i32) -> Self { self.uptime = v; self }
|
||||
pub fn add_compute_slice(mut self, s: ComputeSlice) -> Self { self.computeslices.push(s); self }
|
||||
pub fn add_storage_slice(mut self, s: StorageSlice) -> Self { self.storageslices.push(s); self }
|
||||
pub fn devices(mut self, d: DeviceInfo) -> Self { self.devices = d; self }
|
||||
pub fn country(mut self, c: impl ToString) -> Self { self.country = c.to_string(); self }
|
||||
pub fn capacity(mut self, c: NodeCapacity) -> Self { self.capacity = c; self }
|
||||
pub fn provisiontime(mut self, t: u32) -> Self { self.provisiontime = t; self }
|
||||
pub fn nodegroupid(mut self, v: i32) -> Self {
|
||||
self.nodegroupid = v;
|
||||
self
|
||||
}
|
||||
pub fn uptime(mut self, v: i32) -> Self {
|
||||
self.uptime = v;
|
||||
self
|
||||
}
|
||||
pub fn add_compute_slice(mut self, s: ComputeSlice) -> Self {
|
||||
self.computeslices.push(s);
|
||||
self
|
||||
}
|
||||
pub fn add_storage_slice(mut self, s: StorageSlice) -> Self {
|
||||
self.storageslices.push(s);
|
||||
self
|
||||
}
|
||||
pub fn devices(mut self, d: DeviceInfo) -> Self {
|
||||
self.devices = d;
|
||||
self
|
||||
}
|
||||
pub fn country(mut self, c: impl ToString) -> Self {
|
||||
self.country = c.to_string();
|
||||
self
|
||||
}
|
||||
pub fn capacity(mut self, c: NodeCapacity) -> Self {
|
||||
self.capacity = c;
|
||||
self
|
||||
}
|
||||
pub fn birthtime(mut self, t: u32) -> Self {
|
||||
self.birthtime = t;
|
||||
self
|
||||
}
|
||||
|
||||
pub fn pubkey(mut self, v: impl ToString) -> Self {
|
||||
self.pubkey = v.to_string();
|
||||
self
|
||||
}
|
||||
pub fn signature_node(mut self, v: impl ToString) -> Self {
|
||||
self.signature_node = v.to_string();
|
||||
self
|
||||
}
|
||||
pub fn signature_farmer(mut self, v: impl ToString) -> Self {
|
||||
self.signature_farmer = v.to_string();
|
||||
self
|
||||
}
|
||||
|
||||
/// Placeholder for capacity recalculation out of the devices on the Node
|
||||
pub fn recalc_capacity(mut self) -> Self {
|
||||
pub fn check(self) -> Self {
|
||||
// TODO: calculate NodeCapacity out of the devices on the Node
|
||||
self
|
||||
}
|
||||
|
52
heromodels/src/models/grid4/nodegroup.rs
Normal file
52
heromodels/src/models/grid4/nodegroup.rs
Normal file
@@ -0,0 +1,52 @@
|
||||
use heromodels_core::BaseModelData;
|
||||
use heromodels_derive::model;
|
||||
use rhai::{CustomType, TypeBuilder};
|
||||
use serde::{Deserialize, Serialize};
|
||||
|
||||
use super::common::{PricingPolicy, SLAPolicy};
|
||||
|
||||
/// Grid4 NodeGroup model (root object for farmer configuration)
|
||||
#[model]
|
||||
#[derive(Debug, Clone, Serialize, Deserialize, PartialEq, Default, CustomType)]
|
||||
pub struct NodeGroup {
|
||||
pub base_data: BaseModelData,
|
||||
/// link back to farmer who owns the nodegroup, is a user?
|
||||
#[index]
|
||||
pub farmerid: u32,
|
||||
/// only visible by farmer, in future encrypted, used to boot a node
|
||||
pub secret: String,
|
||||
pub description: String,
|
||||
pub slapolicy: SLAPolicy,
|
||||
pub pricingpolicy: PricingPolicy,
|
||||
/// pricing in CC - cloud credit, per 2GB node slice
|
||||
pub compute_slice_normalized_pricing_cc: f64,
|
||||
/// pricing in CC - cloud credit, per 1GB storage slice
|
||||
pub storage_slice_normalized_pricing_cc: f64,
|
||||
/// signature as done by farmers to validate that they created this group
|
||||
pub signature_farmer: String,
|
||||
}
|
||||
|
||||
impl NodeGroup {
|
||||
pub fn new() -> Self {
|
||||
Self {
|
||||
base_data: BaseModelData::new(),
|
||||
farmerid: 0,
|
||||
secret: String::new(),
|
||||
description: String::new(),
|
||||
slapolicy: SLAPolicy::default(),
|
||||
pricingpolicy: PricingPolicy::new(),
|
||||
compute_slice_normalized_pricing_cc: 0.0,
|
||||
storage_slice_normalized_pricing_cc: 0.0,
|
||||
signature_farmer: String::new(),
|
||||
}
|
||||
}
|
||||
|
||||
pub fn farmerid(mut self, v: u32) -> Self { self.farmerid = v; self }
|
||||
pub fn secret(mut self, v: impl ToString) -> Self { self.secret = v.to_string(); self }
|
||||
pub fn description(mut self, v: impl ToString) -> Self { self.description = v.to_string(); self }
|
||||
pub fn slapolicy(mut self, v: SLAPolicy) -> Self { self.slapolicy = v; self }
|
||||
pub fn pricingpolicy(mut self, v: PricingPolicy) -> Self { self.pricingpolicy = v; self }
|
||||
pub fn compute_slice_normalized_pricing_cc(mut self, v: f64) -> Self { self.compute_slice_normalized_pricing_cc = v; self }
|
||||
pub fn storage_slice_normalized_pricing_cc(mut self, v: f64) -> Self { self.storage_slice_normalized_pricing_cc = v; self }
|
||||
pub fn signature_farmer(mut self, v: impl ToString) -> Self { self.signature_farmer = v.to_string(); self }
|
||||
}
|
85
heromodels/src/models/grid4/reputation.rs
Normal file
85
heromodels/src/models/grid4/reputation.rs
Normal file
@@ -0,0 +1,85 @@
|
||||
use heromodels_core::BaseModelData;
|
||||
use heromodels_derive::model;
|
||||
use rhai::{CustomType, TypeBuilder};
|
||||
use serde::{Deserialize, Serialize};
|
||||
|
||||
/// Node reputation information
|
||||
#[derive(Debug, Clone, Serialize, Deserialize, PartialEq, Default, CustomType)]
|
||||
pub struct NodeReputation {
|
||||
pub node_id: u32,
|
||||
/// between 0 and 100, earned over time
|
||||
pub reputation: i32,
|
||||
/// between 0 and 100, set by system, farmer has no ability to set this
|
||||
pub uptime: i32,
|
||||
}
|
||||
|
||||
/// NodeGroup reputation model
|
||||
#[model]
|
||||
#[derive(Debug, Clone, Serialize, Deserialize, PartialEq, Default, CustomType)]
|
||||
pub struct NodeGroupReputation {
|
||||
pub base_data: BaseModelData,
|
||||
#[index]
|
||||
pub nodegroup_id: u32,
|
||||
/// between 0 and 100, earned over time
|
||||
pub reputation: i32,
|
||||
/// between 0 and 100, set by system, farmer has no ability to set this
|
||||
pub uptime: i32,
|
||||
pub nodes: Vec<NodeReputation>,
|
||||
}
|
||||
|
||||
impl NodeGroupReputation {
|
||||
pub fn new() -> Self {
|
||||
Self {
|
||||
base_data: BaseModelData::new(),
|
||||
nodegroup_id: 0,
|
||||
reputation: 50, // default as per spec
|
||||
uptime: 0,
|
||||
nodes: Vec::new(),
|
||||
}
|
||||
}
|
||||
|
||||
pub fn nodegroup_id(mut self, v: u32) -> Self {
|
||||
self.nodegroup_id = v;
|
||||
self
|
||||
}
|
||||
|
||||
pub fn reputation(mut self, v: i32) -> Self {
|
||||
self.reputation = v;
|
||||
self
|
||||
}
|
||||
|
||||
pub fn uptime(mut self, v: i32) -> Self {
|
||||
self.uptime = v;
|
||||
self
|
||||
}
|
||||
|
||||
pub fn add_node_reputation(mut self, node_rep: NodeReputation) -> Self {
|
||||
self.nodes.push(node_rep);
|
||||
self
|
||||
}
|
||||
}
|
||||
|
||||
impl NodeReputation {
|
||||
pub fn new() -> Self {
|
||||
Self {
|
||||
node_id: 0,
|
||||
reputation: 50, // default as per spec
|
||||
uptime: 0,
|
||||
}
|
||||
}
|
||||
|
||||
pub fn node_id(mut self, v: u32) -> Self {
|
||||
self.node_id = v;
|
||||
self
|
||||
}
|
||||
|
||||
pub fn reputation(mut self, v: i32) -> Self {
|
||||
self.reputation = v;
|
||||
self
|
||||
}
|
||||
|
||||
pub fn uptime(mut self, v: i32) -> Self {
|
||||
self.uptime = v;
|
||||
self
|
||||
}
|
||||
}
|
58
heromodels/src/models/grid4/reservation.rs
Normal file
58
heromodels/src/models/grid4/reservation.rs
Normal file
@@ -0,0 +1,58 @@
|
||||
use heromodels_core::BaseModelData;
|
||||
use heromodels_derive::model;
|
||||
use rhai::{CustomType, TypeBuilder};
|
||||
use serde::{Deserialize, Serialize};
|
||||
|
||||
/// Reservation status as per V spec
|
||||
#[derive(Debug, Clone, Serialize, Deserialize, PartialEq, Default)]
|
||||
pub enum ReservationStatus {
|
||||
#[default]
|
||||
Pending,
|
||||
Confirmed,
|
||||
Assigned,
|
||||
Cancelled,
|
||||
Done,
|
||||
}
|
||||
|
||||
/// Grid4 Reservation model
|
||||
#[model]
|
||||
#[derive(Debug, Clone, Serialize, Deserialize, PartialEq, Default, CustomType)]
|
||||
pub struct Reservation {
|
||||
pub base_data: BaseModelData,
|
||||
/// links back to customer for this capacity
|
||||
#[index]
|
||||
pub customer_id: u32,
|
||||
pub compute_slices: Vec<u32>,
|
||||
pub storage_slices: Vec<u32>,
|
||||
pub status: ReservationStatus,
|
||||
/// if obligation then will be charged and money needs to be in escrow, otherwise its an intent
|
||||
pub obligation: bool,
|
||||
/// epoch
|
||||
pub start_date: u32,
|
||||
pub end_date: u32,
|
||||
}
|
||||
|
||||
impl Reservation {
|
||||
pub fn new() -> Self {
|
||||
Self {
|
||||
base_data: BaseModelData::new(),
|
||||
customer_id: 0,
|
||||
compute_slices: Vec::new(),
|
||||
storage_slices: Vec::new(),
|
||||
status: ReservationStatus::Pending,
|
||||
obligation: false,
|
||||
start_date: 0,
|
||||
end_date: 0,
|
||||
}
|
||||
}
|
||||
|
||||
pub fn customer_id(mut self, v: u32) -> Self { self.customer_id = v; self }
|
||||
pub fn add_compute_slice(mut self, id: u32) -> Self { self.compute_slices.push(id); self }
|
||||
pub fn compute_slices(mut self, v: Vec<u32>) -> Self { self.compute_slices = v; self }
|
||||
pub fn add_storage_slice(mut self, id: u32) -> Self { self.storage_slices.push(id); self }
|
||||
pub fn storage_slices(mut self, v: Vec<u32>) -> Self { self.storage_slices = v; self }
|
||||
pub fn status(mut self, v: ReservationStatus) -> Self { self.status = v; self }
|
||||
pub fn obligation(mut self, v: bool) -> Self { self.obligation = v; self }
|
||||
pub fn start_date(mut self, v: u32) -> Self { self.start_date = v; self }
|
||||
pub fn end_date(mut self, v: u32) -> Self { self.end_date = v; self }
|
||||
}
|
194
heromodels/src/models/grid4/specs/README.md
Normal file
194
heromodels/src/models/grid4/specs/README.md
Normal file
@@ -0,0 +1,194 @@
|
||||
|
||||
# Grid4 Data Model
|
||||
|
||||
This module defines data models for nodes, groups, and slices in a cloud/grid infrastructure. Each root object is marked with `@[heap]` and can be indexed for efficient querying.
|
||||
|
||||
## Root Objects Overview
|
||||
|
||||
| Object | Description | Index Fields |
|
||||
| ----------- | --------------------------------------------- | ------------------------------ |
|
||||
| `Node` | Represents a single node in the grid | `id`, `nodegroupid`, `country` |
|
||||
| `NodeGroup` | Represents a group of nodes owned by a farmer | `id`, `farmerid` |
|
||||
|
||||
---
|
||||
|
||||
## Node
|
||||
|
||||
Represents a single node in the grid with slices, devices, and capacity.
|
||||
|
||||
| Field | Type | Description | Indexed |
|
||||
| --------------- | ---------------- | -------------------------------------------- | ------- |
|
||||
| `id` | `int` | Unique node ID | ✅ |
|
||||
| `nodegroupid` | `int` | ID of the owning node group | ✅ |
|
||||
| `uptime` | `int` | Uptime percentage (0-100) | ✅ |
|
||||
| `computeslices` | `[]ComputeSlice` | List of compute slices | ❌ |
|
||||
| `storageslices` | `[]StorageSlice` | List of storage slices | ❌ |
|
||||
| `devices` | `DeviceInfo` | Hardware device info (storage, memory, etc.) | ❌ |
|
||||
| `country` | `string` | 2-letter country code | ✅ |
|
||||
| `capacity` | `NodeCapacity` | Aggregated hardware capacity | ❌ |
|
||||
| `provisiontime` | `u32` | Provisioning time (simple/compatible format) | ✅ |
|
||||
|
||||
---
|
||||
|
||||
## NodeGroup
|
||||
|
||||
Represents a group of nodes owned by a farmer, with policies.
|
||||
|
||||
| Field | Type | Description | Indexed |
|
||||
| ------------------------------------- | --------------- | ---------------------------------------------- | ------- |
|
||||
| `id` | `u32` | Unique group ID | ✅ |
|
||||
| `farmerid` | `u32` | Farmer/user ID | ✅ |
|
||||
| `secret` | `string` | Encrypted secret for booting nodes | ❌ |
|
||||
| `description` | `string` | Group description | ❌ |
|
||||
| `slapolicy` | `SLAPolicy` | SLA policy details | ❌ |
|
||||
| `pricingpolicy` | `PricingPolicy` | Pricing policy details | ❌ |
|
||||
| `compute_slice_normalized_pricing_cc` | `f64` | Pricing per 2GB compute slice in cloud credits | ❌ |
|
||||
| `storage_slice_normalized_pricing_cc` | `f64` | Pricing per 1GB storage slice in cloud credits | ❌ |
|
||||
| `reputation` | `int` | Reputation (0-100) | ✅ |
|
||||
| `uptime` | `int` | Uptime (0-100) | ✅ |
|
||||
|
||||
---
|
||||
|
||||
## ComputeSlice
|
||||
|
||||
Represents a compute slice (e.g., 1GB memory unit).
|
||||
|
||||
| Field | Type | Description |
|
||||
| -------------------------- | --------------- | -------------------------------- |
|
||||
| `nodeid` | `u32` | Owning node ID |
|
||||
| `id` | `int` | Slice ID in node |
|
||||
| `mem_gb` | `f64` | Memory in GB |
|
||||
| `storage_gb` | `f64` | Storage in GB |
|
||||
| `passmark` | `int` | Passmark score |
|
||||
| `vcores` | `int` | Virtual cores |
|
||||
| `cpu_oversubscription` | `int` | CPU oversubscription ratio |
|
||||
| `storage_oversubscription` | `int` | Storage oversubscription ratio |
|
||||
| `price_range` | `[]f64` | Price range [min, max] |
|
||||
| `gpus` | `u8` | Number of GPUs |
|
||||
| `price_cc` | `f64` | Price per slice in cloud credits |
|
||||
| `pricing_policy` | `PricingPolicy` | Pricing policy |
|
||||
| `sla_policy` | `SLAPolicy` | SLA policy |
|
||||
|
||||
---
|
||||
|
||||
## StorageSlice
|
||||
|
||||
Represents a 1GB storage slice.
|
||||
|
||||
| Field | Type | Description |
|
||||
| ---------------- | --------------- | -------------------------------- |
|
||||
| `nodeid` | `u32` | Owning node ID |
|
||||
| `id` | `int` | Slice ID in node |
|
||||
| `price_cc` | `f64` | Price per slice in cloud credits |
|
||||
| `pricing_policy` | `PricingPolicy` | Pricing policy |
|
||||
| `sla_policy` | `SLAPolicy` | SLA policy |
|
||||
|
||||
---
|
||||
|
||||
## DeviceInfo
|
||||
|
||||
Hardware device information for a node.
|
||||
|
||||
| Field | Type | Description |
|
||||
| --------- | ----------------- | ----------------------- |
|
||||
| `vendor` | `string` | Vendor of the node |
|
||||
| `storage` | `[]StorageDevice` | List of storage devices |
|
||||
| `memory` | `[]MemoryDevice` | List of memory devices |
|
||||
| `cpu` | `[]CPUDevice` | List of CPU devices |
|
||||
| `gpu` | `[]GPUDevice` | List of GPU devices |
|
||||
| `network` | `[]NetworkDevice` | List of network devices |
|
||||
|
||||
---
|
||||
|
||||
## StorageDevice
|
||||
|
||||
| Field | Type | Description |
|
||||
| ------------- | -------- | --------------------- |
|
||||
| `id` | `string` | Unique ID for device |
|
||||
| `size_gb` | `f64` | Size in GB |
|
||||
| `description` | `string` | Description of device |
|
||||
|
||||
---
|
||||
|
||||
## MemoryDevice
|
||||
|
||||
| Field | Type | Description |
|
||||
| ------------- | -------- | --------------------- |
|
||||
| `id` | `string` | Unique ID for device |
|
||||
| `size_gb` | `f64` | Size in GB |
|
||||
| `description` | `string` | Description of device |
|
||||
|
||||
---
|
||||
|
||||
## CPUDevice
|
||||
|
||||
| Field | Type | Description |
|
||||
| ------------- | -------- | ------------------------ |
|
||||
| `id` | `string` | Unique ID for device |
|
||||
| `cores` | `int` | Number of CPU cores |
|
||||
| `passmark` | `int` | Passmark benchmark score |
|
||||
| `description` | `string` | Description of device |
|
||||
| `cpu_brand` | `string` | Brand of the CPU |
|
||||
| `cpu_version` | `string` | Version of the CPU |
|
||||
|
||||
---
|
||||
|
||||
## GPUDevice
|
||||
|
||||
| Field | Type | Description |
|
||||
| ------------- | -------- | --------------------- |
|
||||
| `id` | `string` | Unique ID for device |
|
||||
| `cores` | `int` | Number of GPU cores |
|
||||
| `memory_gb` | `f64` | GPU memory in GB |
|
||||
| `description` | `string` | Description of device |
|
||||
| `gpu_brand` | `string` | Brand of the GPU |
|
||||
| `gpu_version` | `string` | Version of the GPU |
|
||||
|
||||
---
|
||||
|
||||
## NetworkDevice
|
||||
|
||||
| Field | Type | Description |
|
||||
| ------------- | -------- | --------------------- |
|
||||
| `id` | `string` | Unique ID for device |
|
||||
| `speed_mbps` | `int` | Network speed in Mbps |
|
||||
| `description` | `string` | Description of device |
|
||||
|
||||
---
|
||||
|
||||
## NodeCapacity
|
||||
|
||||
Aggregated hardware capacity for a node.
|
||||
|
||||
| Field | Type | Description |
|
||||
| ------------ | ----- | ---------------------- |
|
||||
| `storage_gb` | `f64` | Total storage in GB |
|
||||
| `mem_gb` | `f64` | Total memory in GB |
|
||||
| `mem_gb_gpu` | `f64` | Total GPU memory in GB |
|
||||
| `passmark` | `int` | Total passmark score |
|
||||
| `vcores` | `int` | Total virtual cores |
|
||||
|
||||
---
|
||||
|
||||
## SLAPolicy
|
||||
|
||||
Service Level Agreement policy for slices or node groups.
|
||||
|
||||
| Field | Type | Description |
|
||||
| -------------------- | ----- | --------------------------------------- |
|
||||
| `sla_uptime` | `int` | Required uptime % (e.g., 90) |
|
||||
| `sla_bandwidth_mbit` | `int` | Guaranteed bandwidth in Mbps (0 = none) |
|
||||
| `sla_penalty` | `int` | Penalty % if SLA is breached (0-100) |
|
||||
|
||||
---
|
||||
|
||||
## PricingPolicy
|
||||
|
||||
Pricing policy for slices or node groups.
|
||||
|
||||
| Field | Type | Description |
|
||||
| ---------------------------- | ------- | --------------------------------------------------------- |
|
||||
| `marketplace_year_discounts` | `[]int` | Discounts for 1Y, 2Y, 3Y prepaid usage (e.g. [30,40,50]) |
|
||||
| `volume_discounts` | `[]int` | Volume discounts based on purchase size (e.g. [10,20,30]) |
|
||||
|
||||
|
37
heromodels/src/models/grid4/specs/model_bid.v
Normal file
37
heromodels/src/models/grid4/specs/model_bid.v
Normal file
@@ -0,0 +1,37 @@
|
||||
module datamodel
|
||||
|
||||
// I can bid for infra, and optionally get accepted
|
||||
@[heap]
|
||||
pub struct Bid {
|
||||
pub mut:
|
||||
id u32
|
||||
customer_id u32 // links back to customer for this capacity (user on ledger)
|
||||
compute_slices_nr int // nr of slices I need in 1 machine
|
||||
compute_slice_price f64 // price per 1 GB slice I want to accept
|
||||
storage_slices_nr int
|
||||
storage_slice_price f64 // price per 1 GB storage slice I want to accept
|
||||
storage_slices_nr int
|
||||
status BidStatus
|
||||
obligation bool // if obligation then will be charged and money needs to be in escrow, otherwise its an intent
|
||||
start_date u32 // epoch
|
||||
end_date u32
|
||||
signature_user string // signature as done by a user/consumer to validate their identity and intent
|
||||
billing_period BillingPeriod
|
||||
}
|
||||
|
||||
pub enum BidStatus {
|
||||
pending
|
||||
confirmed
|
||||
assigned
|
||||
cancelled
|
||||
done
|
||||
}
|
||||
|
||||
|
||||
pub enum BillingPeriod {
|
||||
hourly
|
||||
monthly
|
||||
yearly
|
||||
biannually
|
||||
triannually
|
||||
}
|
52
heromodels/src/models/grid4/specs/model_contract.v
Normal file
52
heromodels/src/models/grid4/specs/model_contract.v
Normal file
@@ -0,0 +1,52 @@
|
||||
module datamodel
|
||||
|
||||
// I can bid for infra, and optionally get accepted
|
||||
@[heap]
|
||||
pub struct Contract {
|
||||
pub mut:
|
||||
id u32
|
||||
customer_id u32 // links back to customer for this capacity (user on ledger)
|
||||
compute_slices []ComputeSliceProvisioned
|
||||
storage_slices []StorageSliceProvisioned
|
||||
compute_slice_price f64 // price per 1 GB agreed upon
|
||||
storage_slice_price f64 // price per 1 GB agreed upon
|
||||
network_slice_price f64 // price per 1 GB agreed upon (transfer)
|
||||
status ContractStatus
|
||||
start_date u32 // epoch
|
||||
end_date u32
|
||||
signature_user string // signature as done by a user/consumer to validate their identity and intent
|
||||
signature_hoster string // signature as done by the hoster
|
||||
billing_period BillingPeriod
|
||||
}
|
||||
|
||||
pub enum ConctractStatus {
|
||||
active
|
||||
cancelled
|
||||
error
|
||||
paused
|
||||
}
|
||||
|
||||
|
||||
// typically 1GB of memory, but can be adjusted based based on size of machine
|
||||
pub struct ComputeSliceProvisioned {
|
||||
pub mut:
|
||||
node_id u32
|
||||
id u16 // the id of the slice in the node
|
||||
mem_gb f64
|
||||
storage_gb f64
|
||||
passmark int
|
||||
vcores int
|
||||
cpu_oversubscription int
|
||||
tags string
|
||||
}
|
||||
|
||||
// 1GB of storage
|
||||
pub struct StorageSliceProvisioned {
|
||||
pub mut:
|
||||
node_id u32
|
||||
id u16 // the id of the slice in the node, are tracked in the node itself
|
||||
storage_size_gb int
|
||||
tags string
|
||||
}
|
||||
|
||||
|
104
heromodels/src/models/grid4/specs/model_node.v
Normal file
104
heromodels/src/models/grid4/specs/model_node.v
Normal file
@@ -0,0 +1,104 @@
|
||||
module datamodel
|
||||
|
||||
//ACCESS ONLY TF
|
||||
|
||||
@[heap]
|
||||
pub struct Node {
|
||||
pub mut:
|
||||
id int
|
||||
nodegroupid int
|
||||
uptime int // 0..100
|
||||
computeslices []ComputeSlice
|
||||
storageslices []StorageSlice
|
||||
devices DeviceInfo
|
||||
country string // 2 letter code as specified in lib/data/countries/data/countryInfo.txt, use that library for validation
|
||||
capacity NodeCapacity // Hardware capacity details
|
||||
birthtime u32 // first time node was active
|
||||
pubkey string
|
||||
signature_node string // signature done on node to validate pubkey with privkey
|
||||
signature_farmer string // signature as done by farmers to validate their identity
|
||||
}
|
||||
|
||||
pub struct DeviceInfo {
|
||||
pub mut:
|
||||
vendor string
|
||||
storage []StorageDevice
|
||||
memory []MemoryDevice
|
||||
cpu []CPUDevice
|
||||
gpu []GPUDevice
|
||||
network []NetworkDevice
|
||||
}
|
||||
|
||||
pub struct StorageDevice {
|
||||
pub mut:
|
||||
id string // can be used in node
|
||||
size_gb f64 // Size of the storage device in gigabytes
|
||||
description string // Description of the storage device
|
||||
}
|
||||
|
||||
pub struct MemoryDevice {
|
||||
pub mut:
|
||||
id string // can be used in node
|
||||
size_gb f64 // Size of the memory device in gigabytes
|
||||
description string // Description of the memory device
|
||||
}
|
||||
|
||||
pub struct CPUDevice {
|
||||
pub mut:
|
||||
id string // can be used in node
|
||||
cores int // Number of CPU cores
|
||||
passmark int
|
||||
description string // Description of the CPU
|
||||
cpu_brand string // Brand of the CPU
|
||||
cpu_version string // Version of the CPU
|
||||
}
|
||||
|
||||
pub struct GPUDevice {
|
||||
pub mut:
|
||||
id string // can be used in node
|
||||
cores int // Number of GPU cores
|
||||
memory_gb f64 // Size of the GPU memory in gigabytes
|
||||
description string // Description of the GPU
|
||||
gpu_brand string
|
||||
gpu_version string
|
||||
}
|
||||
|
||||
pub struct NetworkDevice {
|
||||
pub mut:
|
||||
id string // can be used in node
|
||||
speed_mbps int // Network speed in Mbps
|
||||
description string // Description of the network device
|
||||
}
|
||||
|
||||
// NodeCapacity represents the hardware capacity details of a node.
|
||||
pub struct NodeCapacity {
|
||||
pub mut:
|
||||
storage_gb f64 // Total storage in gigabytes
|
||||
mem_gb f64 // Total memory in gigabytes
|
||||
mem_gb_gpu f64 // Total GPU memory in gigabytes
|
||||
passmark int // Passmark score for the node
|
||||
vcores int // Total virtual cores
|
||||
}
|
||||
|
||||
// typically 1GB of memory, but can be adjusted based based on size of machine
|
||||
pub struct ComputeSlice {
|
||||
pub mut:
|
||||
u16 int // the id of the slice in the node
|
||||
mem_gb f64
|
||||
storage_gb f64
|
||||
passmark int
|
||||
vcores int
|
||||
cpu_oversubscription int
|
||||
storage_oversubscription int
|
||||
gpus u8 // nr of GPU's see node to know what GPU's are
|
||||
}
|
||||
|
||||
// 1GB of storage
|
||||
pub struct StorageSlice {
|
||||
pub mut:
|
||||
u16 int // the id of the slice in the node, are tracked in the node itself
|
||||
}
|
||||
|
||||
fn (mut n Node) check() ! {
|
||||
// todo calculate NodeCapacity out of the devices on the Node
|
||||
}
|
33
heromodels/src/models/grid4/specs/model_nodegroup.v
Normal file
33
heromodels/src/models/grid4/specs/model_nodegroup.v
Normal file
@@ -0,0 +1,33 @@
|
||||
module datamodel
|
||||
|
||||
// is a root object, is the only obj farmer needs to configure in the UI, this defines how slices will be created
|
||||
@[heap]
|
||||
pub struct NodeGroup {
|
||||
pub mut:
|
||||
id u32
|
||||
farmerid u32 // link back to farmer who owns the nodegroup, is a user?
|
||||
secret string // only visible by farmer, in future encrypted, used to boot a node
|
||||
description string
|
||||
slapolicy SLAPolicy
|
||||
pricingpolicy PricingPolicy
|
||||
compute_slice_normalized_pricing_cc f64 // pricing in CC - cloud credit, per 2GB node slice
|
||||
storage_slice_normalized_pricing_cc f64 // pricing in CC - cloud credit, per 1GB storage slice
|
||||
signature_farmer string // signature as done by farmers to validate that they created this group
|
||||
}
|
||||
|
||||
pub struct SLAPolicy {
|
||||
pub mut:
|
||||
sla_uptime int // should +90
|
||||
sla_bandwidth_mbit int // minimal mbits we can expect avg over 1h per node, 0 means we don't guarantee
|
||||
sla_penalty int // 0-100, percent of money given back in relation to month if sla breached, e.g. 200 means we return 2 months worth of rev if sla missed
|
||||
}
|
||||
|
||||
pub struct PricingPolicy {
|
||||
pub mut:
|
||||
marketplace_year_discounts []int = [30, 40, 50] // e.g. 30,40,50 means if user has more CC in wallet than 1 year utilization on all his purchaes then this provider gives 30%, 2Y 40%, ...
|
||||
// volume_discounts []int = [10, 20, 30] // e.g. 10,20,30
|
||||
}
|
||||
|
||||
|
||||
|
||||
|
19
heromodels/src/models/grid4/specs/model_reputation.v
Normal file
19
heromodels/src/models/grid4/specs/model_reputation.v
Normal file
@@ -0,0 +1,19 @@
|
||||
|
||||
@[heap]
|
||||
pub struct NodeGroupReputation {
|
||||
pub mut:
|
||||
nodegroup_id u32
|
||||
reputation int = 50 // between 0 and 100, earned over time
|
||||
uptime int // between 0 and 100, set by system, farmer has no ability to set this
|
||||
nodes []NodeReputation
|
||||
}
|
||||
|
||||
pub struct NodeReputation {
|
||||
pub mut:
|
||||
node_id u32
|
||||
reputation int = 50 // between 0 and 100, earned over time
|
||||
uptime int // between 0 and 100, set by system, farmer has no ability to set this
|
||||
}
|
||||
|
||||
|
||||
|
@@ -1,4 +1,4 @@
|
||||
use heromodels_core::{Model, BaseModelData, IndexKey};
|
||||
use heromodels_core::{BaseModelData, IndexKey, Model};
|
||||
use heromodels_derive::model;
|
||||
use serde::{Deserialize, Serialize};
|
||||
use std::collections::HashMap;
|
||||
@@ -209,10 +209,13 @@ pub struct DNSZone {
|
||||
pub base_data: BaseModelData,
|
||||
#[index]
|
||||
pub domain: String,
|
||||
#[index(path = "subdomain")]
|
||||
#[index(path = "record_type")]
|
||||
pub dnsrecords: Vec<DNSRecord>,
|
||||
pub administrators: Vec<u32>,
|
||||
pub status: DNSZoneStatus,
|
||||
pub metadata: HashMap<String, String>,
|
||||
#[index(path = "primary_ns")]
|
||||
pub soarecord: Vec<SOARecord>,
|
||||
}
|
||||
|
||||
@@ -297,5 +300,3 @@ impl DNSZone {
|
||||
self
|
||||
}
|
||||
}
|
||||
|
||||
|
||||
|
@@ -1,4 +1,4 @@
|
||||
use heromodels_core::{Model, BaseModelData, IndexKey};
|
||||
use heromodels_core::{BaseModelData, IndexKey, Model};
|
||||
use heromodels_derive::model;
|
||||
use serde::{Deserialize, Serialize};
|
||||
|
||||
@@ -184,8 +184,6 @@ impl Group {
|
||||
}
|
||||
}
|
||||
|
||||
|
||||
|
||||
/// Represents the membership relationship between users and groups
|
||||
#[model]
|
||||
#[derive(Debug, Clone, Serialize, Deserialize, PartialEq, Default)]
|
||||
@@ -232,5 +230,3 @@ impl UserGroupMembership {
|
||||
self
|
||||
}
|
||||
}
|
||||
|
||||
|
||||
|
@@ -1,4 +1,4 @@
|
||||
use heromodels_core::{Model, BaseModelData, IndexKey};
|
||||
use heromodels_core::{BaseModelData, IndexKey, Model};
|
||||
use heromodels_derive::model;
|
||||
use serde::{Deserialize, Serialize};
|
||||
|
||||
@@ -111,5 +111,3 @@ impl Member {
|
||||
self
|
||||
}
|
||||
}
|
||||
|
||||
|
||||
|
@@ -1,20 +1,10 @@
|
||||
// Export all heroledger model modules
|
||||
pub mod user;
|
||||
pub mod group;
|
||||
pub mod money;
|
||||
pub mod membership;
|
||||
pub mod dnsrecord;
|
||||
pub mod group;
|
||||
pub mod membership;
|
||||
pub mod money;
|
||||
pub mod rhai;
|
||||
pub mod secretbox;
|
||||
pub mod signature;
|
||||
pub mod user;
|
||||
pub mod user_kvs;
|
||||
pub mod rhai;
|
||||
|
||||
// Re-export key types for convenience
|
||||
pub use user::{User, UserStatus, UserProfile, KYCInfo, KYCStatus, SecretBox};
|
||||
pub use group::{Group, UserGroupMembership, GroupStatus, Visibility, GroupConfig};
|
||||
pub use money::{Account, Asset, AccountPolicy, AccountPolicyItem, Transaction, AccountStatus, TransactionType, Signature as TransactionSignature};
|
||||
pub use membership::{Member, MemberRole, MemberStatus};
|
||||
pub use dnsrecord::{DNSZone, DNSRecord, SOARecord, NameType, NameCat, DNSZoneStatus};
|
||||
pub use secretbox::{Notary, NotaryStatus, SecretBoxCategory};
|
||||
pub use signature::{Signature, SignatureStatus, ObjectType};
|
||||
pub use user_kvs::{UserKVS, UserKVSItem};
|
||||
|
@@ -1,4 +1,4 @@
|
||||
use heromodels_core::{Model, BaseModelData, IndexKey};
|
||||
use heromodels_core::{BaseModelData, IndexKey, Model};
|
||||
use heromodels_derive::model;
|
||||
use serde::{Deserialize, Serialize};
|
||||
use std::collections::HashMap;
|
||||
@@ -223,8 +223,6 @@ impl Account {
|
||||
}
|
||||
}
|
||||
|
||||
|
||||
|
||||
/// Represents an asset in the financial system
|
||||
#[model]
|
||||
#[derive(Debug, Clone, Serialize, Deserialize, PartialEq, Default)]
|
||||
@@ -342,8 +340,6 @@ impl Asset {
|
||||
}
|
||||
}
|
||||
|
||||
|
||||
|
||||
/// Represents account policies for various operations
|
||||
#[model]
|
||||
#[derive(Debug, Clone, Serialize, Deserialize, PartialEq, Default)]
|
||||
@@ -400,8 +396,6 @@ impl AccountPolicy {
|
||||
}
|
||||
}
|
||||
|
||||
|
||||
|
||||
/// Represents a financial transaction
|
||||
#[model]
|
||||
#[derive(Debug, Clone, Serialize, Deserialize, PartialEq, Default)]
|
||||
@@ -511,5 +505,3 @@ impl Transaction {
|
||||
self
|
||||
}
|
||||
}
|
||||
|
||||
|
||||
|
@@ -1,8 +1,13 @@
|
||||
use ::rhai::plugin::*;
|
||||
use ::rhai::{Array, Dynamic, Engine, EvalAltResult, Map, Module};
|
||||
use ::rhai::{Dynamic, Engine, EvalAltResult, Module};
|
||||
use std::mem;
|
||||
|
||||
use crate::models::heroledger::*;
|
||||
use crate::models::heroledger::{
|
||||
dnsrecord::DNSZone,
|
||||
group::{Group, Visibility},
|
||||
money::Account,
|
||||
user::{User, UserStatus},
|
||||
};
|
||||
|
||||
// ============================================================================
|
||||
// User Module
|
||||
@@ -12,6 +17,8 @@ type RhaiUser = User;
|
||||
|
||||
#[export_module]
|
||||
mod rhai_user_module {
|
||||
use crate::models::heroledger::user::User;
|
||||
|
||||
use super::RhaiUser;
|
||||
|
||||
#[rhai_fn(name = "new_user", return_raw)]
|
||||
@@ -30,30 +37,21 @@ mod rhai_user_module {
|
||||
}
|
||||
|
||||
#[rhai_fn(name = "add_email", return_raw)]
|
||||
pub fn add_email(
|
||||
user: &mut RhaiUser,
|
||||
email: String,
|
||||
) -> Result<RhaiUser, Box<EvalAltResult>> {
|
||||
pub fn add_email(user: &mut RhaiUser, email: String) -> Result<RhaiUser, Box<EvalAltResult>> {
|
||||
let owned = std::mem::take(user);
|
||||
*user = owned.add_email(email);
|
||||
Ok(user.clone())
|
||||
}
|
||||
|
||||
#[rhai_fn(name = "pubkey", return_raw)]
|
||||
pub fn set_pubkey(
|
||||
user: &mut RhaiUser,
|
||||
pubkey: String,
|
||||
) -> Result<RhaiUser, Box<EvalAltResult>> {
|
||||
pub fn set_pubkey(user: &mut RhaiUser, pubkey: String) -> Result<RhaiUser, Box<EvalAltResult>> {
|
||||
let owned = std::mem::take(user);
|
||||
*user = owned.pubkey(pubkey);
|
||||
Ok(user.clone())
|
||||
}
|
||||
|
||||
#[rhai_fn(name = "status", return_raw)]
|
||||
pub fn set_status(
|
||||
user: &mut RhaiUser,
|
||||
status: String,
|
||||
) -> Result<RhaiUser, Box<EvalAltResult>> {
|
||||
pub fn set_status(user: &mut RhaiUser, status: String) -> Result<RhaiUser, Box<EvalAltResult>> {
|
||||
let status_enum = match status.as_str() {
|
||||
"Active" => UserStatus::Active,
|
||||
"Inactive" => UserStatus::Inactive,
|
||||
@@ -115,10 +113,7 @@ mod rhai_group_module {
|
||||
}
|
||||
|
||||
#[rhai_fn(name = "name", return_raw)]
|
||||
pub fn set_name(
|
||||
group: &mut RhaiGroup,
|
||||
name: String,
|
||||
) -> Result<RhaiGroup, Box<EvalAltResult>> {
|
||||
pub fn set_name(group: &mut RhaiGroup, name: String) -> Result<RhaiGroup, Box<EvalAltResult>> {
|
||||
let owned = std::mem::take(group);
|
||||
*group = owned.name(name);
|
||||
Ok(group.clone())
|
||||
@@ -263,15 +258,11 @@ mod rhai_dns_zone_module {
|
||||
Ok(zone.clone())
|
||||
}
|
||||
|
||||
|
||||
|
||||
#[rhai_fn(name = "save_dns_zone", return_raw)]
|
||||
pub fn save_dns_zone(zone: &mut RhaiDNSZone) -> Result<RhaiDNSZone, Box<EvalAltResult>> {
|
||||
Ok(zone.clone())
|
||||
}
|
||||
|
||||
|
||||
|
||||
// Getters
|
||||
#[rhai_fn(name = "get_id")]
|
||||
pub fn get_id(zone: &mut RhaiDNSZone) -> i64 {
|
||||
|
@@ -1,4 +1,4 @@
|
||||
use heromodels_core::{Model, BaseModelData, IndexKey};
|
||||
use heromodels_core::{BaseModelData, IndexKey, Model};
|
||||
use heromodels_derive::model;
|
||||
use serde::{Deserialize, Serialize};
|
||||
|
||||
@@ -138,5 +138,3 @@ impl Notary {
|
||||
self
|
||||
}
|
||||
}
|
||||
|
||||
|
||||
|
@@ -1,4 +1,4 @@
|
||||
use heromodels_core::{Model, BaseModelData, IndexKey};
|
||||
use heromodels_core::{BaseModelData, IndexKey, Model};
|
||||
use heromodels_derive::model;
|
||||
use serde::{Deserialize, Serialize};
|
||||
|
||||
@@ -116,5 +116,3 @@ impl Signature {
|
||||
self
|
||||
}
|
||||
}
|
||||
|
||||
|
||||
|
@@ -1,4 +1,4 @@
|
||||
use heromodels_core::{Model, BaseModelData, IndexKey};
|
||||
use heromodels_core::{BaseModelData, IndexKey, Model};
|
||||
use heromodels_derive::model;
|
||||
use serde::{Deserialize, Serialize};
|
||||
use std::collections::HashMap;
|
||||
@@ -366,5 +366,3 @@ impl User {
|
||||
self
|
||||
}
|
||||
}
|
||||
|
||||
|
||||
|
@@ -1,7 +1,7 @@
|
||||
use heromodels_core::{Model, BaseModelData, IndexKey};
|
||||
use super::secretbox::SecretBox;
|
||||
use heromodels_core::{BaseModelData, IndexKey, Model};
|
||||
use heromodels_derive::model;
|
||||
use serde::{Deserialize, Serialize};
|
||||
use super::secretbox::SecretBox;
|
||||
|
||||
/// Represents a per-user key-value store
|
||||
#[model]
|
||||
@@ -44,8 +44,6 @@ impl UserKVS {
|
||||
}
|
||||
}
|
||||
|
||||
|
||||
|
||||
/// Represents an item in a user's key-value store
|
||||
#[model]
|
||||
#[derive(Debug, Clone, Serialize, Deserialize, PartialEq, Default)]
|
||||
@@ -116,5 +114,3 @@ impl UserKVSItem {
|
||||
self
|
||||
}
|
||||
}
|
||||
|
||||
|
||||
|
@@ -46,4 +46,4 @@ pub struct IdenfyVerificationData {
|
||||
pub doc_issuing_country: Option<String>,
|
||||
#[serde(rename = "manuallyDataChanged")]
|
||||
pub manually_data_changed: Option<bool>,
|
||||
}
|
||||
}
|
||||
|
@@ -2,4 +2,4 @@
|
||||
|
||||
pub mod kyc;
|
||||
|
||||
pub use kyc::*;
|
||||
pub use kyc::*;
|
||||
|
@@ -8,4 +8,4 @@ pub struct Address {
|
||||
pub postal_code: String,
|
||||
pub country: String,
|
||||
pub company: Option<String>,
|
||||
}
|
||||
}
|
||||
|
@@ -10,16 +10,16 @@ pub mod contact;
|
||||
pub mod finance;
|
||||
pub mod flow;
|
||||
pub mod governance;
|
||||
pub mod grid4;
|
||||
pub mod heroledger;
|
||||
pub mod identity;
|
||||
pub mod legal;
|
||||
pub mod library;
|
||||
pub mod location;
|
||||
pub mod object;
|
||||
pub mod projects;
|
||||
pub mod payment;
|
||||
pub mod identity;
|
||||
pub mod tfmarketplace;
|
||||
pub mod grid4;
|
||||
pub mod projects;
|
||||
// pub mod tfmarketplace;
|
||||
|
||||
// Re-export key types for convenience
|
||||
pub use core::Comment;
|
||||
@@ -39,3 +39,4 @@ pub use legal::{Contract, ContractRevision, ContractSigner, ContractStatus, Sign
|
||||
pub use library::collection::Collection;
|
||||
pub use library::items::{Image, Markdown, Pdf};
|
||||
pub use projects::{Project, Status};
|
||||
pub use heroledger::*;
|
@@ -1,6 +1,6 @@
|
||||
use super::Object;
|
||||
use rhai::plugin::*;
|
||||
use rhai::{CustomType, Dynamic, Engine, EvalAltResult, Module};
|
||||
use super::Object;
|
||||
|
||||
type RhaiObject = Object;
|
||||
|
||||
@@ -16,10 +16,7 @@ pub mod generated_rhai_module {
|
||||
|
||||
/// Set the title of an Object
|
||||
#[rhai_fn(name = "object_title")]
|
||||
pub fn object_title(
|
||||
object: &mut RhaiObject,
|
||||
title: String,
|
||||
) -> RhaiObject {
|
||||
pub fn object_title(object: &mut RhaiObject, title: String) -> RhaiObject {
|
||||
let mut result = object.clone();
|
||||
result.title = title;
|
||||
result
|
||||
@@ -27,10 +24,7 @@ pub mod generated_rhai_module {
|
||||
|
||||
/// Set the description of an Object
|
||||
#[rhai_fn(name = "object_description")]
|
||||
pub fn object_description(
|
||||
object: &mut RhaiObject,
|
||||
description: String,
|
||||
) -> RhaiObject {
|
||||
pub fn object_description(object: &mut RhaiObject, description: String) -> RhaiObject {
|
||||
let mut result = object.clone();
|
||||
result.description = description;
|
||||
result
|
||||
|
@@ -2,4 +2,4 @@
|
||||
|
||||
pub mod stripe;
|
||||
|
||||
pub use stripe::*;
|
||||
pub use stripe::*;
|
||||
|
@@ -27,4 +27,4 @@ pub struct StripeEventData {
|
||||
pub struct StripeEventRequest {
|
||||
pub id: Option<String>,
|
||||
pub idempotency_key: Option<String>,
|
||||
}
|
||||
}
|
||||
|
43
heromodels/test.sh
Executable file
43
heromodels/test.sh
Executable file
@@ -0,0 +1,43 @@
|
||||
#!/usr/bin/env bash
|
||||
set -euo pipefail
|
||||
|
||||
# Config matches examples/tests
|
||||
PGHOST=${PGHOST:-localhost}
|
||||
PGPORT=${PGPORT:-5432}
|
||||
PGUSER=${PGUSER:-postgres}
|
||||
PGPASSWORD=${PGPASSWORD:-test123}
|
||||
export PGPASSWORD
|
||||
|
||||
echo "[test.sh] Checking Postgres at ${PGHOST}:${PGPORT} (user=${PGUSER})..."
|
||||
|
||||
# Require pg_isready
|
||||
if ! command -v pg_isready >/dev/null 2>&1; then
|
||||
echo "[test.sh] ERROR: pg_isready not found. Install PostgreSQL client tools (e.g., brew install libpq && brew link --force libpq)." >&2
|
||||
exit 1
|
||||
fi
|
||||
|
||||
# Wait for Postgres to be ready (30s timeout)
|
||||
ATTEMPTS=30
|
||||
until pg_isready -h "$PGHOST" -p "$PGPORT" -U "$PGUSER" >/dev/null 2>&1; do
|
||||
((ATTEMPTS--)) || {
|
||||
echo "[test.sh] ERROR: Postgres not ready after 30s. Ensure it's running with user=$PGUSER password=$PGPASSWORD host=$PGHOST port=$PGPORT." >&2
|
||||
exit 1
|
||||
}
|
||||
sleep 1
|
||||
echo "[test.sh] Waiting for Postgres..."
|
||||
done
|
||||
|
||||
echo "[test.sh] Postgres is ready. Running tests..."
|
||||
|
||||
# Run fast OurDB test first (no Postgres dependency)
|
||||
echo "[test.sh] Running OurDB test: grid4_ourdb"
|
||||
cargo test -p heromodels --test grid4_ourdb
|
||||
|
||||
# Run Postgres-backed tests (marked ignored)
|
||||
echo "[test.sh] Running Postgres test: heroledger_postgres (ignored)"
|
||||
cargo test -p heromodels --test heroledger_postgres -- --ignored
|
||||
|
||||
echo "[test.sh] Running Postgres test: grid4_postgres (ignored)"
|
||||
cargo test -p heromodels --test grid4_postgres -- --ignored
|
||||
|
||||
echo "[test.sh] Done."
|
117
heromodels/tests/grid4_models.rs
Normal file
117
heromodels/tests/grid4_models.rs
Normal file
@@ -0,0 +1,117 @@
|
||||
use serde_json;
|
||||
|
||||
use heromodels::models::grid4::{
|
||||
ComputeSlice, DeviceInfo, Node, NodeCapacity, PricingPolicy, Reservation, ReservationStatus,
|
||||
SLAPolicy, StorageDevice, StorageSlice,
|
||||
};
|
||||
|
||||
#[test]
|
||||
fn build_and_serde_roundtrip_compute_storage_slices() {
|
||||
let pricing = PricingPolicy::new()
|
||||
.marketplace_year_discounts(vec![20, 30, 40])
|
||||
.volume_discounts(vec![5, 10, 15])
|
||||
.build();
|
||||
|
||||
let sla = SLAPolicy::new()
|
||||
.sla_uptime(99)
|
||||
.sla_bandwidth_mbit(1000)
|
||||
.sla_penalty(150)
|
||||
.build();
|
||||
|
||||
let cs = ComputeSlice::new()
|
||||
.nodeid(42)
|
||||
.slice_id(1)
|
||||
.mem_gb(16.0)
|
||||
.storage_gb(200.0)
|
||||
.passmark(5000)
|
||||
.vcores(8)
|
||||
.cpu_oversubscription(2)
|
||||
.storage_oversubscription(1)
|
||||
.price_range(vec![0.5, 2.0])
|
||||
.gpus(1)
|
||||
.price_cc(1.25)
|
||||
.pricing_policy(pricing.clone())
|
||||
.sla_policy(sla.clone());
|
||||
|
||||
let ss = StorageSlice::new()
|
||||
.nodeid(42)
|
||||
.slice_id(2)
|
||||
.price_cc(0.15)
|
||||
.pricing_policy(pricing)
|
||||
.sla_policy(sla);
|
||||
|
||||
// serde roundtrip compute slice
|
||||
let s = serde_json::to_string(&cs).expect("serialize compute slice");
|
||||
let cs2: ComputeSlice = serde_json::from_str(&s).expect("deserialize compute slice");
|
||||
assert_eq!(cs, cs2);
|
||||
|
||||
// serde roundtrip storage slice
|
||||
let s2 = serde_json::to_string(&ss).expect("serialize storage slice");
|
||||
let ss2: StorageSlice = serde_json::from_str(&s2).expect("deserialize storage slice");
|
||||
assert_eq!(ss, ss2);
|
||||
}
|
||||
|
||||
#[test]
|
||||
fn build_and_serde_roundtrip_node() {
|
||||
let dev = DeviceInfo {
|
||||
vendor: "AcmeVendor".into(),
|
||||
storage: vec![StorageDevice { id: "sda".into(), size_gb: 512.0, description: "NVMe".into() }],
|
||||
memory: vec![],
|
||||
cpu: vec![],
|
||||
gpu: vec![],
|
||||
network: vec![],
|
||||
};
|
||||
|
||||
let cap = NodeCapacity { storage_gb: 2048.0, mem_gb: 128.0, mem_gb_gpu: 24.0, passmark: 12000, vcores: 32 };
|
||||
|
||||
let cs = ComputeSlice::new().nodeid(1).slice_id(1).mem_gb(8.0).storage_gb(100.0).passmark(2500).vcores(4);
|
||||
let ss = StorageSlice::new().nodeid(1).slice_id(2).price_cc(0.2);
|
||||
|
||||
let node = Node::new()
|
||||
.nodegroupid(7)
|
||||
.uptime(99)
|
||||
.add_compute_slice(cs)
|
||||
.add_storage_slice(ss)
|
||||
.devices(dev)
|
||||
.country("NL")
|
||||
.capacity(cap)
|
||||
.provisiontime(1710000000)
|
||||
.pubkey("node_pubkey")
|
||||
.signature_node("sig_node")
|
||||
.signature_farmer("sig_farmer");
|
||||
|
||||
let s = serde_json::to_string(&node).expect("serialize node");
|
||||
let node2: Node = serde_json::from_str(&s).expect("deserialize node");
|
||||
|
||||
assert_eq!(node.nodegroupid, node2.nodegroupid);
|
||||
assert_eq!(node.uptime, node2.uptime);
|
||||
assert_eq!(node.country, node2.country);
|
||||
assert_eq!(node.pubkey, node2.pubkey);
|
||||
assert_eq!(node.signature_node, node2.signature_node);
|
||||
assert_eq!(node.signature_farmer, node2.signature_farmer);
|
||||
assert_eq!(node.computeslices.len(), node2.computeslices.len());
|
||||
assert_eq!(node.storageslices.len(), node2.storageslices.len());
|
||||
}
|
||||
|
||||
#[test]
|
||||
fn build_and_serde_roundtrip_reservation() {
|
||||
let reservation = Reservation::new()
|
||||
.customer_id(1234)
|
||||
.add_compute_slice(11)
|
||||
.add_storage_slice(22)
|
||||
.status(ReservationStatus::Confirmed)
|
||||
.obligation(true)
|
||||
.start_date(1_710_000_000)
|
||||
.end_date(1_720_000_000);
|
||||
|
||||
let s = serde_json::to_string(&reservation).expect("serialize reservation");
|
||||
let reservation2: Reservation = serde_json::from_str(&s).expect("deserialize reservation");
|
||||
|
||||
assert_eq!(reservation.customer_id, reservation2.customer_id);
|
||||
assert_eq!(reservation.status, reservation2.status);
|
||||
assert_eq!(reservation.obligation, reservation2.obligation);
|
||||
assert_eq!(reservation.start_date, reservation2.start_date);
|
||||
assert_eq!(reservation.end_date, reservation2.end_date);
|
||||
assert_eq!(reservation.compute_slices, reservation2.compute_slices);
|
||||
assert_eq!(reservation.storage_slices, reservation2.storage_slices);
|
||||
}
|
82
heromodels/tests/grid4_ourdb.rs
Normal file
82
heromodels/tests/grid4_ourdb.rs
Normal file
@@ -0,0 +1,82 @@
|
||||
use heromodels::db::hero::OurDB;
|
||||
use heromodels::db::{Collection, Db};
|
||||
use heromodels::models::grid4::node::node_index::{country, nodegroupid, pubkey};
|
||||
use heromodels::models::grid4::node::{ComputeSlice, DeviceInfo, Node};
|
||||
use heromodels_core::Model;
|
||||
use std::sync::Arc;
|
||||
|
||||
fn create_test_db() -> Arc<OurDB> {
|
||||
let ts = std::time::SystemTime::now()
|
||||
.duration_since(std::time::UNIX_EPOCH)
|
||||
.unwrap()
|
||||
.as_nanos();
|
||||
let path = format!("/tmp/grid4_node_test_{}", ts);
|
||||
let _ = std::fs::remove_dir_all(&path);
|
||||
Arc::new(OurDB::new(path, true).expect("create OurDB"))
|
||||
}
|
||||
|
||||
#[test]
|
||||
fn grid4_node_basic_roundtrip_and_indexes() {
|
||||
let db = create_test_db();
|
||||
let nodes = db.collection::<Node>().expect("open node collection");
|
||||
|
||||
// Clean any leftover
|
||||
if let Ok(existing) = nodes.get_all() {
|
||||
for n in existing {
|
||||
let _ = nodes.delete_by_id(n.get_id());
|
||||
}
|
||||
}
|
||||
|
||||
// Build a node with some compute slices and device info
|
||||
let cs = ComputeSlice::new()
|
||||
.nodeid(1)
|
||||
.slice_id(1)
|
||||
.mem_gb(32.0)
|
||||
.storage_gb(512.0)
|
||||
.passmark(5000)
|
||||
.vcores(16)
|
||||
.gpus(1)
|
||||
.price_cc(0.25);
|
||||
|
||||
let dev = DeviceInfo {
|
||||
vendor: "ACME".into(),
|
||||
..Default::default()
|
||||
};
|
||||
|
||||
let n = Node::new()
|
||||
.nodegroupid(42)
|
||||
.uptime(99)
|
||||
.add_compute_slice(cs)
|
||||
.devices(dev)
|
||||
.country("BE")
|
||||
.pubkey("PUB_NODE_1")
|
||||
.build();
|
||||
|
||||
let (id, stored) = nodes.set(&n).expect("store node");
|
||||
assert!(id > 0);
|
||||
assert_eq!(stored.country, "BE");
|
||||
|
||||
// get by id
|
||||
let fetched = nodes.get_by_id(id).expect("get by id").expect("exists");
|
||||
assert_eq!(fetched.pubkey, "PUB_NODE_1");
|
||||
|
||||
// query by top-level indexes
|
||||
let by_country = nodes.get::<country, _>("BE").expect("query country");
|
||||
assert_eq!(by_country.len(), 1);
|
||||
assert_eq!(by_country[0].get_id(), id);
|
||||
|
||||
let by_group = nodes.get::<nodegroupid, _>(&42).expect("query group");
|
||||
assert_eq!(by_group.len(), 1);
|
||||
|
||||
let by_pubkey = nodes.get::<pubkey, _>("PUB_NODE_1").expect("query pubkey");
|
||||
assert_eq!(by_pubkey.len(), 1);
|
||||
|
||||
// update
|
||||
let updated = fetched.clone().country("NL");
|
||||
let (_, back) = nodes.set(&updated).expect("update node");
|
||||
assert_eq!(back.country, "NL");
|
||||
|
||||
// delete
|
||||
nodes.delete_by_id(id).expect("delete");
|
||||
assert!(nodes.get_by_id(id).expect("get after delete").is_none());
|
||||
}
|
125
heromodels/tests/grid4_postgres.rs
Normal file
125
heromodels/tests/grid4_postgres.rs
Normal file
@@ -0,0 +1,125 @@
|
||||
use heromodels::db::postgres::{Config, Postgres};
|
||||
use heromodels::db::{Collection, Db};
|
||||
use heromodels::models::grid4::node::node_index::{country, nodegroupid, pubkey};
|
||||
use heromodels::models::grid4::node::{ComputeSlice, DeviceInfo, Node};
|
||||
use heromodels_core::Model;
|
||||
|
||||
// Requires local Postgres (user=postgres password=test123 host=localhost port=5432)
|
||||
// Run with: cargo test -p heromodels --test grid4_postgres -- --ignored
|
||||
#[test]
|
||||
#[ignore]
|
||||
fn grid4_node_postgres_roundtrip_like_example() {
|
||||
let db = Postgres::new(
|
||||
Config::new()
|
||||
.user(Some("postgres".into()))
|
||||
.password(Some("test123".into()))
|
||||
.host(Some("localhost".into()))
|
||||
.port(Some(5432)),
|
||||
)
|
||||
.expect("can connect to Postgres");
|
||||
|
||||
let nodes = db.collection::<Node>().expect("open node collection");
|
||||
|
||||
// Clean existing
|
||||
if let Ok(existing) = nodes.get_all() {
|
||||
for n in existing {
|
||||
let _ = nodes.delete_by_id(n.get_id());
|
||||
}
|
||||
}
|
||||
|
||||
// Build and store multiple nodes via builder and then persist via collection.set(), like examples
|
||||
let cs1 = ComputeSlice::new()
|
||||
.nodeid(10)
|
||||
.slice_id(1)
|
||||
.mem_gb(32.0)
|
||||
.storage_gb(512.0)
|
||||
.passmark(5000)
|
||||
.vcores(16)
|
||||
.gpus(1)
|
||||
.price_cc(0.25);
|
||||
let cs2 = ComputeSlice::new()
|
||||
.nodeid(10)
|
||||
.slice_id(2)
|
||||
.mem_gb(64.0)
|
||||
.storage_gb(2048.0)
|
||||
.passmark(7000)
|
||||
.vcores(24)
|
||||
.gpus(2)
|
||||
.price_cc(0.50);
|
||||
let cs3 = ComputeSlice::new()
|
||||
.nodeid(11)
|
||||
.slice_id(1)
|
||||
.mem_gb(16.0)
|
||||
.storage_gb(256.0)
|
||||
.passmark(3000)
|
||||
.vcores(8)
|
||||
.gpus(0)
|
||||
.price_cc(0.10);
|
||||
|
||||
let dev = DeviceInfo { vendor: "ACME".into(), ..Default::default() };
|
||||
|
||||
let n1 = Node::new()
|
||||
.nodegroupid(99)
|
||||
.uptime(97)
|
||||
.add_compute_slice(cs1)
|
||||
.devices(dev.clone())
|
||||
.country("BE")
|
||||
.pubkey("PG_NODE_1")
|
||||
.build();
|
||||
let n2 = Node::new()
|
||||
.nodegroupid(99)
|
||||
.uptime(96)
|
||||
.add_compute_slice(cs2)
|
||||
.devices(dev.clone())
|
||||
.country("NL")
|
||||
.pubkey("PG_NODE_2")
|
||||
.build();
|
||||
let n3 = Node::new()
|
||||
.nodegroupid(7)
|
||||
.uptime(95)
|
||||
.add_compute_slice(cs3)
|
||||
.devices(dev)
|
||||
.country("BE")
|
||||
.pubkey("PG_NODE_3")
|
||||
.build();
|
||||
|
||||
let (id1, s1) = nodes.set(&n1).expect("store n1");
|
||||
let (id2, s2) = nodes.set(&n2).expect("store n2");
|
||||
let (id3, s3) = nodes.set(&n3).expect("store n3");
|
||||
assert!(id1 > 0 && id2 > 0 && id3 > 0);
|
||||
|
||||
// Query by top-level indexes similar to the example style
|
||||
let be_nodes = nodes.get::<country, _>("BE").expect("by country");
|
||||
assert_eq!(be_nodes.len(), 2);
|
||||
|
||||
let grp_99 = nodes.get::<nodegroupid, _>(&99).expect("by group");
|
||||
assert_eq!(grp_99.len(), 2);
|
||||
|
||||
let by_key = nodes.get::<pubkey, _>("PG_NODE_2").expect("by pubkey");
|
||||
assert_eq!(by_key.len(), 1);
|
||||
assert_eq!(by_key[0].get_id(), id2);
|
||||
|
||||
// Update: change country of n1
|
||||
let updated = s1.clone().country("DE");
|
||||
let (_, back) = nodes.set(&updated).expect("update n1");
|
||||
assert_eq!(back.country, "DE");
|
||||
|
||||
// Cardinality after update
|
||||
let de_nodes = nodes.get::<country, _>("DE").expect("by country DE");
|
||||
assert_eq!(de_nodes.len(), 1);
|
||||
|
||||
// Delete by id and by index
|
||||
nodes.delete_by_id(id2).expect("delete n2 by id");
|
||||
assert!(nodes.get_by_id(id2).unwrap().is_none());
|
||||
|
||||
nodes.delete::<pubkey, _>("PG_NODE_3").expect("delete n3 by pubkey");
|
||||
assert!(nodes.get_by_id(id3).unwrap().is_none());
|
||||
|
||||
// Remaining should be updated n1 only; verify via targeted queries
|
||||
let de_nodes = nodes.get::<country, _>("DE").expect("country DE after deletes");
|
||||
assert_eq!(de_nodes.len(), 1);
|
||||
assert_eq!(de_nodes[0].get_id(), id1);
|
||||
let by_key = nodes.get::<pubkey, _>("PG_NODE_1").expect("by pubkey PG_NODE_1");
|
||||
assert_eq!(by_key.len(), 1);
|
||||
assert_eq!(by_key[0].get_id(), id1);
|
||||
}
|
97
heromodels/tests/heroledger_postgres.rs
Normal file
97
heromodels/tests/heroledger_postgres.rs
Normal file
@@ -0,0 +1,97 @@
|
||||
use heromodels::db::postgres::{Config, Postgres};
|
||||
use heromodels::db::{Collection, Db};
|
||||
use heromodels::models::heroledger::user::user_index::username;
|
||||
use heromodels::models::heroledger::user::User;
|
||||
use heromodels_core::Model;
|
||||
|
||||
// NOTE: Requires a local Postgres running with user=postgres password=test123 host=localhost port=5432
|
||||
// Marked ignored by default. Run with: cargo test -p heromodels --test heroledger_postgres -- --ignored
|
||||
#[test]
|
||||
#[ignore]
|
||||
fn heroledger_user_postgres_roundtrip() {
|
||||
// Connect
|
||||
let db = Postgres::new(
|
||||
Config::new()
|
||||
.user(Some("postgres".into()))
|
||||
.password(Some("test123".into()))
|
||||
.host(Some("localhost".into()))
|
||||
.port(Some(5432)),
|
||||
)
|
||||
.expect("can connect to Postgres");
|
||||
|
||||
// Open collection (will create table and indexes for top-level fields)
|
||||
let users = db.collection::<User>().expect("can open user collection");
|
||||
|
||||
// Clean slate
|
||||
if let Ok(existing) = users.get_all() {
|
||||
for u in existing {
|
||||
let _ = users.delete_by_id(u.get_id());
|
||||
}
|
||||
}
|
||||
|
||||
// Unique suffix to avoid collisions with any pre-existing rows
|
||||
let uniq = format!("{}", std::time::SystemTime::now()
|
||||
.duration_since(std::time::UNIX_EPOCH)
|
||||
.unwrap()
|
||||
.as_nanos());
|
||||
let alice = format!("alice_{}", uniq);
|
||||
let bob = format!("bob_{}", uniq);
|
||||
let carol = format!("carol_{}", uniq);
|
||||
|
||||
// Build and store multiple users
|
||||
let u1 = User::new(0)
|
||||
.username(&alice)
|
||||
.pubkey("PUBKEY_A")
|
||||
.add_email("alice@example.com")
|
||||
.build();
|
||||
let u2 = User::new(0)
|
||||
.username(&bob)
|
||||
.pubkey("PUBKEY_B")
|
||||
.add_email("bob@example.com")
|
||||
.build();
|
||||
let u3 = User::new(0)
|
||||
.username(&carol)
|
||||
.pubkey("PUBKEY_C")
|
||||
.add_email("carol@example.com")
|
||||
.build();
|
||||
|
||||
let (id1, db_u1) = users.set(&u1).expect("store u1");
|
||||
let (id2, db_u2) = users.set(&u2).expect("store u2");
|
||||
let (id3, db_u3) = users.set(&u3).expect("store u3");
|
||||
assert!(id1 > 0 && id2 > 0 && id3 > 0);
|
||||
|
||||
// Fetch by id
|
||||
assert_eq!(users.get_by_id(id1).unwrap().unwrap().username, alice);
|
||||
assert_eq!(users.get_by_id(id2).unwrap().unwrap().username, bob);
|
||||
assert_eq!(users.get_by_id(id3).unwrap().unwrap().username, carol);
|
||||
|
||||
// Fetch by index (top-level username)
|
||||
let by_username = users.get::<username, _>(&alice).expect("by username");
|
||||
assert_eq!(by_username.len(), 1);
|
||||
assert_eq!(by_username[0].get_id(), id1);
|
||||
|
||||
// Update one
|
||||
let updated = db_u1.clone().add_email("work@alice.example");
|
||||
let (id1b, updated_back) = users.set(&updated).expect("update alice");
|
||||
assert_eq!(id1b, id1);
|
||||
assert!(updated_back.email.len() >= 2);
|
||||
|
||||
// Targeted queries to avoid legacy rows in the same table
|
||||
// Verify three users exist via index queries
|
||||
assert_eq!(users.get::<username, _>(&alice).unwrap().len(), 1);
|
||||
assert_eq!(users.get::<username, _>(&bob).unwrap().len(), 1);
|
||||
assert_eq!(users.get::<username, _>(&carol).unwrap().len(), 1);
|
||||
|
||||
// Delete by id
|
||||
users.delete_by_id(id2).expect("delete bob by id");
|
||||
assert!(users.get_by_id(id2).unwrap().is_none());
|
||||
|
||||
// Delete by index (username)
|
||||
users.delete::<username, _>(&carol).expect("delete carol by username");
|
||||
assert!(users.get_by_id(id3).unwrap().is_none());
|
||||
|
||||
// Remaining should be just alice; verify via index
|
||||
let remain = users.get::<username, _>(&alice).expect("get alice after delete");
|
||||
assert_eq!(remain.len(), 1);
|
||||
assert_eq!(remain[0].get_id(), id1);
|
||||
}
|
@@ -1,4 +1,5 @@
|
||||
use heromodels::db::Collection;
|
||||
use heromodels::db::Db;
|
||||
use heromodels::db::hero::OurDB;
|
||||
use heromodels::models::biz::{BusinessType, Company, CompanyStatus, Payment, PaymentStatus};
|
||||
use heromodels_core::Model;
|
||||
@@ -197,12 +198,18 @@ fn test_payment_database_persistence() {
|
||||
);
|
||||
|
||||
// Save payment
|
||||
let (payment_id, saved_payment) = db.set(&payment).expect("Failed to save payment");
|
||||
let (payment_id, saved_payment) = db
|
||||
.collection::<Payment>()
|
||||
.expect("open payment collection")
|
||||
.set(&payment)
|
||||
.expect("Failed to save payment");
|
||||
assert!(payment_id > 0);
|
||||
assert_eq!(saved_payment.payment_intent_id, "pi_db_test");
|
||||
|
||||
// Retrieve payment
|
||||
let retrieved_payment: Payment = db
|
||||
.collection::<Payment>()
|
||||
.expect("open payment collection")
|
||||
.get_by_id(payment_id)
|
||||
.expect("Failed to get payment")
|
||||
.unwrap();
|
||||
@@ -224,20 +231,34 @@ fn test_payment_status_transitions() {
|
||||
1360.0,
|
||||
);
|
||||
|
||||
let (payment_id, mut payment) = db.set(&payment).expect("Failed to save payment");
|
||||
let (payment_id, mut payment) = db
|
||||
.collection::<Payment>()
|
||||
.expect("open payment collection")
|
||||
.set(&payment)
|
||||
.expect("Failed to save payment");
|
||||
|
||||
// Test pending -> completed
|
||||
payment = payment.complete_payment(Some("cus_transition_test".to_string()));
|
||||
let (_, mut payment) = db.set(&payment).expect("Failed to update payment");
|
||||
let (_, mut payment) = db
|
||||
.collection::<Payment>()
|
||||
.expect("open payment collection")
|
||||
.set(&payment)
|
||||
.expect("Failed to update payment");
|
||||
assert!(payment.is_completed());
|
||||
|
||||
// Test completed -> refunded
|
||||
payment = payment.refund_payment();
|
||||
let (_, payment) = db.set(&payment).expect("Failed to update payment");
|
||||
let (_, payment) = db
|
||||
.collection::<Payment>()
|
||||
.expect("open payment collection")
|
||||
.set(&payment)
|
||||
.expect("Failed to update payment");
|
||||
assert!(payment.is_refunded());
|
||||
|
||||
// Verify final state in database
|
||||
let final_payment: Payment = db
|
||||
.collection::<Payment>()
|
||||
.expect("open payment collection")
|
||||
.get_by_id(payment_id)
|
||||
.expect("Failed to get payment")
|
||||
.unwrap();
|
||||
@@ -270,15 +291,18 @@ fn test_company_payment_integration() {
|
||||
let db = create_test_db();
|
||||
|
||||
// Create company with default PendingPayment status
|
||||
let company = Company::new(
|
||||
"Integration Test Corp".to_string(),
|
||||
"ITC-001".to_string(),
|
||||
chrono::Utc::now().timestamp(),
|
||||
)
|
||||
.email("test@integration.com".to_string())
|
||||
.business_type(BusinessType::Starter);
|
||||
let company = Company::new()
|
||||
.name("Integration Test Corp")
|
||||
.registration_number("ITC-001")
|
||||
.incorporation_date(chrono::Utc::now().timestamp())
|
||||
.email("test@integration.com")
|
||||
.business_type(BusinessType::Starter);
|
||||
|
||||
let (company_id, company) = db.set(&company).expect("Failed to save company");
|
||||
let (company_id, company) = db
|
||||
.collection::<Company>()
|
||||
.expect("open company collection")
|
||||
.set(&company)
|
||||
.expect("Failed to save company");
|
||||
assert_eq!(company.status, CompanyStatus::PendingPayment);
|
||||
|
||||
// Create payment for the company
|
||||
@@ -291,18 +315,28 @@ fn test_company_payment_integration() {
|
||||
305.0,
|
||||
);
|
||||
|
||||
let (_payment_id, payment) = db.set(&payment).expect("Failed to save payment");
|
||||
let (_payment_id, payment) = db
|
||||
.collection::<Payment>()
|
||||
.expect("open payment collection")
|
||||
.set(&payment)
|
||||
.expect("Failed to save payment");
|
||||
assert_eq!(payment.company_id, company_id);
|
||||
|
||||
// Complete payment
|
||||
let completed_payment = payment.complete_payment(Some("cus_integration_test".to_string()));
|
||||
let (_, completed_payment) = db
|
||||
.collection::<Payment>()
|
||||
.expect("open payment collection")
|
||||
.set(&completed_payment)
|
||||
.expect("Failed to update payment");
|
||||
|
||||
// Update company status to Active
|
||||
let active_company = company.status(CompanyStatus::Active);
|
||||
let (_, active_company) = db.set(&active_company).expect("Failed to update company");
|
||||
let (_, active_company) = db
|
||||
.collection::<Company>()
|
||||
.expect("open company collection")
|
||||
.set(&active_company)
|
||||
.expect("Failed to update company");
|
||||
|
||||
// Verify final states
|
||||
assert!(completed_payment.is_completed());
|
||||
|
Reference in New Issue
Block a user