Compare commits
	
		
			10 Commits
		
	
	
		
			main
			...
			developmen
		
	
	| Author | SHA1 | Date | |
|---|---|---|---|
|  | 53e9a2d4f0 | ||
| cb1fb0f0ec | |||
|  | a7c978efd4 | ||
|  | 0b0d546b4e | ||
|  | 2f5e18df98 | ||
|  | 77169c073c | ||
|  | ce12f26a91 | ||
|  | 130822b69b | ||
|  | 7439980b33 | ||
|  | 453e86edd2 | 
							
								
								
									
										114
									
								
								Cargo.lock
									
									
									
										generated
									
									
									
								
							
							
						
						
									
										114
									
								
								Cargo.lock
									
									
									
										generated
									
									
									
								
							| @@ -65,13 +65,13 @@ checksum = "7c02d123df017efcdfbd739ef81735b36c5ba83ec3c59c80a9d7ecc718f92e50" | ||||
|  | ||||
| [[package]] | ||||
| name = "async-trait" | ||||
| version = "0.1.88" | ||||
| version = "0.1.89" | ||||
| source = "registry+https://github.com/rust-lang/crates.io-index" | ||||
| checksum = "e539d3fca749fcee5236ab05e93a52867dd549cc157c8cb7f99595f3cedffdb5" | ||||
| checksum = "9035ad2d096bed7955a320ee7e2230574d28fd3c3a0f186cbea1ff3c7eed5dbb" | ||||
| dependencies = [ | ||||
|  "proc-macro2", | ||||
|  "quote", | ||||
|  "syn 2.0.104", | ||||
|  "syn 2.0.106", | ||||
| ] | ||||
|  | ||||
| [[package]] | ||||
| @@ -123,9 +123,9 @@ dependencies = [ | ||||
|  | ||||
| [[package]] | ||||
| name = "bitflags" | ||||
| version = "2.9.1" | ||||
| version = "2.9.2" | ||||
| source = "registry+https://github.com/rust-lang/crates.io-index" | ||||
| checksum = "1b8e56985ec62d17e9c1001dc89c88ecd7dc08e47eba5ec7c29c7b5eeecde967" | ||||
| checksum = "6a65b545ab31d687cff52899d4890855fec459eb6afe0da6417b8a18da87aa29" | ||||
|  | ||||
| [[package]] | ||||
| name = "bitvec" | ||||
| @@ -168,7 +168,7 @@ dependencies = [ | ||||
|  "proc-macro-crate", | ||||
|  "proc-macro2", | ||||
|  "quote", | ||||
|  "syn 2.0.104", | ||||
|  "syn 2.0.106", | ||||
| ] | ||||
|  | ||||
| [[package]] | ||||
| @@ -213,18 +213,18 @@ checksum = "d71b6127be86fdcfddb610f7182ac57211d4b18a3e9c82eb2d17662f2227ad6a" | ||||
|  | ||||
| [[package]] | ||||
| name = "cc" | ||||
| version = "1.2.31" | ||||
| version = "1.2.33" | ||||
| source = "registry+https://github.com/rust-lang/crates.io-index" | ||||
| checksum = "c3a42d84bb6b69d3a8b3eaacf0d88f179e1929695e1ad012b6cf64d9caaa5fd2" | ||||
| checksum = "3ee0f8803222ba5a7e2777dd72ca451868909b1ac410621b676adf07280e9b5f" | ||||
| dependencies = [ | ||||
|  "shlex", | ||||
| ] | ||||
|  | ||||
| [[package]] | ||||
| name = "cfg-if" | ||||
| version = "1.0.1" | ||||
| version = "1.0.3" | ||||
| source = "registry+https://github.com/rust-lang/crates.io-index" | ||||
| checksum = "9555578bc9e57714c812a1f84e4fc5b4d21fcb063490c624de019f7464c91268" | ||||
| checksum = "2fd1289c04a9ea8cb22300a459a72a385d7c73d3259e2ed7dcb2af674838cfa9" | ||||
|  | ||||
| [[package]] | ||||
| name = "cfg_aliases" | ||||
| @@ -372,7 +372,7 @@ checksum = "162ee34ebcb7c64a8abebc059ce0fee27c2262618d7b60ed8faf72fef13c3650" | ||||
| dependencies = [ | ||||
|  "proc-macro2", | ||||
|  "quote", | ||||
|  "syn 2.0.104", | ||||
|  "syn 2.0.106", | ||||
| ] | ||||
|  | ||||
| [[package]] | ||||
| @@ -452,9 +452,9 @@ dependencies = [ | ||||
|  | ||||
| [[package]] | ||||
| name = "hashbrown" | ||||
| version = "0.15.4" | ||||
| version = "0.15.5" | ||||
| source = "registry+https://github.com/rust-lang/crates.io-index" | ||||
| checksum = "5971ac85611da7067dbfcabef3c70ebb5606018acd9e2a3903a0da507521e0d5" | ||||
| checksum = "9229cfe53dfd69f0609a49f65461bd93001ea1ef889cd5529dd176593f5338a1" | ||||
|  | ||||
| [[package]] | ||||
| name = "heck" | ||||
| @@ -491,10 +491,12 @@ dependencies = [ | ||||
| name = "heromodels-derive" | ||||
| version = "0.1.0" | ||||
| dependencies = [ | ||||
|  "heromodels_core", | ||||
|  "proc-macro2", | ||||
|  "quote", | ||||
|  "serde", | ||||
|  "syn 2.0.104", | ||||
|  "serde_json", | ||||
|  "syn 2.0.106", | ||||
| ] | ||||
|  | ||||
| [[package]] | ||||
| @@ -545,7 +547,7 @@ source = "registry+https://github.com/rust-lang/crates.io-index" | ||||
| checksum = "fe4cd85333e22411419a0bcae1297d25e58c9443848b11dc6a86fefe8c78a661" | ||||
| dependencies = [ | ||||
|  "equivalent", | ||||
|  "hashbrown 0.15.4", | ||||
|  "hashbrown 0.15.5", | ||||
| ] | ||||
|  | ||||
| [[package]] | ||||
| @@ -597,7 +599,7 @@ checksum = "03343451ff899767262ec32146f6d559dd759fdadf42ff0e227c7c48f72594b4" | ||||
| dependencies = [ | ||||
|  "proc-macro2", | ||||
|  "quote", | ||||
|  "syn 2.0.104", | ||||
|  "syn 2.0.106", | ||||
| ] | ||||
|  | ||||
| [[package]] | ||||
| @@ -627,9 +629,9 @@ dependencies = [ | ||||
|  | ||||
| [[package]] | ||||
| name = "jsonb" | ||||
| version = "0.5.3" | ||||
| version = "0.5.4" | ||||
| source = "registry+https://github.com/rust-lang/crates.io-index" | ||||
| checksum = "96cbb4fba292867a2d86ed83dbe5f9d036f423bf6a491b7d884058b2fde42fcd" | ||||
| checksum = "a452366d21e8d3cbca680c41388e01d6a88739afef7877961946a6da409f9ccd" | ||||
| dependencies = [ | ||||
|  "byteorder", | ||||
|  "ethnum", | ||||
| @@ -647,9 +649,20 @@ dependencies = [ | ||||
|  | ||||
| [[package]] | ||||
| name = "libc" | ||||
| version = "0.2.174" | ||||
| version = "0.2.175" | ||||
| source = "registry+https://github.com/rust-lang/crates.io-index" | ||||
| checksum = "1171693293099992e19cddea4e8b849964e9846f4acee11b3948bcc337be8776" | ||||
| checksum = "6a82ae493e598baaea5209805c49bbf2ea7de956d50d7da0da1164f9c6d28543" | ||||
|  | ||||
| [[package]] | ||||
| name = "libredox" | ||||
| version = "0.1.9" | ||||
| source = "registry+https://github.com/rust-lang/crates.io-index" | ||||
| checksum = "391290121bad3d37fbddad76d8f5d1c1c314cfc646d143d7e07a3086ddff0ce3" | ||||
| dependencies = [ | ||||
|  "bitflags", | ||||
|  "libc", | ||||
|  "redox_syscall", | ||||
| ] | ||||
|  | ||||
| [[package]] | ||||
| name = "lock_api" | ||||
| @@ -760,6 +773,7 @@ dependencies = [ | ||||
| [[package]] | ||||
| name = "ourdb" | ||||
| version = "0.1.0" | ||||
| source = "git+https://git.ourworld.tf/herocode/herolib_rust#aa0248ef17cb0117bb69f1d9f278f995bb417f16" | ||||
| dependencies = [ | ||||
|  "crc32fast", | ||||
|  "log", | ||||
| @@ -792,9 +806,9 @@ dependencies = [ | ||||
|  | ||||
| [[package]] | ||||
| name = "percent-encoding" | ||||
| version = "2.3.1" | ||||
| version = "2.3.2" | ||||
| source = "registry+https://github.com/rust-lang/crates.io-index" | ||||
| checksum = "e3148f5046208a5d56bcfc03053e3ca6334e51da8dfb19b6cdc8b306fae3283e" | ||||
| checksum = "9b4f627cb1b25917193a259e49bdad08f671f8d9708acfd5fe0a8c1455d87220" | ||||
|  | ||||
| [[package]] | ||||
| name = "phf" | ||||
| @@ -906,9 +920,9 @@ dependencies = [ | ||||
|  | ||||
| [[package]] | ||||
| name = "proc-macro2" | ||||
| version = "1.0.95" | ||||
| version = "1.0.101" | ||||
| source = "registry+https://github.com/rust-lang/crates.io-index" | ||||
| checksum = "02b3e5e68a3a1a02aad3ec490a98007cbc13c37cbe84a3cd7b8e406d76e7f778" | ||||
| checksum = "89ae43fd86e4158d6db51ad8e2b80f313af9cc74f5c0e03ccb87de09998732de" | ||||
| dependencies = [ | ||||
|  "unicode-ident", | ||||
| ] | ||||
| @@ -1079,12 +1093,13 @@ checksum = "a5a11a05ee1ce44058fa3d5961d05194fdbe3ad6b40f904af764d81b86450e6b" | ||||
| dependencies = [ | ||||
|  "proc-macro2", | ||||
|  "quote", | ||||
|  "syn 2.0.104", | ||||
|  "syn 2.0.106", | ||||
| ] | ||||
|  | ||||
| [[package]] | ||||
| name = "rhailib-macros" | ||||
| version = "0.1.0" | ||||
| source = "git+https://git.ourworld.tf/herocode/herolib_rust#aa0248ef17cb0117bb69f1d9f278f995bb417f16" | ||||
| dependencies = [ | ||||
|  "rhai", | ||||
|  "serde", | ||||
| @@ -1143,9 +1158,9 @@ checksum = "56f7d92ca342cea22a06f2121d944b4fd82af56988c270852495420f961d4ace" | ||||
|  | ||||
| [[package]] | ||||
| name = "rustversion" | ||||
| version = "1.0.21" | ||||
| version = "1.0.22" | ||||
| source = "registry+https://github.com/rust-lang/crates.io-index" | ||||
| checksum = "8a0d197bd2c9dc6e53b84da9556a69ba4cdfab8619eb41a8bd1cc2027a0f6b1d" | ||||
| checksum = "b39cdef0fa800fc44525c84ccb54a029961a8215f9619753635a9c0d2538d46d" | ||||
|  | ||||
| [[package]] | ||||
| name = "ryu" | ||||
| @@ -1191,14 +1206,14 @@ checksum = "5b0276cf7f2c73365f7157c8123c21cd9a50fbbd844757af28ca1f5925fc2a00" | ||||
| dependencies = [ | ||||
|  "proc-macro2", | ||||
|  "quote", | ||||
|  "syn 2.0.104", | ||||
|  "syn 2.0.106", | ||||
| ] | ||||
|  | ||||
| [[package]] | ||||
| name = "serde_json" | ||||
| version = "1.0.142" | ||||
| version = "1.0.143" | ||||
| source = "registry+https://github.com/rust-lang/crates.io-index" | ||||
| checksum = "030fedb782600dcbd6f02d479bf0d817ac3bb40d644745b769d6a96bc3afc5a7" | ||||
| checksum = "d401abef1d108fbd9cbaebc3e46611f4b1021f714a0597a71f41ee463f5f4a5a" | ||||
| dependencies = [ | ||||
|  "indexmap", | ||||
|  "itoa", | ||||
| @@ -1247,9 +1262,9 @@ checksum = "56199f7ddabf13fe5074ce809e7d3f42b42ae711800501b5b16ea82ad029c39d" | ||||
|  | ||||
| [[package]] | ||||
| name = "slab" | ||||
| version = "0.4.10" | ||||
| version = "0.4.11" | ||||
| source = "registry+https://github.com/rust-lang/crates.io-index" | ||||
| checksum = "04dc19736151f35336d325007ac991178d504a119863a2fcb3758cdb5e52c50d" | ||||
| checksum = "7a2ae44ef20feb57a68b23d846850f861394c2e02dc425a50098ae8c90267589" | ||||
|  | ||||
| [[package]] | ||||
| name = "smallvec" | ||||
| @@ -1327,7 +1342,7 @@ dependencies = [ | ||||
|  "proc-macro2", | ||||
|  "quote", | ||||
|  "rustversion", | ||||
|  "syn 2.0.104", | ||||
|  "syn 2.0.106", | ||||
| ] | ||||
|  | ||||
| [[package]] | ||||
| @@ -1349,9 +1364,9 @@ dependencies = [ | ||||
|  | ||||
| [[package]] | ||||
| name = "syn" | ||||
| version = "2.0.104" | ||||
| version = "2.0.106" | ||||
| source = "registry+https://github.com/rust-lang/crates.io-index" | ||||
| checksum = "17b6f705963418cdb9927482fa304bc562ece2fdd4f616084c50b7023b435a40" | ||||
| checksum = "ede7c438028d4436d71104916910f5bb611972c5cfd7f89b8300a8186e6fada6" | ||||
| dependencies = [ | ||||
|  "proc-macro2", | ||||
|  "quote", | ||||
| @@ -1387,7 +1402,7 @@ checksum = "4fee6c4efc90059e10f81e6d42c60a18f76588c3d74cb83a0b242a2b6c7504c1" | ||||
| dependencies = [ | ||||
|  "proc-macro2", | ||||
|  "quote", | ||||
|  "syn 2.0.104", | ||||
|  "syn 2.0.106", | ||||
| ] | ||||
|  | ||||
| [[package]] | ||||
| @@ -1401,9 +1416,9 @@ dependencies = [ | ||||
|  | ||||
| [[package]] | ||||
| name = "tinyvec" | ||||
| version = "1.9.0" | ||||
| version = "1.10.0" | ||||
| source = "registry+https://github.com/rust-lang/crates.io-index" | ||||
| checksum = "09b3661f17e86524eccd4371ab0429194e0d7c008abb45f7a7495b1719463c71" | ||||
| checksum = "bfa5fdc3bce6191a1dbc8c02d5c8bffcf557bafa17c124c5264a458f1b0613fa" | ||||
| dependencies = [ | ||||
|  "tinyvec_macros", | ||||
| ] | ||||
| @@ -1442,7 +1457,7 @@ checksum = "6e06d43f1345a3bcd39f6a56dbb7dcab2ba47e68e8ac134855e7e2bdbaf8cab8" | ||||
| dependencies = [ | ||||
|  "proc-macro2", | ||||
|  "quote", | ||||
|  "syn 2.0.104", | ||||
|  "syn 2.0.106", | ||||
| ] | ||||
|  | ||||
| [[package]] | ||||
| @@ -1504,6 +1519,7 @@ dependencies = [ | ||||
| [[package]] | ||||
| name = "tst" | ||||
| version = "0.1.0" | ||||
| source = "git+https://git.ourworld.tf/herocode/herolib_rust#aa0248ef17cb0117bb69f1d9f278f995bb417f16" | ||||
| dependencies = [ | ||||
|  "ourdb", | ||||
|  "thiserror", | ||||
| @@ -1550,9 +1566,9 @@ checksum = "6d49784317cd0d1ee7ec5c716dd598ec5b4483ea832a2dced265471cc0f690ae" | ||||
|  | ||||
| [[package]] | ||||
| name = "uuid" | ||||
| version = "1.17.0" | ||||
| version = "1.18.0" | ||||
| source = "registry+https://github.com/rust-lang/crates.io-index" | ||||
| checksum = "3cf4199d1e5d15ddd86a694e4d0dffa9c323ce759fea589f00fef9d81cc1931d" | ||||
| checksum = "f33196643e165781c20a5ead5582283a7dacbb87855d867fbc2df3f81eddc1be" | ||||
| dependencies = [ | ||||
|  "getrandom 0.3.3", | ||||
|  "js-sys", | ||||
| @@ -1614,7 +1630,7 @@ dependencies = [ | ||||
|  "log", | ||||
|  "proc-macro2", | ||||
|  "quote", | ||||
|  "syn 2.0.104", | ||||
|  "syn 2.0.106", | ||||
|  "wasm-bindgen-shared", | ||||
| ] | ||||
|  | ||||
| @@ -1636,7 +1652,7 @@ checksum = "8ae87ea40c9f689fc23f209965b6fb8a99ad69aeeb0231408be24920604395de" | ||||
| dependencies = [ | ||||
|  "proc-macro2", | ||||
|  "quote", | ||||
|  "syn 2.0.104", | ||||
|  "syn 2.0.106", | ||||
|  "wasm-bindgen-backend", | ||||
|  "wasm-bindgen-shared", | ||||
| ] | ||||
| @@ -1662,11 +1678,11 @@ dependencies = [ | ||||
|  | ||||
| [[package]] | ||||
| name = "whoami" | ||||
| version = "1.6.0" | ||||
| version = "1.6.1" | ||||
| source = "registry+https://github.com/rust-lang/crates.io-index" | ||||
| checksum = "6994d13118ab492c3c80c1f81928718159254c53c472bf9ce36f8dae4add02a7" | ||||
| checksum = "5d4a4db5077702ca3015d3d02d74974948aba2ad9e12ab7df718ee64ccd7e97d" | ||||
| dependencies = [ | ||||
|  "redox_syscall", | ||||
|  "libredox", | ||||
|  "wasite", | ||||
|  "web-sys", | ||||
| ] | ||||
| @@ -1692,7 +1708,7 @@ checksum = "a47fddd13af08290e67f4acabf4b459f647552718f683a7b415d290ac744a836" | ||||
| dependencies = [ | ||||
|  "proc-macro2", | ||||
|  "quote", | ||||
|  "syn 2.0.104", | ||||
|  "syn 2.0.106", | ||||
| ] | ||||
|  | ||||
| [[package]] | ||||
| @@ -1703,7 +1719,7 @@ checksum = "bd9211b69f8dcdfa817bfd14bf1c97c9188afa36f4750130fcdf3f400eca9fa8" | ||||
| dependencies = [ | ||||
|  "proc-macro2", | ||||
|  "quote", | ||||
|  "syn 2.0.104", | ||||
|  "syn 2.0.106", | ||||
| ] | ||||
|  | ||||
| [[package]] | ||||
| @@ -1856,5 +1872,5 @@ checksum = "9ecf5b4cc5364572d7f4c329661bcc82724222973f2cab6f050a4e5c22f75181" | ||||
| dependencies = [ | ||||
|  "proc-macro2", | ||||
|  "quote", | ||||
|  "syn 2.0.104", | ||||
|  "syn 2.0.106", | ||||
| ] | ||||
|   | ||||
| @@ -14,4 +14,6 @@ quote = "1.0" | ||||
| proc-macro2 = "1.0" | ||||
|  | ||||
| [dev-dependencies] | ||||
| serde = { version = "1.0", features = ["derive"] } | ||||
| serde = { version = "1.0", features = ["derive"] } | ||||
| serde_json = "1.0" | ||||
| heromodels_core = { path = "../heromodels_core" } | ||||
| @@ -1,6 +1,6 @@ | ||||
| use proc_macro::TokenStream; | ||||
| use quote::{format_ident, quote}; | ||||
| use syn::{Data, DeriveInput, Fields, parse_macro_input}; | ||||
| use syn::{parse_macro_input, Data, DeriveInput, Fields, Lit, Meta, MetaList, MetaNameValue}; | ||||
|  | ||||
| /// Convert a string to snake_case | ||||
| fn to_snake_case(s: &str) -> String { | ||||
| @@ -47,86 +47,165 @@ pub fn model(_attr: TokenStream, item: TokenStream) -> TokenStream { | ||||
|     let db_prefix = to_snake_case(&name_str); | ||||
|  | ||||
|     // Extract fields with #[index] attribute | ||||
|     let mut indexed_fields = Vec::new(); | ||||
|     let mut custom_index_names = std::collections::HashMap::new(); | ||||
|     // Supports both top-level (no args) and nested path-based indexes declared on a field | ||||
|     #[derive(Clone)] | ||||
|     enum IndexDecl { | ||||
|         TopLevel { | ||||
|             field_ident: syn::Ident, | ||||
|             field_ty: syn::Type, | ||||
|         }, | ||||
|         NestedPath { | ||||
|             on_field_ident: syn::Ident, | ||||
|             path: String, // dotted path relative to the field | ||||
|         }, | ||||
|     } | ||||
|  | ||||
|     let mut index_decls: Vec<IndexDecl> = Vec::new(); | ||||
|  | ||||
|     if let Data::Struct(ref mut data_struct) = input.data { | ||||
|         if let Fields::Named(ref mut fields_named) = data_struct.fields { | ||||
|             for field in &mut fields_named.named { | ||||
|                 let mut attr_idx = None; | ||||
|                 let mut to_remove: Vec<usize> = Vec::new(); | ||||
|                 for (i, attr) in field.attrs.iter().enumerate() { | ||||
|                     if attr.path().is_ident("index") { | ||||
|                         attr_idx = Some(i); | ||||
|                         if let Some(ref field_name) = field.ident { | ||||
|                             // Check if the attribute has parameters | ||||
|                             let mut custom_name = None; | ||||
|                     if !attr.path().is_ident("index") { | ||||
|                         continue; | ||||
|                     } | ||||
|                     to_remove.push(i); | ||||
|  | ||||
|                             // Parse attribute arguments if any | ||||
|                             let meta = attr.meta.clone(); | ||||
|                             if let syn::Meta::List(list) = meta { | ||||
|                                 if let Ok(nested) = list.parse_args_with(syn::punctuated::Punctuated::<syn::Meta, syn::Token![,]>::parse_terminated) { | ||||
|                                         for meta in nested { | ||||
|                                             if let syn::Meta::NameValue(name_value) = meta { | ||||
|                                                 if name_value.path.is_ident("name") { | ||||
|                                                     if let syn::Expr::Lit(syn::ExprLit { lit: syn::Lit::Str(lit_str), .. }) = name_value.value { | ||||
|                                                         custom_name = Some(lit_str.value()); | ||||
|                                                     } | ||||
|                     if let Some(ref field_name) = field.ident { | ||||
|                         match &attr.meta { | ||||
|                             Meta::Path(_) => { | ||||
|                                 // Simple top-level index on this field | ||||
|                                 index_decls.push(IndexDecl::TopLevel { | ||||
|                                     field_ident: field_name.clone(), | ||||
|                                     field_ty: field.ty.clone(), | ||||
|                                 }); | ||||
|                             } | ||||
|                             Meta::List(MetaList { .. }) => { | ||||
|                                 // Parse for path = "..."; name is assumed equal to path | ||||
|                                 // We support syntax: #[index(path = "a.b.c")] | ||||
|                                 if let Ok(nested) = attr.parse_args_with( | ||||
|                                     syn::punctuated::Punctuated::<Meta, syn::Token![,]>::parse_terminated, | ||||
|                                 ) { | ||||
|                                     for meta in nested { | ||||
|                                         if let Meta::NameValue(MetaNameValue { path, value, .. }) = meta { | ||||
|                                             if path.is_ident("path") { | ||||
|                                                 if let syn::Expr::Lit(syn::ExprLit { lit: Lit::Str(lit_str), .. }) = value { | ||||
|                                                     let p = lit_str.value(); | ||||
|                                                     index_decls.push(IndexDecl::NestedPath { | ||||
|                                                         on_field_ident: field_name.clone(), | ||||
|                                                         path: p, | ||||
|                                                     }); | ||||
|                                                 } | ||||
|                                             } | ||||
|                                         } | ||||
|                                     } | ||||
|                                 } | ||||
|                             } | ||||
|  | ||||
|                             indexed_fields.push((field_name.clone(), field.ty.clone())); | ||||
|  | ||||
|                             if let Some(name) = custom_name { | ||||
|                                 custom_index_names.insert(field_name.to_string(), name); | ||||
|                             } | ||||
|                             _ => {} | ||||
|                         } | ||||
|                     } | ||||
|                 } | ||||
|  | ||||
|                 if let Some(idx) = attr_idx { | ||||
|                 // remove all #[index] attributes we processed | ||||
|                 // remove from the back to keep indices valid | ||||
|                 to_remove.sort_unstable(); | ||||
|                 to_remove.drain(..).rev().for_each(|idx| { | ||||
|                     field.attrs.remove(idx); | ||||
|                 } | ||||
|                 }); | ||||
|             } | ||||
|         } | ||||
|     } | ||||
|  | ||||
|     // Generate Model trait implementation | ||||
|     let db_keys_impl = if indexed_fields.is_empty() { | ||||
|     let db_keys_impl = if index_decls.is_empty() { | ||||
|         quote! { | ||||
|             fn db_keys(&self) -> Vec<heromodels_core::IndexKey> { | ||||
|                 Vec::new() | ||||
|             } | ||||
|         } | ||||
|     } else { | ||||
|         let field_keys = indexed_fields.iter().map(|(field_name, _)| { | ||||
|             let name_str = custom_index_names | ||||
|                 .get(&field_name.to_string()) | ||||
|                 .cloned() | ||||
|                 .unwrap_or(field_name.to_string()); | ||||
|             quote! { | ||||
|                 heromodels_core::IndexKey { | ||||
|                     name: #name_str, | ||||
|                     value: self.#field_name.to_string(), | ||||
|         // Build code for keys from each index declaration | ||||
|         let mut key_snippets: Vec<proc_macro2::TokenStream> = Vec::new(); | ||||
|  | ||||
|         for decl in &index_decls { | ||||
|             match decl.clone() { | ||||
|                 IndexDecl::TopLevel { field_ident, .. } => { | ||||
|                     let name_str = field_ident.to_string(); | ||||
|                     key_snippets.push(quote! { | ||||
|                         keys.push(heromodels_core::IndexKey { | ||||
|                             name: #name_str, | ||||
|                             value: self.#field_ident.to_string(), | ||||
|                         }); | ||||
|                     }); | ||||
|                 } | ||||
|                 IndexDecl::NestedPath { on_field_ident, path } => { | ||||
|                     // Name is equal to provided path | ||||
|                     let name_str = path.clone(); | ||||
|                     // Generate traversal code using serde_json to support arrays and objects generically | ||||
|                     // Split the path into static segs for iteration | ||||
|                     let segs: Vec<String> = path.split('.').map(|s| s.to_string()).collect(); | ||||
|                     let segs_iter = segs.iter().map(|s| s.as_str()); | ||||
|                     let segs_array = quote! { [ #( #segs_iter ),* ] }; | ||||
|  | ||||
|                     key_snippets.push(quote! { | ||||
|                         // Serialize the target field to JSON for generic traversal | ||||
|                         let __hm_json_val = ::serde_json::to_value(&self.#on_field_ident).unwrap_or(::serde_json::Value::Null); | ||||
|                         let mut __hm_stack: Vec<&::serde_json::Value> = vec![&__hm_json_val]; | ||||
|                         for __hm_seg in #segs_array.iter() { | ||||
|                             let mut __hm_next: Vec<&::serde_json::Value> = Vec::new(); | ||||
|                             for __hm_v in &__hm_stack { | ||||
|                                 match __hm_v { | ||||
|                                     ::serde_json::Value::Array(arr) => { | ||||
|                                         for __hm_e in arr { | ||||
|                                             if let ::serde_json::Value::Object(map) = __hm_e { | ||||
|                                                 if let Some(x) = map.get(*__hm_seg) { __hm_next.push(x); } | ||||
|                                             } | ||||
|                                         } | ||||
|                                     } | ||||
|                                     ::serde_json::Value::Object(map) => { | ||||
|                                         if let Some(x) = map.get(*__hm_seg) { __hm_next.push(x); } | ||||
|                                     } | ||||
|                                     _ => {} | ||||
|                                 } | ||||
|                             } | ||||
|                             __hm_stack = __hm_next; | ||||
|                             if __hm_stack.is_empty() { break; } | ||||
|                         } | ||||
|                         for __hm_leaf in __hm_stack { | ||||
|                             match __hm_leaf { | ||||
|                                 ::serde_json::Value::Null => {}, | ||||
|                                 ::serde_json::Value::Array(_) => {}, | ||||
|                                 ::serde_json::Value::Object(_) => {}, | ||||
|                                 other => { | ||||
|                                     // Convert primitives to string without surrounding quotes for strings | ||||
|                                     let mut s = other.to_string(); | ||||
|                                     if let ::serde_json::Value::String(_) = other { s = s.trim_matches('"').to_string(); } | ||||
|                                     keys.push(heromodels_core::IndexKey { name: #name_str, value: s }); | ||||
|                                 } | ||||
|                             } | ||||
|                         } | ||||
|                     }); | ||||
|                 } | ||||
|             } | ||||
|         }); | ||||
|         } | ||||
|  | ||||
|         quote! { | ||||
|             fn db_keys(&self) -> Vec<heromodels_core::IndexKey> { | ||||
|                 vec![ | ||||
|                     #(#field_keys),* | ||||
|                 ] | ||||
|                 let mut keys: Vec<heromodels_core::IndexKey> = Vec::new(); | ||||
|                 #(#key_snippets)* | ||||
|                 keys | ||||
|             } | ||||
|         } | ||||
|     }; | ||||
|  | ||||
|     let indexed_field_names = indexed_fields | ||||
|     let indexed_field_names: Vec<String> = index_decls | ||||
|         .iter() | ||||
|         .map(|f| f.0.to_string()) | ||||
|         .collect::<Vec<_>>(); | ||||
|         .map(|d| match d { | ||||
|             IndexDecl::TopLevel { field_ident, .. } => field_ident.to_string(), | ||||
|             IndexDecl::NestedPath { path, .. } => path.clone(), | ||||
|         }) | ||||
|         .collect(); | ||||
|  | ||||
|     let model_impl = quote! { | ||||
|         impl heromodels_core::Model for #struct_name { | ||||
| @@ -152,51 +231,33 @@ pub fn model(_attr: TokenStream, item: TokenStream) -> TokenStream { | ||||
|         } | ||||
|     }; | ||||
|  | ||||
|     // Generate Index trait implementations | ||||
|     // Generate Index trait implementations only for top-level fields, keep existing behavior | ||||
|     let mut index_impls = proc_macro2::TokenStream::new(); | ||||
|     for decl in &index_decls { | ||||
|         if let IndexDecl::TopLevel { field_ident, field_ty } = decl { | ||||
|             let name_str = field_ident.to_string(); | ||||
|             let index_struct_name = format_ident!("{}", &name_str); | ||||
|             let field_type = field_ty.clone(); | ||||
|  | ||||
|     for (field_name, field_type) in &indexed_fields { | ||||
|         let name_str = field_name.to_string(); | ||||
|             let index_impl = quote! { | ||||
|                 pub struct #index_struct_name; | ||||
|  | ||||
|         // Get custom index name if specified, otherwise use field name | ||||
|         let index_key = match custom_index_names.get(&name_str) { | ||||
|             Some(custom_name) => custom_name.clone(), | ||||
|             None => name_str.clone(), | ||||
|         }; | ||||
|                 impl heromodels_core::Index for #index_struct_name { | ||||
|                     type Model = super::#struct_name; | ||||
|                     type Key = #field_type; | ||||
|  | ||||
|         // Convert field name to PascalCase for struct name | ||||
|         // let struct_name_str = to_pascal_case(&name_str); | ||||
|         // let index_struct_name = format_ident!("{}", struct_name_str); | ||||
|         let index_struct_name = format_ident!("{}", &name_str); | ||||
|                     fn key() -> &'static str { #name_str } | ||||
|  | ||||
|         // Default to str for key type | ||||
|         let index_impl = quote! { | ||||
|             pub struct #index_struct_name; | ||||
|  | ||||
|             impl heromodels_core::Index for #index_struct_name { | ||||
|                 type Model = super::#struct_name; | ||||
|                 type Key = #field_type; | ||||
|  | ||||
|                 fn key() -> &'static str { | ||||
|                     #index_key | ||||
|                     fn field_name() -> &'static str { #name_str } | ||||
|                 } | ||||
|  | ||||
|                 fn field_name() -> &'static str { | ||||
|                     #name_str | ||||
|                 } | ||||
|             } | ||||
|         }; | ||||
|  | ||||
|         index_impls.extend(index_impl); | ||||
|             }; | ||||
|             index_impls.extend(index_impl); | ||||
|         } | ||||
|     } | ||||
|  | ||||
|     if !index_impls.is_empty() { | ||||
|         let index_mod_name = format_ident!("{}_index", db_prefix); | ||||
|         index_impls = quote! { | ||||
|             pub mod #index_mod_name { | ||||
|                 #index_impls | ||||
|             } | ||||
|         } | ||||
|         index_impls = quote! { pub mod #index_mod_name { #index_impls } }; | ||||
|     } | ||||
|  | ||||
|     // Combine the original struct with the generated implementations | ||||
|   | ||||
| @@ -1,7 +1,38 @@ | ||||
| use heromodels_derive::model; | ||||
| use serde::{Deserialize, Serialize}; | ||||
|  | ||||
| // Define the necessary structs and traits for testing | ||||
| // Make the current crate visible as an extern crate named `heromodels_core` | ||||
| extern crate self as heromodels_core; | ||||
| extern crate serde_json; // ensure ::serde_json path resolves | ||||
|  | ||||
| // Mock the heromodels_core API at crate root (visible via the alias above) | ||||
| #[derive(Debug, Clone, PartialEq, Eq)] | ||||
| pub struct IndexKey { | ||||
|     pub name: &'static str, | ||||
|     pub value: String, | ||||
| } | ||||
|  | ||||
| pub trait Model: std::fmt::Debug + Clone + Serialize + for<'de> Deserialize<'de> + Send + Sync + 'static { | ||||
|     fn db_prefix() -> &'static str | ||||
|     where | ||||
|         Self: Sized; | ||||
|     fn get_id(&self) -> u32; | ||||
|     fn base_data_mut(&mut self) -> &mut BaseModelData; | ||||
|     fn db_keys(&self) -> Vec<IndexKey> { | ||||
|         Vec::new() | ||||
|     } | ||||
|     fn indexed_fields() -> Vec<&'static str> { | ||||
|         Vec::new() | ||||
|     } | ||||
| } | ||||
|  | ||||
| pub trait Index { | ||||
|     type Model: Model; | ||||
|     type Key: ToString + ?Sized; | ||||
|     fn key() -> &'static str; | ||||
|     fn field_name() -> &'static str; | ||||
| } | ||||
|  | ||||
| #[derive(Debug, Clone, Serialize, Deserialize)] | ||||
| pub struct BaseModelData { | ||||
|     pub id: u32, | ||||
| @@ -11,41 +42,18 @@ pub struct BaseModelData { | ||||
| } | ||||
|  | ||||
| impl BaseModelData { | ||||
|     pub fn new(id: u32) -> Self { | ||||
|         let now = 1000; // Mock timestamp | ||||
|         Self { | ||||
|             id, | ||||
|             created_at: now, | ||||
|             modified_at: now, | ||||
|             comments: Vec::new(), | ||||
|         } | ||||
|     pub fn new() -> Self { | ||||
|         let now = 1000; | ||||
|         Self { id: 0, created_at: now, modified_at: now, comments: Vec::new() } | ||||
|     } | ||||
|     pub fn update_modified(&mut self) { self.modified_at += 1; } | ||||
| } | ||||
|  | ||||
| #[derive(Debug, Clone, PartialEq, Eq)] | ||||
| pub struct IndexKey { | ||||
|     pub name: &'static str, | ||||
|     pub value: String, | ||||
| } | ||||
|  | ||||
| pub trait Model: std::fmt::Debug + Clone { | ||||
|     fn db_prefix() -> &'static str; | ||||
|     fn get_id(&self) -> u32; | ||||
|     fn base_data_mut(&mut self) -> &mut BaseModelData; | ||||
|     fn db_keys(&self) -> Vec<IndexKey>; | ||||
| } | ||||
|  | ||||
| pub trait Index { | ||||
|     type Model: Model; | ||||
|     type Key: ?Sized; | ||||
|     fn key() -> &'static str; | ||||
| } | ||||
|  | ||||
| // Test struct using the model macro | ||||
| // Top-level field index tests | ||||
| #[derive(Debug, Clone, Serialize, Deserialize)] | ||||
| #[model] | ||||
| struct TestUser { | ||||
|     base_data: BaseModelData, | ||||
| pub struct TestUser { | ||||
|     base_data: heromodels_core::BaseModelData, | ||||
|  | ||||
|     #[index] | ||||
|     username: String, | ||||
| @@ -54,25 +62,12 @@ struct TestUser { | ||||
|     is_active: bool, | ||||
| } | ||||
|  | ||||
| // Test struct with custom index name | ||||
| #[derive(Debug, Clone, Serialize, Deserialize)] | ||||
| #[model] | ||||
| struct TestUserWithCustomIndex { | ||||
|     base_data: BaseModelData, | ||||
|  | ||||
|     #[index(name = "custom_username")] | ||||
|     username: String, | ||||
|  | ||||
|     #[index] | ||||
|     is_active: bool, | ||||
| } | ||||
|  | ||||
| #[test] | ||||
| fn test_basic_model() { | ||||
|     assert_eq!(TestUser::db_prefix(), "test_user"); | ||||
|  | ||||
|     let user = TestUser { | ||||
|         base_data: BaseModelData::new(1), | ||||
|         base_data: heromodels_core::BaseModelData::new(), | ||||
|         username: "test".to_string(), | ||||
|         is_active: true, | ||||
|     }; | ||||
| @@ -85,22 +80,47 @@ fn test_basic_model() { | ||||
|     assert_eq!(keys[1].value, "true"); | ||||
| } | ||||
|  | ||||
| // Nested path index tests (including vector traversal) | ||||
| #[derive(Debug, Clone, Serialize, Deserialize, Default)] | ||||
| struct GPU { gpu_brand: String } | ||||
|  | ||||
| #[derive(Debug, Clone, Serialize, Deserialize, Default)] | ||||
| struct CPU { cpu_brand: String } | ||||
|  | ||||
| #[derive(Debug, Clone, Serialize, Deserialize, Default)] | ||||
| struct DeviceInfo { vendor: String, cpu: Vec<CPU>, gpu: Vec<GPU> } | ||||
|  | ||||
| #[derive(Debug, Clone, Serialize, Deserialize)] | ||||
| #[model] | ||||
| pub struct NodeLike { | ||||
|     base_data: heromodels_core::BaseModelData, | ||||
|  | ||||
|     #[index(path = "vendor")] | ||||
|     #[index(path = "cpu.cpu_brand")] | ||||
|     #[index(path = "gpu.gpu_brand")] | ||||
|     devices: DeviceInfo, | ||||
| } | ||||
|  | ||||
| #[test] | ||||
| fn test_custom_index_name() { | ||||
|     let user = TestUserWithCustomIndex { | ||||
|         base_data: BaseModelData::new(1), | ||||
|         username: "test".to_string(), | ||||
|         is_active: true, | ||||
| fn test_nested_indexes() { | ||||
|     let n = NodeLike { | ||||
|         base_data: heromodels_core::BaseModelData::new(), | ||||
|         devices: DeviceInfo { | ||||
|             vendor: "SuperVendor".to_string(), | ||||
|             cpu: vec![CPU { cpu_brand: "Intel".into() }, CPU { cpu_brand: "AMD".into() }], | ||||
|             gpu: vec![GPU { gpu_brand: "NVIDIA".into() }, GPU { gpu_brand: "AMD".into() }], | ||||
|         }, | ||||
|     }; | ||||
|  | ||||
|     // Check that the Username struct uses the custom index name | ||||
|     assert_eq!(Username::key(), "custom_username"); | ||||
|     let mut keys = n.db_keys(); | ||||
|     // Sort for deterministic assertions | ||||
|     keys.sort_by(|a,b| a.name.cmp(b.name).then(a.value.cmp(&b.value))); | ||||
|  | ||||
|     // Check that the db_keys method returns the correct keys | ||||
|     let keys = user.db_keys(); | ||||
|     assert_eq!(keys.len(), 2); | ||||
|     assert_eq!(keys[0].name, "custom_username"); | ||||
|     assert_eq!(keys[0].value, "test"); | ||||
|     assert_eq!(keys[1].name, "is_active"); | ||||
|     assert_eq!(keys[1].value, "true"); | ||||
|     // Expect 1 (vendor) + 2 (cpu brands) + 2 (gpu brands) = 5 keys | ||||
|     assert_eq!(keys.len(), 5); | ||||
|     assert!(keys.iter().any(|k| k.name == "vendor" && k.value == "SuperVendor")); | ||||
|     assert!(keys.iter().any(|k| k.name == "cpu.cpu_brand" && k.value == "Intel")); | ||||
|     assert!(keys.iter().any(|k| k.name == "cpu.cpu_brand" && k.value == "AMD")); | ||||
|     assert!(keys.iter().any(|k| k.name == "gpu.gpu_brand" && k.value == "NVIDIA")); | ||||
|     assert!(keys.iter().any(|k| k.name == "gpu.gpu_brand" && k.value == "AMD")); | ||||
| } | ||||
|   | ||||
							
								
								
									
										1637
									
								
								heromodels/Cargo.lock
									
									
									
										generated
									
									
									
								
							
							
						
						
									
										1637
									
								
								heromodels/Cargo.lock
									
									
									
										generated
									
									
									
								
							
										
											
												File diff suppressed because it is too large
												Load Diff
											
										
									
								
							| @@ -10,11 +10,11 @@ serde = { version = "1.0", features = ["derive"] } | ||||
| serde_json = "1.0" | ||||
| bincode = { version = "2", features = ["serde"] } | ||||
| chrono = { version = "0.4", features = ["serde"] } | ||||
| ourdb = { path = "../../herolib_rust/packages/data/ourdb" } | ||||
| tst = { path = "../../herolib_rust/packages/data/tst" } | ||||
| ourdb = { git = "https://git.ourworld.tf/herocode/herolib_rust", package = "ourdb" } | ||||
| tst = { git = "https://git.ourworld.tf/herocode/herolib_rust", package = "tst" } | ||||
| heromodels-derive = { path = "../heromodels-derive" } | ||||
| heromodels_core = { path = "../heromodels_core" } | ||||
| rhailib-macros = { path = "../../herolib_rust/rhailib/src/macros" } | ||||
| rhailib-macros = { git = "https://git.ourworld.tf/herocode/herolib_rust", package = "rhailib-macros" } | ||||
| rhai = { version = "1.21.0", features = [ | ||||
|   "std", | ||||
|   "sync", | ||||
| @@ -53,11 +53,19 @@ path = "examples/finance_example/main.rs" | ||||
| name = "flow_example" | ||||
| path = "examples/flow_example.rs" | ||||
|  | ||||
| [[example]] | ||||
| name = "biz_rhai" | ||||
| path = "examples/biz_rhai/example.rs" | ||||
| required-features = ["rhai"] | ||||
| # [[example]] | ||||
| # name = "biz_rhai" | ||||
| # path = "examples/biz_rhai/example.rs" | ||||
| # required-features = ["rhai"] | ||||
|  | ||||
| [[example]] | ||||
| name = "postgres_model_example" | ||||
| path = "examples/postgres_example/example.rs" | ||||
|  | ||||
| [[example]] | ||||
| name = "heroledger_example" | ||||
| path = "examples/heroledger_example/example.rs" | ||||
|  | ||||
| [[example]] | ||||
| name = "grid4_example" | ||||
| path = "examples/grid4_example/example.rs" | ||||
|   | ||||
| @@ -1,10 +1,25 @@ | ||||
| use chrono::{Duration, Utc}; | ||||
| use chrono::{Duration, Utc, NaiveDateTime}; | ||||
| use heromodels::db::{Collection, Db}; | ||||
| use heromodels::models::User; | ||||
| use heromodels::models::calendar::{AttendanceStatus, Attendee, Calendar, Event, EventStatus}; | ||||
| use heromodels_core::Model; | ||||
|  | ||||
| fn main() { | ||||
|     // Helper to format i64 timestamps | ||||
|     let fmt_time = |ts: i64| -> String { | ||||
|         let ndt = NaiveDateTime::from_timestamp_opt(ts, 0) | ||||
|             .unwrap_or(NaiveDateTime::from_timestamp_opt(0, 0).unwrap()); | ||||
|         chrono::DateTime::<Utc>::from_utc(ndt, Utc) | ||||
|             .format("%Y-%m-%d %H:%M") | ||||
|             .to_string() | ||||
|     }; | ||||
|     let fmt_date = |ts: i64| -> String { | ||||
|         let ndt = NaiveDateTime::from_timestamp_opt(ts, 0) | ||||
|             .unwrap_or(NaiveDateTime::from_timestamp_opt(0, 0).unwrap()); | ||||
|         chrono::DateTime::<Utc>::from_utc(ndt, Utc) | ||||
|             .format("%Y-%m-%d") | ||||
|             .to_string() | ||||
|     }; | ||||
|     // Create a new DB instance, reset before every run | ||||
|     let db_path = "/tmp/ourdb_calendar_example"; | ||||
|     let db = heromodels::db::hero::OurDB::new(db_path, true).expect("Can create DB"); | ||||
| @@ -47,50 +62,21 @@ fn main() { | ||||
|     println!("- User 2 (ID: {}): {}", user2_id, stored_user2.full_name); | ||||
|     println!("- User 3 (ID: {}): {}", user3_id, stored_user3.full_name); | ||||
|  | ||||
|     // --- Create Attendees --- | ||||
|     // --- Create Attendees (embedded in events, not stored separately) --- | ||||
|     println!("\n--- Creating Attendees ---"); | ||||
|     let attendee1 = Attendee::new(user1_id).status(AttendanceStatus::Accepted); | ||||
|     let attendee2 = Attendee::new(user2_id).status(AttendanceStatus::Tentative); | ||||
|     let attendee3 = Attendee::new(user3_id); // Default NoResponse | ||||
|  | ||||
|     // Store attendees in database and get their IDs | ||||
|     let attendee_collection = db | ||||
|         .collection::<Attendee>() | ||||
|         .expect("can open attendee collection"); | ||||
|  | ||||
|     let (attendee1_id, stored_attendee1) = attendee_collection | ||||
|         .set(&attendee1) | ||||
|         .expect("can set attendee1"); | ||||
|     let (attendee2_id, stored_attendee2) = attendee_collection | ||||
|         .set(&attendee2) | ||||
|         .expect("can set attendee2"); | ||||
|     let (attendee3_id, stored_attendee3) = attendee_collection | ||||
|         .set(&attendee3) | ||||
|         .expect("can set attendee3"); | ||||
|  | ||||
|     println!("Created attendees:"); | ||||
|     println!( | ||||
|         "- Attendee 1 (ID: {}): Contact ID {}, Status: {:?}", | ||||
|         attendee1_id, stored_attendee1.contact_id, stored_attendee1.status | ||||
|     ); | ||||
|     println!( | ||||
|         "- Attendee 2 (ID: {}): Contact ID {}, Status: {:?}", | ||||
|         attendee2_id, stored_attendee2.contact_id, stored_attendee2.status | ||||
|     ); | ||||
|     println!( | ||||
|         "- Attendee 3 (ID: {}): Contact ID {}, Status: {:?}", | ||||
|         attendee3_id, stored_attendee3.contact_id, stored_attendee3.status | ||||
|     ); | ||||
|  | ||||
|     // --- Create Events with Attendees --- | ||||
|     println!("\n--- Creating Events with Enhanced Features ---"); | ||||
|     let now = Utc::now(); | ||||
|     let event1_start = (now + Duration::hours(1)).timestamp(); | ||||
|     let event1_end = (now + Duration::hours(2)).timestamp(); | ||||
|  | ||||
|     let event1 = Event::new( | ||||
|         "Team Meeting", | ||||
|         now + Duration::hours(1), | ||||
|         now + Duration::hours(2), | ||||
|     ) | ||||
|     let event1 = Event::new() | ||||
|     .title("Team Meeting") | ||||
|     .reschedule(event1_start, event1_end) | ||||
|     .description("Weekly sync-up meeting to discuss project progress.") | ||||
|     .location("Conference Room A") | ||||
|     .color("#FF5722") // Red-orange color | ||||
| @@ -99,14 +85,14 @@ fn main() { | ||||
|     .category("Work") | ||||
|     .reminder_minutes(15) | ||||
|     .timezone("UTC") | ||||
|     .add_attendee(attendee1_id) | ||||
|     .add_attendee(attendee2_id); | ||||
|     .add_attendee(attendee1.clone()) | ||||
|     .add_attendee(attendee2.clone()); | ||||
|  | ||||
|     let event2 = Event::new( | ||||
|         "Project Brainstorm", | ||||
|         now + Duration::days(1), | ||||
|         now + Duration::days(1) + Duration::minutes(90), | ||||
|     ) | ||||
|     let event2_start = (now + Duration::days(1)).timestamp(); | ||||
|     let event2_end = (now + Duration::days(1) + Duration::minutes(90)).timestamp(); | ||||
|     let event2 = Event::new() | ||||
|     .title("Project Brainstorm") | ||||
|     .reschedule(event2_start, event2_end) | ||||
|     .description("Brainstorming session for new project features.") | ||||
|     .location("Innovation Lab") | ||||
|     .color("#4CAF50") // Green color | ||||
| @@ -115,28 +101,28 @@ fn main() { | ||||
|     .category("Planning") | ||||
|     .reminder_minutes(30) | ||||
|     .is_recurring(true) | ||||
|     .add_attendee(attendee1_id) | ||||
|     .add_attendee(attendee3_id); | ||||
|     .add_attendee(attendee1.clone()) | ||||
|     .add_attendee(attendee3.clone()); | ||||
|  | ||||
|     let event3 = Event::new( | ||||
|         "Client Call", | ||||
|         now + Duration::days(2), | ||||
|         now + Duration::days(2) + Duration::hours(1), | ||||
|     ) | ||||
|     let event3_start = (now + Duration::days(2)).timestamp(); | ||||
|     let event3_end = (now + Duration::days(2) + Duration::hours(1)).timestamp(); | ||||
|     let event3 = Event::new() | ||||
|     .title("Client Call") | ||||
|     .reschedule(event3_start, event3_end) | ||||
|     .description("Quarterly review with key client.") | ||||
|     .color("#9C27B0") // Purple color | ||||
|     .created_by(user3_id) | ||||
|     .status(EventStatus::Published) | ||||
|     .category("Client") | ||||
|     .reminder_minutes(60) | ||||
|     .add_attendee(attendee2_id); | ||||
|     .add_attendee(attendee2.clone()); | ||||
|  | ||||
|     // Create an all-day event | ||||
|     let event4 = Event::new( | ||||
|         "Company Holiday", | ||||
|         now + Duration::days(7), | ||||
|         now + Duration::days(7) + Duration::hours(24), | ||||
|     ) | ||||
|     let event4_start = (now + Duration::days(7)).timestamp(); | ||||
|     let event4_end = (now + Duration::days(7) + Duration::hours(24)).timestamp(); | ||||
|     let event4 = Event::new() | ||||
|     .title("Company Holiday") | ||||
|     .reschedule(event4_start, event4_end) | ||||
|     .description("National holiday - office closed.") | ||||
|     .color("#FFC107") // Amber color | ||||
|     .all_day(true) | ||||
| @@ -148,7 +134,7 @@ fn main() { | ||||
|     println!( | ||||
|         "- Event 1: '{}' at {} with {} attendees", | ||||
|         event1.title, | ||||
|         event1.start_time.format("%Y-%m-%d %H:%M"), | ||||
|         fmt_time(event1.start_time), | ||||
|         event1.attendees.len() | ||||
|     ); | ||||
|     println!( | ||||
| @@ -174,12 +160,19 @@ fn main() { | ||||
|     ); | ||||
|     println!("  All-day: {}", event1.all_day); | ||||
|     println!("  Recurring: {}", event1.is_recurring); | ||||
|     println!("  Attendee IDs: {:?}", event1.attendees); | ||||
|     println!( | ||||
|         "  Attendee IDs: {:?}", | ||||
|         event1 | ||||
|             .attendees | ||||
|             .iter() | ||||
|             .map(|a| a.contact_id) | ||||
|             .collect::<Vec<u32>>() | ||||
|     ); | ||||
|  | ||||
|     println!( | ||||
|         "- Event 2: '{}' at {} with {} attendees", | ||||
|         event2.title, | ||||
|         event2.start_time.format("%Y-%m-%d %H:%M"), | ||||
|         fmt_time(event2.start_time), | ||||
|         event2.attendees.len() | ||||
|     ); | ||||
|     println!( | ||||
| @@ -205,12 +198,19 @@ fn main() { | ||||
|     ); | ||||
|     println!("  All-day: {}", event2.all_day); | ||||
|     println!("  Recurring: {}", event2.is_recurring); | ||||
|     println!("  Attendee IDs: {:?}", event2.attendees); | ||||
|     println!( | ||||
|         "  Attendee IDs: {:?}", | ||||
|         event2 | ||||
|             .attendees | ||||
|             .iter() | ||||
|             .map(|a| a.contact_id) | ||||
|             .collect::<Vec<u32>>() | ||||
|     ); | ||||
|  | ||||
|     println!( | ||||
|         "- Event 3: '{}' at {} with {} attendees", | ||||
|         event3.title, | ||||
|         event3.start_time.format("%Y-%m-%d %H:%M"), | ||||
|         fmt_time(event3.start_time), | ||||
|         event3.attendees.len() | ||||
|     ); | ||||
|     println!( | ||||
| @@ -236,12 +236,19 @@ fn main() { | ||||
|     ); | ||||
|     println!("  All-day: {}", event3.all_day); | ||||
|     println!("  Recurring: {}", event3.is_recurring); | ||||
|     println!("  Attendee IDs: {:?}", event3.attendees); | ||||
|     println!( | ||||
|         "  Attendee IDs: {:?}", | ||||
|         event3 | ||||
|             .attendees | ||||
|             .iter() | ||||
|             .map(|a| a.contact_id) | ||||
|             .collect::<Vec<u32>>() | ||||
|     ); | ||||
|  | ||||
|     println!( | ||||
|         "- Event 4: '{}' at {} (All-day: {})", | ||||
|         event4.title, | ||||
|         event4.start_time.format("%Y-%m-%d"), | ||||
|         fmt_date(event4.start_time), | ||||
|         event4.all_day | ||||
|     ); | ||||
|     println!( | ||||
| @@ -262,25 +269,37 @@ fn main() { | ||||
|     let new_start = now + Duration::hours(2); | ||||
|     let new_end = now + Duration::hours(3); | ||||
|     let mut updated_event1 = event1.clone(); | ||||
|     updated_event1 = updated_event1.reschedule(new_start, new_end); | ||||
|     updated_event1 = updated_event1.reschedule(new_start.timestamp(), new_end.timestamp()); | ||||
|     println!( | ||||
|         "Rescheduled '{}' to {}", | ||||
|         updated_event1.title, | ||||
|         new_start.format("%Y-%m-%d %H:%M") | ||||
|         fmt_time(new_start.timestamp()) | ||||
|     ); | ||||
|  | ||||
|     // Remove an attendee | ||||
|     updated_event1 = updated_event1.remove_attendee(attendee1_id); | ||||
|     updated_event1 = updated_event1.remove_attendee(user1_id); | ||||
|     println!( | ||||
|         "Removed attendee {} from '{}'. Remaining attendee IDs: {:?}", | ||||
|         attendee1_id, updated_event1.title, updated_event1.attendees | ||||
|         user1_id, | ||||
|         updated_event1.title, | ||||
|         updated_event1 | ||||
|             .attendees | ||||
|             .iter() | ||||
|             .map(|a| a.contact_id) | ||||
|             .collect::<Vec<u32>>() | ||||
|     ); | ||||
|  | ||||
|     // Add a new attendee | ||||
|     updated_event1 = updated_event1.add_attendee(attendee3_id); | ||||
|     updated_event1 = updated_event1.add_attendee(attendee3.clone()); | ||||
|     println!( | ||||
|         "Added attendee {} to '{}'. Current attendee IDs: {:?}", | ||||
|         attendee3_id, updated_event1.title, updated_event1.attendees | ||||
|         user3_id, | ||||
|         updated_event1.title, | ||||
|         updated_event1 | ||||
|             .attendees | ||||
|             .iter() | ||||
|             .map(|a| a.contact_id) | ||||
|             .collect::<Vec<u32>>() | ||||
|     ); | ||||
|  | ||||
|     // --- Demonstrate Event Status Changes --- | ||||
| @@ -300,11 +319,11 @@ fn main() { | ||||
|     println!("Cancelled event: '{}'", cancelled_event.title); | ||||
|  | ||||
|     // Update event with new features | ||||
|     let enhanced_event = Event::new( | ||||
|         "Enhanced Meeting", | ||||
|         now + Duration::days(5), | ||||
|         now + Duration::days(5) + Duration::hours(2), | ||||
|     ) | ||||
|     let enhanced_start = (now + Duration::days(5)).timestamp(); | ||||
|     let enhanced_end = (now + Duration::days(5) + Duration::hours(2)).timestamp(); | ||||
|     let enhanced_event = Event::new() | ||||
|     .title("Enhanced Meeting") | ||||
|     .reschedule(enhanced_start, enhanced_end) | ||||
|     .description("Meeting with all new features demonstrated.") | ||||
|     .location("Virtual - Zoom") | ||||
|     .color("#673AB7") // Deep purple | ||||
| @@ -314,9 +333,9 @@ fn main() { | ||||
|     .reminder_minutes(45) | ||||
|     .timezone("America/New_York") | ||||
|     .is_recurring(true) | ||||
|     .add_attendee(attendee1_id) | ||||
|     .add_attendee(attendee2_id) | ||||
|     .add_attendee(attendee3_id); | ||||
|     .add_attendee(attendee1) | ||||
|     .add_attendee(attendee2) | ||||
|     .add_attendee(attendee3); | ||||
|  | ||||
|     println!("Created enhanced event with all features:"); | ||||
|     println!("  Title: {}", enhanced_event.title); | ||||
| @@ -485,13 +504,13 @@ fn main() { | ||||
|     println!("\n--- Modifying Calendar ---"); | ||||
|  | ||||
|     // Create and store a new event | ||||
|     let new_event = Event::new( | ||||
|         "1-on-1 Meeting", | ||||
|         now + Duration::days(3), | ||||
|         now + Duration::days(3) + Duration::minutes(30), | ||||
|     ) | ||||
|     .description("One-on-one meeting with team member.") | ||||
|     .location("Office"); | ||||
|     let ne_start = (now + Duration::days(3)).timestamp(); | ||||
|     let ne_end = (now + Duration::days(3) + Duration::minutes(30)).timestamp(); | ||||
|     let new_event = Event::new() | ||||
|         .title("1-on-1 Meeting") | ||||
|         .reschedule(ne_start, ne_end) | ||||
|         .description("One-on-one meeting with team member.") | ||||
|         .location("Office"); | ||||
|  | ||||
|     let (new_event_id, _stored_new_event) = | ||||
|         event_collection.set(&new_event).expect("can set new event"); | ||||
| @@ -565,7 +584,7 @@ fn main() { | ||||
|             "- Event ID: {}, Title: '{}', Start: {}, Attendees: {}", | ||||
|             event.get_id(), | ||||
|             event.title, | ||||
|             event.start_time.format("%Y-%m-%d %H:%M"), | ||||
|             fmt_time(event.start_time), | ||||
|             event.attendees.len() | ||||
|         ); | ||||
|     } | ||||
| @@ -583,22 +602,16 @@ fn main() { | ||||
|             retrieved_event1.attendees.len() | ||||
|         ); | ||||
|  | ||||
|         // Look up attendee details for each attendee ID | ||||
|         for &attendee_id in &retrieved_event1.attendees { | ||||
|             if let Some(attendee) = attendee_collection | ||||
|                 .get_by_id(attendee_id) | ||||
|                 .expect("can try to get attendee") | ||||
|         // Look up attendee details directly from embedded attendees | ||||
|         for attendee in &retrieved_event1.attendees { | ||||
|             if let Some(user) = user_collection | ||||
|                 .get_by_id(attendee.contact_id) | ||||
|                 .expect("can try to get user") | ||||
|             { | ||||
|                 // Look up user details for the attendee's contact_id | ||||
|                 if let Some(user) = user_collection | ||||
|                     .get_by_id(attendee.contact_id) | ||||
|                     .expect("can try to get user") | ||||
|                 { | ||||
|                     println!( | ||||
|                         "  - Attendee ID {}: {} (User: {}, Status: {:?})", | ||||
|                         attendee_id, user.full_name, attendee.contact_id, attendee.status | ||||
|                     ); | ||||
|                 } | ||||
|                 println!( | ||||
|                     "  - User {}: {} (Status: {:?})", | ||||
|                     attendee.contact_id, user.full_name, attendee.status | ||||
|                 ); | ||||
|             } | ||||
|         } | ||||
|     } | ||||
|   | ||||
| @@ -1,26 +1,26 @@ | ||||
| use circles_launcher::{new_launcher}; | ||||
| use heromodels::models::circle::circle::{new_circle}; | ||||
| use secp256k1::{Secp256k1, SecretKey, PublicKey}; | ||||
| use circles_launcher::new_launcher; | ||||
| use heromodels::models::circle::circle::new_circle; | ||||
| use rand::rngs::OsRng; | ||||
| use secp256k1::{PublicKey, Secp256k1, SecretKey}; | ||||
|  | ||||
| #[tokio::main] | ||||
| async fn main() -> Result<(), Box<dyn std::error::Error>> {     | ||||
| async fn main() -> Result<(), Box<dyn std::error::Error>> { | ||||
|     // Generate valid secp256k1 keypairs for testing | ||||
|     let secp = Secp256k1::new(); | ||||
|     let mut rng = OsRng; | ||||
|      | ||||
|  | ||||
|     let secret_key1 = SecretKey::new(&mut rng); | ||||
|     let public_key1 = PublicKey::from_secret_key(&secp, &secret_key1); | ||||
|     let pk1_hex = hex::encode(public_key1.serialize()); | ||||
|      | ||||
|  | ||||
|     let secret_key2 = SecretKey::new(&mut rng); | ||||
|     let public_key2 = PublicKey::from_secret_key(&secp, &secret_key2); | ||||
|     let pk2_hex = hex::encode(public_key2.serialize()); | ||||
|      | ||||
|  | ||||
|     let secret_key3 = SecretKey::new(&mut rng); | ||||
|     let public_key3 = PublicKey::from_secret_key(&secp, &secret_key3); | ||||
|     let pk3_hex = hex::encode(public_key3.serialize()); | ||||
|      | ||||
|  | ||||
|     println!("Generated test public keys:"); | ||||
|     println!("  PK1: {}", pk1_hex); | ||||
|     println!("  PK2: {}", pk2_hex); | ||||
| @@ -36,4 +36,4 @@ async fn main() -> Result<(), Box<dyn std::error::Error>> { | ||||
|         .save(); | ||||
|  | ||||
|     Ok(()) | ||||
| } | ||||
| } | ||||
|   | ||||
							
								
								
									
										199
									
								
								heromodels/examples/grid4_bid_example.rs
									
									
									
									
									
										Normal file
									
								
							
							
						
						
									
										199
									
								
								heromodels/examples/grid4_bid_example.rs
									
									
									
									
									
										Normal file
									
								
							| @@ -0,0 +1,199 @@ | ||||
| use heromodels::db::{Collection, Db}; | ||||
| use heromodels::models::grid4::{Bid, BidStatus, BillingPeriod}; | ||||
| use heromodels::models::grid4::bid::bid_index::customer_id; | ||||
| use heromodels_core::Model; | ||||
|  | ||||
| // Helper function to print bid details | ||||
| fn print_bid_details(bid: &Bid) { | ||||
|     println!("\n--- Bid Details ---"); | ||||
|     println!("ID: {}", bid.get_id()); | ||||
|     println!("Customer ID: {}", bid.customer_id); | ||||
|     println!("Compute Slices: {}", bid.compute_slices_nr); | ||||
|     println!("Compute Slice Price: ${:.2}", bid.compute_slice_price); | ||||
|     println!("Storage Slices: {}", bid.storage_slices_nr); | ||||
|     println!("Storage Slice Price: ${:.2}", bid.storage_slice_price); | ||||
|     println!("Status: {:?}", bid.status); | ||||
|     println!("Obligation: {}", bid.obligation); | ||||
|     println!("Start Date: {}", bid.start_date); | ||||
|     println!("End Date: {}", bid.end_date); | ||||
|     println!("Billing Period: {:?}", bid.billing_period); | ||||
|     println!("Signature User: {}", bid.signature_user); | ||||
|     println!("Created At: {}", bid.base_data.created_at); | ||||
|     println!("Modified At: {}", bid.base_data.modified_at); | ||||
| } | ||||
|  | ||||
| fn main() { | ||||
|     // Create a new DB instance in /tmp/grid4_db, and reset before every run | ||||
|     let db = heromodels::db::hero::OurDB::new("/tmp/grid4_db", true).expect("Can create DB"); | ||||
|  | ||||
|     println!("Grid4 Bid Models - Basic Usage Example"); | ||||
|     println!("====================================="); | ||||
|  | ||||
|     // Create bids with different configurations | ||||
|  | ||||
|     // Bid 1 - Small compute request | ||||
|     let bid1 = Bid::new() | ||||
|         .customer_id(101) | ||||
|         .compute_slices_nr(4) | ||||
|         .compute_slice_price(0.05) | ||||
|         .storage_slices_nr(10) | ||||
|         .storage_slice_price(0.02) | ||||
|         .status(BidStatus::Pending) | ||||
|         .obligation(false) | ||||
|         .start_date(1640995200) // 2022-01-01 | ||||
|         .end_date(1672531200)   // 2023-01-01 | ||||
|         .billing_period(BillingPeriod::Monthly) | ||||
|         .signature_user("sig_user_101_abc123".to_string()); | ||||
|  | ||||
|     // Bid 2 - Large compute request with obligation | ||||
|     let bid2 = Bid::new() | ||||
|         .customer_id(102) | ||||
|         .compute_slices_nr(16) | ||||
|         .compute_slice_price(0.04) | ||||
|         .storage_slices_nr(50) | ||||
|         .storage_slice_price(0.015) | ||||
|         .status(BidStatus::Confirmed) | ||||
|         .obligation(true) | ||||
|         .start_date(1640995200) | ||||
|         .end_date(1704067200)   // 2024-01-01 | ||||
|         .billing_period(BillingPeriod::Yearly) | ||||
|         .signature_user("sig_user_102_def456".to_string()); | ||||
|  | ||||
|     // Bid 3 - Storage-heavy request | ||||
|     let bid3 = Bid::new() | ||||
|         .customer_id(103) | ||||
|         .compute_slices_nr(2) | ||||
|         .compute_slice_price(0.06) | ||||
|         .storage_slices_nr(100) | ||||
|         .storage_slice_price(0.01) | ||||
|         .status(BidStatus::Assigned) | ||||
|         .obligation(true) | ||||
|         .start_date(1640995200) | ||||
|         .end_date(1672531200) | ||||
|         .billing_period(BillingPeriod::Hourly) | ||||
|         .signature_user("sig_user_103_ghi789".to_string()); | ||||
|  | ||||
|     // Bid 4 - Cancelled bid | ||||
|     let bid4 = Bid::new() | ||||
|         .customer_id(104) | ||||
|         .compute_slices_nr(8) | ||||
|         .compute_slice_price(0.055) | ||||
|         .storage_slices_nr(25) | ||||
|         .storage_slice_price(0.018) | ||||
|         .status(BidStatus::Cancelled) | ||||
|         .obligation(false) | ||||
|         .start_date(1640995200) | ||||
|         .end_date(1672531200) | ||||
|         .billing_period(BillingPeriod::Monthly) | ||||
|         .signature_user("sig_user_104_jkl012".to_string()); | ||||
|  | ||||
|     // Save all bids to database and get their assigned IDs and updated models | ||||
|     let (bid1_id, db_bid1) = db | ||||
|         .collection() | ||||
|         .expect("can open bid collection") | ||||
|         .set(&bid1) | ||||
|         .expect("can set bid"); | ||||
|     let (bid2_id, db_bid2) = db | ||||
|         .collection() | ||||
|         .expect("can open bid collection") | ||||
|         .set(&bid2) | ||||
|         .expect("can set bid"); | ||||
|     let (bid3_id, db_bid3) = db | ||||
|         .collection() | ||||
|         .expect("can open bid collection") | ||||
|         .set(&bid3) | ||||
|         .expect("can set bid"); | ||||
|     let (bid4_id, db_bid4) = db | ||||
|         .collection() | ||||
|         .expect("can open bid collection") | ||||
|         .set(&bid4) | ||||
|         .expect("can set bid"); | ||||
|  | ||||
|     println!("Bid 1 assigned ID: {}", bid1_id); | ||||
|     println!("Bid 2 assigned ID: {}", bid2_id); | ||||
|     println!("Bid 3 assigned ID: {}", bid3_id); | ||||
|     println!("Bid 4 assigned ID: {}", bid4_id); | ||||
|  | ||||
|     // Print all bids retrieved from database | ||||
|     println!("\n--- Bids Retrieved from Database ---"); | ||||
|     println!("\n1. Small compute bid:"); | ||||
|     print_bid_details(&db_bid1); | ||||
|  | ||||
|     println!("\n2. Large compute bid with obligation:"); | ||||
|     print_bid_details(&db_bid2); | ||||
|  | ||||
|     println!("\n3. Storage-heavy bid:"); | ||||
|     print_bid_details(&db_bid3); | ||||
|  | ||||
|     println!("\n4. Cancelled bid:"); | ||||
|     print_bid_details(&db_bid4); | ||||
|  | ||||
|     // Demonstrate different ways to retrieve bids from the database | ||||
|     println!("\n--- Retrieving Bids by Different Methods ---"); | ||||
|     println!("\n1. By Customer ID Index (Customer 102):"); | ||||
|      | ||||
|     let customer_bids = db | ||||
|         .collection::<Bid>() | ||||
|         .expect("can open bid collection") | ||||
|         .get::<customer_id, _>(&102u32) | ||||
|         .expect("can load bids by customer"); | ||||
|  | ||||
|     assert_eq!(customer_bids.len(), 1); | ||||
|     print_bid_details(&customer_bids[0]); | ||||
|  | ||||
|     println!("\n2. Updating Bid Status:"); | ||||
|     let mut updated_bid = db_bid1.clone(); | ||||
|     updated_bid.status = BidStatus::Confirmed; | ||||
|  | ||||
|     let (_, confirmed_bid) = db | ||||
|         .collection::<Bid>() | ||||
|         .expect("can open bid collection") | ||||
|         .set(&updated_bid) | ||||
|         .expect("can update bid"); | ||||
|  | ||||
|     println!("Updated bid status to Confirmed:"); | ||||
|     print_bid_details(&confirmed_bid); | ||||
|  | ||||
|     // 3. Delete a bid and show the updated results | ||||
|     println!("\n3. After Deleting a Bid:"); | ||||
|     println!("Deleting bid with ID: {}", bid4_id); | ||||
|     db.collection::<Bid>() | ||||
|         .expect("can open bid collection") | ||||
|         .delete_by_id(bid4_id) | ||||
|         .expect("can delete existing bid"); | ||||
|  | ||||
|     // Show remaining bids | ||||
|     let all_bids = db | ||||
|         .collection::<Bid>() | ||||
|         .expect("can open bid collection") | ||||
|         .get_all() | ||||
|         .expect("can load all bids"); | ||||
|  | ||||
|     println!("Remaining bids count: {}", all_bids.len()); | ||||
|     assert_eq!(all_bids.len(), 3); | ||||
|  | ||||
|     // Calculate total compute and storage requested | ||||
|     println!("\n--- Bid Analytics ---"); | ||||
|     let total_compute_slices: i32 = all_bids.iter().map(|b| b.compute_slices_nr).sum(); | ||||
|     let total_storage_slices: i32 = all_bids.iter().map(|b| b.storage_slices_nr).sum(); | ||||
|     let avg_compute_price: f64 = all_bids.iter().map(|b| b.compute_slice_price).sum::<f64>() / all_bids.len() as f64; | ||||
|     let avg_storage_price: f64 = all_bids.iter().map(|b| b.storage_slice_price).sum::<f64>() / all_bids.len() as f64; | ||||
|  | ||||
|     println!("Total Compute Slices Requested: {}", total_compute_slices); | ||||
|     println!("Total Storage Slices Requested: {}", total_storage_slices); | ||||
|     println!("Average Compute Price: ${:.3}", avg_compute_price); | ||||
|     println!("Average Storage Price: ${:.3}", avg_storage_price); | ||||
|  | ||||
|     // Count bids by status | ||||
|     let confirmed_count = all_bids.iter().filter(|b| matches!(b.status, BidStatus::Confirmed)).count(); | ||||
|     let assigned_count = all_bids.iter().filter(|b| matches!(b.status, BidStatus::Assigned)).count(); | ||||
|     let pending_count = all_bids.iter().filter(|b| matches!(b.status, BidStatus::Pending)).count(); | ||||
|  | ||||
|     println!("\nBids by Status:"); | ||||
|     println!("  Confirmed: {}", confirmed_count); | ||||
|     println!("  Assigned: {}", assigned_count); | ||||
|     println!("  Pending: {}", pending_count); | ||||
|  | ||||
|     println!("\n--- Model Information ---"); | ||||
|     println!("Bid DB Prefix: {}", Bid::db_prefix()); | ||||
| } | ||||
							
								
								
									
										301
									
								
								heromodels/examples/grid4_contract_example.rs
									
									
									
									
									
										Normal file
									
								
							
							
						
						
									
										301
									
								
								heromodels/examples/grid4_contract_example.rs
									
									
									
									
									
										Normal file
									
								
							| @@ -0,0 +1,301 @@ | ||||
| use heromodels::db::{Collection, Db}; | ||||
| use heromodels::models::grid4::{Contract, ContractStatus}; | ||||
| use heromodels::models::grid4::contract::contract_index::customer_id; | ||||
| use heromodels_core::Model; | ||||
|  | ||||
| // Helper function to print contract details | ||||
| fn print_contract_details(contract: &Contract) { | ||||
|     println!("\n--- Contract Details ---"); | ||||
|     println!("ID: {}", contract.get_id()); | ||||
|     println!("Customer ID: {}", contract.customer_id); | ||||
|     println!("Compute Slices: {}", contract.compute_slices.len()); | ||||
|     println!("Storage Slices: {}", contract.storage_slices.len()); | ||||
|     println!("Compute Slice Price: ${:.2}", contract.compute_slice_price); | ||||
|     println!("Storage Slice Price: ${:.2}", contract.storage_slice_price); | ||||
|     println!("Network Slice Price: ${:.2}", contract.network_slice_price); | ||||
|     println!("Status: {:?}", contract.status); | ||||
|     println!("Start Date: {}", contract.start_date); | ||||
|     println!("End Date: {}", contract.end_date); | ||||
|     println!("Billing Period: {:?}", contract.billing_period); | ||||
|     println!("Signature User: {}", contract.signature_user); | ||||
|     println!("Signature Hoster: {}", contract.signature_hoster); | ||||
|     println!("Created At: {}", contract.base_data.created_at); | ||||
|     println!("Modified At: {}", contract.base_data.modified_at); | ||||
|      | ||||
|     // Print compute slices details | ||||
|     if !contract.compute_slices.is_empty() { | ||||
|         println!("  Compute Slices:"); | ||||
|         for (i, slice) in contract.compute_slices.iter().enumerate() { | ||||
|             println!("    {}. Node: {}, ID: {}, Memory: {:.1}GB, Storage: {:.1}GB, Passmark: {}, vCores: {}",  | ||||
|                 i + 1, slice.node_id, slice.id, slice.mem_gb, slice.storage_gb, slice.passmark, slice.vcores); | ||||
|         } | ||||
|     } | ||||
|      | ||||
|     // Print storage slices details | ||||
|     if !contract.storage_slices.is_empty() { | ||||
|         println!("  Storage Slices:"); | ||||
|         for (i, slice) in contract.storage_slices.iter().enumerate() { | ||||
|             println!("    {}. Node: {}, ID: {}, Size: {}GB",  | ||||
|                 i + 1, slice.node_id, slice.id, slice.storage_size_gb); | ||||
|         } | ||||
|     } | ||||
| } | ||||
|  | ||||
| fn main() { | ||||
|     // Create a new DB instance in /tmp/grid4_contracts_db, and reset before every run | ||||
|     let db = heromodels::db::hero::OurDB::new("/tmp/grid4_contracts_db", true).expect("Can create DB"); | ||||
|  | ||||
|     println!("Grid4 Contract Models - Basic Usage Example"); | ||||
|     println!("=========================================="); | ||||
|  | ||||
|     // Create compute slices for contracts | ||||
|     let compute_slice1 = ComputeSliceProvisioned::new() | ||||
|         .node_id(1001) | ||||
|         .id(1) | ||||
|         .mem_gb(2.0) | ||||
|         .storage_gb(20.0) | ||||
|         .passmark(2500) | ||||
|         .vcores(2) | ||||
|         .cpu_oversubscription(150) | ||||
|         .tags("web-server,production".to_string()); | ||||
|  | ||||
|     let compute_slice2 = ComputeSliceProvisioned::new() | ||||
|         .node_id(1002) | ||||
|         .id(2) | ||||
|         .mem_gb(4.0) | ||||
|         .storage_gb(40.0) | ||||
|         .passmark(5000) | ||||
|         .vcores(4) | ||||
|         .cpu_oversubscription(120) | ||||
|         .tags("database,high-performance".to_string()); | ||||
|  | ||||
|     let compute_slice3 = ComputeSliceProvisioned::new() | ||||
|         .node_id(1003) | ||||
|         .id(1) | ||||
|         .mem_gb(8.0) | ||||
|         .storage_gb(80.0) | ||||
|         .passmark(10000) | ||||
|         .vcores(8) | ||||
|         .cpu_oversubscription(100) | ||||
|         .tags("ml-training,gpu-enabled".to_string()); | ||||
|  | ||||
|     // Create storage slices for contracts | ||||
|     let storage_slice1 = StorageSliceProvisioned::new() | ||||
|         .node_id(2001) | ||||
|         .id(1) | ||||
|         .storage_size_gb(100) | ||||
|         .tags("backup,cold-storage".to_string()); | ||||
|  | ||||
|     let storage_slice2 = StorageSliceProvisioned::new() | ||||
|         .node_id(2002) | ||||
|         .id(2) | ||||
|         .storage_size_gb(500) | ||||
|         .tags("data-lake,analytics".to_string()); | ||||
|  | ||||
|     let storage_slice3 = StorageSliceProvisioned::new() | ||||
|         .node_id(2003) | ||||
|         .id(1) | ||||
|         .storage_size_gb(1000) | ||||
|         .tags("archive,long-term".to_string()); | ||||
|  | ||||
|     // Create contracts with different configurations | ||||
|  | ||||
|     // Contract 1 - Small web hosting contract | ||||
|     let contract1 = Contract::new() | ||||
|         .customer_id(201) | ||||
|         .add_compute_slice(compute_slice1.clone()) | ||||
|         .add_storage_slice(storage_slice1.clone()) | ||||
|         .compute_slice_price(0.05) | ||||
|         .storage_slice_price(0.02) | ||||
|         .network_slice_price(0.01) | ||||
|         .status(ContractStatus::Active) | ||||
|         .start_date(1640995200) // 2022-01-01 | ||||
|         .end_date(1672531200)   // 2023-01-01 | ||||
|         .billing_period(BillingPeriod::Monthly) | ||||
|         .signature_user("contract_user_201_abc123".to_string()) | ||||
|         .signature_hoster("hoster_node1001_xyz789".to_string()); | ||||
|  | ||||
|     // Contract 2 - Database hosting contract | ||||
|     let contract2 = Contract::new() | ||||
|         .customer_id(202) | ||||
|         .add_compute_slice(compute_slice2.clone()) | ||||
|         .add_storage_slice(storage_slice2.clone()) | ||||
|         .compute_slice_price(0.04) | ||||
|         .storage_slice_price(0.015) | ||||
|         .network_slice_price(0.008) | ||||
|         .status(ContractStatus::Active) | ||||
|         .start_date(1640995200) | ||||
|         .end_date(1704067200)   // 2024-01-01 | ||||
|         .billing_period(BillingPeriod::Yearly) | ||||
|         .signature_user("contract_user_202_def456".to_string()) | ||||
|         .signature_hoster("hoster_node1002_uvw123".to_string()); | ||||
|  | ||||
|     // Contract 3 - ML training contract (paused) | ||||
|     let contract3 = Contract::new() | ||||
|         .customer_id(203) | ||||
|         .add_compute_slice(compute_slice3.clone()) | ||||
|         .add_storage_slice(storage_slice3.clone()) | ||||
|         .compute_slice_price(0.08) | ||||
|         .storage_slice_price(0.01) | ||||
|         .network_slice_price(0.015) | ||||
|         .status(ContractStatus::Paused) | ||||
|         .start_date(1640995200) | ||||
|         .end_date(1672531200) | ||||
|         .billing_period(BillingPeriod::Hourly) | ||||
|         .signature_user("contract_user_203_ghi789".to_string()) | ||||
|         .signature_hoster("hoster_node1003_rst456".to_string()); | ||||
|  | ||||
|     // Contract 4 - Multi-slice enterprise contract | ||||
|     let contract4 = Contract::new() | ||||
|         .customer_id(204) | ||||
|         .add_compute_slice(compute_slice1.clone()) | ||||
|         .add_compute_slice(compute_slice2.clone()) | ||||
|         .add_storage_slice(storage_slice1.clone()) | ||||
|         .add_storage_slice(storage_slice2.clone()) | ||||
|         .compute_slice_price(0.045) | ||||
|         .storage_slice_price(0.018) | ||||
|         .network_slice_price(0.012) | ||||
|         .status(ContractStatus::Active) | ||||
|         .start_date(1640995200) | ||||
|         .end_date(1735689600)   // 2025-01-01 | ||||
|         .billing_period(BillingPeriod::Monthly) | ||||
|         .signature_user("contract_user_204_jkl012".to_string()) | ||||
|         .signature_hoster("hoster_enterprise_mno345".to_string()); | ||||
|  | ||||
|     // Save all contracts to database and get their assigned IDs and updated models | ||||
|     let (contract1_id, db_contract1) = db | ||||
|         .collection() | ||||
|         .expect("can open contract collection") | ||||
|         .set(&contract1) | ||||
|         .expect("can set contract"); | ||||
|     let (contract2_id, db_contract2) = db | ||||
|         .collection() | ||||
|         .expect("can open contract collection") | ||||
|         .set(&contract2) | ||||
|         .expect("can set contract"); | ||||
|     let (contract3_id, db_contract3) = db | ||||
|         .collection() | ||||
|         .expect("can open contract collection") | ||||
|         .set(&contract3) | ||||
|         .expect("can set contract"); | ||||
|     let (contract4_id, db_contract4) = db | ||||
|         .collection() | ||||
|         .expect("can open contract collection") | ||||
|         .set(&contract4) | ||||
|         .expect("can set contract"); | ||||
|  | ||||
|     println!("Contract 1 assigned ID: {}", contract1_id); | ||||
|     println!("Contract 2 assigned ID: {}", contract2_id); | ||||
|     println!("Contract 3 assigned ID: {}", contract3_id); | ||||
|     println!("Contract 4 assigned ID: {}", contract4_id); | ||||
|  | ||||
|     // Print all contracts retrieved from database | ||||
|     println!("\n--- Contracts Retrieved from Database ---"); | ||||
|     println!("\n1. Web hosting contract:"); | ||||
|     print_contract_details(&db_contract1); | ||||
|  | ||||
|     println!("\n2. Database hosting contract:"); | ||||
|     print_contract_details(&db_contract2); | ||||
|  | ||||
|     println!("\n3. ML training contract (paused):"); | ||||
|     print_contract_details(&db_contract3); | ||||
|  | ||||
|     println!("\n4. Enterprise multi-slice contract:"); | ||||
|     print_contract_details(&db_contract4); | ||||
|  | ||||
|     // Demonstrate different ways to retrieve contracts from the database | ||||
|  | ||||
|     // 1. Retrieve by customer ID index | ||||
|     println!("\n--- Retrieving Contracts by Different Methods ---"); | ||||
|     println!("\n1. By Customer ID Index (Customer 202):"); | ||||
|     let customer_contracts = db | ||||
|         .collection::<Contract>() | ||||
|         .expect("can open contract collection") | ||||
|         .get::<customer_id, _>(&202u32) | ||||
|         .expect("can load contracts by customer"); | ||||
|  | ||||
|     assert_eq!(customer_contracts.len(), 1); | ||||
|     print_contract_details(&customer_contracts[0]); | ||||
|  | ||||
|     // 2. Update contract status | ||||
|     println!("\n2. Resuming Paused Contract:"); | ||||
|     let mut updated_contract = db_contract3.clone(); | ||||
|     updated_contract.status = ContractStatus::Active; | ||||
|  | ||||
|     let (_, resumed_contract) = db | ||||
|         .collection::<Contract>() | ||||
|         .expect("can open contract collection") | ||||
|         .set(&updated_contract) | ||||
|         .expect("can update contract"); | ||||
|  | ||||
|     println!("Updated contract status to Active:"); | ||||
|     print_contract_details(&resumed_contract); | ||||
|  | ||||
|     // 3. Cancel a contract | ||||
|     println!("\n3. Cancelling a Contract:"); | ||||
|     let mut cancelled_contract = db_contract1.clone(); | ||||
|     cancelled_contract.status = ContractStatus::Cancelled; | ||||
|  | ||||
|     let (_, final_contract) = db | ||||
|         .collection::<Contract>() | ||||
|         .expect("can open contract collection") | ||||
|         .set(&cancelled_contract) | ||||
|         .expect("can update contract"); | ||||
|  | ||||
|     println!("Cancelled contract:"); | ||||
|     print_contract_details(&final_contract); | ||||
|  | ||||
|     // Show remaining active contracts | ||||
|     let all_contracts = db | ||||
|         .collection::<Contract>() | ||||
|         .expect("can open contract collection") | ||||
|         .get_all() | ||||
|         .expect("can load all contracts"); | ||||
|  | ||||
|     println!("\n--- Contract Analytics ---"); | ||||
|     let active_contracts: Vec<_> = all_contracts.iter() | ||||
|         .filter(|c| matches!(c.status, ContractStatus::Active)) | ||||
|         .collect(); | ||||
|     let paused_contracts: Vec<_> = all_contracts.iter() | ||||
|         .filter(|c| matches!(c.status, ContractStatus::Paused)) | ||||
|         .collect(); | ||||
|     let cancelled_contracts: Vec<_> = all_contracts.iter() | ||||
|         .filter(|c| matches!(c.status, ContractStatus::Cancelled)) | ||||
|         .collect(); | ||||
|  | ||||
|     println!("Total Contracts: {}", all_contracts.len()); | ||||
|     println!("Active Contracts: {}", active_contracts.len()); | ||||
|     println!("Paused Contracts: {}", paused_contracts.len()); | ||||
|     println!("Cancelled Contracts: {}", cancelled_contracts.len()); | ||||
|  | ||||
|     // Calculate total provisioned resources | ||||
|     let total_compute_slices: usize = all_contracts.iter().map(|c| c.compute_slices.len()).sum(); | ||||
|     let total_storage_slices: usize = all_contracts.iter().map(|c| c.storage_slices.len()).sum(); | ||||
|     let total_memory_gb: f64 = all_contracts.iter() | ||||
|         .flat_map(|c| &c.compute_slices) | ||||
|         .map(|s| s.mem_gb) | ||||
|         .sum(); | ||||
|     let total_storage_gb: i32 = all_contracts.iter() | ||||
|         .flat_map(|c| &c.storage_slices) | ||||
|         .map(|s| s.storage_size_gb) | ||||
|         .sum(); | ||||
|  | ||||
|     println!("\nProvisioned Resources:"); | ||||
|     println!("  Total Compute Slices: {}", total_compute_slices); | ||||
|     println!("  Total Storage Slices: {}", total_storage_slices); | ||||
|     println!("  Total Memory: {:.1} GB", total_memory_gb); | ||||
|     println!("  Total Storage: {} GB", total_storage_gb); | ||||
|  | ||||
|     // Calculate average pricing | ||||
|     let avg_compute_price: f64 = all_contracts.iter().map(|c| c.compute_slice_price).sum::<f64>() / all_contracts.len() as f64; | ||||
|     let avg_storage_price: f64 = all_contracts.iter().map(|c| c.storage_slice_price).sum::<f64>() / all_contracts.len() as f64; | ||||
|     let avg_network_price: f64 = all_contracts.iter().map(|c| c.network_slice_price).sum::<f64>() / all_contracts.len() as f64; | ||||
|  | ||||
|     println!("\nAverage Pricing:"); | ||||
|     println!("  Compute: ${:.3} per slice", avg_compute_price); | ||||
|     println!("  Storage: ${:.3} per slice", avg_storage_price); | ||||
|     println!("  Network: ${:.3} per slice", avg_network_price); | ||||
|  | ||||
|     println!("\n--- Model Information ---"); | ||||
|     println!("Contract DB Prefix: {}", Contract::db_prefix()); | ||||
| } | ||||
							
								
								
									
										12
									
								
								heromodels/examples/grid4_example/README.md
									
									
									
									
									
										Normal file
									
								
							
							
						
						
									
										12
									
								
								heromodels/examples/grid4_example/README.md
									
									
									
									
									
										Normal file
									
								
							| @@ -0,0 +1,12 @@ | ||||
| # Grid4 Node Example (OurDB) | ||||
|  | ||||
| This example demonstrates how to use the Grid4 `Node` model against the embedded OurDB backend. | ||||
|  | ||||
| - Creates an in-memory/on-disk OurDB under `/tmp`. | ||||
| - Demonstrates CRUD and simple index lookups on `country`, `nodegroupid`, and `pubkey`. | ||||
|  | ||||
| Run it: | ||||
|  | ||||
| ```bash | ||||
| cargo run -p heromodels --example grid4_example | ||||
| ``` | ||||
							
								
								
									
										66
									
								
								heromodels/examples/grid4_example/example.rs
									
									
									
									
									
										Normal file
									
								
							
							
						
						
									
										66
									
								
								heromodels/examples/grid4_example/example.rs
									
									
									
									
									
										Normal file
									
								
							| @@ -0,0 +1,66 @@ | ||||
| use heromodels::db::hero::OurDB; | ||||
| use heromodels::db::{Collection, Db}; | ||||
| use heromodels::models::grid4::node::node_index::{country, nodegroupid, pubkey}; | ||||
| use heromodels::models::grid4::node::{ComputeSlice, DeviceInfo, Node}; | ||||
| use std::sync::Arc; | ||||
|  | ||||
| fn main() { | ||||
|     // Create a temp OurDB | ||||
|     let ts = std::time::SystemTime::now() | ||||
|         .duration_since(std::time::UNIX_EPOCH) | ||||
|         .unwrap() | ||||
|         .as_nanos(); | ||||
|     let path = format!("/tmp/grid4_example_{}", ts); | ||||
|     let _ = std::fs::remove_dir_all(&path); | ||||
|     let db = Arc::new(OurDB::new(&path, true).expect("create OurDB")); | ||||
|  | ||||
|     let nodes = db.collection::<Node>().expect("open node collection"); | ||||
|  | ||||
|     // Build a node | ||||
|     let cs = ComputeSlice::new() | ||||
|         .nodeid(1) | ||||
|         .slice_id(1) | ||||
|         .mem_gb(64.0) | ||||
|         .storage_gb(1024.0) | ||||
|         .passmark(8000) | ||||
|         .vcores(24) | ||||
|         .gpus(2) | ||||
|         .price_cc(0.5); | ||||
|  | ||||
|     let dev = DeviceInfo { | ||||
|         vendor: "ACME".into(), | ||||
|         ..Default::default() | ||||
|     }; | ||||
|  | ||||
|     let n = Node::new() | ||||
|         .nodegroupid(7) | ||||
|         .uptime(98) | ||||
|         .add_compute_slice(cs) | ||||
|         .devices(dev) | ||||
|         .country("BE") | ||||
|         .pubkey("PUB_NODE_X") | ||||
|         .build(); | ||||
|  | ||||
|     // Store | ||||
|     let (id, stored) = nodes.set(&n).expect("store node"); | ||||
|     println!("Stored node id={id} pubkey={} country={}", stored.pubkey, stored.country); | ||||
|  | ||||
|     // Query by indexes | ||||
|     let by_country = nodes.get::<country, _>("BE").expect("query country"); | ||||
|     println!("Found {} nodes in country=BE", by_country.len()); | ||||
|  | ||||
|     let by_group = nodes.get::<nodegroupid, _>(&7).expect("query group"); | ||||
|     println!("Found {} nodes in group=7", by_group.len()); | ||||
|  | ||||
|     let by_key = nodes.get::<pubkey, _>("PUB_NODE_X").expect("query pubkey"); | ||||
|     println!("Found {} with pubkey PUB_NODE_X", by_key.len()); | ||||
|  | ||||
|     // Update | ||||
|     let updated = stored.clone().country("NL"); | ||||
|     let (_, back) = nodes.set(&updated).expect("update node"); | ||||
|     println!("Updated node country={}", back.country); | ||||
|  | ||||
|     // Delete | ||||
|     nodes.delete_by_id(id).expect("delete node"); | ||||
|     println!("Deleted node id={id}"); | ||||
| } | ||||
							
								
								
									
										390
									
								
								heromodels/examples/grid4_node_example.rs
									
									
									
									
									
										Normal file
									
								
							
							
						
						
									
										390
									
								
								heromodels/examples/grid4_node_example.rs
									
									
									
									
									
										Normal file
									
								
							| @@ -0,0 +1,390 @@ | ||||
| use heromodels::db::{Collection, Db}; | ||||
| use heromodels::models::grid4::{Node, NodeDevice, ComputeSlice, StorageSlice}; | ||||
| use heromodels::models::grid4::node::node_index::{nodegroupid, country}; | ||||
| use heromodels_core::Model; | ||||
|  | ||||
| // Helper function to print node details | ||||
| fn print_node_details(node: &Node) { | ||||
|     println!("\n--- Node Details ---"); | ||||
|     println!("ID: {}", node.get_id()); | ||||
|     println!("NodeGroup ID: {}", node.nodegroupid); | ||||
|     println!("Uptime: {}%", node.uptime); | ||||
|     println!("Country: {}", node.country); | ||||
|     println!("Birth Time: {}", node.birthtime); | ||||
|     println!("Public Key: {}", node.pubkey); | ||||
|     println!("Compute Slices: {}", node.computeslices.len()); | ||||
|     println!("Storage Slices: {}", node.storageslices.len()); | ||||
|     println!("Created At: {}", node.base_data.created_at); | ||||
|     println!("Modified At: {}", node.base_data.modified_at); | ||||
|      | ||||
|     // Print capacity details | ||||
|     println!("  Capacity:"); | ||||
|     println!("    Storage: {:.1} GB", node.capacity.storage_gb); | ||||
|     println!("    Memory: {:.1} GB", node.capacity.mem_gb); | ||||
|     println!("    GPU Memory: {:.1} GB", node.capacity.mem_gb_gpu); | ||||
|     println!("    Passmark: {}", node.capacity.passmark); | ||||
|     println!("    vCores: {}", node.capacity.vcores); | ||||
|      | ||||
|     // Print device info | ||||
|     println!("  Devices:"); | ||||
|     println!("    Vendor: {}", node.devices.vendor); | ||||
|     println!("    CPUs: {}", node.devices.cpu.len()); | ||||
|     println!("    GPUs: {}", node.devices.gpu.len()); | ||||
|     println!("    Memory: {}", node.devices.memory.len()); | ||||
|     println!("    Storage: {}", node.devices.storage.len()); | ||||
|     println!("    Network: {}", node.devices.network.len()); | ||||
|      | ||||
|     // Print compute slices | ||||
|     if !node.computeslices.is_empty() { | ||||
|         println!("  Compute Slices:"); | ||||
|         for (i, slice) in node.computeslices.iter().enumerate() { | ||||
|             println!("    {}. ID: {}, Memory: {:.1}GB, Storage: {:.1}GB, vCores: {}, GPUs: {}",  | ||||
|                 i + 1, slice.id, slice.mem_gb, slice.storage_gb, slice.vcores, slice.gpus); | ||||
|         } | ||||
|     } | ||||
|      | ||||
|     // Print storage slices | ||||
|     if !node.storageslices.is_empty() { | ||||
|         println!("  Storage Slices:"); | ||||
|         for (i, slice) in node.storageslices.iter().enumerate() { | ||||
|             println!("    {}. ID: {}", i + 1, slice.id); | ||||
|         } | ||||
|     } | ||||
| } | ||||
|  | ||||
| fn main() { | ||||
|     // Create a new DB instance in /tmp/grid4_nodes_db, and reset before every run | ||||
|     let db = heromodels::db::hero::OurDB::new("/tmp/grid4_nodes_db", true).expect("Can create DB"); | ||||
|  | ||||
|     println!("Grid4 Node Models - Basic Usage Example"); | ||||
|     println!("======================================"); | ||||
|  | ||||
|     // Create device components for nodes | ||||
|      | ||||
|     // CPU devices | ||||
|     let cpu1 = CPUDevice { | ||||
|         id: "cpu_intel_i7_12700k".to_string(), | ||||
|         cores: 12, | ||||
|         passmark: 28500, | ||||
|         description: "Intel Core i7-12700K".to_string(), | ||||
|         cpu_brand: "Intel".to_string(), | ||||
|         cpu_version: "12th Gen".to_string(), | ||||
|     }; | ||||
|      | ||||
|     let cpu2 = CPUDevice { | ||||
|         id: "cpu_amd_ryzen_9_5900x".to_string(), | ||||
|         cores: 12, | ||||
|         passmark: 32000, | ||||
|         description: "AMD Ryzen 9 5900X".to_string(), | ||||
|         cpu_brand: "AMD".to_string(), | ||||
|         cpu_version: "Zen 3".to_string(), | ||||
|     }; | ||||
|  | ||||
|     // GPU devices | ||||
|     let gpu1 = GPUDevice { | ||||
|         id: "gpu_rtx_3080".to_string(), | ||||
|         cores: 8704, | ||||
|         memory_gb: 10.0, | ||||
|         description: "NVIDIA GeForce RTX 3080".to_string(), | ||||
|         gpu_brand: "NVIDIA".to_string(), | ||||
|         gpu_version: "RTX 30 Series".to_string(), | ||||
|     }; | ||||
|      | ||||
|     let gpu2 = GPUDevice { | ||||
|         id: "gpu_rtx_4090".to_string(), | ||||
|         cores: 16384, | ||||
|         memory_gb: 24.0, | ||||
|         description: "NVIDIA GeForce RTX 4090".to_string(), | ||||
|         gpu_brand: "NVIDIA".to_string(), | ||||
|         gpu_version: "RTX 40 Series".to_string(), | ||||
|     }; | ||||
|  | ||||
|     // Memory devices | ||||
|     let memory1 = MemoryDevice { | ||||
|         id: "mem_ddr4_32gb".to_string(), | ||||
|         size_gb: 32.0, | ||||
|         description: "DDR4-3200 32GB Kit".to_string(), | ||||
|     }; | ||||
|      | ||||
|     let memory2 = MemoryDevice { | ||||
|         id: "mem_ddr5_64gb".to_string(), | ||||
|         size_gb: 64.0, | ||||
|         description: "DDR5-5600 64GB Kit".to_string(), | ||||
|     }; | ||||
|  | ||||
|     // Storage devices | ||||
|     let storage1 = StorageDevice { | ||||
|         id: "ssd_nvme_1tb".to_string(), | ||||
|         size_gb: 1000.0, | ||||
|         description: "NVMe SSD 1TB".to_string(), | ||||
|     }; | ||||
|      | ||||
|     let storage2 = StorageDevice { | ||||
|         id: "hdd_sata_4tb".to_string(), | ||||
|         size_gb: 4000.0, | ||||
|         description: "SATA HDD 4TB".to_string(), | ||||
|     }; | ||||
|  | ||||
|     // Network devices | ||||
|     let network1 = NetworkDevice { | ||||
|         id: "eth_1gbit".to_string(), | ||||
|         speed_mbps: 1000, | ||||
|         description: "Gigabit Ethernet".to_string(), | ||||
|     }; | ||||
|      | ||||
|     let network2 = NetworkDevice { | ||||
|         id: "eth_10gbit".to_string(), | ||||
|         speed_mbps: 10000, | ||||
|         description: "10 Gigabit Ethernet".to_string(), | ||||
|     }; | ||||
|  | ||||
|     // Create device info configurations | ||||
|     let devices1 = DeviceInfo { | ||||
|         vendor: "Dell".to_string(), | ||||
|         cpu: vec![cpu1.clone()], | ||||
|         gpu: vec![gpu1.clone()], | ||||
|         memory: vec![memory1.clone()], | ||||
|         storage: vec![storage1.clone(), storage2.clone()], | ||||
|         network: vec![network1.clone()], | ||||
|     }; | ||||
|      | ||||
|     let devices2 = DeviceInfo { | ||||
|         vendor: "HP".to_string(), | ||||
|         cpu: vec![cpu2.clone()], | ||||
|         gpu: vec![gpu2.clone()], | ||||
|         memory: vec![memory2.clone()], | ||||
|         storage: vec![storage1.clone()], | ||||
|         network: vec![network2.clone()], | ||||
|     }; | ||||
|  | ||||
|     // Create node capacities | ||||
|     let capacity1 = NodeCapacity { | ||||
|         storage_gb: 5000.0, | ||||
|         mem_gb: 32.0, | ||||
|         mem_gb_gpu: 10.0, | ||||
|         passmark: 28500, | ||||
|         vcores: 24, | ||||
|     }; | ||||
|      | ||||
|     let capacity2 = NodeCapacity { | ||||
|         storage_gb: 1000.0, | ||||
|         mem_gb: 64.0, | ||||
|         mem_gb_gpu: 24.0, | ||||
|         passmark: 32000, | ||||
|         vcores: 24, | ||||
|     }; | ||||
|  | ||||
|     // Create compute slices | ||||
|     let compute_slice1 = ComputeSlice::new() | ||||
|         .id(1) | ||||
|         .mem_gb(4.0) | ||||
|         .storage_gb(100.0) | ||||
|         .passmark(3000) | ||||
|         .vcores(2) | ||||
|         .cpu_oversubscription(150) | ||||
|         .storage_oversubscription(120) | ||||
|         .gpus(0); | ||||
|      | ||||
|     let compute_slice2 = ComputeSlice::new() | ||||
|         .id(2) | ||||
|         .mem_gb(8.0) | ||||
|         .storage_gb(200.0) | ||||
|         .passmark(6000) | ||||
|         .vcores(4) | ||||
|         .cpu_oversubscription(130) | ||||
|         .storage_oversubscription(110) | ||||
|         .gpus(1); | ||||
|      | ||||
|     let compute_slice3 = ComputeSlice::new() | ||||
|         .id(1) | ||||
|         .mem_gb(16.0) | ||||
|         .storage_gb(400.0) | ||||
|         .passmark(12000) | ||||
|         .vcores(8) | ||||
|         .cpu_oversubscription(110) | ||||
|         .storage_oversubscription(100) | ||||
|         .gpus(1); | ||||
|  | ||||
|     // Create storage slices | ||||
|     let storage_slice1 = StorageSlice::new().id(1); | ||||
|     let storage_slice2 = StorageSlice::new().id(2); | ||||
|     let storage_slice3 = StorageSlice::new().id(3); | ||||
|  | ||||
|     // Create nodes with different configurations | ||||
|  | ||||
|     // Node 1 - Web hosting node | ||||
|     let node1 = Node::new() | ||||
|         .nodegroupid(1001) | ||||
|         .uptime(98) | ||||
|         .add_compute_slice(compute_slice1.clone()) | ||||
|         .add_compute_slice(compute_slice2.clone()) | ||||
|         .add_storage_slice(storage_slice1.clone()) | ||||
|         .add_storage_slice(storage_slice2.clone()) | ||||
|         .devices(devices1.clone()) | ||||
|         .country("US".to_string()) | ||||
|         .capacity(capacity1.clone()) | ||||
|         .birthtime(1640995200) // 2022-01-01 | ||||
|         .pubkey("node1_pubkey_abc123xyz789".to_string()) | ||||
|         .signature_node("node1_signature_def456".to_string()) | ||||
|         .signature_farmer("farmer1_signature_ghi789".to_string()); | ||||
|  | ||||
|     // Node 2 - High-performance computing node | ||||
|     let node2 = Node::new() | ||||
|         .nodegroupid(1002) | ||||
|         .uptime(99) | ||||
|         .add_compute_slice(compute_slice3.clone()) | ||||
|         .add_storage_slice(storage_slice3.clone()) | ||||
|         .devices(devices2.clone()) | ||||
|         .country("DE".to_string()) | ||||
|         .capacity(capacity2.clone()) | ||||
|         .birthtime(1672531200) // 2023-01-01 | ||||
|         .pubkey("node2_pubkey_jkl012mno345".to_string()) | ||||
|         .signature_node("node2_signature_pqr678".to_string()) | ||||
|         .signature_farmer("farmer2_signature_stu901".to_string()); | ||||
|  | ||||
|     // Node 3 - Storage-focused node | ||||
|     let node3 = Node::new() | ||||
|         .nodegroupid(1001) | ||||
|         .uptime(95) | ||||
|         .add_storage_slice(storage_slice1.clone()) | ||||
|         .add_storage_slice(storage_slice2.clone()) | ||||
|         .add_storage_slice(storage_slice3.clone()) | ||||
|         .devices(devices1.clone()) | ||||
|         .country("NL".to_string()) | ||||
|         .capacity(capacity1.clone()) | ||||
|         .birthtime(1704067200) // 2024-01-01 | ||||
|         .pubkey("node3_pubkey_vwx234yzab567".to_string()) | ||||
|         .signature_node("node3_signature_cde890".to_string()) | ||||
|         .signature_farmer("farmer1_signature_fgh123".to_string()); | ||||
|  | ||||
|     // Save all nodes to database and get their assigned IDs and updated models | ||||
|     let (node1_id, db_node1) = db | ||||
|         .collection() | ||||
|         .expect("can open node collection") | ||||
|         .set(&node1) | ||||
|         .expect("can set node"); | ||||
|     let (node2_id, db_node2) = db | ||||
|         .collection() | ||||
|         .expect("can open node collection") | ||||
|         .set(&node2) | ||||
|         .expect("can set node"); | ||||
|     let (node3_id, db_node3) = db | ||||
|         .collection() | ||||
|         .expect("can open node collection") | ||||
|         .set(&node3) | ||||
|         .expect("can set node"); | ||||
|  | ||||
|     println!("Node 1 assigned ID: {}", node1_id); | ||||
|     println!("Node 2 assigned ID: {}", node2_id); | ||||
|     println!("Node 3 assigned ID: {}", node3_id); | ||||
|  | ||||
|     // Print all nodes retrieved from database | ||||
|     println!("\n--- Nodes Retrieved from Database ---"); | ||||
|     println!("\n1. Web hosting node:"); | ||||
|     print_node_details(&db_node1); | ||||
|  | ||||
|     println!("\n2. High-performance computing node:"); | ||||
|     print_node_details(&db_node2); | ||||
|  | ||||
|     println!("\n3. Storage-focused node:"); | ||||
|     print_node_details(&db_node3); | ||||
|  | ||||
|     // Demonstrate different ways to retrieve nodes from the database | ||||
|  | ||||
|     // 1. Retrieve by nodegroup ID index | ||||
|     println!("\n--- Retrieving Nodes by Different Methods ---"); | ||||
|     println!("\n1. By NodeGroup ID Index (NodeGroup 1001):"); | ||||
|     let nodegroup_nodes = db | ||||
|         .collection::<Node>() | ||||
|         .expect("can open node collection") | ||||
|         .get::<nodegroupid, _>(&1001i32) | ||||
|         .expect("can load nodes by nodegroup"); | ||||
|  | ||||
|     assert_eq!(nodegroup_nodes.len(), 2); | ||||
|     for (i, node) in nodegroup_nodes.iter().enumerate() { | ||||
|         println!("  Node {}: ID {}, Country: {}, Uptime: {}%",  | ||||
|             i + 1, node.get_id(), node.country, node.uptime); | ||||
|     } | ||||
|  | ||||
|     // 2. Retrieve by country index | ||||
|     println!("\n2. By Country Index (Germany - DE):"); | ||||
|     let country_nodes = db | ||||
|         .collection::<Node>() | ||||
|         .expect("can open node collection") | ||||
|         .get::<country, _>("DE") | ||||
|         .expect("can load nodes by country"); | ||||
|  | ||||
|     assert_eq!(country_nodes.len(), 1); | ||||
|     print_node_details(&country_nodes[0]); | ||||
|  | ||||
|     // 3. Update node uptime | ||||
|     println!("\n3. Updating Node Uptime:"); | ||||
|     let mut updated_node = db_node1.clone(); | ||||
|     updated_node.uptime = 99; | ||||
|  | ||||
|     let (_, uptime_updated_node) = db | ||||
|         .collection::<Node>() | ||||
|         .expect("can open node collection") | ||||
|         .set(&updated_node) | ||||
|         .expect("can update node"); | ||||
|  | ||||
|     println!("Updated node uptime to 99%:"); | ||||
|     println!("  Node ID: {}, New Uptime: {}%", uptime_updated_node.get_id(), uptime_updated_node.uptime); | ||||
|  | ||||
|     // Show all nodes and calculate analytics | ||||
|     let all_nodes = db | ||||
|         .collection::<Node>() | ||||
|         .expect("can open node collection") | ||||
|         .get_all() | ||||
|         .expect("can load all nodes"); | ||||
|  | ||||
|     println!("\n--- Node Analytics ---"); | ||||
|     println!("Total Nodes: {}", all_nodes.len()); | ||||
|  | ||||
|     // Calculate total capacity | ||||
|     let total_storage_gb: f64 = all_nodes.iter().map(|n| n.capacity.storage_gb).sum(); | ||||
|     let total_memory_gb: f64 = all_nodes.iter().map(|n| n.capacity.mem_gb).sum(); | ||||
|     let total_gpu_memory_gb: f64 = all_nodes.iter().map(|n| n.capacity.mem_gb_gpu).sum(); | ||||
|     let total_vcores: i32 = all_nodes.iter().map(|n| n.capacity.vcores).sum(); | ||||
|     let avg_uptime: f64 = all_nodes.iter().map(|n| n.uptime as f64).sum::<f64>() / all_nodes.len() as f64; | ||||
|  | ||||
|     println!("Total Capacity:"); | ||||
|     println!("  Storage: {:.1} GB", total_storage_gb); | ||||
|     println!("  Memory: {:.1} GB", total_memory_gb); | ||||
|     println!("  GPU Memory: {:.1} GB", total_gpu_memory_gb); | ||||
|     println!("  vCores: {}", total_vcores); | ||||
|     println!("  Average Uptime: {:.1}%", avg_uptime); | ||||
|  | ||||
|     // Count nodes by country | ||||
|     let mut country_counts = std::collections::HashMap::new(); | ||||
|     for node in &all_nodes { | ||||
|         *country_counts.entry(&node.country).or_insert(0) += 1; | ||||
|     } | ||||
|  | ||||
|     println!("\nNodes by Country:"); | ||||
|     for (country, count) in country_counts { | ||||
|         println!("  {}: {}", country, count); | ||||
|     } | ||||
|  | ||||
|     // Count total slices | ||||
|     let total_compute_slices: usize = all_nodes.iter().map(|n| n.computeslices.len()).sum(); | ||||
|     let total_storage_slices: usize = all_nodes.iter().map(|n| n.storageslices.len()).sum(); | ||||
|  | ||||
|     println!("\nTotal Slices:"); | ||||
|     println!("  Compute Slices: {}", total_compute_slices); | ||||
|     println!("  Storage Slices: {}", total_storage_slices); | ||||
|  | ||||
|     // Vendor distribution | ||||
|     let mut vendor_counts = std::collections::HashMap::new(); | ||||
|     for node in &all_nodes { | ||||
|         *vendor_counts.entry(&node.devices.vendor).or_insert(0) += 1; | ||||
|     } | ||||
|  | ||||
|     println!("\nNodes by Vendor:"); | ||||
|     for (vendor, count) in vendor_counts { | ||||
|         println!("  {}: {}", vendor, count); | ||||
|     } | ||||
|  | ||||
|     println!("\n--- Model Information ---"); | ||||
|     println!("Node DB Prefix: {}", Node::db_prefix()); | ||||
| } | ||||
							
								
								
									
										284
									
								
								heromodels/examples/grid4_nodegroup_example.rs
									
									
									
									
									
										Normal file
									
								
							
							
						
						
									
										284
									
								
								heromodels/examples/grid4_nodegroup_example.rs
									
									
									
									
									
										Normal file
									
								
							| @@ -0,0 +1,284 @@ | ||||
| use heromodels::db::{Collection, Db}; | ||||
| use heromodels::models::grid4::{NodeGroup, PricingPolicy, SLAPolicy}; | ||||
| use heromodels_core::Model; | ||||
|  | ||||
| // Helper function to print nodegroup details | ||||
| fn print_nodegroup_details(nodegroup: &NodeGroup) { | ||||
|     println!("\n--- NodeGroup Details ---"); | ||||
|     println!("ID: {}", nodegroup.get_id()); | ||||
|     println!("Farmer ID: {}", nodegroup.farmerid); | ||||
|     println!("Description: {}", nodegroup.description); | ||||
|     println!("Secret: {}", nodegroup.secret); | ||||
|     println!("Compute Slice Pricing (CC): {:.4}", nodegroup.compute_slice_normalized_pricing_cc); | ||||
|     println!("Storage Slice Pricing (CC): {:.4}", nodegroup.storage_slice_normalized_pricing_cc); | ||||
|     println!("Signature Farmer: {}", nodegroup.signature_farmer); | ||||
|     println!("Created At: {}", nodegroup.base_data.created_at); | ||||
|     println!("Modified At: {}", nodegroup.base_data.modified_at); | ||||
|      | ||||
|     // Print SLA Policy details | ||||
|     println!("  SLA Policy:"); | ||||
|     println!("    Uptime: {}%", nodegroup.slapolicy.sla_uptime); | ||||
|     println!("    Bandwidth: {} Mbit/s", nodegroup.slapolicy.sla_bandwidth_mbit); | ||||
|     println!("    Penalty: {}%", nodegroup.slapolicy.sla_penalty); | ||||
|      | ||||
|     // Print Pricing Policy details | ||||
|     println!("  Pricing Policy:"); | ||||
|     println!("    Marketplace Year Discounts: {:?}%", nodegroup.pricingpolicy.marketplace_year_discounts); | ||||
| } | ||||
|  | ||||
| fn main() { | ||||
|     // Create a new DB instance in /tmp/grid4_nodegroups_db, and reset before every run | ||||
|     let db = heromodels::db::hero::OurDB::new("/tmp/grid4_nodegroups_db", true).expect("Can create DB"); | ||||
|  | ||||
|     println!("Grid4 NodeGroup Models - Basic Usage Example"); | ||||
|     println!("==========================================="); | ||||
|  | ||||
|     // Create SLA policies | ||||
|     let sla_policy_premium = SLAPolicy { | ||||
|         sla_uptime: 99, | ||||
|         sla_bandwidth_mbit: 1000, | ||||
|         sla_penalty: 200, | ||||
|     }; | ||||
|      | ||||
|     let sla_policy_standard = SLAPolicy { | ||||
|         sla_uptime: 95, | ||||
|         sla_bandwidth_mbit: 100, | ||||
|         sla_penalty: 100, | ||||
|     }; | ||||
|      | ||||
|     let sla_policy_basic = SLAPolicy { | ||||
|         sla_uptime: 90, | ||||
|         sla_bandwidth_mbit: 50, | ||||
|         sla_penalty: 50, | ||||
|     }; | ||||
|  | ||||
|     // Create pricing policies | ||||
|     let pricing_policy_aggressive = PricingPolicy { | ||||
|         marketplace_year_discounts: vec![40, 50, 60], | ||||
|     }; | ||||
|      | ||||
|     let pricing_policy_standard = PricingPolicy { | ||||
|         marketplace_year_discounts: vec![30, 40, 50], | ||||
|     }; | ||||
|      | ||||
|     let pricing_policy_conservative = PricingPolicy { | ||||
|         marketplace_year_discounts: vec![20, 30, 40], | ||||
|     }; | ||||
|  | ||||
|     // Create nodegroups with different configurations | ||||
|  | ||||
|     // NodeGroup 1 - Premium hosting provider | ||||
|     let nodegroup1 = NodeGroup::new() | ||||
|         .farmerid(501) | ||||
|         .secret("encrypted_boot_secret_premium_abc123".to_string()) | ||||
|         .description("Premium hosting with 99% uptime SLA and high-speed connectivity".to_string()) | ||||
|         .slapolicy(sla_policy_premium.clone()) | ||||
|         .pricingpolicy(pricing_policy_aggressive.clone()) | ||||
|         .compute_slice_normalized_pricing_cc(0.0450) | ||||
|         .storage_slice_normalized_pricing_cc(0.0180) | ||||
|         .signature_farmer("farmer_501_premium_signature_xyz789".to_string()); | ||||
|  | ||||
|     // NodeGroup 2 - Standard business provider | ||||
|     let nodegroup2 = NodeGroup::new() | ||||
|         .farmerid(502) | ||||
|         .secret("encrypted_boot_secret_standard_def456".to_string()) | ||||
|         .description("Standard business hosting with reliable performance".to_string()) | ||||
|         .slapolicy(sla_policy_standard.clone()) | ||||
|         .pricingpolicy(pricing_policy_standard.clone()) | ||||
|         .compute_slice_normalized_pricing_cc(0.0350) | ||||
|         .storage_slice_normalized_pricing_cc(0.0150) | ||||
|         .signature_farmer("farmer_502_standard_signature_uvw012".to_string()); | ||||
|  | ||||
|     // NodeGroup 3 - Budget-friendly provider | ||||
|     let nodegroup3 = NodeGroup::new() | ||||
|         .farmerid(503) | ||||
|         .secret("encrypted_boot_secret_budget_ghi789".to_string()) | ||||
|         .description("Cost-effective hosting for development and testing".to_string()) | ||||
|         .slapolicy(sla_policy_basic.clone()) | ||||
|         .pricingpolicy(pricing_policy_conservative.clone()) | ||||
|         .compute_slice_normalized_pricing_cc(0.0250) | ||||
|         .storage_slice_normalized_pricing_cc(0.0120) | ||||
|         .signature_farmer("farmer_503_budget_signature_rst345".to_string()); | ||||
|  | ||||
|     // NodeGroup 4 - Enterprise provider | ||||
|     let nodegroup4 = NodeGroup::new() | ||||
|         .farmerid(504) | ||||
|         .secret("encrypted_boot_secret_enterprise_jkl012".to_string()) | ||||
|         .description("Enterprise-grade infrastructure with maximum reliability".to_string()) | ||||
|         .slapolicy(sla_policy_premium.clone()) | ||||
|         .pricingpolicy(pricing_policy_standard.clone()) | ||||
|         .compute_slice_normalized_pricing_cc(0.0500) | ||||
|         .storage_slice_normalized_pricing_cc(0.0200) | ||||
|         .signature_farmer("farmer_504_enterprise_signature_mno678".to_string()); | ||||
|  | ||||
|     // Save all nodegroups to database and get their assigned IDs and updated models | ||||
|     let (nodegroup1_id, db_nodegroup1) = db | ||||
|         .collection() | ||||
|         .expect("can open nodegroup collection") | ||||
|         .set(&nodegroup1) | ||||
|         .expect("can set nodegroup"); | ||||
|     let (nodegroup2_id, db_nodegroup2) = db | ||||
|         .collection() | ||||
|         .expect("can open nodegroup collection") | ||||
|         .set(&nodegroup2) | ||||
|         .expect("can set nodegroup"); | ||||
|     let (nodegroup3_id, db_nodegroup3) = db | ||||
|         .collection() | ||||
|         .expect("can open nodegroup collection") | ||||
|         .set(&nodegroup3) | ||||
|         .expect("can set nodegroup"); | ||||
|     let (nodegroup4_id, db_nodegroup4) = db | ||||
|         .collection() | ||||
|         .expect("can open nodegroup collection") | ||||
|         .set(&nodegroup4) | ||||
|         .expect("can set nodegroup"); | ||||
|  | ||||
|     println!("NodeGroup 1 assigned ID: {}", nodegroup1_id); | ||||
|     println!("NodeGroup 2 assigned ID: {}", nodegroup2_id); | ||||
|     println!("NodeGroup 3 assigned ID: {}", nodegroup3_id); | ||||
|     println!("NodeGroup 4 assigned ID: {}", nodegroup4_id); | ||||
|  | ||||
|     // Print all nodegroups retrieved from database | ||||
|     println!("\n--- NodeGroups Retrieved from Database ---"); | ||||
|     println!("\n1. Premium hosting provider:"); | ||||
|     print_nodegroup_details(&db_nodegroup1); | ||||
|  | ||||
|     println!("\n2. Standard business provider:"); | ||||
|     print_nodegroup_details(&db_nodegroup2); | ||||
|  | ||||
|     println!("\n3. Budget-friendly provider:"); | ||||
|     print_nodegroup_details(&db_nodegroup3); | ||||
|  | ||||
|     println!("\n4. Enterprise provider:"); | ||||
|     print_nodegroup_details(&db_nodegroup4); | ||||
|  | ||||
|     // Demonstrate different ways to retrieve nodegroups from the database | ||||
|  | ||||
|     // 1. Retrieve by farmer ID index | ||||
|     println!("\n--- Retrieving NodeGroups by Different Methods ---"); | ||||
|     println!("\n1. By Farmer ID Index (Farmer 502):"); | ||||
|     let farmer_nodegroups = db | ||||
|         .collection::<NodeGroup>() | ||||
|         .expect("can open nodegroup collection") | ||||
|         .get_by_index("farmerid", &502u32) | ||||
|         .expect("can load nodegroups by farmer"); | ||||
|  | ||||
|     assert_eq!(farmer_nodegroups.len(), 1); | ||||
|     print_nodegroup_details(&farmer_nodegroups[0]); | ||||
|  | ||||
|     // 2. Update nodegroup pricing | ||||
|     println!("\n2. Updating NodeGroup Pricing:"); | ||||
|     let mut updated_nodegroup = db_nodegroup3.clone(); | ||||
|     updated_nodegroup.compute_slice_normalized_pricing_cc = 0.0280; | ||||
|     updated_nodegroup.storage_slice_normalized_pricing_cc = 0.0130; | ||||
|  | ||||
|     let (_, price_updated_nodegroup) = db | ||||
|         .collection::<NodeGroup>() | ||||
|         .expect("can open nodegroup collection") | ||||
|         .set(&updated_nodegroup) | ||||
|         .expect("can update nodegroup"); | ||||
|  | ||||
|     println!("Updated pricing for budget provider:"); | ||||
|     println!("  Compute: {:.4} CC", price_updated_nodegroup.compute_slice_normalized_pricing_cc); | ||||
|     println!("  Storage: {:.4} CC", price_updated_nodegroup.storage_slice_normalized_pricing_cc); | ||||
|  | ||||
|     // 3. Update SLA policy | ||||
|     println!("\n3. Updating SLA Policy:"); | ||||
|     let mut sla_updated_nodegroup = db_nodegroup2.clone(); | ||||
|     sla_updated_nodegroup.slapolicy.sla_uptime = 98; | ||||
|     sla_updated_nodegroup.slapolicy.sla_bandwidth_mbit = 500; | ||||
|  | ||||
|     let (_, sla_updated_nodegroup) = db | ||||
|         .collection::<NodeGroup>() | ||||
|         .expect("can open nodegroup collection") | ||||
|         .set(&sla_updated_nodegroup) | ||||
|         .expect("can update nodegroup"); | ||||
|  | ||||
|     println!("Updated SLA policy for standard provider:"); | ||||
|     println!("  Uptime: {}%", sla_updated_nodegroup.slapolicy.sla_uptime); | ||||
|     println!("  Bandwidth: {} Mbit/s", sla_updated_nodegroup.slapolicy.sla_bandwidth_mbit); | ||||
|  | ||||
|     // Show all nodegroups and calculate analytics | ||||
|     let all_nodegroups = db | ||||
|         .collection::<NodeGroup>() | ||||
|         .expect("can open nodegroup collection") | ||||
|         .get_all() | ||||
|         .expect("can load all nodegroups"); | ||||
|  | ||||
|     println!("\n--- NodeGroup Analytics ---"); | ||||
|     println!("Total NodeGroups: {}", all_nodegroups.len()); | ||||
|  | ||||
|     // Calculate pricing statistics | ||||
|     let avg_compute_price: f64 = all_nodegroups.iter() | ||||
|         .map(|ng| ng.compute_slice_normalized_pricing_cc) | ||||
|         .sum::<f64>() / all_nodegroups.len() as f64; | ||||
|     let avg_storage_price: f64 = all_nodegroups.iter() | ||||
|         .map(|ng| ng.storage_slice_normalized_pricing_cc) | ||||
|         .sum::<f64>() / all_nodegroups.len() as f64; | ||||
|  | ||||
|     let min_compute_price = all_nodegroups.iter() | ||||
|         .map(|ng| ng.compute_slice_normalized_pricing_cc) | ||||
|         .fold(f64::INFINITY, f64::min); | ||||
|     let max_compute_price = all_nodegroups.iter() | ||||
|         .map(|ng| ng.compute_slice_normalized_pricing_cc) | ||||
|         .fold(f64::NEG_INFINITY, f64::max); | ||||
|  | ||||
|     println!("Pricing Statistics:"); | ||||
|     println!("  Average Compute Price: {:.4} CC", avg_compute_price); | ||||
|     println!("  Average Storage Price: {:.4} CC", avg_storage_price); | ||||
|     println!("  Compute Price Range: {:.4} - {:.4} CC", min_compute_price, max_compute_price); | ||||
|  | ||||
|     // Calculate SLA statistics | ||||
|     let avg_uptime: f64 = all_nodegroups.iter() | ||||
|         .map(|ng| ng.slapolicy.sla_uptime as f64) | ||||
|         .sum::<f64>() / all_nodegroups.len() as f64; | ||||
|     let avg_bandwidth: f64 = all_nodegroups.iter() | ||||
|         .map(|ng| ng.slapolicy.sla_bandwidth_mbit as f64) | ||||
|         .sum::<f64>() / all_nodegroups.len() as f64; | ||||
|     let avg_penalty: f64 = all_nodegroups.iter() | ||||
|         .map(|ng| ng.slapolicy.sla_penalty as f64) | ||||
|         .sum::<f64>() / all_nodegroups.len() as f64; | ||||
|  | ||||
|     println!("\nSLA Statistics:"); | ||||
|     println!("  Average Uptime Guarantee: {:.1}%", avg_uptime); | ||||
|     println!("  Average Bandwidth Guarantee: {:.0} Mbit/s", avg_bandwidth); | ||||
|     println!("  Average Penalty Rate: {:.0}%", avg_penalty); | ||||
|  | ||||
|     // Count farmers | ||||
|     let unique_farmers: std::collections::HashSet<_> = all_nodegroups.iter() | ||||
|         .map(|ng| ng.farmerid) | ||||
|         .collect(); | ||||
|  | ||||
|     println!("\nFarmer Statistics:"); | ||||
|     println!("  Unique Farmers: {}", unique_farmers.len()); | ||||
|     println!("  NodeGroups per Farmer: {:.1}", all_nodegroups.len() as f64 / unique_farmers.len() as f64); | ||||
|  | ||||
|     // Analyze discount policies | ||||
|     let total_discount_tiers: usize = all_nodegroups.iter() | ||||
|         .map(|ng| ng.pricingpolicy.marketplace_year_discounts.len()) | ||||
|         .sum(); | ||||
|     let avg_discount_tiers: f64 = total_discount_tiers as f64 / all_nodegroups.len() as f64; | ||||
|  | ||||
|     println!("\nDiscount Policy Statistics:"); | ||||
|     println!("  Average Discount Tiers: {:.1}", avg_discount_tiers); | ||||
|  | ||||
|     // Find best value providers (high SLA, low price) | ||||
|     println!("\n--- Provider Rankings ---"); | ||||
|     let mut providers_with_scores: Vec<_> = all_nodegroups.iter() | ||||
|         .map(|ng| { | ||||
|             let value_score = (ng.slapolicy.sla_uptime as f64) / ng.compute_slice_normalized_pricing_cc; | ||||
|             (ng, value_score) | ||||
|         }) | ||||
|         .collect(); | ||||
|      | ||||
|     providers_with_scores.sort_by(|a, b| b.1.partial_cmp(&a.1).unwrap()); | ||||
|  | ||||
|     println!("Best Value Providers (Uptime/Price ratio):"); | ||||
|     for (i, (ng, score)) in providers_with_scores.iter().enumerate() { | ||||
|         println!("  {}. Farmer {}: {:.0} ({}% uptime, {:.4} CC)",  | ||||
|             i + 1, ng.farmerid, score, ng.slapolicy.sla_uptime, ng.compute_slice_normalized_pricing_cc); | ||||
|     } | ||||
|  | ||||
|     println!("\n--- Model Information ---"); | ||||
|     println!("NodeGroup DB Prefix: {}", NodeGroup::db_prefix()); | ||||
| } | ||||
							
								
								
									
										311
									
								
								heromodels/examples/grid4_reputation_example.rs
									
									
									
									
									
										Normal file
									
								
							
							
						
						
									
										311
									
								
								heromodels/examples/grid4_reputation_example.rs
									
									
									
									
									
										Normal file
									
								
							| @@ -0,0 +1,311 @@ | ||||
| use heromodels::db::{Collection, Db}; | ||||
| use heromodels::models::grid4::{NodeGroupReputation, NodeReputation}; | ||||
| use heromodels_core::Model; | ||||
|  | ||||
| // Helper function to print nodegroup reputation details | ||||
| fn print_nodegroup_reputation_details(reputation: &NodeGroupReputation) { | ||||
|     println!("\n--- NodeGroup Reputation Details ---"); | ||||
|     println!("ID: {}", reputation.get_id()); | ||||
|     println!("NodeGroup ID: {}", reputation.nodegroup_id); | ||||
|     println!("Reputation Score: {}/100", reputation.reputation); | ||||
|     println!("Uptime: {}%", reputation.uptime); | ||||
|     println!("Node Count: {}", reputation.nodes.len()); | ||||
|     println!("Created At: {}", reputation.base_data.created_at); | ||||
|     println!("Modified At: {}", reputation.base_data.modified_at); | ||||
|      | ||||
|     // Print individual node reputations | ||||
|     if !reputation.nodes.is_empty() { | ||||
|         println!("  Individual Node Reputations:"); | ||||
|         for (i, node_rep) in reputation.nodes.iter().enumerate() { | ||||
|             println!("    {}. Node {}: Reputation {}/100, Uptime {}%",  | ||||
|                 i + 1, node_rep.node_id, node_rep.reputation, node_rep.uptime); | ||||
|         } | ||||
|          | ||||
|         // Calculate average node reputation and uptime | ||||
|         let avg_node_reputation: f64 = reputation.nodes.iter() | ||||
|             .map(|n| n.reputation as f64) | ||||
|             .sum::<f64>() / reputation.nodes.len() as f64; | ||||
|         let avg_node_uptime: f64 = reputation.nodes.iter() | ||||
|             .map(|n| n.uptime as f64) | ||||
|             .sum::<f64>() / reputation.nodes.len() as f64; | ||||
|          | ||||
|         println!("    Average Node Reputation: {:.1}/100", avg_node_reputation); | ||||
|         println!("    Average Node Uptime: {:.1}%", avg_node_uptime); | ||||
|     } | ||||
| } | ||||
|  | ||||
| fn main() { | ||||
|     // Create a new DB instance in /tmp/grid4_reputation_db, and reset before every run | ||||
|     let db = heromodels::db::hero::OurDB::new("/tmp/grid4_reputation_db", true).expect("Can create DB"); | ||||
|  | ||||
|     println!("Grid4 Reputation Models - Basic Usage Example"); | ||||
|     println!("============================================"); | ||||
|  | ||||
|     // Create individual node reputations | ||||
|      | ||||
|     // High-performing nodes | ||||
|     let node_rep1 = NodeReputation::new() | ||||
|         .node_id(1001) | ||||
|         .reputation(85) | ||||
|         .uptime(99); | ||||
|      | ||||
|     let node_rep2 = NodeReputation::new() | ||||
|         .node_id(1002) | ||||
|         .reputation(92) | ||||
|         .uptime(98); | ||||
|      | ||||
|     let node_rep3 = NodeReputation::new() | ||||
|         .node_id(1003) | ||||
|         .reputation(78) | ||||
|         .uptime(97); | ||||
|  | ||||
|     // Medium-performing nodes | ||||
|     let node_rep4 = NodeReputation::new() | ||||
|         .node_id(2001) | ||||
|         .reputation(65) | ||||
|         .uptime(94); | ||||
|      | ||||
|     let node_rep5 = NodeReputation::new() | ||||
|         .node_id(2002) | ||||
|         .reputation(72) | ||||
|         .uptime(96); | ||||
|  | ||||
|     // Lower-performing nodes | ||||
|     let node_rep6 = NodeReputation::new() | ||||
|         .node_id(3001) | ||||
|         .reputation(45) | ||||
|         .uptime(88); | ||||
|      | ||||
|     let node_rep7 = NodeReputation::new() | ||||
|         .node_id(3002) | ||||
|         .reputation(38) | ||||
|         .uptime(85); | ||||
|  | ||||
|     // New nodes with default reputation | ||||
|     let node_rep8 = NodeReputation::new() | ||||
|         .node_id(4001) | ||||
|         .reputation(50) // default | ||||
|         .uptime(0);     // just started | ||||
|      | ||||
|     let node_rep9 = NodeReputation::new() | ||||
|         .node_id(4002) | ||||
|         .reputation(50) // default | ||||
|         .uptime(0);     // just started | ||||
|  | ||||
|     // Create nodegroup reputations with different performance profiles | ||||
|  | ||||
|     // NodeGroup 1 - High-performance provider | ||||
|     let nodegroup_rep1 = NodeGroupReputation::new() | ||||
|         .nodegroup_id(1001) | ||||
|         .reputation(85) // high reputation earned over time | ||||
|         .uptime(98)     // excellent uptime | ||||
|         .add_node_reputation(node_rep1.clone()) | ||||
|         .add_node_reputation(node_rep2.clone()) | ||||
|         .add_node_reputation(node_rep3.clone()); | ||||
|  | ||||
|     // NodeGroup 2 - Medium-performance provider | ||||
|     let nodegroup_rep2 = NodeGroupReputation::new() | ||||
|         .nodegroup_id(1002) | ||||
|         .reputation(68) // decent reputation | ||||
|         .uptime(95)     // good uptime | ||||
|         .add_node_reputation(node_rep4.clone()) | ||||
|         .add_node_reputation(node_rep5.clone()); | ||||
|  | ||||
|     // NodeGroup 3 - Struggling provider | ||||
|     let nodegroup_rep3 = NodeGroupReputation::new() | ||||
|         .nodegroup_id(1003) | ||||
|         .reputation(42) // below average reputation | ||||
|         .uptime(87)     // poor uptime | ||||
|         .add_node_reputation(node_rep6.clone()) | ||||
|         .add_node_reputation(node_rep7.clone()); | ||||
|  | ||||
|     // NodeGroup 4 - New provider (default reputation) | ||||
|     let nodegroup_rep4 = NodeGroupReputation::new() | ||||
|         .nodegroup_id(1004) | ||||
|         .reputation(50) // default starting reputation | ||||
|         .uptime(0)      // no history yet | ||||
|         .add_node_reputation(node_rep8.clone()) | ||||
|         .add_node_reputation(node_rep9.clone()); | ||||
|  | ||||
|     // Save all nodegroup reputations to database and get their assigned IDs and updated models | ||||
|     let (rep1_id, db_rep1) = db | ||||
|         .collection() | ||||
|         .expect("can open reputation collection") | ||||
|         .set(&nodegroup_rep1) | ||||
|         .expect("can set reputation"); | ||||
|     let (rep2_id, db_rep2) = db | ||||
|         .collection() | ||||
|         .expect("can open reputation collection") | ||||
|         .set(&nodegroup_rep2) | ||||
|         .expect("can set reputation"); | ||||
|     let (rep3_id, db_rep3) = db | ||||
|         .collection() | ||||
|         .expect("can open reputation collection") | ||||
|         .set(&nodegroup_rep3) | ||||
|         .expect("can set reputation"); | ||||
|     let (rep4_id, db_rep4) = db | ||||
|         .collection() | ||||
|         .expect("can open reputation collection") | ||||
|         .set(&nodegroup_rep4) | ||||
|         .expect("can set reputation"); | ||||
|  | ||||
|     println!("NodeGroup Reputation 1 assigned ID: {}", rep1_id); | ||||
|     println!("NodeGroup Reputation 2 assigned ID: {}", rep2_id); | ||||
|     println!("NodeGroup Reputation 3 assigned ID: {}", rep3_id); | ||||
|     println!("NodeGroup Reputation 4 assigned ID: {}", rep4_id); | ||||
|  | ||||
|     // Print all reputation records retrieved from database | ||||
|     println!("\n--- Reputation Records Retrieved from Database ---"); | ||||
|     println!("\n1. High-performance provider:"); | ||||
|     print_nodegroup_reputation_details(&db_rep1); | ||||
|  | ||||
|     println!("\n2. Medium-performance provider:"); | ||||
|     print_nodegroup_reputation_details(&db_rep2); | ||||
|  | ||||
|     println!("\n3. Struggling provider:"); | ||||
|     print_nodegroup_reputation_details(&db_rep3); | ||||
|  | ||||
|     println!("\n4. New provider:"); | ||||
|     print_nodegroup_reputation_details(&db_rep4); | ||||
|  | ||||
|     // Demonstrate different ways to retrieve reputation records from the database | ||||
|  | ||||
|     // 1. Retrieve by nodegroup ID index | ||||
|     println!("\n--- Retrieving Reputation by Different Methods ---"); | ||||
|     println!("\n1. By NodeGroup ID Index (NodeGroup 1002):"); | ||||
|     let nodegroup_reps = db | ||||
|         .collection::<NodeGroupReputation>() | ||||
|         .expect("can open reputation collection") | ||||
|         .get_by_index("nodegroup_id", &1002u32) | ||||
|         .expect("can load reputation by nodegroup"); | ||||
|  | ||||
|     assert_eq!(nodegroup_reps.len(), 1); | ||||
|     print_nodegroup_reputation_details(&nodegroup_reps[0]); | ||||
|  | ||||
|     // 2. Update reputation scores (simulate performance improvement) | ||||
|     println!("\n2. Updating Reputation Scores (Performance Improvement):"); | ||||
|     let mut improved_rep = db_rep3.clone(); | ||||
|     improved_rep.reputation = 55; // improved from 42 | ||||
|     improved_rep.uptime = 92;     // improved from 87 | ||||
|      | ||||
|     // Also improve individual node reputations | ||||
|     for node_rep in &mut improved_rep.nodes { | ||||
|         node_rep.reputation += 10; // boost each node's reputation | ||||
|         node_rep.uptime += 5;      // improve uptime | ||||
|     } | ||||
|  | ||||
|     let (_, updated_rep) = db | ||||
|         .collection::<NodeGroupReputation>() | ||||
|         .expect("can open reputation collection") | ||||
|         .set(&improved_rep) | ||||
|         .expect("can update reputation"); | ||||
|  | ||||
|     println!("Updated reputation for struggling provider:"); | ||||
|     print_nodegroup_reputation_details(&updated_rep); | ||||
|  | ||||
|     // 3. Add new node to existing nodegroup reputation | ||||
|     println!("\n3. Adding New Node to Existing NodeGroup:"); | ||||
|     let new_node_rep = NodeReputation::new() | ||||
|         .node_id(1004) | ||||
|         .reputation(88) | ||||
|         .uptime(99); | ||||
|  | ||||
|     let mut expanded_rep = db_rep1.clone(); | ||||
|     expanded_rep.add_node_reputation(new_node_rep); | ||||
|     // Recalculate nodegroup reputation based on node average | ||||
|     let total_node_rep: i32 = expanded_rep.nodes.iter().map(|n| n.reputation).sum(); | ||||
|     expanded_rep.reputation = total_node_rep / expanded_rep.nodes.len() as i32; | ||||
|  | ||||
|     let (_, expanded_rep) = db | ||||
|         .collection::<NodeGroupReputation>() | ||||
|         .expect("can open reputation collection") | ||||
|         .set(&expanded_rep) | ||||
|         .expect("can update reputation"); | ||||
|  | ||||
|     println!("Added new high-performing node to top provider:"); | ||||
|     print_nodegroup_reputation_details(&expanded_rep); | ||||
|  | ||||
|     // Show all reputation records and calculate analytics | ||||
|     let all_reps = db | ||||
|         .collection::<NodeGroupReputation>() | ||||
|         .expect("can open reputation collection") | ||||
|         .get_all() | ||||
|         .expect("can load all reputations"); | ||||
|  | ||||
|     println!("\n--- Reputation Analytics ---"); | ||||
|     println!("Total NodeGroup Reputations: {}", all_reps.len()); | ||||
|  | ||||
|     // Calculate overall statistics | ||||
|     let avg_nodegroup_reputation: f64 = all_reps.iter() | ||||
|         .map(|r| r.reputation as f64) | ||||
|         .sum::<f64>() / all_reps.len() as f64; | ||||
|     let avg_nodegroup_uptime: f64 = all_reps.iter() | ||||
|         .filter(|r| r.uptime > 0) // exclude new providers with 0 uptime | ||||
|         .map(|r| r.uptime as f64) | ||||
|         .sum::<f64>() / all_reps.iter().filter(|r| r.uptime > 0).count() as f64; | ||||
|  | ||||
|     println!("Overall Statistics:"); | ||||
|     println!("  Average NodeGroup Reputation: {:.1}/100", avg_nodegroup_reputation); | ||||
|     println!("  Average NodeGroup Uptime: {:.1}%", avg_nodegroup_uptime); | ||||
|  | ||||
|     // Count reputation tiers | ||||
|     let excellent_reps = all_reps.iter().filter(|r| r.reputation >= 80).count(); | ||||
|     let good_reps = all_reps.iter().filter(|r| r.reputation >= 60 && r.reputation < 80).count(); | ||||
|     let average_reps = all_reps.iter().filter(|r| r.reputation >= 40 && r.reputation < 60).count(); | ||||
|     let poor_reps = all_reps.iter().filter(|r| r.reputation < 40).count(); | ||||
|  | ||||
|     println!("\nReputation Distribution:"); | ||||
|     println!("  Excellent (80-100): {}", excellent_reps); | ||||
|     println!("  Good (60-79): {}", good_reps); | ||||
|     println!("  Average (40-59): {}", average_reps); | ||||
|     println!("  Poor (0-39): {}", poor_reps); | ||||
|  | ||||
|     // Calculate total nodes and their statistics | ||||
|     let total_nodes: usize = all_reps.iter().map(|r| r.nodes.len()).sum(); | ||||
|     let all_node_reps: Vec<i32> = all_reps.iter() | ||||
|         .flat_map(|r| &r.nodes) | ||||
|         .map(|n| n.reputation) | ||||
|         .collect(); | ||||
|     let all_node_uptimes: Vec<i32> = all_reps.iter() | ||||
|         .flat_map(|r| &r.nodes) | ||||
|         .filter(|n| n.uptime > 0) | ||||
|         .map(|n| n.uptime) | ||||
|         .collect(); | ||||
|  | ||||
|     let avg_node_reputation: f64 = all_node_reps.iter().sum::<i32>() as f64 / all_node_reps.len() as f64; | ||||
|     let avg_node_uptime: f64 = all_node_uptimes.iter().sum::<i32>() as f64 / all_node_uptimes.len() as f64; | ||||
|  | ||||
|     println!("\nNode-Level Statistics:"); | ||||
|     println!("  Total Nodes: {}", total_nodes); | ||||
|     println!("  Average Node Reputation: {:.1}/100", avg_node_reputation); | ||||
|     println!("  Average Node Uptime: {:.1}%", avg_node_uptime); | ||||
|  | ||||
|     // Find best and worst performing nodegroups | ||||
|     let best_nodegroup = all_reps.iter().max_by_key(|r| r.reputation).unwrap(); | ||||
|     let worst_nodegroup = all_reps.iter().min_by_key(|r| r.reputation).unwrap(); | ||||
|  | ||||
|     println!("\nPerformance Leaders:"); | ||||
|     println!("  Best NodeGroup: {} (Reputation: {}, Uptime: {}%)",  | ||||
|         best_nodegroup.nodegroup_id, best_nodegroup.reputation, best_nodegroup.uptime); | ||||
|     println!("  Worst NodeGroup: {} (Reputation: {}, Uptime: {}%)",  | ||||
|         worst_nodegroup.nodegroup_id, worst_nodegroup.reputation, worst_nodegroup.uptime); | ||||
|  | ||||
|     // Rank nodegroups by reputation | ||||
|     let mut ranked_nodegroups: Vec<_> = all_reps.iter().collect(); | ||||
|     ranked_nodegroups.sort_by(|a, b| b.reputation.cmp(&a.reputation)); | ||||
|  | ||||
|     println!("\nNodeGroup Rankings (by Reputation):"); | ||||
|     for (i, rep) in ranked_nodegroups.iter().enumerate() { | ||||
|         let status = match rep.reputation { | ||||
|             80..=100 => "Excellent", | ||||
|             60..=79 => "Good", | ||||
|             40..=59 => "Average", | ||||
|             _ => "Poor", | ||||
|         }; | ||||
|         println!("  {}. NodeGroup {}: {} ({}/100, {}% uptime)",  | ||||
|             i + 1, rep.nodegroup_id, status, rep.reputation, rep.uptime); | ||||
|     } | ||||
|  | ||||
|     println!("\n--- Model Information ---"); | ||||
|     println!("NodeGroupReputation DB Prefix: {}", NodeGroupReputation::db_prefix()); | ||||
| } | ||||
| @@ -1,6 +1,6 @@ | ||||
| use heromodels::models::heroledger::rhai::register_heroledger_rhai_modules; | ||||
| use heromodels_core::db::hero::OurDB; | ||||
| use rhai::{Dynamic, Engine}; | ||||
| use heromodels::models::heroledger::rhai::register_heroledger_rhai_modules; | ||||
| use std::sync::Arc; | ||||
| use std::{fs, path::Path}; | ||||
|  | ||||
|   | ||||
							
								
								
									
										15
									
								
								heromodels/examples/heroledger_example/README.md
									
									
									
									
									
										Normal file
									
								
							
							
						
						
									
										15
									
								
								heromodels/examples/heroledger_example/README.md
									
									
									
									
									
										Normal file
									
								
							| @@ -0,0 +1,15 @@ | ||||
| # Heroledger Postgres Example | ||||
|  | ||||
| This example demonstrates how to use the Heroledger `User` model against Postgres using the `heromodels::db::postgres` backend. | ||||
|  | ||||
| - Connects to Postgres with user `postgres` and password `test123` on `localhost:5432`. | ||||
| - Creates the table and indexes automatically on first use. | ||||
| - Shows basic CRUD and an index lookup on `username`. | ||||
|  | ||||
| Run it: | ||||
|  | ||||
| ```bash | ||||
| cargo run -p heromodels --example heroledger_example | ||||
| ``` | ||||
|  | ||||
| Make sure Postgres is running locally and accessible with the credentials above. | ||||
							
								
								
									
										54
									
								
								heromodels/examples/heroledger_example/example.rs
									
									
									
									
									
										Normal file
									
								
							
							
						
						
									
										54
									
								
								heromodels/examples/heroledger_example/example.rs
									
									
									
									
									
										Normal file
									
								
							| @@ -0,0 +1,54 @@ | ||||
| use heromodels::db::postgres::{Config, Postgres}; | ||||
| use heromodels::db::{Collection, Db}; | ||||
| use heromodels::models::heroledger::user::user_index::username; | ||||
| use heromodels::models::heroledger::user::{SecretBox, User}; | ||||
|  | ||||
| fn main() { | ||||
|     let db = Postgres::new( | ||||
|         Config::new() | ||||
|             .user(Some("postgres".into())) | ||||
|             .password(Some("test123".into())) | ||||
|             .host(Some("localhost".into())) | ||||
|             .port(Some(5432)), | ||||
|     ) | ||||
|     .expect("Can connect to Postgres"); | ||||
|  | ||||
|     println!("Heroledger User - Postgres Example"); | ||||
|     println!("=================================="); | ||||
|  | ||||
|     let users = db.collection::<User>().expect("open user collection"); | ||||
|  | ||||
|     // Clean | ||||
|     if let Ok(existing) = users.get_all() { | ||||
|         for u in existing { | ||||
|             let _ = users.delete_by_id(u.get_id()); | ||||
|         } | ||||
|     } | ||||
|  | ||||
|     let sb = SecretBox::new().data(vec![1, 2, 3]).nonce(vec![9, 9, 9]).build(); | ||||
|  | ||||
|     let u = User::new(0) | ||||
|         .username("alice") | ||||
|         .pubkey("PUBKEY_A") | ||||
|         .add_email("alice@example.com") | ||||
|         .add_userprofile(sb) | ||||
|         .build(); | ||||
|  | ||||
|     let (id, stored) = users.set(&u).expect("store user"); | ||||
|     println!("Stored user id={id} username={} pubkey={}", stored.username, stored.pubkey); | ||||
|  | ||||
|     let by_idx = users.get::<username, _>("alice").expect("by username"); | ||||
|     println!("Found {} user(s) with username=alice", by_idx.len()); | ||||
|  | ||||
|     let fetched = users.get_by_id(id).expect("get by id").expect("exists"); | ||||
|     println!("Fetched by id={} username={} emails={:?}", id, fetched.username, fetched.email); | ||||
|  | ||||
|     // Update | ||||
|     let updated = fetched.clone().add_email("work@alice.example"); | ||||
|     let (_, back) = users.set(&updated).expect("update user"); | ||||
|     println!("Updated emails = {:?}", back.email); | ||||
|  | ||||
|     // Delete | ||||
|     users.delete_by_id(id).expect("delete user"); | ||||
|     println!("Deleted user id={id}"); | ||||
| } | ||||
| @@ -1,8 +1,11 @@ | ||||
| use heromodels::db::postgres::Config; | ||||
| use heromodels::db::{Collection, Db}; | ||||
| use heromodels::models::userexample::user::user_index::{is_active, username}; | ||||
| use heromodels::models::userexample::user::user_index::{email, username}; | ||||
| use heromodels::models::{Comment, User}; | ||||
| use heromodels_core::Model; | ||||
| // For demonstrating embedded/nested indexes | ||||
| use heromodels::models::grid4::node::{ComputeSlice, DeviceInfo, Node}; | ||||
| use heromodels::models::grid4::node::node_index::{country as node_country, pubkey as node_pubkey}; | ||||
|  | ||||
| // Helper function to print user details | ||||
| fn print_user_details(user: &User) { | ||||
| @@ -37,14 +40,31 @@ fn main() { | ||||
|     ) | ||||
|     .expect("Can connect to postgress"); | ||||
|  | ||||
|     // Unique suffix to avoid collisions with legacy rows from prior runs | ||||
|     use std::time::{SystemTime, UNIX_EPOCH}; | ||||
|     let ts = SystemTime::now() | ||||
|         .duration_since(UNIX_EPOCH) | ||||
|         .unwrap() | ||||
|         .as_secs(); | ||||
|     let user1_name = format!("johndoe_{}", ts); | ||||
|     let user2_name = format!("janesmith_{}", ts); | ||||
|     let user3_name = format!("willism_{}", ts); | ||||
|     let user4_name = format!("carrols_{}", ts); | ||||
|     let user1_email = format!("john.doe+{}@example.com", ts); | ||||
|     let user2_email = format!("jane.smith+{}@example.com", ts); | ||||
|     let user3_email = format!("willis.masters+{}@example.com", ts); | ||||
|     let user4_email = format!("carrol.smith+{}@example.com", ts); | ||||
|  | ||||
|     println!("Hero Models - Basic Usage Example"); | ||||
|     println!("================================"); | ||||
|  | ||||
|     // Clean up any existing data to ensure consistent results | ||||
|     println!("Cleaning up existing data..."); | ||||
|     let user_collection = db.collection::<User>().expect("can open user collection"); | ||||
|     let comment_collection = db.collection::<Comment>().expect("can open comment collection"); | ||||
|      | ||||
|     let comment_collection = db | ||||
|         .collection::<Comment>() | ||||
|         .expect("can open comment collection"); | ||||
|  | ||||
|     // Clear all existing users and comments | ||||
|     if let Ok(existing_users) = user_collection.get_all() { | ||||
|         for user in existing_users { | ||||
| @@ -62,32 +82,32 @@ fn main() { | ||||
|  | ||||
|     // User 1 | ||||
|     let user1 = User::new() | ||||
|         .username("johndoe") | ||||
|         .email("john.doe@example.com") | ||||
|         .username(&user1_name) | ||||
|         .email(&user1_email) | ||||
|         .full_name("John Doe") | ||||
|         .is_active(false) | ||||
|         .build(); | ||||
|  | ||||
|     // User 2 | ||||
|     let user2 = User::new() | ||||
|         .username("janesmith") | ||||
|         .email("jane.smith@example.com") | ||||
|         .username(&user2_name) | ||||
|         .email(&user2_email) | ||||
|         .full_name("Jane Smith") | ||||
|         .is_active(true) | ||||
|         .build(); | ||||
|  | ||||
|     // User 3 | ||||
|     let user3 = User::new() | ||||
|         .username("willism") | ||||
|         .email("willis.masters@example.com") | ||||
|         .username(&user3_name) | ||||
|         .email(&user3_email) | ||||
|         .full_name("Willis Masters") | ||||
|         .is_active(true) | ||||
|         .build(); | ||||
|  | ||||
|     // User 4 | ||||
|     let user4 = User::new() | ||||
|         .username("carrols") | ||||
|         .email("carrol.smith@example.com") | ||||
|         .username(&user4_name) | ||||
|         .email(&user4_email) | ||||
|         .full_name("Carrol Smith") | ||||
|         .is_active(false) | ||||
|         .build(); | ||||
| @@ -143,66 +163,95 @@ fn main() { | ||||
|     let stored_users = db | ||||
|         .collection::<User>() | ||||
|         .expect("can open user collection") | ||||
|         .get::<username, _>("johndoe") | ||||
|         .get::<username, _>(&user1_name) | ||||
|         .expect("can load stored user"); | ||||
|  | ||||
|     assert_eq!(stored_users.len(), 1); | ||||
|     print_user_details(&stored_users[0]); | ||||
|  | ||||
|     // 2. Retrieve by active status | ||||
|     println!("\n2. By Active Status (Active = true):"); | ||||
|     let active_users = db | ||||
|     // 2. Retrieve by email index | ||||
|     println!("\n2. By Email Index:"); | ||||
|     let by_email = db | ||||
|         .collection::<User>() | ||||
|         .expect("can open user collection") | ||||
|         .get::<is_active, _>(&true) | ||||
|         .expect("can load stored users"); | ||||
|  | ||||
|     assert_eq!(active_users.len(), 2); | ||||
|     for active_user in active_users.iter() { | ||||
|         print_user_details(active_user); | ||||
|     } | ||||
|         .get::<email, _>(&user2_email) | ||||
|         .expect("can load stored user by email"); | ||||
|     assert_eq!(by_email.len(), 1); | ||||
|     print_user_details(&by_email[0]); | ||||
|  | ||||
|     // 3. Delete a user and show the updated results | ||||
|     println!("\n3. After Deleting a User:"); | ||||
|     let user_to_delete_id = active_users[0].get_id(); | ||||
|     let user_to_delete_id = stored_users[0].get_id(); | ||||
|     println!("Deleting user with ID: {user_to_delete_id}"); | ||||
|     db.collection::<User>() | ||||
|         .expect("can open user collection") | ||||
|         .delete_by_id(user_to_delete_id) | ||||
|         .expect("can delete existing user"); | ||||
|  | ||||
|     // Show remaining active users | ||||
|     let active_users = db | ||||
|     // Verify deletion by querying the same username again | ||||
|     let should_be_empty = db | ||||
|         .collection::<User>() | ||||
|         .expect("can open user collection") | ||||
|         .get::<is_active, _>(&true) | ||||
|         .expect("can load stored users"); | ||||
|  | ||||
|     println!("   a. Remaining Active Users:"); | ||||
|     assert_eq!(active_users.len(), 1); | ||||
|     for active_user in active_users.iter() { | ||||
|         print_user_details(active_user); | ||||
|     } | ||||
|  | ||||
|     // Show inactive users | ||||
|     let inactive_users = db | ||||
|         .collection::<User>() | ||||
|         .expect("can open user collection") | ||||
|         .get::<is_active, _>(&false) | ||||
|         .expect("can load stored users"); | ||||
|  | ||||
|     println!("   b. Inactive Users:"); | ||||
|     assert_eq!(inactive_users.len(), 2); | ||||
|     for inactive_user in inactive_users.iter() { | ||||
|         print_user_details(inactive_user); | ||||
|     } | ||||
|         .get::<username, _>(&user1_name) | ||||
|         .expect("can query by username after delete"); | ||||
|     println!("   a. Query by username '{}' after delete -> {} results", user1_name, should_be_empty.len()); | ||||
|     assert_eq!(should_be_empty.len(), 0); | ||||
|  | ||||
|     // Delete a user based on an index for good measure | ||||
|     db.collection::<User>() | ||||
|         .expect("can open user collection") | ||||
|         .delete::<username, _>("janesmith") | ||||
|         .delete::<username, _>(&user4_name) | ||||
|         .expect("can delete existing user"); | ||||
|  | ||||
|     // Demonstrate embedded/nested indexes with Grid4 Node | ||||
|     println!("\n--- Demonstrating Embedded/Nested Indexes (Grid4::Node) ---"); | ||||
|     println!("Node indexed fields: {:?}", Node::indexed_fields()); | ||||
|  | ||||
|     // Build a minimal node with nested data and persist it | ||||
|     let cs = ComputeSlice::new() | ||||
|         .nodeid(42) | ||||
|         .slice_id(1) | ||||
|         .mem_gb(32.0) | ||||
|         .storage_gb(512.0) | ||||
|         .passmark(6000) | ||||
|         .vcores(16) | ||||
|         .gpus(1) | ||||
|         .price_cc(0.33); | ||||
|     let dev = DeviceInfo { vendor: "ACME".into(), ..Default::default() }; | ||||
|     let node = Node::new() | ||||
|         .nodegroupid(101) | ||||
|         .uptime(99) | ||||
|         .add_compute_slice(cs) | ||||
|         .devices(dev) | ||||
|         .country("BE") | ||||
|         .pubkey("EX_NODE_PK_1") | ||||
|         .build(); | ||||
|     let (node_id, _stored_node) = db | ||||
|         .collection::<Node>() | ||||
|         .expect("can open node collection") | ||||
|         .set(&node) | ||||
|         .expect("can set node"); | ||||
|     println!("Stored node id: {}", node_id); | ||||
|  | ||||
|     // Query by top-level indexes | ||||
|     let be_nodes = db | ||||
|         .collection::<Node>() | ||||
|         .expect("can open node collection") | ||||
|         .get::<node_country, _>("BE") | ||||
|         .expect("can query nodes by country"); | ||||
|     println!("Nodes in BE (count may include legacy rows): {}", be_nodes.len()); | ||||
|  | ||||
|     let by_pk = db | ||||
|         .collection::<Node>() | ||||
|         .expect("can open node collection") | ||||
|         .get::<node_pubkey, _>("EX_NODE_PK_1") | ||||
|         .expect("can query node by pubkey"); | ||||
|     assert!(by_pk.iter().any(|n| n.get_id() == node_id)); | ||||
|  | ||||
|     // Note: Nested path indexes (e.g., devices.vendor, computeslices.passmark) are created and used | ||||
|     // for DB-side indexing, but are not yet exposed as typed Index keys in the API. They appear in | ||||
|     // Node::indexed_fields() and contribute to Model::db_keys(), enabling performant JSONB GIN indexes. | ||||
|  | ||||
|     println!("\n--- User Model Information ---"); | ||||
|     println!("User DB Prefix: {}", User::db_prefix()); | ||||
|  | ||||
| @@ -212,7 +261,7 @@ fn main() { | ||||
|     // 1. Create and save a comment | ||||
|     println!("\n1. Creating a Comment:"); | ||||
|     let comment = Comment::new() | ||||
|         .user_id(db_user1.get_id()) // commenter's user ID | ||||
|         .user_id(db_user2.get_id()) // commenter's user ID (use an existing user) | ||||
|         .content("This is a comment on the user") | ||||
|         .build(); | ||||
|  | ||||
| @@ -230,7 +279,7 @@ fn main() { | ||||
|  | ||||
|     // 3. Associate the comment with a user | ||||
|     println!("\n2. Associating Comment with User:"); | ||||
|     let mut updated_user = db_user1.clone(); | ||||
|     let mut updated_user = db_user2.clone(); | ||||
|     updated_user.base_data.add_comment(db_comment.get_id()); | ||||
|  | ||||
|     // Save the updated user and get the new version | ||||
|   | ||||
| @@ -8,8 +8,8 @@ use std::{ | ||||
|     collections::HashSet, | ||||
|     path::PathBuf, | ||||
|     sync::{ | ||||
|         atomic::{AtomicU32, Ordering}, | ||||
|         Arc, Mutex, | ||||
|         atomic::{AtomicU32, Ordering}, | ||||
|     }, | ||||
| }; | ||||
|  | ||||
|   | ||||
| @@ -119,4 +119,4 @@ impl Circle { | ||||
| /// Creates a new circle builder | ||||
| pub fn new_circle() -> Circle { | ||||
|     Circle::new() | ||||
| } | ||||
| } | ||||
|   | ||||
| @@ -1,16 +1,17 @@ | ||||
| use crate::db::Db; | ||||
| use rhailib_macros::{ | ||||
|     register_authorized_create_by_id_fn, register_authorized_delete_by_id_fn, register_authorized_get_by_id_fn, | ||||
| }; | ||||
| use rhai::plugin::*; | ||||
| use rhai::{Array, Dynamic, Engine, EvalAltResult, Map, Module}; | ||||
| use rhailib_macros::{ | ||||
|     register_authorized_create_by_id_fn, register_authorized_delete_by_id_fn, | ||||
|     register_authorized_get_by_id_fn, | ||||
| }; | ||||
| use std::collections::HashMap; | ||||
| use std::sync::Arc; | ||||
|  | ||||
| use crate::models::circle::Circle; | ||||
| type RhaiCircle = Circle; | ||||
| use crate::db::hero::OurDB; | ||||
| use crate::db::Collection; | ||||
| use crate::db::hero::OurDB; | ||||
| use crate::models::circle::ThemeData; | ||||
|  | ||||
| #[export_module] | ||||
|   | ||||
							
								
								
									
										128
									
								
								heromodels/src/models/grid4/bid.rs
									
									
									
									
									
										Normal file
									
								
							
							
						
						
									
										128
									
								
								heromodels/src/models/grid4/bid.rs
									
									
									
									
									
										Normal file
									
								
							| @@ -0,0 +1,128 @@ | ||||
| use heromodels_core::BaseModelData; | ||||
| use heromodels_derive::model; | ||||
| use rhai::{CustomType, TypeBuilder}; | ||||
| use serde::{Deserialize, Serialize}; | ||||
|  | ||||
| /// Bid status enumeration | ||||
| #[derive(Debug, Clone, Serialize, Deserialize, PartialEq, Default)] | ||||
| pub enum BidStatus { | ||||
|     #[default] | ||||
|     Pending, | ||||
|     Confirmed, | ||||
|     Assigned, | ||||
|     Cancelled, | ||||
|     Done, | ||||
| } | ||||
|  | ||||
| /// Billing period enumeration | ||||
| #[derive(Debug, Clone, Serialize, Deserialize, PartialEq, Default)] | ||||
| pub enum BillingPeriod { | ||||
|     #[default] | ||||
|     Hourly, | ||||
|     Monthly, | ||||
|     Yearly, | ||||
|     Biannually, | ||||
|     Triannually, | ||||
| } | ||||
|  | ||||
| /// I can bid for infra, and optionally get accepted | ||||
| #[model] | ||||
| #[derive(Debug, Clone, Serialize, Deserialize, PartialEq, Default, CustomType)] | ||||
| pub struct Bid { | ||||
|     pub base_data: BaseModelData, | ||||
|     /// links back to customer for this capacity (user on ledger) | ||||
|     #[index] | ||||
|     pub customer_id: u32, | ||||
|     /// nr of slices I need in 1 machine | ||||
|     pub compute_slices_nr: i32, | ||||
|     /// price per 1 GB slice I want to accept | ||||
|     pub compute_slice_price: f64, | ||||
|     /// nr of storage slices needed | ||||
|     pub storage_slices_nr: i32, | ||||
|     /// price per 1 GB storage slice I want to accept | ||||
|     pub storage_slice_price: f64, | ||||
|     pub status: BidStatus, | ||||
|     /// if obligation then will be charged and money needs to be in escrow, otherwise its an intent | ||||
|     pub obligation: bool, | ||||
|     /// epoch timestamp | ||||
|     pub start_date: u32, | ||||
|     /// epoch timestamp | ||||
|     pub end_date: u32, | ||||
|     /// signature as done by a user/consumer to validate their identity and intent | ||||
|     pub signature_user: String, | ||||
|     pub billing_period: BillingPeriod, | ||||
| } | ||||
|  | ||||
| impl Bid { | ||||
|     pub fn new() -> Self { | ||||
|         Self { | ||||
|             base_data: BaseModelData::new(), | ||||
|             customer_id: 0, | ||||
|             compute_slices_nr: 0, | ||||
|             compute_slice_price: 0.0, | ||||
|             storage_slices_nr: 0, | ||||
|             storage_slice_price: 0.0, | ||||
|             status: BidStatus::default(), | ||||
|             obligation: false, | ||||
|             start_date: 0, | ||||
|             end_date: 0, | ||||
|             signature_user: String::new(), | ||||
|             billing_period: BillingPeriod::default(), | ||||
|         } | ||||
|     } | ||||
|  | ||||
|     pub fn customer_id(mut self, v: u32) -> Self { | ||||
|         self.customer_id = v; | ||||
|         self | ||||
|     } | ||||
|      | ||||
|     pub fn compute_slices_nr(mut self, v: i32) -> Self { | ||||
|         self.compute_slices_nr = v; | ||||
|         self | ||||
|     } | ||||
|      | ||||
|     pub fn compute_slice_price(mut self, v: f64) -> Self { | ||||
|         self.compute_slice_price = v; | ||||
|         self | ||||
|     } | ||||
|      | ||||
|     pub fn storage_slices_nr(mut self, v: i32) -> Self { | ||||
|         self.storage_slices_nr = v; | ||||
|         self | ||||
|     } | ||||
|      | ||||
|     pub fn storage_slice_price(mut self, v: f64) -> Self { | ||||
|         self.storage_slice_price = v; | ||||
|         self | ||||
|     } | ||||
|      | ||||
|     pub fn status(mut self, v: BidStatus) -> Self { | ||||
|         self.status = v; | ||||
|         self | ||||
|     } | ||||
|      | ||||
|     pub fn obligation(mut self, v: bool) -> Self { | ||||
|         self.obligation = v; | ||||
|         self | ||||
|     } | ||||
|      | ||||
|     pub fn start_date(mut self, v: u32) -> Self { | ||||
|         self.start_date = v; | ||||
|         self | ||||
|     } | ||||
|      | ||||
|     pub fn end_date(mut self, v: u32) -> Self { | ||||
|         self.end_date = v; | ||||
|         self | ||||
|     } | ||||
|      | ||||
|     pub fn signature_user(mut self, v: impl ToString) -> Self { | ||||
|         self.signature_user = v.to_string(); | ||||
|         self | ||||
|     } | ||||
|      | ||||
|     pub fn billing_period(mut self, v: BillingPeriod) -> Self { | ||||
|         self.billing_period = v; | ||||
|         self | ||||
|     } | ||||
| } | ||||
							
								
								
									
										39
									
								
								heromodels/src/models/grid4/common.rs
									
									
									
									
									
										Normal file
									
								
							
							
						
						
									
										39
									
								
								heromodels/src/models/grid4/common.rs
									
									
									
									
									
										Normal file
									
								
							| @@ -0,0 +1,39 @@ | ||||
| use rhai::{CustomType, TypeBuilder}; | ||||
| use serde::{Deserialize, Serialize}; | ||||
|  | ||||
| /// SLA policy matching the V spec `SLAPolicy` | ||||
| #[derive(Debug, Clone, Serialize, Deserialize, PartialEq, Default, CustomType)] | ||||
| pub struct SLAPolicy { | ||||
|     /// should +90 | ||||
|     pub sla_uptime: i32, | ||||
|     /// minimal mbits we can expect avg over 1h per node, 0 means we don't guarantee | ||||
|     pub sla_bandwidth_mbit: i32, | ||||
|     /// 0-100, percent of money given back in relation to month if sla breached, | ||||
|     /// e.g. 200 means we return 2 months worth of rev if sla missed | ||||
|     pub sla_penalty: i32, | ||||
| } | ||||
|  | ||||
| impl SLAPolicy { | ||||
|     pub fn new() -> Self { Self::default() } | ||||
|     pub fn sla_uptime(mut self, v: i32) -> Self { self.sla_uptime = v; self } | ||||
|     pub fn sla_bandwidth_mbit(mut self, v: i32) -> Self { self.sla_bandwidth_mbit = v; self } | ||||
|     pub fn sla_penalty(mut self, v: i32) -> Self { self.sla_penalty = v; self } | ||||
|     pub fn build(self) -> Self { self } | ||||
| } | ||||
|  | ||||
| /// Pricing policy matching the V spec `PricingPolicy` | ||||
| #[derive(Debug, Clone, Serialize, Deserialize, PartialEq, Default, CustomType)] | ||||
| pub struct PricingPolicy { | ||||
|     /// e.g. 30,40,50 means if user has more CC in wallet than 1 year utilization | ||||
|     /// then this provider gives 30%, 2Y 40%, ... | ||||
|     pub marketplace_year_discounts: Vec<i32>, | ||||
|     /// e.g. 10,20,30 | ||||
|     pub volume_discounts: Vec<i32>, | ||||
| } | ||||
|  | ||||
| impl PricingPolicy { | ||||
|     pub fn new() -> Self { Self { marketplace_year_discounts: vec![30, 40, 50], volume_discounts: vec![10, 20, 30] } } | ||||
|     pub fn marketplace_year_discounts(mut self, v: Vec<i32>) -> Self { self.marketplace_year_discounts = v; self } | ||||
|     pub fn volume_discounts(mut self, v: Vec<i32>) -> Self { self.volume_discounts = v; self } | ||||
|     pub fn build(self) -> Self { self } | ||||
| } | ||||
							
								
								
									
										219
									
								
								heromodels/src/models/grid4/contract.rs
									
									
									
									
									
										Normal file
									
								
							
							
						
						
									
										219
									
								
								heromodels/src/models/grid4/contract.rs
									
									
									
									
									
										Normal file
									
								
							| @@ -0,0 +1,219 @@ | ||||
| use heromodels_core::BaseModelData; | ||||
| use heromodels_derive::model; | ||||
| use rhai::{CustomType, TypeBuilder}; | ||||
| use serde::{Deserialize, Serialize}; | ||||
| use super::bid::BillingPeriod; | ||||
|  | ||||
| /// Contract status enumeration | ||||
| #[derive(Debug, Clone, Serialize, Deserialize, PartialEq, Default)] | ||||
| pub enum ContractStatus { | ||||
|     #[default] | ||||
|     Active, | ||||
|     Cancelled, | ||||
|     Error, | ||||
|     Paused, | ||||
| } | ||||
|  | ||||
| /// Compute slice provisioned for a contract | ||||
| #[derive(Debug, Clone, Serialize, Deserialize, PartialEq, Default, CustomType)] | ||||
| pub struct ComputeSliceProvisioned { | ||||
|     pub node_id: u32, | ||||
|     /// the id of the slice in the node | ||||
|     pub id: u16, | ||||
|     pub mem_gb: f64, | ||||
|     pub storage_gb: f64, | ||||
|     pub passmark: i32, | ||||
|     pub vcores: i32, | ||||
|     pub cpu_oversubscription: i32, | ||||
|     pub tags: String, | ||||
| } | ||||
|  | ||||
| /// Storage slice provisioned for a contract | ||||
| #[derive(Debug, Clone, Serialize, Deserialize, PartialEq, Default, CustomType)] | ||||
| pub struct StorageSliceProvisioned { | ||||
|     pub node_id: u32, | ||||
|     /// the id of the slice in the node, are tracked in the node itself | ||||
|     pub id: u16, | ||||
|     pub storage_size_gb: i32, | ||||
|     pub tags: String, | ||||
| } | ||||
|  | ||||
| /// Contract for provisioned infrastructure | ||||
| #[model] | ||||
| #[derive(Debug, Clone, Serialize, Deserialize, PartialEq, Default, CustomType)] | ||||
| pub struct Contract { | ||||
|     pub base_data: BaseModelData, | ||||
|     /// links back to customer for this capacity (user on ledger) | ||||
|     #[index] | ||||
|     pub customer_id: u32, | ||||
|     pub compute_slices: Vec<ComputeSliceProvisioned>, | ||||
|     pub storage_slices: Vec<StorageSliceProvisioned>, | ||||
|     /// price per 1 GB agreed upon | ||||
|     pub compute_slice_price: f64, | ||||
|     /// price per 1 GB agreed upon | ||||
|     pub storage_slice_price: f64, | ||||
|     /// price per 1 GB agreed upon (transfer) | ||||
|     pub network_slice_price: f64, | ||||
|     pub status: ContractStatus, | ||||
|     /// epoch timestamp | ||||
|     pub start_date: u32, | ||||
|     /// epoch timestamp | ||||
|     pub end_date: u32, | ||||
|     /// signature as done by a user/consumer to validate their identity and intent | ||||
|     pub signature_user: String, | ||||
|     /// signature as done by the hoster | ||||
|     pub signature_hoster: String, | ||||
|     pub billing_period: BillingPeriod, | ||||
| } | ||||
|  | ||||
| impl Contract { | ||||
|     pub fn new() -> Self { | ||||
|         Self { | ||||
|             base_data: BaseModelData::new(), | ||||
|             customer_id: 0, | ||||
|             compute_slices: Vec::new(), | ||||
|             storage_slices: Vec::new(), | ||||
|             compute_slice_price: 0.0, | ||||
|             storage_slice_price: 0.0, | ||||
|             network_slice_price: 0.0, | ||||
|             status: ContractStatus::default(), | ||||
|             start_date: 0, | ||||
|             end_date: 0, | ||||
|             signature_user: String::new(), | ||||
|             signature_hoster: String::new(), | ||||
|             billing_period: BillingPeriod::default(), | ||||
|         } | ||||
|     } | ||||
|  | ||||
|     pub fn customer_id(mut self, v: u32) -> Self { | ||||
|         self.customer_id = v; | ||||
|         self | ||||
|     } | ||||
|      | ||||
|     pub fn add_compute_slice(mut self, slice: ComputeSliceProvisioned) -> Self { | ||||
|         self.compute_slices.push(slice); | ||||
|         self | ||||
|     } | ||||
|      | ||||
|     pub fn add_storage_slice(mut self, slice: StorageSliceProvisioned) -> Self { | ||||
|         self.storage_slices.push(slice); | ||||
|         self | ||||
|     } | ||||
|      | ||||
|     pub fn compute_slice_price(mut self, v: f64) -> Self { | ||||
|         self.compute_slice_price = v; | ||||
|         self | ||||
|     } | ||||
|      | ||||
|     pub fn storage_slice_price(mut self, v: f64) -> Self { | ||||
|         self.storage_slice_price = v; | ||||
|         self | ||||
|     } | ||||
|      | ||||
|     pub fn network_slice_price(mut self, v: f64) -> Self { | ||||
|         self.network_slice_price = v; | ||||
|         self | ||||
|     } | ||||
|      | ||||
|     pub fn status(mut self, v: ContractStatus) -> Self { | ||||
|         self.status = v; | ||||
|         self | ||||
|     } | ||||
|      | ||||
|     pub fn start_date(mut self, v: u32) -> Self { | ||||
|         self.start_date = v; | ||||
|         self | ||||
|     } | ||||
|      | ||||
|     pub fn end_date(mut self, v: u32) -> Self { | ||||
|         self.end_date = v; | ||||
|         self | ||||
|     } | ||||
|      | ||||
|     pub fn signature_user(mut self, v: impl ToString) -> Self { | ||||
|         self.signature_user = v.to_string(); | ||||
|         self | ||||
|     } | ||||
|      | ||||
|     pub fn signature_hoster(mut self, v: impl ToString) -> Self { | ||||
|         self.signature_hoster = v.to_string(); | ||||
|         self | ||||
|     } | ||||
|      | ||||
|     pub fn billing_period(mut self, v: BillingPeriod) -> Self { | ||||
|         self.billing_period = v; | ||||
|         self | ||||
|     } | ||||
| } | ||||
|  | ||||
| impl ComputeSliceProvisioned { | ||||
|     pub fn new() -> Self { | ||||
|         Self::default() | ||||
|     } | ||||
|      | ||||
|     pub fn node_id(mut self, v: u32) -> Self { | ||||
|         self.node_id = v; | ||||
|         self | ||||
|     } | ||||
|      | ||||
|     pub fn id(mut self, v: u16) -> Self { | ||||
|         self.id = v; | ||||
|         self | ||||
|     } | ||||
|      | ||||
|     pub fn mem_gb(mut self, v: f64) -> Self { | ||||
|         self.mem_gb = v; | ||||
|         self | ||||
|     } | ||||
|      | ||||
|     pub fn storage_gb(mut self, v: f64) -> Self { | ||||
|         self.storage_gb = v; | ||||
|         self | ||||
|     } | ||||
|      | ||||
|     pub fn passmark(mut self, v: i32) -> Self { | ||||
|         self.passmark = v; | ||||
|         self | ||||
|     } | ||||
|      | ||||
|     pub fn vcores(mut self, v: i32) -> Self { | ||||
|         self.vcores = v; | ||||
|         self | ||||
|     } | ||||
|      | ||||
|     pub fn cpu_oversubscription(mut self, v: i32) -> Self { | ||||
|         self.cpu_oversubscription = v; | ||||
|         self | ||||
|     } | ||||
|      | ||||
|     pub fn tags(mut self, v: impl ToString) -> Self { | ||||
|         self.tags = v.to_string(); | ||||
|         self | ||||
|     } | ||||
| } | ||||
|  | ||||
| impl StorageSliceProvisioned { | ||||
|     pub fn new() -> Self { | ||||
|         Self::default() | ||||
|     } | ||||
|      | ||||
|     pub fn node_id(mut self, v: u32) -> Self { | ||||
|         self.node_id = v; | ||||
|         self | ||||
|     } | ||||
|      | ||||
|     pub fn id(mut self, v: u16) -> Self { | ||||
|         self.id = v; | ||||
|         self | ||||
|     } | ||||
|      | ||||
|     pub fn storage_size_gb(mut self, v: i32) -> Self { | ||||
|         self.storage_size_gb = v; | ||||
|         self | ||||
|     } | ||||
|      | ||||
|     pub fn tags(mut self, v: impl ToString) -> Self { | ||||
|         self.tags = v.to_string(); | ||||
|         self | ||||
|     } | ||||
| } | ||||
| @@ -1,16 +1,18 @@ | ||||
| pub mod bid; | ||||
| pub mod common; | ||||
| pub mod contract; | ||||
| pub mod node; | ||||
| pub mod nodegroup; | ||||
| pub mod reputation; | ||||
| pub mod reservation; | ||||
|  | ||||
| pub use bid::{Bid, BidStatus, BillingPeriod}; | ||||
| pub use common::{PricingPolicy, SLAPolicy}; | ||||
| pub use contract::{Contract, ContractStatus, ComputeSliceProvisioned, StorageSliceProvisioned}; | ||||
| pub use node::{ | ||||
|     Node, | ||||
|     DeviceInfo, | ||||
|     StorageDevice, | ||||
|     MemoryDevice, | ||||
|     CPUDevice, | ||||
|     GPUDevice, | ||||
|     NetworkDevice, | ||||
|     NodeCapacity, | ||||
|     ComputeSlice, | ||||
|     StorageSlice, | ||||
|     PricingPolicy, | ||||
|     SLAPolicy, | ||||
| }; | ||||
|     CPUDevice, ComputeSlice, DeviceInfo, GPUDevice, MemoryDevice, NetworkDevice, Node, | ||||
|     NodeCapacity, StorageDevice, StorageSlice, | ||||
| }; | ||||
| pub use nodegroup::NodeGroup; | ||||
| pub use reputation::{NodeGroupReputation, NodeReputation}; | ||||
| pub use reservation::{Reservation, ReservationStatus}; | ||||
|   | ||||
| @@ -1,7 +1,8 @@ | ||||
| use heromodels_core::BaseModelData; | ||||
| use heromodels_derive::model; | ||||
| use rhai::CustomType; | ||||
| use rhai::{CustomType, TypeBuilder}; | ||||
| use serde::{Deserialize, Serialize}; | ||||
| use super::common::{PricingPolicy, SLAPolicy}; | ||||
|  | ||||
| /// Storage device information | ||||
| #[derive(Debug, Clone, Serialize, Deserialize, PartialEq, Default, CustomType)] | ||||
| @@ -94,57 +95,26 @@ pub struct NodeCapacity { | ||||
|     pub vcores: i32, | ||||
| } | ||||
|  | ||||
| /// Pricing policy for slices (minimal version until full spec available) | ||||
| #[derive(Debug, Clone, Serialize, Deserialize, PartialEq, Default, CustomType)] | ||||
| pub struct PricingPolicy { | ||||
|     /// Human friendly policy name (e.g. "fixed", "market") | ||||
|     pub name: String, | ||||
|     /// Optional free-form details as JSON-encoded string | ||||
|     pub details: Option<String>, | ||||
| } | ||||
|  | ||||
| /// SLA policy for slices (minimal version until full spec available) | ||||
| #[derive(Debug, Clone, Serialize, Deserialize, PartialEq, Default, CustomType)] | ||||
| pub struct SLAPolicy { | ||||
|     /// Uptime in percentage (0..100) | ||||
|     pub uptime: f32, | ||||
|     /// Max response time in ms | ||||
|     pub max_response_time_ms: u32, | ||||
| } | ||||
| // PricingPolicy and SLAPolicy moved to `common.rs` to be shared across models. | ||||
|  | ||||
| /// Compute slice (typically represents a base unit of compute) | ||||
| #[model] | ||||
| #[derive(Debug, Clone, Serialize, Deserialize, PartialEq, Default, CustomType)] | ||||
| pub struct ComputeSlice { | ||||
|     pub base_data: BaseModelData, | ||||
|     /// the node in the grid, there is an object describing the node | ||||
|     #[index] | ||||
|     pub nodeid: u32, | ||||
|     /// the id of the slice in the node | ||||
|     #[index] | ||||
|     pub id: i32, | ||||
|     pub id: u16, | ||||
|     pub mem_gb: f64, | ||||
|     pub storage_gb: f64, | ||||
|     pub passmark: i32, | ||||
|     pub vcores: i32, | ||||
|     pub cpu_oversubscription: i32, | ||||
|     pub storage_oversubscription: i32, | ||||
|     /// Min/max allowed price range for validation | ||||
|     #[serde(default)] | ||||
|     pub price_range: Vec<f64>, | ||||
|     /// nr of GPU's see node to know what GPU's are | ||||
|     pub gpus: u8, | ||||
|     /// price per slice (even if the grouped one) | ||||
|     pub price_cc: f64, | ||||
|     pub pricing_policy: PricingPolicy, | ||||
|     pub sla_policy: SLAPolicy, | ||||
| } | ||||
|  | ||||
| impl ComputeSlice { | ||||
|     pub fn new() -> Self { | ||||
|         Self { | ||||
|             base_data: BaseModelData::new(), | ||||
|             nodeid: 0, | ||||
|             id: 0, | ||||
|             mem_gb: 0.0, | ||||
|             storage_gb: 0.0, | ||||
| @@ -152,63 +122,62 @@ impl ComputeSlice { | ||||
|             vcores: 0, | ||||
|             cpu_oversubscription: 0, | ||||
|             storage_oversubscription: 0, | ||||
|             price_range: vec![0.0, 0.0], | ||||
|             gpus: 0, | ||||
|             price_cc: 0.0, | ||||
|             pricing_policy: PricingPolicy::default(), | ||||
|             sla_policy: SLAPolicy::default(), | ||||
|         } | ||||
|     } | ||||
|  | ||||
|     pub fn nodeid(mut self, nodeid: u32) -> Self { self.nodeid = nodeid; self } | ||||
|     pub fn slice_id(mut self, id: i32) -> Self { self.id = id; self } | ||||
|     pub fn mem_gb(mut self, v: f64) -> Self { self.mem_gb = v; self } | ||||
|     pub fn storage_gb(mut self, v: f64) -> Self { self.storage_gb = v; self } | ||||
|     pub fn passmark(mut self, v: i32) -> Self { self.passmark = v; self } | ||||
|     pub fn vcores(mut self, v: i32) -> Self { self.vcores = v; self } | ||||
|     pub fn cpu_oversubscription(mut self, v: i32) -> Self { self.cpu_oversubscription = v; self } | ||||
|     pub fn storage_oversubscription(mut self, v: i32) -> Self { self.storage_oversubscription = v; self } | ||||
|     pub fn price_range(mut self, min_max: Vec<f64>) -> Self { self.price_range = min_max; self } | ||||
|     pub fn gpus(mut self, v: u8) -> Self { self.gpus = v; self } | ||||
|     pub fn price_cc(mut self, v: f64) -> Self { self.price_cc = v; self } | ||||
|     pub fn pricing_policy(mut self, p: PricingPolicy) -> Self { self.pricing_policy = p; self } | ||||
|     pub fn sla_policy(mut self, p: SLAPolicy) -> Self { self.sla_policy = p; self } | ||||
|     pub fn id(mut self, id: u16) -> Self { | ||||
|         self.id = id; | ||||
|         self | ||||
|     } | ||||
|     pub fn mem_gb(mut self, v: f64) -> Self { | ||||
|         self.mem_gb = v; | ||||
|         self | ||||
|     } | ||||
|     pub fn storage_gb(mut self, v: f64) -> Self { | ||||
|         self.storage_gb = v; | ||||
|         self | ||||
|     } | ||||
|     pub fn passmark(mut self, v: i32) -> Self { | ||||
|         self.passmark = v; | ||||
|         self | ||||
|     } | ||||
|     pub fn vcores(mut self, v: i32) -> Self { | ||||
|         self.vcores = v; | ||||
|         self | ||||
|     } | ||||
|     pub fn cpu_oversubscription(mut self, v: i32) -> Self { | ||||
|         self.cpu_oversubscription = v; | ||||
|         self | ||||
|     } | ||||
|     pub fn storage_oversubscription(mut self, v: i32) -> Self { | ||||
|         self.storage_oversubscription = v; | ||||
|         self | ||||
|     } | ||||
|     pub fn gpus(mut self, v: u8) -> Self { | ||||
|         self.gpus = v; | ||||
|         self | ||||
|     } | ||||
| } | ||||
|  | ||||
| /// Storage slice (typically 1GB of storage) | ||||
| #[model] | ||||
| #[derive(Debug, Clone, Serialize, Deserialize, PartialEq, Default, CustomType)] | ||||
| pub struct StorageSlice { | ||||
|     pub base_data: BaseModelData, | ||||
|     /// the node in the grid | ||||
|     #[index] | ||||
|     pub nodeid: u32, | ||||
|     /// the id of the slice in the node, are tracked in the node itself | ||||
|     #[index] | ||||
|     pub id: i32, | ||||
|     /// price per slice (even if the grouped one) | ||||
|     pub price_cc: f64, | ||||
|     pub pricing_policy: PricingPolicy, | ||||
|     pub sla_policy: SLAPolicy, | ||||
|     pub id: u16, | ||||
| } | ||||
|  | ||||
| impl StorageSlice { | ||||
|     pub fn new() -> Self { | ||||
|         Self { | ||||
|             base_data: BaseModelData::new(), | ||||
|             nodeid: 0, | ||||
|             id: 0, | ||||
|             price_cc: 0.0, | ||||
|             pricing_policy: PricingPolicy::default(), | ||||
|             sla_policy: SLAPolicy::default(), | ||||
|         } | ||||
|     } | ||||
|  | ||||
|     pub fn nodeid(mut self, nodeid: u32) -> Self { self.nodeid = nodeid; self } | ||||
|     pub fn slice_id(mut self, id: i32) -> Self { self.id = id; self } | ||||
|     pub fn price_cc(mut self, v: f64) -> Self { self.price_cc = v; self } | ||||
|     pub fn pricing_policy(mut self, p: PricingPolicy) -> Self { self.pricing_policy = p; self } | ||||
|     pub fn sla_policy(mut self, p: SLAPolicy) -> Self { self.sla_policy = p; self } | ||||
|     pub fn id(mut self, id: u16) -> Self { | ||||
|         self.id = id; | ||||
|         self | ||||
|     } | ||||
| } | ||||
|  | ||||
| /// Grid4 Node model | ||||
| @@ -224,13 +193,20 @@ pub struct Node { | ||||
|     pub computeslices: Vec<ComputeSlice>, | ||||
|     pub storageslices: Vec<StorageSlice>, | ||||
|     pub devices: DeviceInfo, | ||||
|     /// 2 letter code | ||||
|     /// 2 letter code as specified in lib/data/countries/data/countryInfo.txt | ||||
|     #[index] | ||||
|     pub country: String, | ||||
|     /// Hardware capacity details | ||||
|     pub capacity: NodeCapacity, | ||||
|     /// lets keep it simple and compatible | ||||
|     pub provisiontime: u32, | ||||
|     /// first time node was active | ||||
|     pub birthtime: u32, | ||||
|     /// node public key | ||||
|     #[index] | ||||
|     pub pubkey: String, | ||||
|     /// signature done on node to validate pubkey with privkey | ||||
|     pub signature_node: String, | ||||
|     /// signature as done by farmers to validate their identity | ||||
|     pub signature_farmer: String, | ||||
| } | ||||
|  | ||||
| impl Node { | ||||
| @@ -244,21 +220,61 @@ impl Node { | ||||
|             devices: DeviceInfo::default(), | ||||
|             country: String::new(), | ||||
|             capacity: NodeCapacity::default(), | ||||
|             provisiontime: 0, | ||||
|             birthtime: 0, | ||||
|             pubkey: String::new(), | ||||
|             signature_node: String::new(), | ||||
|             signature_farmer: String::new(), | ||||
|         } | ||||
|     } | ||||
|  | ||||
|     pub fn nodegroupid(mut self, v: i32) -> Self { self.nodegroupid = v; self } | ||||
|     pub fn uptime(mut self, v: i32) -> Self { self.uptime = v; self } | ||||
|     pub fn add_compute_slice(mut self, s: ComputeSlice) -> Self { self.computeslices.push(s); self } | ||||
|     pub fn add_storage_slice(mut self, s: StorageSlice) -> Self { self.storageslices.push(s); self } | ||||
|     pub fn devices(mut self, d: DeviceInfo) -> Self { self.devices = d; self } | ||||
|     pub fn country(mut self, c: impl ToString) -> Self { self.country = c.to_string(); self } | ||||
|     pub fn capacity(mut self, c: NodeCapacity) -> Self { self.capacity = c; self } | ||||
|     pub fn provisiontime(mut self, t: u32) -> Self { self.provisiontime = t; self } | ||||
|     pub fn nodegroupid(mut self, v: i32) -> Self { | ||||
|         self.nodegroupid = v; | ||||
|         self | ||||
|     } | ||||
|     pub fn uptime(mut self, v: i32) -> Self { | ||||
|         self.uptime = v; | ||||
|         self | ||||
|     } | ||||
|     pub fn add_compute_slice(mut self, s: ComputeSlice) -> Self { | ||||
|         self.computeslices.push(s); | ||||
|         self | ||||
|     } | ||||
|     pub fn add_storage_slice(mut self, s: StorageSlice) -> Self { | ||||
|         self.storageslices.push(s); | ||||
|         self | ||||
|     } | ||||
|     pub fn devices(mut self, d: DeviceInfo) -> Self { | ||||
|         self.devices = d; | ||||
|         self | ||||
|     } | ||||
|     pub fn country(mut self, c: impl ToString) -> Self { | ||||
|         self.country = c.to_string(); | ||||
|         self | ||||
|     } | ||||
|     pub fn capacity(mut self, c: NodeCapacity) -> Self { | ||||
|         self.capacity = c; | ||||
|         self | ||||
|     } | ||||
|     pub fn birthtime(mut self, t: u32) -> Self { | ||||
|         self.birthtime = t; | ||||
|         self | ||||
|     } | ||||
|  | ||||
|     pub fn pubkey(mut self, v: impl ToString) -> Self { | ||||
|         self.pubkey = v.to_string(); | ||||
|         self | ||||
|     } | ||||
|     pub fn signature_node(mut self, v: impl ToString) -> Self { | ||||
|         self.signature_node = v.to_string(); | ||||
|         self | ||||
|     } | ||||
|     pub fn signature_farmer(mut self, v: impl ToString) -> Self { | ||||
|         self.signature_farmer = v.to_string(); | ||||
|         self | ||||
|     } | ||||
|  | ||||
|     /// Placeholder for capacity recalculation out of the devices on the Node | ||||
|     pub fn recalc_capacity(mut self) -> Self { | ||||
|     pub fn check(self) -> Self { | ||||
|         // TODO: calculate NodeCapacity out of the devices on the Node | ||||
|         self | ||||
|     } | ||||
|   | ||||
							
								
								
									
										52
									
								
								heromodels/src/models/grid4/nodegroup.rs
									
									
									
									
									
										Normal file
									
								
							
							
						
						
									
										52
									
								
								heromodels/src/models/grid4/nodegroup.rs
									
									
									
									
									
										Normal file
									
								
							| @@ -0,0 +1,52 @@ | ||||
| use heromodels_core::BaseModelData; | ||||
| use heromodels_derive::model; | ||||
| use rhai::{CustomType, TypeBuilder}; | ||||
| use serde::{Deserialize, Serialize}; | ||||
|  | ||||
| use super::common::{PricingPolicy, SLAPolicy}; | ||||
|  | ||||
| /// Grid4 NodeGroup model (root object for farmer configuration) | ||||
| #[model] | ||||
| #[derive(Debug, Clone, Serialize, Deserialize, PartialEq, Default, CustomType)] | ||||
| pub struct NodeGroup { | ||||
|     pub base_data: BaseModelData, | ||||
|     /// link back to farmer who owns the nodegroup, is a user? | ||||
|     #[index] | ||||
|     pub farmerid: u32, | ||||
|     /// only visible by farmer, in future encrypted, used to boot a node | ||||
|     pub secret: String, | ||||
|     pub description: String, | ||||
|     pub slapolicy: SLAPolicy, | ||||
|     pub pricingpolicy: PricingPolicy, | ||||
|     /// pricing in CC - cloud credit, per 2GB node slice | ||||
|     pub compute_slice_normalized_pricing_cc: f64, | ||||
|     /// pricing in CC - cloud credit, per 1GB storage slice | ||||
|     pub storage_slice_normalized_pricing_cc: f64, | ||||
|     /// signature as done by farmers to validate that they created this group | ||||
|     pub signature_farmer: String, | ||||
| } | ||||
|  | ||||
| impl NodeGroup { | ||||
|     pub fn new() -> Self { | ||||
|         Self { | ||||
|             base_data: BaseModelData::new(), | ||||
|             farmerid: 0, | ||||
|             secret: String::new(), | ||||
|             description: String::new(), | ||||
|             slapolicy: SLAPolicy::default(), | ||||
|             pricingpolicy: PricingPolicy::new(), | ||||
|             compute_slice_normalized_pricing_cc: 0.0, | ||||
|             storage_slice_normalized_pricing_cc: 0.0, | ||||
|             signature_farmer: String::new(), | ||||
|         } | ||||
|     } | ||||
|  | ||||
|     pub fn farmerid(mut self, v: u32) -> Self { self.farmerid = v; self } | ||||
|     pub fn secret(mut self, v: impl ToString) -> Self { self.secret = v.to_string(); self } | ||||
|     pub fn description(mut self, v: impl ToString) -> Self { self.description = v.to_string(); self } | ||||
|     pub fn slapolicy(mut self, v: SLAPolicy) -> Self { self.slapolicy = v; self } | ||||
|     pub fn pricingpolicy(mut self, v: PricingPolicy) -> Self { self.pricingpolicy = v; self } | ||||
|     pub fn compute_slice_normalized_pricing_cc(mut self, v: f64) -> Self { self.compute_slice_normalized_pricing_cc = v; self } | ||||
|     pub fn storage_slice_normalized_pricing_cc(mut self, v: f64) -> Self { self.storage_slice_normalized_pricing_cc = v; self } | ||||
|     pub fn signature_farmer(mut self, v: impl ToString) -> Self { self.signature_farmer = v.to_string(); self } | ||||
| } | ||||
							
								
								
									
										85
									
								
								heromodels/src/models/grid4/reputation.rs
									
									
									
									
									
										Normal file
									
								
							
							
						
						
									
										85
									
								
								heromodels/src/models/grid4/reputation.rs
									
									
									
									
									
										Normal file
									
								
							| @@ -0,0 +1,85 @@ | ||||
| use heromodels_core::BaseModelData; | ||||
| use heromodels_derive::model; | ||||
| use rhai::{CustomType, TypeBuilder}; | ||||
| use serde::{Deserialize, Serialize}; | ||||
|  | ||||
| /// Node reputation information | ||||
| #[derive(Debug, Clone, Serialize, Deserialize, PartialEq, Default, CustomType)] | ||||
| pub struct NodeReputation { | ||||
|     pub node_id: u32, | ||||
|     /// between 0 and 100, earned over time | ||||
|     pub reputation: i32, | ||||
|     /// between 0 and 100, set by system, farmer has no ability to set this | ||||
|     pub uptime: i32, | ||||
| } | ||||
|  | ||||
| /// NodeGroup reputation model | ||||
| #[model] | ||||
| #[derive(Debug, Clone, Serialize, Deserialize, PartialEq, Default, CustomType)] | ||||
| pub struct NodeGroupReputation { | ||||
|     pub base_data: BaseModelData, | ||||
|     #[index] | ||||
|     pub nodegroup_id: u32, | ||||
|     /// between 0 and 100, earned over time | ||||
|     pub reputation: i32, | ||||
|     /// between 0 and 100, set by system, farmer has no ability to set this | ||||
|     pub uptime: i32, | ||||
|     pub nodes: Vec<NodeReputation>, | ||||
| } | ||||
|  | ||||
| impl NodeGroupReputation { | ||||
|     pub fn new() -> Self { | ||||
|         Self { | ||||
|             base_data: BaseModelData::new(), | ||||
|             nodegroup_id: 0, | ||||
|             reputation: 50, // default as per spec | ||||
|             uptime: 0, | ||||
|             nodes: Vec::new(), | ||||
|         } | ||||
|     } | ||||
|  | ||||
|     pub fn nodegroup_id(mut self, v: u32) -> Self { | ||||
|         self.nodegroup_id = v; | ||||
|         self | ||||
|     } | ||||
|      | ||||
|     pub fn reputation(mut self, v: i32) -> Self { | ||||
|         self.reputation = v; | ||||
|         self | ||||
|     } | ||||
|      | ||||
|     pub fn uptime(mut self, v: i32) -> Self { | ||||
|         self.uptime = v; | ||||
|         self | ||||
|     } | ||||
|      | ||||
|     pub fn add_node_reputation(mut self, node_rep: NodeReputation) -> Self { | ||||
|         self.nodes.push(node_rep); | ||||
|         self | ||||
|     } | ||||
| } | ||||
|  | ||||
| impl NodeReputation { | ||||
|     pub fn new() -> Self { | ||||
|         Self { | ||||
|             node_id: 0, | ||||
|             reputation: 50, // default as per spec | ||||
|             uptime: 0, | ||||
|         } | ||||
|     } | ||||
|      | ||||
|     pub fn node_id(mut self, v: u32) -> Self { | ||||
|         self.node_id = v; | ||||
|         self | ||||
|     } | ||||
|      | ||||
|     pub fn reputation(mut self, v: i32) -> Self { | ||||
|         self.reputation = v; | ||||
|         self | ||||
|     } | ||||
|      | ||||
|     pub fn uptime(mut self, v: i32) -> Self { | ||||
|         self.uptime = v; | ||||
|         self | ||||
|     } | ||||
| } | ||||
							
								
								
									
										58
									
								
								heromodels/src/models/grid4/reservation.rs
									
									
									
									
									
										Normal file
									
								
							
							
						
						
									
										58
									
								
								heromodels/src/models/grid4/reservation.rs
									
									
									
									
									
										Normal file
									
								
							| @@ -0,0 +1,58 @@ | ||||
| use heromodels_core::BaseModelData; | ||||
| use heromodels_derive::model; | ||||
| use rhai::{CustomType, TypeBuilder}; | ||||
| use serde::{Deserialize, Serialize}; | ||||
|  | ||||
| /// Reservation status as per V spec | ||||
| #[derive(Debug, Clone, Serialize, Deserialize, PartialEq, Default)] | ||||
| pub enum ReservationStatus { | ||||
|     #[default] | ||||
|     Pending, | ||||
|     Confirmed, | ||||
|     Assigned, | ||||
|     Cancelled, | ||||
|     Done, | ||||
| } | ||||
|  | ||||
| /// Grid4 Reservation model | ||||
| #[model] | ||||
| #[derive(Debug, Clone, Serialize, Deserialize, PartialEq, Default, CustomType)] | ||||
| pub struct Reservation { | ||||
|     pub base_data: BaseModelData, | ||||
|     /// links back to customer for this capacity | ||||
|     #[index] | ||||
|     pub customer_id: u32, | ||||
|     pub compute_slices: Vec<u32>, | ||||
|     pub storage_slices: Vec<u32>, | ||||
|     pub status: ReservationStatus, | ||||
|     /// if obligation then will be charged and money needs to be in escrow, otherwise its an intent | ||||
|     pub obligation: bool, | ||||
|     /// epoch | ||||
|     pub start_date: u32, | ||||
|     pub end_date: u32, | ||||
| } | ||||
|  | ||||
| impl Reservation { | ||||
|     pub fn new() -> Self { | ||||
|         Self { | ||||
|             base_data: BaseModelData::new(), | ||||
|             customer_id: 0, | ||||
|             compute_slices: Vec::new(), | ||||
|             storage_slices: Vec::new(), | ||||
|             status: ReservationStatus::Pending, | ||||
|             obligation: false, | ||||
|             start_date: 0, | ||||
|             end_date: 0, | ||||
|         } | ||||
|     } | ||||
|  | ||||
|     pub fn customer_id(mut self, v: u32) -> Self { self.customer_id = v; self } | ||||
|     pub fn add_compute_slice(mut self, id: u32) -> Self { self.compute_slices.push(id); self } | ||||
|     pub fn compute_slices(mut self, v: Vec<u32>) -> Self { self.compute_slices = v; self } | ||||
|     pub fn add_storage_slice(mut self, id: u32) -> Self { self.storage_slices.push(id); self } | ||||
|     pub fn storage_slices(mut self, v: Vec<u32>) -> Self { self.storage_slices = v; self } | ||||
|     pub fn status(mut self, v: ReservationStatus) -> Self { self.status = v; self } | ||||
|     pub fn obligation(mut self, v: bool) -> Self { self.obligation = v; self } | ||||
|     pub fn start_date(mut self, v: u32) -> Self { self.start_date = v; self } | ||||
|     pub fn end_date(mut self, v: u32) -> Self { self.end_date = v; self } | ||||
| } | ||||
| @@ -1,4 +1,4 @@ | ||||
| use heromodels_core::{Model, BaseModelData, IndexKey}; | ||||
| use heromodels_core::{BaseModelData, IndexKey, Model}; | ||||
| use heromodels_derive::model; | ||||
| use serde::{Deserialize, Serialize}; | ||||
| use std::collections::HashMap; | ||||
| @@ -209,10 +209,13 @@ pub struct DNSZone { | ||||
|     pub base_data: BaseModelData, | ||||
|     #[index] | ||||
|     pub domain: String, | ||||
|     #[index(path = "subdomain")] | ||||
|     #[index(path = "record_type")] | ||||
|     pub dnsrecords: Vec<DNSRecord>, | ||||
|     pub administrators: Vec<u32>, | ||||
|     pub status: DNSZoneStatus, | ||||
|     pub metadata: HashMap<String, String>, | ||||
|     #[index(path = "primary_ns")] | ||||
|     pub soarecord: Vec<SOARecord>, | ||||
| } | ||||
|  | ||||
| @@ -297,5 +300,3 @@ impl DNSZone { | ||||
|         self | ||||
|     } | ||||
| } | ||||
|  | ||||
|  | ||||
|   | ||||
| @@ -1,4 +1,4 @@ | ||||
| use heromodels_core::{Model, BaseModelData, IndexKey}; | ||||
| use heromodels_core::{BaseModelData, IndexKey, Model}; | ||||
| use heromodels_derive::model; | ||||
| use serde::{Deserialize, Serialize}; | ||||
|  | ||||
| @@ -184,8 +184,6 @@ impl Group { | ||||
|     } | ||||
| } | ||||
|  | ||||
|  | ||||
|  | ||||
| /// Represents the membership relationship between users and groups | ||||
| #[model] | ||||
| #[derive(Debug, Clone, Serialize, Deserialize, PartialEq, Default)] | ||||
| @@ -232,5 +230,3 @@ impl UserGroupMembership { | ||||
|         self | ||||
|     } | ||||
| } | ||||
|  | ||||
|  | ||||
|   | ||||
| @@ -1,4 +1,4 @@ | ||||
| use heromodels_core::{Model, BaseModelData, IndexKey}; | ||||
| use heromodels_core::{BaseModelData, IndexKey, Model}; | ||||
| use heromodels_derive::model; | ||||
| use serde::{Deserialize, Serialize}; | ||||
|  | ||||
| @@ -111,5 +111,3 @@ impl Member { | ||||
|         self | ||||
|     } | ||||
| } | ||||
|  | ||||
|  | ||||
|   | ||||
| @@ -1,20 +1,10 @@ | ||||
| // Export all heroledger model modules | ||||
| pub mod user; | ||||
| pub mod group; | ||||
| pub mod money; | ||||
| pub mod membership; | ||||
| pub mod dnsrecord; | ||||
| pub mod group; | ||||
| pub mod membership; | ||||
| pub mod money; | ||||
| pub mod rhai; | ||||
| pub mod secretbox; | ||||
| pub mod signature; | ||||
| pub mod user; | ||||
| pub mod user_kvs; | ||||
| pub mod rhai; | ||||
|  | ||||
| // Re-export key types for convenience | ||||
| pub use user::{User, UserStatus, UserProfile, KYCInfo, KYCStatus, SecretBox}; | ||||
| pub use group::{Group, UserGroupMembership, GroupStatus, Visibility, GroupConfig}; | ||||
| pub use money::{Account, Asset, AccountPolicy, AccountPolicyItem, Transaction, AccountStatus, TransactionType, Signature as TransactionSignature}; | ||||
| pub use membership::{Member, MemberRole, MemberStatus}; | ||||
| pub use dnsrecord::{DNSZone, DNSRecord, SOARecord, NameType, NameCat, DNSZoneStatus}; | ||||
| pub use secretbox::{Notary, NotaryStatus, SecretBoxCategory}; | ||||
| pub use signature::{Signature, SignatureStatus, ObjectType}; | ||||
| pub use user_kvs::{UserKVS, UserKVSItem}; | ||||
|   | ||||
| @@ -1,4 +1,4 @@ | ||||
| use heromodels_core::{Model, BaseModelData, IndexKey}; | ||||
| use heromodels_core::{BaseModelData, IndexKey, Model}; | ||||
| use heromodels_derive::model; | ||||
| use serde::{Deserialize, Serialize}; | ||||
| use std::collections::HashMap; | ||||
| @@ -223,8 +223,6 @@ impl Account { | ||||
|     } | ||||
| } | ||||
|  | ||||
|  | ||||
|  | ||||
| /// Represents an asset in the financial system | ||||
| #[model] | ||||
| #[derive(Debug, Clone, Serialize, Deserialize, PartialEq, Default)] | ||||
| @@ -342,8 +340,6 @@ impl Asset { | ||||
|     } | ||||
| } | ||||
|  | ||||
|  | ||||
|  | ||||
| /// Represents account policies for various operations | ||||
| #[model] | ||||
| #[derive(Debug, Clone, Serialize, Deserialize, PartialEq, Default)] | ||||
| @@ -400,8 +396,6 @@ impl AccountPolicy { | ||||
|     } | ||||
| } | ||||
|  | ||||
|  | ||||
|  | ||||
| /// Represents a financial transaction | ||||
| #[model] | ||||
| #[derive(Debug, Clone, Serialize, Deserialize, PartialEq, Default)] | ||||
| @@ -511,5 +505,3 @@ impl Transaction { | ||||
|         self | ||||
|     } | ||||
| } | ||||
|  | ||||
|  | ||||
|   | ||||
| @@ -1,8 +1,13 @@ | ||||
| use ::rhai::plugin::*; | ||||
| use ::rhai::{Array, Dynamic, Engine, EvalAltResult, Map, Module}; | ||||
| use ::rhai::{Dynamic, Engine, EvalAltResult, Module}; | ||||
| use std::mem; | ||||
|  | ||||
| use crate::models::heroledger::*; | ||||
| use crate::models::heroledger::{ | ||||
|     dnsrecord::DNSZone, | ||||
|     group::{Group, Visibility}, | ||||
|     money::Account, | ||||
|     user::{User, UserStatus}, | ||||
| }; | ||||
|  | ||||
| // ============================================================================ | ||||
| // User Module | ||||
| @@ -12,6 +17,8 @@ type RhaiUser = User; | ||||
|  | ||||
| #[export_module] | ||||
| mod rhai_user_module { | ||||
|     use crate::models::heroledger::user::User; | ||||
|  | ||||
|     use super::RhaiUser; | ||||
|  | ||||
|     #[rhai_fn(name = "new_user", return_raw)] | ||||
| @@ -30,30 +37,21 @@ mod rhai_user_module { | ||||
|     } | ||||
|  | ||||
|     #[rhai_fn(name = "add_email", return_raw)] | ||||
|     pub fn add_email( | ||||
|         user: &mut RhaiUser, | ||||
|         email: String, | ||||
|     ) -> Result<RhaiUser, Box<EvalAltResult>> { | ||||
|     pub fn add_email(user: &mut RhaiUser, email: String) -> Result<RhaiUser, Box<EvalAltResult>> { | ||||
|         let owned = std::mem::take(user); | ||||
|         *user = owned.add_email(email); | ||||
|         Ok(user.clone()) | ||||
|     } | ||||
|  | ||||
|     #[rhai_fn(name = "pubkey", return_raw)] | ||||
|     pub fn set_pubkey( | ||||
|         user: &mut RhaiUser, | ||||
|         pubkey: String, | ||||
|     ) -> Result<RhaiUser, Box<EvalAltResult>> { | ||||
|     pub fn set_pubkey(user: &mut RhaiUser, pubkey: String) -> Result<RhaiUser, Box<EvalAltResult>> { | ||||
|         let owned = std::mem::take(user); | ||||
|         *user = owned.pubkey(pubkey); | ||||
|         Ok(user.clone()) | ||||
|     } | ||||
|  | ||||
|     #[rhai_fn(name = "status", return_raw)] | ||||
|     pub fn set_status( | ||||
|         user: &mut RhaiUser, | ||||
|         status: String, | ||||
|     ) -> Result<RhaiUser, Box<EvalAltResult>> { | ||||
|     pub fn set_status(user: &mut RhaiUser, status: String) -> Result<RhaiUser, Box<EvalAltResult>> { | ||||
|         let status_enum = match status.as_str() { | ||||
|             "Active" => UserStatus::Active, | ||||
|             "Inactive" => UserStatus::Inactive, | ||||
| @@ -115,10 +113,7 @@ mod rhai_group_module { | ||||
|     } | ||||
|  | ||||
|     #[rhai_fn(name = "name", return_raw)] | ||||
|     pub fn set_name( | ||||
|         group: &mut RhaiGroup, | ||||
|         name: String, | ||||
|     ) -> Result<RhaiGroup, Box<EvalAltResult>> { | ||||
|     pub fn set_name(group: &mut RhaiGroup, name: String) -> Result<RhaiGroup, Box<EvalAltResult>> { | ||||
|         let owned = std::mem::take(group); | ||||
|         *group = owned.name(name); | ||||
|         Ok(group.clone()) | ||||
| @@ -263,15 +258,11 @@ mod rhai_dns_zone_module { | ||||
|         Ok(zone.clone()) | ||||
|     } | ||||
|  | ||||
|  | ||||
|  | ||||
|     #[rhai_fn(name = "save_dns_zone", return_raw)] | ||||
|     pub fn save_dns_zone(zone: &mut RhaiDNSZone) -> Result<RhaiDNSZone, Box<EvalAltResult>> { | ||||
|         Ok(zone.clone()) | ||||
|     } | ||||
|  | ||||
|  | ||||
|  | ||||
|     // Getters | ||||
|     #[rhai_fn(name = "get_id")] | ||||
|     pub fn get_id(zone: &mut RhaiDNSZone) -> i64 { | ||||
|   | ||||
| @@ -1,4 +1,4 @@ | ||||
| use heromodels_core::{Model, BaseModelData, IndexKey}; | ||||
| use heromodels_core::{BaseModelData, IndexKey, Model}; | ||||
| use heromodels_derive::model; | ||||
| use serde::{Deserialize, Serialize}; | ||||
|  | ||||
| @@ -138,5 +138,3 @@ impl Notary { | ||||
|         self | ||||
|     } | ||||
| } | ||||
|  | ||||
|  | ||||
|   | ||||
| @@ -1,4 +1,4 @@ | ||||
| use heromodels_core::{Model, BaseModelData, IndexKey}; | ||||
| use heromodels_core::{BaseModelData, IndexKey, Model}; | ||||
| use heromodels_derive::model; | ||||
| use serde::{Deserialize, Serialize}; | ||||
|  | ||||
| @@ -116,5 +116,3 @@ impl Signature { | ||||
|         self | ||||
|     } | ||||
| } | ||||
|  | ||||
|  | ||||
|   | ||||
| @@ -1,4 +1,4 @@ | ||||
| use heromodels_core::{Model, BaseModelData, IndexKey}; | ||||
| use heromodels_core::{BaseModelData, IndexKey, Model}; | ||||
| use heromodels_derive::model; | ||||
| use serde::{Deserialize, Serialize}; | ||||
| use std::collections::HashMap; | ||||
| @@ -366,5 +366,3 @@ impl User { | ||||
|         self | ||||
|     } | ||||
| } | ||||
|  | ||||
|  | ||||
|   | ||||
| @@ -1,7 +1,7 @@ | ||||
| use heromodels_core::{Model, BaseModelData, IndexKey}; | ||||
| use super::secretbox::SecretBox; | ||||
| use heromodels_core::{BaseModelData, IndexKey, Model}; | ||||
| use heromodels_derive::model; | ||||
| use serde::{Deserialize, Serialize}; | ||||
| use super::secretbox::SecretBox; | ||||
|  | ||||
| /// Represents a per-user key-value store | ||||
| #[model] | ||||
| @@ -44,8 +44,6 @@ impl UserKVS { | ||||
|     } | ||||
| } | ||||
|  | ||||
|  | ||||
|  | ||||
| /// Represents an item in a user's key-value store | ||||
| #[model] | ||||
| #[derive(Debug, Clone, Serialize, Deserialize, PartialEq, Default)] | ||||
| @@ -116,5 +114,3 @@ impl UserKVSItem { | ||||
|         self | ||||
|     } | ||||
| } | ||||
|  | ||||
|  | ||||
|   | ||||
| @@ -46,4 +46,4 @@ pub struct IdenfyVerificationData { | ||||
|     pub doc_issuing_country: Option<String>, | ||||
|     #[serde(rename = "manuallyDataChanged")] | ||||
|     pub manually_data_changed: Option<bool>, | ||||
| } | ||||
| } | ||||
|   | ||||
| @@ -2,4 +2,4 @@ | ||||
|  | ||||
| pub mod kyc; | ||||
|  | ||||
| pub use kyc::*; | ||||
| pub use kyc::*; | ||||
|   | ||||
| @@ -8,4 +8,4 @@ pub struct Address { | ||||
|     pub postal_code: String, | ||||
|     pub country: String, | ||||
|     pub company: Option<String>, | ||||
| } | ||||
| } | ||||
|   | ||||
| @@ -10,16 +10,16 @@ pub mod contact; | ||||
| pub mod finance; | ||||
| pub mod flow; | ||||
| pub mod governance; | ||||
| pub mod grid4; | ||||
| pub mod heroledger; | ||||
| pub mod identity; | ||||
| pub mod legal; | ||||
| pub mod library; | ||||
| pub mod location; | ||||
| pub mod object; | ||||
| pub mod projects; | ||||
| pub mod payment; | ||||
| pub mod identity; | ||||
| pub mod tfmarketplace; | ||||
| pub mod grid4; | ||||
| pub mod projects; | ||||
| // pub mod tfmarketplace; | ||||
|  | ||||
| // Re-export key types for convenience | ||||
| pub use core::Comment; | ||||
| @@ -39,3 +39,4 @@ pub use legal::{Contract, ContractRevision, ContractSigner, ContractStatus, Sign | ||||
| pub use library::collection::Collection; | ||||
| pub use library::items::{Image, Markdown, Pdf}; | ||||
| pub use projects::{Project, Status}; | ||||
| pub use heroledger::*; | ||||
| @@ -1,6 +1,6 @@ | ||||
| use super::Object; | ||||
| use rhai::plugin::*; | ||||
| use rhai::{CustomType, Dynamic, Engine, EvalAltResult, Module}; | ||||
| use super::Object; | ||||
|  | ||||
| type RhaiObject = Object; | ||||
|  | ||||
| @@ -16,10 +16,7 @@ pub mod generated_rhai_module { | ||||
|  | ||||
|     /// Set the title of an Object | ||||
|     #[rhai_fn(name = "object_title")] | ||||
|     pub fn object_title( | ||||
|         object: &mut RhaiObject, | ||||
|         title: String, | ||||
|     ) -> RhaiObject { | ||||
|     pub fn object_title(object: &mut RhaiObject, title: String) -> RhaiObject { | ||||
|         let mut result = object.clone(); | ||||
|         result.title = title; | ||||
|         result | ||||
| @@ -27,10 +24,7 @@ pub mod generated_rhai_module { | ||||
|  | ||||
|     /// Set the description of an Object | ||||
|     #[rhai_fn(name = "object_description")] | ||||
|     pub fn object_description( | ||||
|         object: &mut RhaiObject, | ||||
|         description: String, | ||||
|     ) -> RhaiObject { | ||||
|     pub fn object_description(object: &mut RhaiObject, description: String) -> RhaiObject { | ||||
|         let mut result = object.clone(); | ||||
|         result.description = description; | ||||
|         result | ||||
|   | ||||
| @@ -2,4 +2,4 @@ | ||||
|  | ||||
| pub mod stripe; | ||||
|  | ||||
| pub use stripe::*; | ||||
| pub use stripe::*; | ||||
|   | ||||
| @@ -27,4 +27,4 @@ pub struct StripeEventData { | ||||
| pub struct StripeEventRequest { | ||||
|     pub id: Option<String>, | ||||
|     pub idempotency_key: Option<String>, | ||||
| } | ||||
| } | ||||
|   | ||||
							
								
								
									
										43
									
								
								heromodels/test.sh
									
									
									
									
									
										Executable file
									
								
							
							
						
						
									
										43
									
								
								heromodels/test.sh
									
									
									
									
									
										Executable file
									
								
							| @@ -0,0 +1,43 @@ | ||||
| #!/usr/bin/env bash | ||||
| set -euo pipefail | ||||
|  | ||||
| # Config matches examples/tests | ||||
| PGHOST=${PGHOST:-localhost} | ||||
| PGPORT=${PGPORT:-5432} | ||||
| PGUSER=${PGUSER:-postgres} | ||||
| PGPASSWORD=${PGPASSWORD:-test123} | ||||
| export PGPASSWORD | ||||
|  | ||||
| echo "[test.sh] Checking Postgres at ${PGHOST}:${PGPORT} (user=${PGUSER})..." | ||||
|  | ||||
| # Require pg_isready | ||||
| if ! command -v pg_isready >/dev/null 2>&1; then | ||||
|   echo "[test.sh] ERROR: pg_isready not found. Install PostgreSQL client tools (e.g., brew install libpq && brew link --force libpq)." >&2 | ||||
|   exit 1 | ||||
| fi | ||||
|  | ||||
| # Wait for Postgres to be ready (30s timeout) | ||||
| ATTEMPTS=30 | ||||
| until pg_isready -h "$PGHOST" -p "$PGPORT" -U "$PGUSER" >/dev/null 2>&1; do | ||||
|   ((ATTEMPTS--)) || { | ||||
|     echo "[test.sh] ERROR: Postgres not ready after 30s. Ensure it's running with user=$PGUSER password=$PGPASSWORD host=$PGHOST port=$PGPORT." >&2 | ||||
|     exit 1 | ||||
|   } | ||||
|   sleep 1 | ||||
|   echo "[test.sh] Waiting for Postgres..." | ||||
| done | ||||
|  | ||||
| echo "[test.sh] Postgres is ready. Running tests..." | ||||
|  | ||||
| # Run fast OurDB test first (no Postgres dependency) | ||||
| echo "[test.sh] Running OurDB test: grid4_ourdb" | ||||
| cargo test -p heromodels --test grid4_ourdb | ||||
|  | ||||
| # Run Postgres-backed tests (marked ignored) | ||||
| echo "[test.sh] Running Postgres test: heroledger_postgres (ignored)" | ||||
| cargo test -p heromodels --test heroledger_postgres -- --ignored | ||||
|  | ||||
| echo "[test.sh] Running Postgres test: grid4_postgres (ignored)" | ||||
| cargo test -p heromodels --test grid4_postgres -- --ignored | ||||
|  | ||||
| echo "[test.sh] Done." | ||||
							
								
								
									
										117
									
								
								heromodels/tests/grid4_models.rs
									
									
									
									
									
										Normal file
									
								
							
							
						
						
									
										117
									
								
								heromodels/tests/grid4_models.rs
									
									
									
									
									
										Normal file
									
								
							| @@ -0,0 +1,117 @@ | ||||
| use serde_json; | ||||
|  | ||||
| use heromodels::models::grid4::{ | ||||
|     ComputeSlice, DeviceInfo, Node, NodeCapacity, PricingPolicy, Reservation, ReservationStatus, | ||||
|     SLAPolicy, StorageDevice, StorageSlice, | ||||
| }; | ||||
|  | ||||
| #[test] | ||||
| fn build_and_serde_roundtrip_compute_storage_slices() { | ||||
|     let pricing = PricingPolicy::new() | ||||
|         .marketplace_year_discounts(vec![20, 30, 40]) | ||||
|         .volume_discounts(vec![5, 10, 15]) | ||||
|         .build(); | ||||
|  | ||||
|     let sla = SLAPolicy::new() | ||||
|         .sla_uptime(99) | ||||
|         .sla_bandwidth_mbit(1000) | ||||
|         .sla_penalty(150) | ||||
|         .build(); | ||||
|  | ||||
|     let cs = ComputeSlice::new() | ||||
|         .nodeid(42) | ||||
|         .slice_id(1) | ||||
|         .mem_gb(16.0) | ||||
|         .storage_gb(200.0) | ||||
|         .passmark(5000) | ||||
|         .vcores(8) | ||||
|         .cpu_oversubscription(2) | ||||
|         .storage_oversubscription(1) | ||||
|         .price_range(vec![0.5, 2.0]) | ||||
|         .gpus(1) | ||||
|         .price_cc(1.25) | ||||
|         .pricing_policy(pricing.clone()) | ||||
|         .sla_policy(sla.clone()); | ||||
|  | ||||
|     let ss = StorageSlice::new() | ||||
|         .nodeid(42) | ||||
|         .slice_id(2) | ||||
|         .price_cc(0.15) | ||||
|         .pricing_policy(pricing) | ||||
|         .sla_policy(sla); | ||||
|  | ||||
|     // serde roundtrip compute slice | ||||
|     let s = serde_json::to_string(&cs).expect("serialize compute slice"); | ||||
|     let cs2: ComputeSlice = serde_json::from_str(&s).expect("deserialize compute slice"); | ||||
|     assert_eq!(cs, cs2); | ||||
|  | ||||
|     // serde roundtrip storage slice | ||||
|     let s2 = serde_json::to_string(&ss).expect("serialize storage slice"); | ||||
|     let ss2: StorageSlice = serde_json::from_str(&s2).expect("deserialize storage slice"); | ||||
|     assert_eq!(ss, ss2); | ||||
| } | ||||
|  | ||||
| #[test] | ||||
| fn build_and_serde_roundtrip_node() { | ||||
|     let dev = DeviceInfo { | ||||
|         vendor: "AcmeVendor".into(), | ||||
|         storage: vec![StorageDevice { id: "sda".into(), size_gb: 512.0, description: "NVMe".into() }], | ||||
|         memory: vec![], | ||||
|         cpu: vec![], | ||||
|         gpu: vec![], | ||||
|         network: vec![], | ||||
|     }; | ||||
|  | ||||
|     let cap = NodeCapacity { storage_gb: 2048.0, mem_gb: 128.0, mem_gb_gpu: 24.0, passmark: 12000, vcores: 32 }; | ||||
|  | ||||
|     let cs = ComputeSlice::new().nodeid(1).slice_id(1).mem_gb(8.0).storage_gb(100.0).passmark(2500).vcores(4); | ||||
|     let ss = StorageSlice::new().nodeid(1).slice_id(2).price_cc(0.2); | ||||
|  | ||||
|     let node = Node::new() | ||||
|         .nodegroupid(7) | ||||
|         .uptime(99) | ||||
|         .add_compute_slice(cs) | ||||
|         .add_storage_slice(ss) | ||||
|         .devices(dev) | ||||
|         .country("NL") | ||||
|         .capacity(cap) | ||||
|         .provisiontime(1710000000) | ||||
|         .pubkey("node_pubkey") | ||||
|         .signature_node("sig_node") | ||||
|         .signature_farmer("sig_farmer"); | ||||
|  | ||||
|     let s = serde_json::to_string(&node).expect("serialize node"); | ||||
|     let node2: Node = serde_json::from_str(&s).expect("deserialize node"); | ||||
|  | ||||
|     assert_eq!(node.nodegroupid, node2.nodegroupid); | ||||
|     assert_eq!(node.uptime, node2.uptime); | ||||
|     assert_eq!(node.country, node2.country); | ||||
|     assert_eq!(node.pubkey, node2.pubkey); | ||||
|     assert_eq!(node.signature_node, node2.signature_node); | ||||
|     assert_eq!(node.signature_farmer, node2.signature_farmer); | ||||
|     assert_eq!(node.computeslices.len(), node2.computeslices.len()); | ||||
|     assert_eq!(node.storageslices.len(), node2.storageslices.len()); | ||||
| } | ||||
|  | ||||
| #[test] | ||||
| fn build_and_serde_roundtrip_reservation() { | ||||
|     let reservation = Reservation::new() | ||||
|         .customer_id(1234) | ||||
|         .add_compute_slice(11) | ||||
|         .add_storage_slice(22) | ||||
|         .status(ReservationStatus::Confirmed) | ||||
|         .obligation(true) | ||||
|         .start_date(1_710_000_000) | ||||
|         .end_date(1_720_000_000); | ||||
|  | ||||
|     let s = serde_json::to_string(&reservation).expect("serialize reservation"); | ||||
|     let reservation2: Reservation = serde_json::from_str(&s).expect("deserialize reservation"); | ||||
|  | ||||
|     assert_eq!(reservation.customer_id, reservation2.customer_id); | ||||
|     assert_eq!(reservation.status, reservation2.status); | ||||
|     assert_eq!(reservation.obligation, reservation2.obligation); | ||||
|     assert_eq!(reservation.start_date, reservation2.start_date); | ||||
|     assert_eq!(reservation.end_date, reservation2.end_date); | ||||
|     assert_eq!(reservation.compute_slices, reservation2.compute_slices); | ||||
|     assert_eq!(reservation.storage_slices, reservation2.storage_slices); | ||||
| } | ||||
							
								
								
									
										82
									
								
								heromodels/tests/grid4_ourdb.rs
									
									
									
									
									
										Normal file
									
								
							
							
						
						
									
										82
									
								
								heromodels/tests/grid4_ourdb.rs
									
									
									
									
									
										Normal file
									
								
							| @@ -0,0 +1,82 @@ | ||||
| use heromodels::db::hero::OurDB; | ||||
| use heromodels::db::{Collection, Db}; | ||||
| use heromodels::models::grid4::node::node_index::{country, nodegroupid, pubkey}; | ||||
| use heromodels::models::grid4::node::{ComputeSlice, DeviceInfo, Node}; | ||||
| use heromodels_core::Model; | ||||
| use std::sync::Arc; | ||||
|  | ||||
| fn create_test_db() -> Arc<OurDB> { | ||||
|     let ts = std::time::SystemTime::now() | ||||
|         .duration_since(std::time::UNIX_EPOCH) | ||||
|         .unwrap() | ||||
|         .as_nanos(); | ||||
|     let path = format!("/tmp/grid4_node_test_{}", ts); | ||||
|     let _ = std::fs::remove_dir_all(&path); | ||||
|     Arc::new(OurDB::new(path, true).expect("create OurDB")) | ||||
| } | ||||
|  | ||||
| #[test] | ||||
| fn grid4_node_basic_roundtrip_and_indexes() { | ||||
|     let db = create_test_db(); | ||||
|     let nodes = db.collection::<Node>().expect("open node collection"); | ||||
|  | ||||
|     // Clean any leftover | ||||
|     if let Ok(existing) = nodes.get_all() { | ||||
|         for n in existing { | ||||
|             let _ = nodes.delete_by_id(n.get_id()); | ||||
|         } | ||||
|     } | ||||
|  | ||||
|     // Build a node with some compute slices and device info | ||||
|     let cs = ComputeSlice::new() | ||||
|         .nodeid(1) | ||||
|         .slice_id(1) | ||||
|         .mem_gb(32.0) | ||||
|         .storage_gb(512.0) | ||||
|         .passmark(5000) | ||||
|         .vcores(16) | ||||
|         .gpus(1) | ||||
|         .price_cc(0.25); | ||||
|  | ||||
|     let dev = DeviceInfo { | ||||
|         vendor: "ACME".into(), | ||||
|         ..Default::default() | ||||
|     }; | ||||
|  | ||||
|     let n = Node::new() | ||||
|         .nodegroupid(42) | ||||
|         .uptime(99) | ||||
|         .add_compute_slice(cs) | ||||
|         .devices(dev) | ||||
|         .country("BE") | ||||
|         .pubkey("PUB_NODE_1") | ||||
|         .build(); | ||||
|  | ||||
|     let (id, stored) = nodes.set(&n).expect("store node"); | ||||
|     assert!(id > 0); | ||||
|     assert_eq!(stored.country, "BE"); | ||||
|  | ||||
|     // get by id | ||||
|     let fetched = nodes.get_by_id(id).expect("get by id").expect("exists"); | ||||
|     assert_eq!(fetched.pubkey, "PUB_NODE_1"); | ||||
|  | ||||
|     // query by top-level indexes | ||||
|     let by_country = nodes.get::<country, _>("BE").expect("query country"); | ||||
|     assert_eq!(by_country.len(), 1); | ||||
|     assert_eq!(by_country[0].get_id(), id); | ||||
|  | ||||
|     let by_group = nodes.get::<nodegroupid, _>(&42).expect("query group"); | ||||
|     assert_eq!(by_group.len(), 1); | ||||
|  | ||||
|     let by_pubkey = nodes.get::<pubkey, _>("PUB_NODE_1").expect("query pubkey"); | ||||
|     assert_eq!(by_pubkey.len(), 1); | ||||
|  | ||||
|     // update | ||||
|     let updated = fetched.clone().country("NL"); | ||||
|     let (_, back) = nodes.set(&updated).expect("update node"); | ||||
|     assert_eq!(back.country, "NL"); | ||||
|  | ||||
|     // delete | ||||
|     nodes.delete_by_id(id).expect("delete"); | ||||
|     assert!(nodes.get_by_id(id).expect("get after delete").is_none()); | ||||
| } | ||||
							
								
								
									
										125
									
								
								heromodels/tests/grid4_postgres.rs
									
									
									
									
									
										Normal file
									
								
							
							
						
						
									
										125
									
								
								heromodels/tests/grid4_postgres.rs
									
									
									
									
									
										Normal file
									
								
							| @@ -0,0 +1,125 @@ | ||||
| use heromodels::db::postgres::{Config, Postgres}; | ||||
| use heromodels::db::{Collection, Db}; | ||||
| use heromodels::models::grid4::node::node_index::{country, nodegroupid, pubkey}; | ||||
| use heromodels::models::grid4::node::{ComputeSlice, DeviceInfo, Node}; | ||||
| use heromodels_core::Model; | ||||
|  | ||||
| // Requires local Postgres (user=postgres password=test123 host=localhost port=5432) | ||||
| // Run with: cargo test -p heromodels --test grid4_postgres -- --ignored | ||||
| #[test] | ||||
| #[ignore] | ||||
| fn grid4_node_postgres_roundtrip_like_example() { | ||||
|     let db = Postgres::new( | ||||
|         Config::new() | ||||
|             .user(Some("postgres".into())) | ||||
|             .password(Some("test123".into())) | ||||
|             .host(Some("localhost".into())) | ||||
|             .port(Some(5432)), | ||||
|     ) | ||||
|     .expect("can connect to Postgres"); | ||||
|  | ||||
|     let nodes = db.collection::<Node>().expect("open node collection"); | ||||
|  | ||||
|     // Clean existing | ||||
|     if let Ok(existing) = nodes.get_all() { | ||||
|         for n in existing { | ||||
|             let _ = nodes.delete_by_id(n.get_id()); | ||||
|         } | ||||
|     } | ||||
|  | ||||
|     // Build and store multiple nodes via builder and then persist via collection.set(), like examples | ||||
|     let cs1 = ComputeSlice::new() | ||||
|         .nodeid(10) | ||||
|         .slice_id(1) | ||||
|         .mem_gb(32.0) | ||||
|         .storage_gb(512.0) | ||||
|         .passmark(5000) | ||||
|         .vcores(16) | ||||
|         .gpus(1) | ||||
|         .price_cc(0.25); | ||||
|     let cs2 = ComputeSlice::new() | ||||
|         .nodeid(10) | ||||
|         .slice_id(2) | ||||
|         .mem_gb(64.0) | ||||
|         .storage_gb(2048.0) | ||||
|         .passmark(7000) | ||||
|         .vcores(24) | ||||
|         .gpus(2) | ||||
|         .price_cc(0.50); | ||||
|     let cs3 = ComputeSlice::new() | ||||
|         .nodeid(11) | ||||
|         .slice_id(1) | ||||
|         .mem_gb(16.0) | ||||
|         .storage_gb(256.0) | ||||
|         .passmark(3000) | ||||
|         .vcores(8) | ||||
|         .gpus(0) | ||||
|         .price_cc(0.10); | ||||
|  | ||||
|     let dev = DeviceInfo { vendor: "ACME".into(), ..Default::default() }; | ||||
|  | ||||
|     let n1 = Node::new() | ||||
|         .nodegroupid(99) | ||||
|         .uptime(97) | ||||
|         .add_compute_slice(cs1) | ||||
|         .devices(dev.clone()) | ||||
|         .country("BE") | ||||
|         .pubkey("PG_NODE_1") | ||||
|         .build(); | ||||
|     let n2 = Node::new() | ||||
|         .nodegroupid(99) | ||||
|         .uptime(96) | ||||
|         .add_compute_slice(cs2) | ||||
|         .devices(dev.clone()) | ||||
|         .country("NL") | ||||
|         .pubkey("PG_NODE_2") | ||||
|         .build(); | ||||
|     let n3 = Node::new() | ||||
|         .nodegroupid(7) | ||||
|         .uptime(95) | ||||
|         .add_compute_slice(cs3) | ||||
|         .devices(dev) | ||||
|         .country("BE") | ||||
|         .pubkey("PG_NODE_3") | ||||
|         .build(); | ||||
|  | ||||
|     let (id1, s1) = nodes.set(&n1).expect("store n1"); | ||||
|     let (id2, s2) = nodes.set(&n2).expect("store n2"); | ||||
|     let (id3, s3) = nodes.set(&n3).expect("store n3"); | ||||
|     assert!(id1 > 0 && id2 > 0 && id3 > 0); | ||||
|  | ||||
|     // Query by top-level indexes similar to the example style | ||||
|     let be_nodes = nodes.get::<country, _>("BE").expect("by country"); | ||||
|     assert_eq!(be_nodes.len(), 2); | ||||
|  | ||||
|     let grp_99 = nodes.get::<nodegroupid, _>(&99).expect("by group"); | ||||
|     assert_eq!(grp_99.len(), 2); | ||||
|  | ||||
|     let by_key = nodes.get::<pubkey, _>("PG_NODE_2").expect("by pubkey"); | ||||
|     assert_eq!(by_key.len(), 1); | ||||
|     assert_eq!(by_key[0].get_id(), id2); | ||||
|  | ||||
|     // Update: change country of n1 | ||||
|     let updated = s1.clone().country("DE"); | ||||
|     let (_, back) = nodes.set(&updated).expect("update n1"); | ||||
|     assert_eq!(back.country, "DE"); | ||||
|  | ||||
|     // Cardinality after update | ||||
|     let de_nodes = nodes.get::<country, _>("DE").expect("by country DE"); | ||||
|     assert_eq!(de_nodes.len(), 1); | ||||
|  | ||||
|     // Delete by id and by index | ||||
|     nodes.delete_by_id(id2).expect("delete n2 by id"); | ||||
|     assert!(nodes.get_by_id(id2).unwrap().is_none()); | ||||
|  | ||||
|     nodes.delete::<pubkey, _>("PG_NODE_3").expect("delete n3 by pubkey"); | ||||
|     assert!(nodes.get_by_id(id3).unwrap().is_none()); | ||||
|  | ||||
|     // Remaining should be updated n1 only; verify via targeted queries | ||||
|     let de_nodes = nodes.get::<country, _>("DE").expect("country DE after deletes"); | ||||
|     assert_eq!(de_nodes.len(), 1); | ||||
|     assert_eq!(de_nodes[0].get_id(), id1); | ||||
|     let by_key = nodes.get::<pubkey, _>("PG_NODE_1").expect("by pubkey PG_NODE_1"); | ||||
|     assert_eq!(by_key.len(), 1); | ||||
|     assert_eq!(by_key[0].get_id(), id1); | ||||
| } | ||||
							
								
								
									
										97
									
								
								heromodels/tests/heroledger_postgres.rs
									
									
									
									
									
										Normal file
									
								
							
							
						
						
									
										97
									
								
								heromodels/tests/heroledger_postgres.rs
									
									
									
									
									
										Normal file
									
								
							| @@ -0,0 +1,97 @@ | ||||
| use heromodels::db::postgres::{Config, Postgres}; | ||||
| use heromodels::db::{Collection, Db}; | ||||
| use heromodels::models::heroledger::user::user_index::username; | ||||
| use heromodels::models::heroledger::user::User; | ||||
| use heromodels_core::Model; | ||||
|  | ||||
| // NOTE: Requires a local Postgres running with user=postgres password=test123 host=localhost port=5432 | ||||
| // Marked ignored by default. Run with: cargo test -p heromodels --test heroledger_postgres -- --ignored | ||||
| #[test] | ||||
| #[ignore] | ||||
| fn heroledger_user_postgres_roundtrip() { | ||||
|     // Connect | ||||
|     let db = Postgres::new( | ||||
|         Config::new() | ||||
|             .user(Some("postgres".into())) | ||||
|             .password(Some("test123".into())) | ||||
|             .host(Some("localhost".into())) | ||||
|             .port(Some(5432)), | ||||
|     ) | ||||
|     .expect("can connect to Postgres"); | ||||
|  | ||||
|     // Open collection (will create table and indexes for top-level fields) | ||||
|     let users = db.collection::<User>().expect("can open user collection"); | ||||
|  | ||||
|     // Clean slate | ||||
|     if let Ok(existing) = users.get_all() { | ||||
|         for u in existing { | ||||
|             let _ = users.delete_by_id(u.get_id()); | ||||
|         } | ||||
|     } | ||||
|  | ||||
|     // Unique suffix to avoid collisions with any pre-existing rows | ||||
|     let uniq = format!("{}", std::time::SystemTime::now() | ||||
|         .duration_since(std::time::UNIX_EPOCH) | ||||
|         .unwrap() | ||||
|         .as_nanos()); | ||||
|     let alice = format!("alice_{}", uniq); | ||||
|     let bob = format!("bob_{}", uniq); | ||||
|     let carol = format!("carol_{}", uniq); | ||||
|  | ||||
|     // Build and store multiple users | ||||
|     let u1 = User::new(0) | ||||
|         .username(&alice) | ||||
|         .pubkey("PUBKEY_A") | ||||
|         .add_email("alice@example.com") | ||||
|         .build(); | ||||
|     let u2 = User::new(0) | ||||
|         .username(&bob) | ||||
|         .pubkey("PUBKEY_B") | ||||
|         .add_email("bob@example.com") | ||||
|         .build(); | ||||
|     let u3 = User::new(0) | ||||
|         .username(&carol) | ||||
|         .pubkey("PUBKEY_C") | ||||
|         .add_email("carol@example.com") | ||||
|         .build(); | ||||
|  | ||||
|     let (id1, db_u1) = users.set(&u1).expect("store u1"); | ||||
|     let (id2, db_u2) = users.set(&u2).expect("store u2"); | ||||
|     let (id3, db_u3) = users.set(&u3).expect("store u3"); | ||||
|     assert!(id1 > 0 && id2 > 0 && id3 > 0); | ||||
|  | ||||
|     // Fetch by id | ||||
|     assert_eq!(users.get_by_id(id1).unwrap().unwrap().username, alice); | ||||
|     assert_eq!(users.get_by_id(id2).unwrap().unwrap().username, bob); | ||||
|     assert_eq!(users.get_by_id(id3).unwrap().unwrap().username, carol); | ||||
|  | ||||
|     // Fetch by index (top-level username) | ||||
|     let by_username = users.get::<username, _>(&alice).expect("by username"); | ||||
|     assert_eq!(by_username.len(), 1); | ||||
|     assert_eq!(by_username[0].get_id(), id1); | ||||
|  | ||||
|     // Update one | ||||
|     let updated = db_u1.clone().add_email("work@alice.example"); | ||||
|     let (id1b, updated_back) = users.set(&updated).expect("update alice"); | ||||
|     assert_eq!(id1b, id1); | ||||
|     assert!(updated_back.email.len() >= 2); | ||||
|  | ||||
|     // Targeted queries to avoid legacy rows in the same table | ||||
|     // Verify three users exist via index queries | ||||
|     assert_eq!(users.get::<username, _>(&alice).unwrap().len(), 1); | ||||
|     assert_eq!(users.get::<username, _>(&bob).unwrap().len(), 1); | ||||
|     assert_eq!(users.get::<username, _>(&carol).unwrap().len(), 1); | ||||
|  | ||||
|     // Delete by id | ||||
|     users.delete_by_id(id2).expect("delete bob by id"); | ||||
|     assert!(users.get_by_id(id2).unwrap().is_none()); | ||||
|  | ||||
|     // Delete by index (username) | ||||
|     users.delete::<username, _>(&carol).expect("delete carol by username"); | ||||
|     assert!(users.get_by_id(id3).unwrap().is_none()); | ||||
|  | ||||
|     // Remaining should be just alice; verify via index | ||||
|     let remain = users.get::<username, _>(&alice).expect("get alice after delete"); | ||||
|     assert_eq!(remain.len(), 1); | ||||
|     assert_eq!(remain[0].get_id(), id1); | ||||
| } | ||||
| @@ -1,4 +1,5 @@ | ||||
| use heromodels::db::Collection; | ||||
| use heromodels::db::Db; | ||||
| use heromodels::db::hero::OurDB; | ||||
| use heromodels::models::biz::{BusinessType, Company, CompanyStatus, Payment, PaymentStatus}; | ||||
| use heromodels_core::Model; | ||||
| @@ -197,12 +198,18 @@ fn test_payment_database_persistence() { | ||||
|     ); | ||||
|  | ||||
|     // Save payment | ||||
|     let (payment_id, saved_payment) = db.set(&payment).expect("Failed to save payment"); | ||||
|     let (payment_id, saved_payment) = db | ||||
|         .collection::<Payment>() | ||||
|         .expect("open payment collection") | ||||
|         .set(&payment) | ||||
|         .expect("Failed to save payment"); | ||||
|     assert!(payment_id > 0); | ||||
|     assert_eq!(saved_payment.payment_intent_id, "pi_db_test"); | ||||
|  | ||||
|     // Retrieve payment | ||||
|     let retrieved_payment: Payment = db | ||||
|         .collection::<Payment>() | ||||
|         .expect("open payment collection") | ||||
|         .get_by_id(payment_id) | ||||
|         .expect("Failed to get payment") | ||||
|         .unwrap(); | ||||
| @@ -224,20 +231,34 @@ fn test_payment_status_transitions() { | ||||
|         1360.0, | ||||
|     ); | ||||
|  | ||||
|     let (payment_id, mut payment) = db.set(&payment).expect("Failed to save payment"); | ||||
|     let (payment_id, mut payment) = db | ||||
|         .collection::<Payment>() | ||||
|         .expect("open payment collection") | ||||
|         .set(&payment) | ||||
|         .expect("Failed to save payment"); | ||||
|  | ||||
|     // Test pending -> completed | ||||
|     payment = payment.complete_payment(Some("cus_transition_test".to_string())); | ||||
|     let (_, mut payment) = db.set(&payment).expect("Failed to update payment"); | ||||
|     let (_, mut payment) = db | ||||
|         .collection::<Payment>() | ||||
|         .expect("open payment collection") | ||||
|         .set(&payment) | ||||
|         .expect("Failed to update payment"); | ||||
|     assert!(payment.is_completed()); | ||||
|  | ||||
|     // Test completed -> refunded | ||||
|     payment = payment.refund_payment(); | ||||
|     let (_, payment) = db.set(&payment).expect("Failed to update payment"); | ||||
|     let (_, payment) = db | ||||
|         .collection::<Payment>() | ||||
|         .expect("open payment collection") | ||||
|         .set(&payment) | ||||
|         .expect("Failed to update payment"); | ||||
|     assert!(payment.is_refunded()); | ||||
|  | ||||
|     // Verify final state in database | ||||
|     let final_payment: Payment = db | ||||
|         .collection::<Payment>() | ||||
|         .expect("open payment collection") | ||||
|         .get_by_id(payment_id) | ||||
|         .expect("Failed to get payment") | ||||
|         .unwrap(); | ||||
| @@ -270,15 +291,18 @@ fn test_company_payment_integration() { | ||||
|     let db = create_test_db(); | ||||
|  | ||||
|     // Create company with default PendingPayment status | ||||
|     let company = Company::new( | ||||
|         "Integration Test Corp".to_string(), | ||||
|         "ITC-001".to_string(), | ||||
|         chrono::Utc::now().timestamp(), | ||||
|     ) | ||||
|     .email("test@integration.com".to_string()) | ||||
|     .business_type(BusinessType::Starter); | ||||
|     let company = Company::new() | ||||
|         .name("Integration Test Corp") | ||||
|         .registration_number("ITC-001") | ||||
|         .incorporation_date(chrono::Utc::now().timestamp()) | ||||
|         .email("test@integration.com") | ||||
|         .business_type(BusinessType::Starter); | ||||
|  | ||||
|     let (company_id, company) = db.set(&company).expect("Failed to save company"); | ||||
|     let (company_id, company) = db | ||||
|         .collection::<Company>() | ||||
|         .expect("open company collection") | ||||
|         .set(&company) | ||||
|         .expect("Failed to save company"); | ||||
|     assert_eq!(company.status, CompanyStatus::PendingPayment); | ||||
|  | ||||
|     // Create payment for the company | ||||
| @@ -291,18 +315,28 @@ fn test_company_payment_integration() { | ||||
|         305.0, | ||||
|     ); | ||||
|  | ||||
|     let (_payment_id, payment) = db.set(&payment).expect("Failed to save payment"); | ||||
|     let (_payment_id, payment) = db | ||||
|         .collection::<Payment>() | ||||
|         .expect("open payment collection") | ||||
|         .set(&payment) | ||||
|         .expect("Failed to save payment"); | ||||
|     assert_eq!(payment.company_id, company_id); | ||||
|  | ||||
|     // Complete payment | ||||
|     let completed_payment = payment.complete_payment(Some("cus_integration_test".to_string())); | ||||
|     let (_, completed_payment) = db | ||||
|         .collection::<Payment>() | ||||
|         .expect("open payment collection") | ||||
|         .set(&completed_payment) | ||||
|         .expect("Failed to update payment"); | ||||
|  | ||||
|     // Update company status to Active | ||||
|     let active_company = company.status(CompanyStatus::Active); | ||||
|     let (_, active_company) = db.set(&active_company).expect("Failed to update company"); | ||||
|     let (_, active_company) = db | ||||
|         .collection::<Company>() | ||||
|         .expect("open company collection") | ||||
|         .set(&active_company) | ||||
|         .expect("Failed to update company"); | ||||
|  | ||||
|     // Verify final states | ||||
|     assert!(completed_payment.is_completed()); | ||||
|   | ||||
		Reference in New Issue
	
	Block a user