Merge branch 'development' of git.ourworld.tf:tfgrid/info_tfgrid into development

* 'development' of git.ourworld.tf:tfgrid/info_tfgrid:
  ...
  ...
  export for mdbook
  s
  s
  fixed
  finalized info_fgrid sync
  updated sync with info_grid
  troubleshootinfg farming
  update
  updated faq, tft
  updated 0133
  updated 0133
  Update collections/projectinca/stories/stories.md
  collection update
This commit is contained in:
despiegk 2024-09-19 06:32:55 +04:00
commit c91950cde5
229 changed files with 4458 additions and 1947 deletions

View File

@ -10,29 +10,35 @@
- [Pricing Calculator](dashboard/deploy/pricing_calculator.md)
- [Node Finder](dashboard/deploy/node_finder.md)
- [Virtual Machines](dashboard/solutions/vm_intro.md)
- [Micro and Full VM Differences ](dashboard/solutions/vm_differences.md)
- [Full Virtual Machine](dashboard/solutions/fullvm.md)
- [Micro Virtual Machine](dashboard/solutions/microvm.md)
- [Nixos MicroVM](dashboard/solutions/nixos_micro.md)
- [Add a Domain](dashboard/solutions/add_domain.md)
- [Micro and Full VM Differences ](dashboard/solutions/vm_differences.md)
- [Add a Domain to a VM](dashboard/solutions/add_domain.md)
- [Orchestrators](dashboard/deploy/orchestrators.md)
- [Kubernetes](dashboard/solutions/k8s.md)
- [Caprover](dashboard/solutions/caprover.md)
- [Caprover Admin](dashboard/solutions/caprover_admin.md)
- [Caprover Worker](dashboard/solutions/caprover_worker.md)
- [CapRover](dashboard/solutions/caprover.md)
- [CapRover Admin](dashboard/solutions/caprover_admin.md)
- [CapRover Worker](dashboard/solutions/caprover_worker.md)
- [Applications](dashboard/deploy/applications.md)
- [Algorand](dashboard/solutions/algorand.md)
- [CasperLabs](dashboard/solutions/casper.md)
- [Discourse](dashboard/solutions/discourse.md)
- [Domains](dashboard/solutions/domains.md)
- [Funkwhale](dashboard/solutions/funkwhale.md)
- [Gitea](dashboard/solutions/gitea.md)
- [Jenkins](dashboard/solutions/jenkins.md)
- [Jitsi](dashboard/solutions/jitsi.md)
- [Mattermost](dashboard/solutions/mattermost.md)
- [Nextcloud](dashboard/solutions/nextcloud.md)
- [Node Pilot](dashboard/solutions/nodepilot.md)
- [Nostr](dashboard/solutions/nostr.md)
- [Peertube](dashboard/solutions/peertube.md)
- [Presearch](dashboard/solutions/presearch.md)
- [Static Website](dashboard/solutions/static_website.md)
- [Subsquid](dashboard/solutions/subsquid.md)
- [Taiga](dashboard/solutions/taiga.md)
- [TFRobot](dashboard/solutions/tfrobot.md)
- [Umbrel](dashboard/solutions/umbrel.md)
- [WordPress](dashboard/solutions/wordpress.md)
- [Your Contracts](dashboard/deploy/your_contracts.md)
@ -60,6 +66,7 @@
- [Installation](developers/javascript/grid3_javascript_installation.md)
- [Loading Client](developers/javascript/grid3_javascript_loadclient.md)
- [Deploy a VM](developers/javascript/grid3_javascript_vm.md)
- [Deploy a VM with Mycelium Network](developers/javascript/grid3_javascript_vm_with_mycelium_network.md)
- [Capacity Planning](developers/javascript/grid3_javascript_capacity_planning.md)
- [Deploy Multiple VMs](developers/javascript/grid3_javascript_vms.md)
- [Deploy CapRover](developers/javascript/grid3_javascript_caprover.md)
@ -155,6 +162,7 @@
- [TFGrid Stacks](developers/grid_deployment/tfgrid_stacks.md)
- [Full VM Grid Deployment](developers/grid_deployment/grid_deployment_full_vm.md)
- [Grid Snapshots](developers/grid_deployment/snapshots.md)
- [Z-OS Boot Generator](developers/grid_deployment/zos_boot_generator/zos_boot_generator.md)
- [Deploy the Dashboard](developers/grid_deployment/deploy_dashboard.md)
- [Application Case Study: Nextcloud](developers/grid_deployment/app_case_study_nextcloud.md)
- [Farmers](farmers/farmers.md)
@ -165,15 +173,16 @@
- [4. Wipe All the Disks](farmers/3node_building/4_wipe_all_disks.md)
- [5. Set the BIOS/UEFI](farmers/3node_building/5_set_bios_uefi.md)
- [6. Boot the 3Node](farmers/3node_building/6_boot_3node.md)
- [Farming Requirements](farmers/farming_requirements.md)
- [Farming Optimization](farmers/farming_optimization/farming_optimization.md)
- [GPU Farming](farmers/3node_building/gpu_farming.md)
- [HDD-Only Nodes](farmers/farming_optimization/hdd_only_nodes.md)
- [Set Additional Fees](farmers/farming_optimization/set_additional_fees.md)
- [Minting Receipts](farmers/3node_building/minting_receipts.md)
- [Minting Periods](farmers/farming_optimization/minting_periods.md)
- [Room Parameters](farmers/farming_optimization/farm_room_parameters.md)
- [Farming Costs](farmers/farming_optimization/farming_costs.md)
- [Calculate Your ROI](farmers/farming_optimization/calculate_roi.md)
- [Farming Requirements](farmers/farming_optimization/farming_requirements.md)
- [Advanced Networking](farmers/advanced_networking/advanced_networking_toc.md)
- [Networking Overview](farmers/advanced_networking/networking_overview.md)
- [Network Considerations](farmers/advanced_networking/network_considerations.md)
@ -184,27 +193,43 @@
- [Minting and the Farmerbot](farmers/farmerbot/farmerbot_minting.md)
- [Node Status Bot](farmers/node_status_bot/node_status_bot.md)
- [Farming Troubleshooting](farmers/farming_troubleshooting.md)
- [Build, Post Build & Memory](farmers/farming_troubleshooting_tips.md)
- [System Administrators](system_administrators/system_administrators.md)
- [Getting Started](system_administrators/getstarted/tfgrid3_getstarted.md)
- [SSH Remote Connection](system_administrators/getstarted/ssh_guide/ssh_guide.md)
- [SSH with OpenSSH](system_administrators/getstarted/ssh_guide/ssh_openssh.md)
- [Advanced Methods](system_administrators/getstarted/ssh_guide/advanced_methods/advanced_methods.md)
- [SSH with PuTTY](system_administrators/getstarted/ssh_guide/ssh_putty.md)
- [SSH with WSL](system_administrators/getstarted/ssh_guide/ssh_wsl.md)
- [WireGuard Access](system_administrators/getstarted/ssh_guide/ssh_wireguard.md)
- [Remote Desktop and GUI](system_administrators/getstarted/remote-desktop_gui/remote-desktop_gui.md)
- [Cockpit: a Web-based Interface for Servers](system_administrators/getstarted/remote-desktop_gui/cockpit_guide/cockpit_guide.md)
- [XRDP: an Open-Source Remote Desktop Protocol](system_administrators/getstarted/remote-desktop_gui/xrdp_guide/xrdp_guide.md)
- [Apache Guacamole: a Clientless Remote Desktop Gateway](system_administrators/getstarted/remote-desktop_gui/guacamole_guide/guacamole_guide.md)
- [Planetary Network](system_administrators/getstarted/planetarynetwork.md)
- [WireGuard](system_administrators/getstarted/ssh_guide/advanced_methods/ssh_wireguard.md)
- [Planetary Network](system_administrators/getstarted/ssh_guide/advanced_methods/planetarynetwork.md)
- [TFGrid Deployments](system_administrators/getstarted/tfgrid_deployments.md)
- [TFGrid Services](system_administrators/getstarted/tfgrid_services/tf_grid_services_readme.md)
- [Mycelium](system_administrators/mycelium/mycelium_toc.md)
- [Overview](system_administrators/mycelium/overview.md)
- [Installation](system_administrators/mycelium/installation.md)
- [Additional Information](system_administrators/mycelium/information.md)
- [Message](system_administrators/mycelium/message.md)
- [Packet](system_administrators/mycelium/packet.md)
- [Data Packet](system_administrators/mycelium/data_packet.md)
- [API YAML](system_administrators/mycelium/api_yaml.md)
- [Pulumi](system_administrators/pulumi/pulumi_readme.md)
- [Introduction to Pulumi](system_administrators/pulumi/pulumi_intro.md)
- [Installing Pulumi](system_administrators/pulumi/pulumi_install.md)
- [Deployment Examples](system_administrators/pulumi/pulumi_examples.md)
- [Deployment Details](system_administrators/pulumi/pulumi_deployment_details.md)
- [Complete Guides](system_administrators/pulumi/pulumi_complete_guides/pulumi_complete_guides_toc.md)
- [Pulumi and YAML](system_administrators/pulumi/pulumi_complete_guides/pulumi_yaml.md)
- [Pulumi and Python](system_administrators/pulumi/pulumi_complete_guides/pulumi_python.md)
- [Pulumi and Go](system_administrators/pulumi/pulumi_complete_guides/pulumi_go.md)
- [GPU](system_administrators/gpu/gpu_toc.md)
- [GPU Support](system_administrators/gpu/gpu.md)
- [Terraform](system_administrators/terraform/terraform_toc.md)
- [Overview](system_administrators/terraform/terraform_readme.md)
- [Introduction to Terraform](system_administrators/terraform/terraform_readme.md)
- [Installing Terraform](system_administrators/terraform/terraform_install.md)
- [Terraform Basics](system_administrators/terraform/terraform_basics.md)
- [Full VM Deployment](system_administrators/terraform/terraform_full_vm.md)
- [GPU Support](system_administrators/terraform/terraform_gpu_support.md)
- [Terrafprm Basics](system_administrators/terraform/terraform_basics.md)
- [Resources](system_administrators/terraform/resources/terraform_resources_readme.md)
- [Using Scheduler](system_administrators/terraform/resources/terraform_scheduler.md)
- [Virtual Machine](system_administrators/terraform/resources/terraform_vm.md)
@ -218,6 +243,7 @@
- [CapRover](system_administrators/terraform/resources/terraform_caprover.md)
- [Advanced](system_administrators/terraform/advanced/terraform_advanced_readme.md)
- [Terraform Provider](system_administrators/terraform/advanced/terraform_provider.md)
- [GPU Support](system_administrators/terraform/terraform_gpu_support.md)
- [Terraform Provisioners](system_administrators/terraform/advanced/terraform_provisioners.md)
- [Mounts](system_administrators/terraform/advanced/terraform_mounts.md)
- [Capacity Planning](system_administrators/terraform/advanced/terraform_capacity_planning.md)
@ -231,23 +257,13 @@
- [Nextcloud Single Deployment](system_administrators/terraform/advanced/terraform_nextcloud_single.md)
- [Nextcloud Redundant Deployment](system_administrators/terraform/advanced/terraform_nextcloud_redundant.md)
- [Nextcloud 2-Node VPN Deployment](system_administrators/terraform/advanced/terraform_nextcloud_vpn.md)
- [Pulumi](system_administrators/pulumi/pulumi_readme.md)
- [Introduction to Pulumi](system_administrators/pulumi/pulumi_intro.md)
- [Installing Pulumi](system_administrators/pulumi/pulumi_install.md)
- [Deployment Examples](system_administrators/pulumi/pulumi_examples.md)
- [Deployment Details](system_administrators/pulumi/pulumi_deployment_details.md)
- [Mycelium](system_administrators/mycelium/mycelium_toc.md)
- [Overview](system_administrators/mycelium/overview.md)
- [Installation](system_administrators/mycelium/installation.md)
- [Additional Information](system_administrators/mycelium/information.md)
- [Message](system_administrators/mycelium/message.md)
- [Packet](system_administrators/mycelium/packet.md)
- [Data Packet](system_administrators/mycelium/data_packet.md)
- [API YAML](system_administrators/mycelium/api_yaml.md)
- [Computer and IT Basics](system_administrators/computer_it_basics/computer_it_basics.md)
- [CLI and Scripts Basics](system_administrators/computer_it_basics/cli_scripts_basics.md)
- [Docker Basics](system_administrators/computer_it_basics/docker_basics.md)
- [Git and GitHub Basics](system_administrators/computer_it_basics/git_github_basics.md)
- [Gitea](system_administrators/computer_it_basics/gitea/gitea_toc.md)
- [Gitea Basics](system_administrators/computer_it_basics/gitea/gitea_basics.md)
- [Gitea API](system_administrators/computer_it_basics/gitea/gitea_api.md)
- [Firewall Basics](system_administrators/computer_it_basics/firewall_basics/firewall_basics.md)
- [UFW Basics](system_administrators/computer_it_basics/firewall_basics/ufw_basics.md)
- [Firewalld Basics](system_administrators/computer_it_basics/firewall_basics/firewalld_basics.md)
@ -265,12 +281,20 @@
- [IPFS on a Micro VM](system_administrators/advanced/ipfs/ipfs_microvm.md)
- [MinIO Operator with Helm3](system_administrators/advanced/minio_helm3.md)
- [Hummingbot](system_administrators/advanced/hummingbot.md)
- [AI & ML Workloads](system_administrators/advanced/ai_ml_workloads.md)
- [AI & ML Workloads](system_administrators/advanced/ai_ml_workloads/ai_ml_workloads_toc.md)
- [CPU and Llama](system_administrators/advanced/ai_ml_workloads/cpu_and_llama.md)
- [GPU and Pytorch](system_administrators/advanced/ai_ml_workloads/gpu_and_pytorch.md)
- [GPU and Fooocus](system_administrators/advanced/ai_ml_workloads/gpu_and_fooocus.md)
- [Ecommerce](system_administrators/advanced/ecommerce/ecommerce.md)
- [WooCommerce](system_administrators/advanced/ecommerce/woocommerce.md)
- [nopCommerce](system_administrators/advanced/ecommerce/nopcommerce.md)
- [HTTPS with Caddy](system_administrators/advanced/https_caddy.md)
- [Node Status Bot](system_administrators/advanced/node_status_bot.md)
- [Minetest](system_administrators/advanced/minetest.md)
- [Remote Desktop and GUI](system_administrators/getstarted/remote-desktop_gui/remote-desktop_gui.md)
- [Cockpit: a Web-based Interface for Servers](system_administrators/getstarted/remote-desktop_gui/cockpit_guide/cockpit_guide.md)
- [XRDP: an Open-Source Remote Desktop Protocol](system_administrators/getstarted/remote-desktop_gui/xrdp_guide/xrdp_guide.md)
- [Apache Guacamole: a Clientless Remote Desktop Gateway](system_administrators/getstarted/remote-desktop_gui/guacamole_guide/guacamole_guide.md)
- [ThreeFold Token](threefold_token/threefold_token.md)
- [TFT Bridges](threefold_token/tft_bridges/tft_bridges.md)
- [TFChain-Stellar Bridge](threefold_token/tft_bridges/tfchain_stellar_bridge.md)
@ -295,12 +319,6 @@
- [FAQ](faq/faq.md)
- [Knowledge Base](manual/knowledge_base.md)
- [About](about/about.md)
- [ThreeFold History](about/threefold_history.md)
- [Token History](about/token_history.md)
- [Genesis Pool](about/genesis_pool.md)
- [Genesis Pool Dubai](about/genesis_pool_dubai.md)
- [Genesis Pool Ghent](about/genesis_pool_ghent.md)
- [Genesis Pool Details](about/genesis_block_pool_details.md)
- [ThreeFold Tech](about/threefold_tech.md)
- [Organisation Structure](about/orgstructure.md)
- [Governance](about/governance.md)
@ -325,7 +343,6 @@
- [TFGrid v3.0.0 Alpha-4](about/roadmap/releasenotes/tfgrid_release_3_0_a4.md)
- [TFGrid v3.0.0 Alpha-2](about/roadmap/releasenotes/tfgrid_release_3_0_a2.md)
- [TFGrid v3.0.0](about/roadmap/releasenotes/tfgrid_release_3_0.md)
- [ThreeFold Token](about/token_overview/token_overview.md)
- [Technology](tech/technology_toc.md)
- [Introduction](tech/technology.md)
- [The Internet Today](tech/internet_today.md)

View File

@ -33,7 +33,7 @@
- [Enterprise Roadmap](tech/enterprise_roadmap.md)
- [Opensource TFGrid Roadmap](tech/tfgrid_roadmap.md)
- [Hero Roadmap](tech/hero_roadmap.md)
- [ThreeFold Core Features](tech/features.md)
- [Core Features](tech/features.md)
- [Compute](tech/compute.md)
- [Zero-OS](tech/zero_os.md)
- [Zero-Deploy](tech/zero_deploy.md)

View File

@ -1,3 +0,0 @@
#!/bin/bash
hero mdbook -u https://git.ourworld.tf/tfgrid/info_tfgrid/src/branch/main/heroscript/tech
rsync -rv ~/hero/www/info/tech/ root@info.ourworld.tf:/root/hero/www/info/tech/

View File

@ -0,0 +1,60 @@
- [Overview](tech/overview.md)
- [The Cloud Today](tech/cloud_today.md)
- [The Internet Today](tech/internet_today.md)
- [History of Computers](tech/history/c64.md)
- [Too Many Layers](tech/history/layers.md)
- [The Cloud Re-invented](tech/cloud_reinvented.md)
- [The Internet Re-invented](tech/internet_reinvented.md)
- [Cloud Beyond Cost](tech/cloud_like_insurance.md)
- [World Records](tech/world_records.md)
- [Architecture](tech/architecture.md)
- [Cloud Engine](tech/cloudengine.md)
- [Empowers Internet](tech/internet_arch.md)
- [Hero as Virtual Administrator](tech/hero_virtual_admin.md)
- [Key Innovations](tech/key_innovations_overview.md)
- [Compute](tech/compute_inno.md)
- [Zero-OS](tech/zos_innovation.md)
- [Zero-Images](tech/zero_image_inno.md)
- [Deterministic Deploy](tech/zero_deploy_inno.md)
- [Zero-Install](tech/zero_install_inno.md)
- [Network](tech/network_inno.md)
- [Mycelium](tech/mycelium_inno.md)
- [Shortest Path Routing](tech/mycelium_shortest_path_routing_inno.md)
- [Whitelists](tech/mycelium_whiltelist.md)
- [Network Wall](tech/network_wall_innovation.md)
- [Virtual Browser](tech/virtual_browser.md)
- [Storage](tech/storage_inno.md)
- [Quantum Safe Storage](tech/zstor_innovation.md)
- [Quantum Safe Filesystem](tech/qsfs_innovation.md)
- [FungiStor](tech/fungistor_innovation.md)
- [Energy Efficient](tech/energy_efficient.md)
- [Status/Roadmap](tech/roadmap.md)
- [Enterprise Roadmap](tech/enterprise_roadmap.md)
- [Opensource TFGrid Roadmap](tech/tfgrid_roadmap.md)
- [ Core Features](tech/features.md)
- [Compute](tech/compute.md)
- [Zero-OS](tech/zero_os.md)
- [Zero-Deploy](tech/zero_deploy.md)
- [Zero-Install](tech/zero_install.md)
- [Zero-Kube](tech/zkube.md)
- [Storage](tech/qsss_home.md)
- [Quantum Safe Storage Algo](tech/qss_algorithm.md)
- [Zero Knowledge proof](tech/qss_zero_knowledge_proof.md)
- [NFT Storage](tech/nft_storage.md)
- [S3 Storage](tech/s3_interface)
- [File System](tech/qss_filesystem.md)
- [Network](tech/networking.md)
- [Mycelium](tech/mycelium.md)
- [Web Gateway](tech/webgw.md)
- [Use Cases](partners_utilization/partners.md)
- [Tier-S DC](partners_utilization/tier_s_datacenter.md)
- [Digital Freezone](partners_utilization/freezone.md)
- [Mkondo](partners_utilization/mkondo.md)
- [Vverse](partners_utilization/vverse.md)
- [Holochain](partners_utilization/holochain.md)
- [TZG](partners_utilization/tanzania.md)
- [Sikana](partners_utilization/sikana.md)
- [Vindo](partners_utilization/vindo.md)
<!-- - [Helium](partners_utilization/helium.md) -->
<!-- - [Elestio](partners_utilization/elestio.md) -->

View File

@ -0,0 +1,60 @@
- [Overview](tech/overview.md)
- [The Cloud Today](tech/cloud_today.md)
- [The Internet Today](tech/internet_today.md)
- [History of Computers](tech/history/c64.md)
- [Too Many Layers](tech/history/layers.md)
- [The Cloud Re-invented](tech/cloud_reinvented.md)
- [The Internet Re-invented](tech/internet_reinvented.md)
- [Cloud Beyond Cost](tech/cloud_like_insurance.md)
- [World Records](tech/world_records.md)
- [Architecture](tech/architecture.md)
- [Cloud Engine](tech/cloudengine.md)
- [Empowers Internet](tech/internet_arch.md)
- [Hero as Virtual Administrator](tech/hero_virtual_admin.md)
- [Key Innovations](tech/key_innovations_overview.md)
- [Compute](tech/compute_inno.md)
- [Zero-OS](tech/zos_innovation.md)
- [Zero-Images](tech/zero_image_inno.md)
- [Deterministic Deploy](tech/zero_deploy_inno.md)
- [Zero-Install](tech/zero_install_inno.md)
- [Network](tech/network_inno.md)
- [Mycelium](tech/mycelium_inno.md)
- [Shortest Path Routing](tech/mycelium_shortest_path_routing_inno.md)
- [Whitelists](tech/mycelium_whiltelist.md)
- [Network Wall](tech/network_wall_innovation.md)
- [Virtual Browser](tech/virtual_browser.md)
- [Storage](tech/storage_inno.md)
- [Quantum Safe Storage](tech/zstor_innovation.md)
- [Quantum Safe Filesystem](tech/qsfs_innovation.md)
- [FungiStor](tech/fungistor_innovation.md)
- [Energy Efficient](tech/energy_efficient.md)
- [Status/Roadmap](tech/roadmap.md)
- [Enterprise Roadmap](tech/enterprise_roadmap.md)
- [Opensource TFGrid Roadmap](tech/tfgrid_roadmap.md)
- [ Core Features](tech/features.md)
- [Compute](tech/compute.md)
- [Zero-OS](tech/zero_os.md)
- [Zero-Deploy](tech/zero_deploy.md)
- [Zero-Install](tech/zero_install.md)
- [Zero-Kube](tech/zkube.md)
- [Storage](tech/qsss_home.md)
- [Quantum Safe Storage Algo](tech/qss_algorithm.md)
- [Zero Knowledge proof](tech/qss_zero_knowledge_proof.md)
- [NFT Storage](tech/nft_storage.md)
- [S3 Storage](tech/s3_interface)
- [File System](tech/qss_filesystem.md)
- [Network](tech/networking.md)
- [Mycelium](tech/mycelium.md)
- [Web Gateway](tech/webgw.md)
- [Use Cases](partners_utilization/partners.md)
- [Tier-S DC](partners_utilization/tier_s_datacenter.md)
- [Digital Freezone](partners_utilization/freezone.md)
- [Mkondo](partners_utilization/mkondo.md)
- [Vverse](partners_utilization/vverse.md)
- [Holochain](partners_utilization/holochain.md)
- [TZG](partners_utilization/tanzania.md)
- [Sikana](partners_utilization/sikana.md)
- [Vindo](partners_utilization/vindo.md)
<!-- - [Helium](partners_utilization/helium.md) -->
<!-- - [Elestio](partners_utilization/elestio.md) -->

View File

@ -1,69 +0,0 @@
<h1> Genesis Pool Details </h1>
<h2>Table of Contents </h2>
- [Genesis Pool](#genesis-pool)
- [Genesis Block](#genesis-block)
- [Genesis Block Value](#genesis-block-value)
- [Calculation](#calculation)
- [Genesis Pool Details](#genesis-pool-details)
***
## Genesis Pool
Genesis pool is the initial capacity with which the network started, was available when the project officially launched (blockchain launch March 2018).
- +-300 computer (all owned by ThreeFold_Dubai)
- Belgium: 117+30 (hosted by BetterToken)
- Dubai: 148 (hosted by TF FZC itself)
- total estimate resource/compute units
- CRU: 4800,
- HRU: 8100000
- MRU: 18600
- SRU: 106000
## Genesis Block
Genesis block is the first block registered in the blockchain. This consists of a number of TFT, in our case 695M TFT.
> Maximum amount of tokens in the ThreeFold Blockchain at launch = 100 Billion (in other words genesis pool < 1% at start of max nr TFT)
### Genesis Block Value
It's hard to define the value of the genesis block when it was calculated, there was no established TFT price.
- If TFT price = USD 0.01: +-7M USD (this token price has not been established but could be 2016-17)
- Summer 2023 the price is back on USD 0.01, which we believe is too low for the value created, lets hope for a better future.
### Calculation
To come up with a reasonable number and show the community that there was hardware available for the genesis block, we made an excel calculation.
- Servers as part of genesis pool calculation
- +-300 computer (all owned by ThreeFold_Dubai)
- Belgium: 117+30 (hosted by BetterToken)
- Dubai: 148 (hosted by TF FZC itself)
- Hardware as used in many years before token launch (March 2018)
- At least 100+ servers over quite some years
- Total estimate resource/compute units
- CRU: 4,800
- HRU: 8,100,000
- MRU: 18,600
- SRU: 106,000
- Cloud Units
- Results in 3,927 CU and 8,225 SU
- The farming rules used were farming/minting rules v1 but with no difficulty level and TFT price 0.01
- Duration
- We took +- 1.5 years in our calculation
- Averaged out, it's for sure not exact science
- But we can say that the amount of capacity listed has been available long enough for our engineers during the pre-launch period. Probably not with those exact listed servers but in general.
- Result: **695M TFT**
_The purpose of this exercise is to demonstrate there is a reasoning behind the 695M TFT and computers which have been available. It's not intended as exact proof nor defense. We believe the value given was in line with the situation at that time._
## Genesis Pool Details
- Block 0: [Block 0 on Explorer](https://explorer2.threefoldtoken.com/hash.html?hash=a2ee0aa706c9de46ec57dbba1af8d352fbcc3cc5e9626fc56337fd5e9ca44c8d)
- Genesis Block Code: [Code of Block 0](https://github.com/threefoldfoundation/tfchain/blob/master/pkg/config/config.go#L103)

View File

@ -1,70 +0,0 @@
<h1> Genesis Pool </h1>
<h2>Table of Contents </h2>
- [Introduction](#introduction)
- [Genesis Pool Token Usage](#genesis-pool-token-usage)
- [Remarks](#remarks)
***
## Introduction
At the end of March 2018, ThreeFold launched the public blockchain.
ThreeFold developed their own blockchain software called Rivine, which was probably the first proof of blockstake blockchain in the world. We did not like the way how the other blockchains at that time were doing proof of work, which is basically burning a lot of energy to prove the validity of their blockchain.
Rivine is a fork from the blockchain work done by the team of SIA and since then a lot of work has been done on it to fulfill our own requirements. The Rivine blockchain will no longer be used after May 2020.
ThreeFold is the result of more than 20 years of work in the Internet space, over a number of companies.
The technology used at start in March 2018 has been developed mainly out of three companies: ThreeFold_Dubai, BetterToken(bettertoken) and GreenIT Globe. Later in 2018, TF Tech was spun off from our incubator.
TF Tech is a company born out of our Incubator called [Incubaid](http://www.incubaid.com/) in Belgium.
TF Tech has a purpose to further develop the software and commercialize the capabilities on a global basis, mainly by working together with tech partners.
The public version of our blockchain was started March 2018. The servers used during development and mining tokens already started years before.
Many hundreds of servers have been used to develop the technology which now makes up our ThreeFold_Grid.
+-300 servers are the foundation of our TF Grid.
Most of the servers are in Dubai and in Ghent (Belgium).
- To see the [genesis pool in Ghent, see here](genesis_pool_ghent.md)
- To see the [genesis pool in Dubai, see here](genesis_pool_dubai.md)
All genesis pools were owned by the foundation. Many of those servers are at this point no longer active. The operations were done by ThreeFold_Dubai and BetterToken as Farming Cooperative.
> For information about genesis pool/block, see [here](genesis_block_pool_details.md).
## Genesis Pool Token Usage
- A lot of the genesis pool tokens went to the original shareholders of a company who created a lot of the technology which was the basis at that time for ThreeFold.
- Most of these tokens are locked up and are not tradeable.
- This was a deal made mid 2018 and provided the ThreeFold Dubai with technology and a global engineering team.
- The other part went to ThreeFold Dubai, to allow the Foundation to promote & further grow the project.
> [See Token Overview](token_overview.md) for more details.
The tokens were used from out of ThreeFold_Dubai to create value for the ThreeFold Grid.
- Initial funding: sell TFT as future IT capacity
- IT capacity delivered amongst others from the computers deployed by genesis pools (+300 servers)
- grants to community, bounties for coders, evangelists, ...
- max bounty given to contributors/founders = 2.5m TFT
- funding for projects like coding, marketing, ...
- There is a token grant program, but not really active yet.
- Fund the day-to-day operation of threefold_dubai
- Fund some development projects for our open source technology
- Public exchange fees
- Operational costs of keeping the genesis pool operational (engineers, data center, bandwidth, ...)
- Reward for the ThreeFold larger community and contributors
## Remarks
- ThreeFold_Dubai is run as a [not-for-profit organization](legal@@definitions_legal)
- All (future) profits generated, tokens=IT capacity sold are used to promote and grow the ThreeFold Project.
- None of the potential profits generated go to the shareholders of the company.
- Investments and loans given will of course be paid back to the relevant investors.

View File

@ -1,7 +0,0 @@
## The Genesis Pool Dubai
![](img/genesispool_1.jpg)
![](img/genesispool_2.jpg)
Read more about ThreeFold Dubai [here](threefold_dubai.md).

View File

@ -1,11 +0,0 @@
## The Genesis Pool (Ghent)
![](img/lochristi_1_.jpg)
![](img/lochristi_2_.jpg)
![](img/lochristi_3.jpg)
![](img/lochristi_4.jpg)
![](img/lochristi_5_.jpg)
![](img/lochristi_6_.jpg)
![](img/lochristi_7.jpg)

View File

@ -1,59 +0,0 @@
<h1> ThreeFold History </h1>
<h2>Table of Contents</h2>
- [Introduction](#introduction)
- [History](#history)
- [ThreeFold Project Funding Total](#threefold-project-funding-total)
- [Status](#status)
- [Genesis Pool](#genesis-pool)
- [History of Tokens](#history-of-tokens)
***
## Introduction
The project, now in its sixth year and is gratefulf or the support of its community and commercial entity [ThreeFold Tech](https://github.com/threefoldtech).
ThreeFold is fundamentally a decentralized initiative. Within this framework, ThreeFold Dubai plays a pivotal role in championing and advancing the ThreeFold Grid and the broader movement.
Our founders have largely retained their tokens, with only minimal sales, if any. Their intent is clear: they plan to hold onto their tokens until the grid achieves global recognition and the token value surpasses 0.2 USD.
## History
In the earlier days of ThreeFold, there were multiple teams collaborating, but the two core teams were located in Dubai and Belgium.
A group of early supporters bought IT capacity (through buying TFT) from our Genesis pool and our early farmers. These buyers could use their TFT to buy IT capacity from [ThreeFold Dubai](threefold_dubai.md) or [BetterToken](bettertoken.md) BV until April 2020, or from the TF Grid directly in a fully decentralized way starting May 2020.
The ThreeFold Grid is the result of many farmers using the open source technology of ThreeFold Tech.
Originally, the technology used was created by three companies: GreenIT Globe, ThreeFold Dubai & ThreeFold Tech. The last two still actively participate in the creation of tech components or content as used by all ThreeFold Farmers today.
## ThreeFold Project Funding Total
How much funding was used to make the ThreeFold project possible?
> +- 50M USD
- +20M USD for all farming (thank you farmers)
- 15M USD in ThreeFold Tech as convertible loan (by 50+ investors)
- 5M USD in early IT capacity purchases (as TFT)
- +10M USD funding from Incubaid/Kristof (estimate)
- ThreeFold Tech was established Oct 2018, from out of Incubaid
- Related to people related to [Incubaid](https://www.incubaid.com)
- Over quite some years, multiple companies/projects
## Status
We have worked with multiple regions over the years to look for appropriate structures, we realize we need more funding as such we have launched a venture creator in mauritius who will hopefully invest 7.5m EUR in TFTech as well as in TF Dubai.
See our [overview of our companies](threefold_companies.md)
## Genesis Pool
To kickstart the ThreeFold Grid back in 2017 the foundation committed large amounts of capacity to the grid. This was called the [Genesis Pool](genesis_pool.md) and the tokens sold as mentioned could be used to use capacity from this pool and more.
## History of Tokens
For more info about history of tokens, see [token history](token_history.md).

View File

@ -1,96 +0,0 @@
<h1> Token History </h1>
<h2>Table of Contents </h2>
- [Introduction](#introduction)
- [Quick History Overview](#quick-history-overview)
- [Organic Growth](#organic-growth)
- [Farming Model Improvements](#farming-model-improvements)
- [TFT Versions](#tft-versions)
- [Migration](#migration)
- [Migration from TFTv1 Rivine to TFTv1 Stellar (2020)](#migration-from-tftv1-rivine-to-tftv1-stellar-2020)
- [Migration from TFTv1 Stellar (Staking Pool=TFTA) to TFTv2 Stellar (Trading or Production Pool=TFT)](#migration-from-tftv1-stellar-staking-pooltfta-to-tftv2-stellar-trading-or-production-pooltft)
- [Technical Information](#technical-information)
***
## Introduction
We present the ThreeFold token history and the path from TFT v1 towards TFT v2.
## Quick History Overview
- More than 10 years ago, this project started from out of our incubator (see [Incubaid](https://www.incubaid.com))
- More than 6 years ago TF Foundation started deploying capacity for development purposes
- This became our ([our genesis pool](genesis_pool.md)), which is the beginning of farming
- In 2017-18, value of the genesis pools were about 7m USD in TFT (tokens did not exist yet)
- Genesis pools are owned by ThreeFold Dubai (ThreeFold_Dubai).
- March 2018 our first-generation blockchain for the ThreeFold_Token saw daylight
- The TFT v1 was launched on a blockchain called Rivine (PTO)
- The genesis pool resulted in the first batch initial TFT
- The blockchain nodes were hosted by +30 different parties completely unrelated to each other
- In Q2 2019, ThreeFold_Dubai launched our generation 1 of our TF Grid
- April 2020 ThreeFold_Dubai launched the TFGrid v2.0 which is now public and usable by the world
- ThreeFold has a new website and a new wiki
- The farmers & TFT holders have at their own will upgraded their wallets, zero-nodes, ...
- May 2020: ThreeFold_Dubai launched our 2nd version of our token called TFT but this time on Stellar
- The original TFTv1 kept all same properties and benefits and is now called TFTA also on Stellar (is technology choice), anyone can move from TFTv1 to TFTv2
- TF Foundation Dubai has provisioned the TFTv1 & TFTv2 on Stellar blockchain, but has no influence or access to any of the wallets or for that matter the 3Nodes (the nodes providing IT capacity)
- See below for more info, this was the result of 12 months of work with our community and of-course consensus to do this.
## Organic Growth
We didn't artificially pump the value of the tokens.
We did not issue (print) tokens and go out onto an exchange to offer these tokens to the market. This is referred to as a public ICO. Some ICOs were not very clean in how they created hype and convinced people to invest. Because of our decision not to do a public ICO, we have not been able to raise much money, but we feel that this was more aligned with our values.
We have sold some TFT over the counter but please note every buyer could at any point in time use these TFTs to buy IT capacity, this makes these TFT purposeful, even from the very start.
## Farming Model Improvements
In Q2 2020 we were launching TF Grid 2.0 with updated minting rules. As part of these farming rules the max number of tokens became 4 billion, which changes the optics of the original size of the genesis token pool.
In Q3 2021 we launched TF Grid 3.0 which has again brought improvement to the farming model. Its up to the farmers to choose if they want to change to the new farming model or not.
### TFT Versions
| | version 1 Rivine | version 1 Stellar | version 2 Stellar |
| ------------------------------- | --------------------------- | ----------------- | --------------------------- |
| blockchain tech | Rivine, proof of blockstake | Public, Stellar | Public, Stellar |
| on public blockchain | march 2018 | 2020 May | 2020 May |
| farmed since | +-2017 | 2020 May | tbd |
| freely transferable (\*) | YES | YES | YES |
| complete blockchain feature set | YES | YES | YES |
| decentralized exchange | YES (atomic swap) | YES (Stellar) | YES (Stellar) |
| public exchange | BTC Alpha till Dec 2019 | Stellar | Stellar, BTC Alpha & Liquid (until August 2022) |
| freely tradable on exchange | YES | YES | YES |
| Name on Blockchain | TFT | TFTA | TFT |
| Purpose | v1 token | Staking Pool | Trading Pool |
## Migration
### Migration from TFTv1 Rivine to TFTv1 Stellar (2020)
- TF Tech decided to no longer support development of Rivine, at this time there are better technologies available as blockchain
- The Foundation investigated many blockchain platforms & recommended to use Stellar
- Jimber (company which maintains the wallet, which is open source code), has made the changes in the wallet to be able to support this new blockchain
- The conversation had to be a mandatory one, because otherwise there would be the potential of double-spending problems over both simultaneously-active blockchains
- What happened here can be compared to a website deciding to change the database backend (change from e.g. MS Sql to Oracle). The users of the website should not have to be aware of this migration
- Every user had to do the transaction themselves, no developer or anyone else had control over this migration step. This was an automatic step
- Everyone can use the validation scripts available to check the correct conversion between two blockchain technologies. The validation scripts prove that every transaction in the conversion happened well
### Migration from TFTv1 Stellar (Staking Pool=TFTA) to TFTv2 Stellar (Trading or Production Pool=TFT)
- See [TFTA to TFT](legal@@tfta_to_tft)
### Technical Information
[TFTA TrustLine](https://stellar.expert/explorer/public/asset/TFTA-GBUT4GP5GJ6B3XW5PXENHQA7TXJI5GOPW3NF4W3ZIW6OOO4ISY6WNLN2)
Accounts that got initial balances migrated them from the previous blockchain, [rivine](https://explorer2.threefoldtoken.com/).
To validate this, each migration transaction contain hash of the rivine lock transaction in their memo in hex format.
[Rivine block explorer](https://explorer2.threefoldtoken.com/) can be used for validation purposes.
> Important note: The ThreeFold Token (TFT) is not an investment instrument. TFTs represent IT capacity on the ThreeFold Grid, farmers create TFT, developers use TFT.

Binary file not shown.

Before

Width:  |  Height:  |  Size: 103 KiB

Binary file not shown.

Before

Width:  |  Height:  |  Size: 24 KiB

Binary file not shown.

Before

Width:  |  Height:  |  Size: 23 KiB

Binary file not shown.

Before

Width:  |  Height:  |  Size: 24 KiB

Binary file not shown.

Before

Width:  |  Height:  |  Size: 24 KiB

Binary file not shown.

Before

Width:  |  Height:  |  Size: 23 KiB

View File

@ -1,105 +0,0 @@
<h1> ThreeFold Special Wallets </h1>
<h2>Table of Contents</h2>
- [Introduction](#introduction)
- [Exchange and OTC Wallets](#exchange-and-otc-wallets)
- [ThreeFold Contribution Wallets](#threefold-contribution-wallets)
- [Wisdom Council Wallets](#wisdom-council-wallets)
- [Important Note](#important-note)
- [Remarks](#remarks)
- [Proof-of-Utilization Wallets](#proof-of-utilization-wallets)
***
## Introduction
We present special wallets that hold a given amount of TFT.
## Exchange and OTC Wallets
| **Description** | **TFT Balance** | **Address** |
| ------------------ | ----------- | -------------------------------------------------------------------------------- |
| Liquid Exchange #1 | !!wiki.include page:'manual:ga7opn4a3jnhlphpewm4pjdoyydynzom7es6yl3o7nc3pry3v3ux6anm.md' | [GA7OPN4A3JNHLPHPEWM4PJDOYYDYNZOM7ES6YL3O7NC3PRY3V3UX6ANM](https://stellar.expert/explorer/public/account/GA7OPN4A3JNHLPHPEWM4PJDOYYDYNZOM7ES6YL3O7NC3PRY3V3UX6ANM) |
| Liquid Exchange #2 | !!wiki.include page:'manual:gdskfynmzwtb3v5an26ceaq27643q3kb4x6my4uto2liidfnd4spqzyu.md' | [GDSKFYNMZWTB3V5AN26CEAQ27643Q3KB4X6MY4UTO2LIIDFND4SPQZYU](https://stellar.expert/explorer/public/account/GDSKFYNMZWTB3V5AN26CEAQ27643Q3KB4X6MY4UTO2LIIDFND4SPQZYU) |
| gettft.com | !!wiki.include page:'manual:gbqhn7rl4lsrpr2tt74id2ujpz2axchqy2wkgctdljm3nxvj7gqhucod.md' | [GBQHN7RL4LSRPR2TT74ID2UJPZ2AXCHQY2WKGCTDLJM3NXVJ7GQHUCOD](https://stellar.expert/explorer/public/account/GBQHN7RL4LSRPR2TT74ID2UJPZ2AXCHQY2WKGCTDLJM3NXVJ7GQHUCOD) |
| BTC-Alpha Exchange | !!wiki.include page:'manual:gbtpaxxp6534upc4mlngfgjwcd6dnsrvippozwxaqawi4fktlojy2a2s.md' | [GBTPAXXP6534UPC4MLNGFGJWCD6DNSRVIPPOZWXAQAWI4FKTLOJY2A2S](https://stellar.expert/explorer/public/account/GBTPAXXP6534UPC4MLNGFGJWCD6DNSRVIPPOZWXAQAWI4FKTLOJY2A2S) |
## ThreeFold Contribution Wallets
| **Description** | **TFT Balance** | **Address** |
| ------------------------------- | ----------- | -------------------------------------------------------------------------------- |
| TF DAY2DAY operations | !!wiki.include page:'manual:gb2c5hczywngvm6jgxdwqbjtmuy4s2hpptcah63hfaqvl2alxdw7ssj7.md' | [GB2C5HCZYWNGVM6JGXDWQBJTMUY4S2HPPTCAH63HFAQVL2ALXDW7SSJ7](https://stellar.expert/explorer/public/account/GB2C5HCZYWNGVM6JGXDWQBJTMUY4S2HPPTCAH63HFAQVL2ALXDW7SSJ7) |
| TF Promotion Wallet | !!wiki.include page:'manual:gdlvib44lvonm5k67lupsfzmsx7g2rlyvbm5mmhuj4naqju7ch4hbjbo.md' | [GDLVIB44LVONM5K67LUPSFZMSX7G2RLYVBM5MMHUJ4NAQJU7CH4HBJBO](https://stellar.expert/explorer/public/account/GDLVIB44LVONM5K67LUPSFZMSX7G2RLYVBM5MMHUJ4NAQJU7CH4HBJBO) |
| TF Grants Wallet | !!wiki.include page:'manual:gdkxtuynw4bjkdm2l7b5xuyfuisv52kuu4g7vpnlf4zsikburm622ypz.md' | [GDKXTUYNW4BJKDM2L7B5XUYFUISV52KUU4G7VPNLF4ZSIKBURM622YPZ](https://stellar.expert/explorer/public/account/GDKXTUYNW4BJKDM2L7B5XUYFUISV52KUU4G7VPNLF4ZSIKBURM622YPZ) |
| ThreeFold Carbon Credit Funding | !!wiki.include page:'manual:gdijy6k2bbrirx423zfuykkfdn66xp2kmsbzfqse2psndz6edvqtrlsu.md' | [GDIJY6K2BBRIRX423ZFUYKKFDN66XP2KMSBZFQSE2PSNDZ6EDVQTRLSU](https://stellar.expert/explorer/public/account/GDIJY6K2BBRIRX423ZFUYKKFDN66XP2KMSBZFQSE2PSNDZ6EDVQTRLSU) |
| TF Team Wallet | !!wiki.include page:'manual:gcwhwdrxypxqaoymqkb66szplm6uankgmsl4sp7lsoia6ottoyq6hbih.md' | [GCWHWDRXYPXQAOYMQKB66SZPLM6UANKGMSL4SP7LSOIA6OTTOYQ6HBIH](https://stellar.expert/explorer/public/account/GCWHWDRXYPXQAOYMQKB66SZPLM6UANKGMSL4SP7LSOIA6OTTOYQ6HBIH) |
## Wisdom Council Wallets
| **Description** | **TFT Balance** | **Address** |
| --------------------------------------- | ----------- | -------------------------------------------------------------------------------- |
| Liquidity/Ecosystem Contribution Wisdom | !!wiki.include page:'manual:gbv734i2sv4yddpvjmyxu3iz2aiu5geajrad4e4bqg7ca2n63nxspmd6.md' | [GBV734I2SV4YDDPVJMYXU3IZ2AIU5GEAJRAD4E4BQG7CA2N63NXSPMD6](https://stellar.expert/explorer/public/account/GBV734I2SV4YDDPVJMYXU3IZ2AIU5GEAJRAD4E4BQG7CA2N63NXSPMD6) |
| TF Promotion Wisdom | !!wiki.include page:'manual:gai4c2bgoa3yhvqzzw7ow4fhoggywtubevnhb6mw4zafg7zaa7d5ipc3.md' | [GAI4C2BGOA3YHVQZZW7OW4FHOGGYWTUBEVNHB6MW4ZAFG7ZAA7D5IPC3](https://stellar.expert/explorer/public/account/GAI4C2BGOA3YHVQZZW7OW4FHOGGYWTUBEVNHB6MW4ZAFG7ZAA7D5IPC3) |
| TF Grants Wisdom | !!wiki.include page:'manual:gcej7dmulftt25uh4faagoz6ker4wxayqgjusiitqd527dgtksxkbqgr.md' | [GCEJ7DMULFTT25UH4FAAGOZ6KER4WXAYQGJUSIITQD527DGTKSXKBQGR](https://stellar.expert/explorer/public/account/GCEJ7DMULFTT25UH4FAAGOZ6KER4WXAYQGJUSIITQD527DGTKSXKBQGR) |
| TF Team Wisdom | !!wiki.include page:'manual:gaqxblfg4bzgivy6dbjvwe5eap3unhmia2pycuvly2juspvwpuf36bw4.md' | [GAQXBLFG4BZGIVY6DBJVWE5EAP3UNHMIA2PYCUVLY2JUSPVWPUF36BW4](https://stellar.expert/explorer/public/account/GAQXBLFG4BZGIVY6DBJVWE5EAP3UNHMIA2PYCUVLY2JUSPVWPUF36BW4) |
| Wisdom Council Locked | !!wiki.include page:'manual:gaugosylcx7jztqyf2k7rimhfwksa3wsi2oq4irkxmdmve6abjijmfqr.md' | [GAUGOSYLCX7JZTQYF2K7RIMHFWKSA3WSI2OQ4IRKXMDMVE6ABJIJMFQR](https://stellar.expert/explorer/public/account/GAUGOSYLCX7JZTQYF2K7RIMHFWKSA3WSI2OQ4IRKXMDMVE6ABJIJMFQR) |
## Important Note
ThreeFold DMCC (Dubai) is in the process of acquiring a substantial number of tokens. While these tokens possess liquidity from a technical standpoint, they are not currently accessible or traded on the open market. This reserve of tokens has been allocated for our upcoming commercial rollout, and their governance will be managed through consensus based system with input from the community.
## Remarks
- All wisdom council wallets are protected by multisignature of the members of the wisdom council
- All foundation wallets are protected by members of the foundation (4 on 6 need to sign)
- Signatures can be checked by going to detail of account and then to the stellar link
- The foundation will never spend tokens if the markets cannot support it and all proceeds are 100% used for the benefit of the ThreeFold project.
## Proof-of-Utilization Wallets
There are some wallets associated with [proof-of-utilization](proof_of_utilization.md). These wallets are on TFChain.
The addresses are the following:
- Mainnet ThreeFold Foundation: 5DCaGQfz2PH35EMJTHFMjc6Tk5SkqhjekVvrycY5M5xiYzis
- Mainnet Default Solution Provider: 5Dd6adUJH8wvqb9SPC96JdZ85nK1671MeMSxkPZ6Q7rE4byc
- Testnet ThreeFold Foundation: 5H6XYX17yJyjazoLVZqxxEPwMdGn99wginjmFBKtjvk8iJ3e
- Testnet Default Solution Provider: 5Esq6iLLBGGJFsCEXpoFhxHhqcaGqTvDasdwy8jPFDH1jYaM
- Staking Pool: 5CNposRewardAccount11111111111111111111111111FSU
To check the balance of any of those wallets, follow those steps:
- Go to the Polkadot API ([Mainnet](https://polkadot.js.org/apps/?rpc=wss%3A%2F%2Ftfchain.grid.tf#/chainstate), [Testnet](https://polkadot.js.org/apps/?rpc=wss%3A%2F%2Ftfchain.test.grid.tf#/chainstate))
- Under `selected state query`, select `system`
- On the right drop down menu, select `account(AccountId32): FrameSystemAccountInfo`
- Under `Option<AccountId32>`, write the wallet address of one of the three accounts displayed above
- Click on the `plus` button on the far right of the `selected state query` line.
As a general example, here's what it looks like:
![Wallet example](./img/polkadot_wallet_example.png)
Here are the outputs for three wallets shown above:
- Mainnet ThreeFold Foundation
![Mainnet TF Foundation Wallet](./img/wallet_tf_foundation_main.png)
- Mainnet Default Solution Provider
![Mainnet Solution Provider Wallet](./img/wallet_solution_provider_main.png)
- Testnet ThreeFold Foundation
![Testnet TF Foundation Wallet](./img/wallet_tf_foundation_test.png)
- Testnet Default Solution Provider
![Testnet Solution Provider Wallet](./img/wallet_solution_provider_test.png)
- Staking Pool
![Staking Pool Wallet](./img/wallet_staking_pool.png)
> Note: To get the proper TFT amount, you need to account fo the fact that TFT uses 7 decimal places. For this reason, to get the proper quantity in TFT, move the decimal place by dividing by 1e7 (i.e. 1x10⁷).

View File

@ -1,89 +0,0 @@
<h1> ThreeFold Token Overview </h1>
<h2>Table of Contents </h2>
- [Introduction to TFT](#introduction-to-tft)
- [Proof-of-Capacity](#proof-of-capacity)
- [Proof-of-Utilization](#proof-of-utilization)
- [Proof-of-Utility Distribution Flow](#proof-of-utility-distribution-flow)
- [TFT Distribution](#tft-distribution)
- [TFT Marketcap and Market Price](#tft-marketcap-and-market-price)
- [Complemetary Information](#complemetary-information)
- [Disclaimer](#disclaimer)
***
## Introduction to TFT
ThreeFold tokens, or TFTs, are exclusively generated when new capacity is added to the TF Grid. There are no centralized issuers. Tokens have not been created out of thin air.
While the ThreeFold Grid can expand, a maximum of 1 billion TFTs can ever be in circulation. This limit ensures stability of value and incentivization for all stakeholders.
TFT lives on the Stellar Blockchain. TFT holders benefit from a big ecosystem of proven wallets and mediums of exchange.
By employing Stellar technology, TFT transactions and smart contracts are powered by one of the most energy-efficient blockchains available. Furthermore, TFT is the medium of exchange on the greenest internet network in the world. The market for farming, cultivating and trading TFT is open to all.
Anyone with internet connection, power supply and necessary hardware can become a Farmer or trade ThreeFold tokens (TFT).
By farming, buying, holding, and utilizing ThreeFold Tokens, you are actively supporting the expansion of the ThreeFold Grid and its use cases — creating a more sustainable, fair, and equally accessible Internet.
## Proof-of-Capacity
ThreeFold uses proof-of-capacity to mint tokens. Since the genenis pool, all tokens that are being minted are the result of farming. Minting will stop during 2024, to keep the total amount of TFT at 1 billion, instead of the previously planned 4 billion. Read more about this [here](https://forum.threefold.io/t/end-feb-2024-update-from-the-team/4233).
> For more details, see [Proof of Capacity](proof_of_capacity.md)
## Proof-of-Utilization
TFT is used on the TFGrid to purchase network, compute and storage resources through the proof-of-utilization protocol.
### Proof-of-Utility Distribution Flow
![](img/token_distribution.png)
> For more details, see [Proof-of-Utilization](proof_of_utilization.md)
## TFT Distribution
The supply distribution of TFT is as follows:
| Supply Distribution | Qty (Millions) |
| ------------------- | -------------- |
| Total supply | 942 |
| TF Foundation Supply | 162 |
| Circulating supply | 780 |
| Maximum supply | 1000 |
The total supply of TFT is distributed as follows:
| Total Supply Distribution | Qty (Millions) |
| ------------------------------------------- | -------------- |
| TF Foundation: Ecosystem Grants | 22 |
| TF Foundation: Promotion & Marketing Effort | 100 |
| TF Foundation: Ecosystem Contribution & Liquidity Exchanges | 40 |
| Genesis Pool & Farming Rewards | 780 |
## TFT Marketcap and Market Price
The TFT market price and marketcap are as follows:
| **Description** | **Value** |
| ------------------------- | ------------- |
| TFT Market Price | 0.028 USD |
| TFT Market Cap | 21,840,000 USD |
The market cap is equal to the product of the TFT market price and the circulating supply.
> Market Cap = (TFT Market Price) X (TFT Circulating Supply)
The values here are subject to change. Check the current market conditions.
## Complemetary Information
- [ThreeFold History](threefold_history.md)
- [Token History](token_history.md)
## Disclaimer
> Important Note: The ThreeFold Token (TFT) is not an investment instrument.
TFTs represent IT capacity on the ThreeFold Grid, farmers create TFT, developers use TFT.

View File

@ -8,9 +8,5 @@ To deploy on the ThreeFold Grid, refer to the [System Administrators](system_adm
- [Cloud Units](cloudunits.md)
- [Pricing](pricing_toc.md)
- [Pricing Overview](pricing.md)
- [Staking Discounts](staking_discount_levels.md)
- [Cloud Pricing Compare](cloud_pricing_compare.md)
- [Grid Billing](grid_billing.md)
- [Resource Units](resource_units_calc_cloudunits.md)
- [Resource Units Advanced](resourceunits_advanced.md)

View File

@ -23,7 +23,7 @@ Resource units are used to measure and convert capacity on the hardware level in
| ------------ | ------------------------------------ | ---- |
| Core Unit | 1 Logical Core (Hyperthreaded Core) | CRU |
| Mem Unit | 1 GB mem | MRU |
| HD Unit | 1 GB | HRU |
| HDD Unit | 1 GB | HRU |
| SSD Unit | 1 GB | SRU |
| Network Unit | 1 GB of bandwidth transmitted in/out | NRU |

View File

@ -19,7 +19,7 @@
## Introduction
The [__ThreeFold Circle Tool__](https://circles.threefold.me ) is our own self-hosted (desktop only) project management tool based on [Taiga](https://www.taiga.io/), an open-source project management tool for cross-functional agile. It offers a lot of different project management kits and features such as the scrum board, kanban board, issues management, and many more.
The [__ThreeFold Circle Tool__](https://circles.threefold.me ) is our own self-hosted (desktop only) project management tool based on [Taiga](https://taiga.io/), an open-source project management tool for cross-functional agile. It offers a lot of different project management kits and features such as the scrum board, kanban board, issues management, and many more.
Our teams at ThreeFold use the Circle Tool to self-manage our tasks, thus it is deemed necessary for the new onboarded team members to learn how to use the tool. Unfortunately we only provide the desktop version of the tool at this moment since we normally manage our projects on the computer.

View File

@ -15,10 +15,10 @@ The backend for the weblets is introduced with the [Javascript Client](developer
<h2> Table of Contents </h2>
- [Wallet Connector](wallet_connector.md)
- [TFGrid](tfgrid.md)
- [Deploy](deploy.md)
- [Farms](farms.md)
- [TFChain](tfchain.md)
- [TFGrid](tfgrid/tfgrid.md)
- [Deploy](deploy/deploy.md)
- [Farms](farms/farms.md)
- [TFChain](tfchain/tfchain.md)
## Advantages
@ -42,11 +42,12 @@ You can access the ThreeFold Dashboard on different TF Chain networks.
- Deploys one thing at a time.
- Might take sometime to deploy a solution like Peertube, so you should wait a little bit until it's fully running.
## Dashboard Backups
## List of Mainnet Backend Stacks
If the main Dashboard URLs are not working for any reason, the following URLs can be used. Those Dashboard URLs are fully independent of the main Dashboard URLs shown above.
We provide independent mainnet backend stacks. Here is the current list:
- [https://dashboard.02.dev.grid.tf](https://dashboard.02.dev.grid.tf) for Dev net
- [https://dashboard.02.qa.grid.tf](https://dashboard.02.qa.grid.tf) for QA net
- [https://dashboard.02.test.grid.tf](https://dashboard.02.test.grid.tf) for Test net
- [https://dashboard.02.grid.tf](https://dashboard.02.grid.tf) for Main net
- [https://dashboard.grid.tf](https://dashboard.grid.tf)
- [https://dashboard.be.grid.tf](https://dashboard.be.grid.tf)
- [https://dashboard.fin.grid.tf](https://dashboard.fin.grid.tf)
- [https://dashboard.sg.grid.tf](https://dashboard.sg.grid.tf)
- [https://dashboard.us.grid.tf](https://dashboard.us.grid.tf)

View File

@ -11,14 +11,20 @@ Easily deploy your favourite applications on the ThreeFold grid with a click of
- [Algorand](algorand.md)
- [CasperLabs](casper.md)
- [Discourse](discourse.md)
- [Domains](domains.md)
- [Funkwhale](funkwhale.md)
- [Gitea](gitea.md)
- [Jenkins](jenkins.md)
- [Jitsi](jitsi.md)
- [Mattermost](mattermost.md)
- [Nextcloud](nextcloud.md)
- [Node Pilot](nodepilot.md)
- [Nostr](nostr.md)
- [Peertube](peertube.md)
- [Presearch](presearch.md)
- [Static Website](static_website.md)
- [Subsquid](subsquid.md)
- [Taiga](taiga.md)
- [TFRobot](tfrobot.md)
- [Umbrel](umbrel.md)
- [WordPress](wordpress.md)

View File

@ -3,3 +3,5 @@
Find or Publish your Flist from [Zero-OS Hub](https://hub.grid.tf/)
![](../img/0_hub.png)
Learn more about the Zero-OS Hub [here](developers@@zos_hub).

View File

@ -45,11 +45,14 @@ If your deployment has some minimum requirements, you can easily filter relevant
## Node Details
You can see all of the node details when you click on its row.
You can see all of the node details when you click on its row:
![](../img/dashboard_node_finder_node_view.png)
Note that the network speed test displayed in the Node Finder is updated every 6 hours.
> Note: The network speed test displayed in the Node Finder is updated every 6 hours.
To access the public Grafana page displaying additional information, click on `Check Node Health`:
![](../img/node_finder_grafana.png)
## Gateway Nodes

View File

@ -2,24 +2,24 @@
This comprehensive guide aims to provide users with detailed instructions and insights into efficiently managing their _Farms_. Farms encompass servers and storage devices contributing computational and storage capabilities to the grid, empowering users to oversee, maintain, and optimize their resources effectively.
- [Getting started](#getting-started)
- [Create a new Farm](#create-a-new-farm)
- [Getting Started](#getting-started)
- [Create a New Farm](#create-a-new-farm)
- [Manage Your Farms](#manage-your-farms)
- [Add a public IP to your Farm](#add-a-public-ip-to-your-farm)
- [Add a Stellar address for payout](#add-a-stellar-address-for-payout)
- [Generate your node bootstrap image](#generate-your-node-bootstrap-image)
- [Additional information](#additional-information)
- [Add a Public IP to Your Farm](#add-a-public-ip-to-your-farm)
- [Add a Stellar Address for Payout](#add-a-stellar-address-for-payout)
- [Generate Your Node Bootstrap Image](#generate-your-node-bootstrap-image)
- [Additional Information](#additional-information)
- [Manage Your Nodes](#manage-your-nodes)
- [Node information](#node-information)
- [Node Information](#node-information)
- [Extra Fees](#extra-fees)
- [Public Configuration](#public-configuration)
- [The Difference Between IPs Assigned to Nodes Versus a Farm](#the-difference-between-ips-assigned-to-nodes-versus-a-farm)
## Getting started
## Getting Started
After logging in to the TF Dashboard, on the sidebar click on **Dashboard** then _Your Farms_ .
## Create a new Farm
## Create a New Farm
If you want to start farming, you need a farmID, the ID of the farm that is owning the hardware node(s) you connect to the TFGrid.
@ -47,7 +47,7 @@ You can browse your Farms in _Farms_ table; Farms table contains all your own fa
![](../img/dashboard_farms_farms_table.png)
### Add a public IP to your Farm
### Add a Public IP to Your Farm
If you have public IPv4 addresses available that can be used for usage on the TFGrid, you can add them in your farm.
Click `ADD IP`, specify the addresses, the gateway and click `CREATE`.
@ -69,7 +69,7 @@ Deleting IPv4 addresses is also possible here. The `Deployed Contract ID` gives
![ ](../img/dashboard_farms_ip_details.png)
### Add a Stellar address for payout
### Add a Stellar Address for Payout
In a first phase, farming of tokens still results in payout on the Stellar network. So to get the farming reward, a Stellar address needs to be provided.
@ -79,7 +79,7 @@ In a first phase, farming of tokens still results in payout on the Stellar netwo
You can read about different ways to store TFT [here](threefold_token@@storing_tft). Make sure to use a Stellar wallet for your farming rewards.
### Generate your node bootstrap image
### Generate Your Node Bootstrap Image
Once you know your farmID, you can set up your node on TFGrid3. Click on `Bootstrap Node Image`.
@ -87,7 +87,7 @@ Once you know your farmID, you can set up your node on TFGrid3. Click on `Bootst
Read more Zero-OS bootstrap image [here](farmers@@2_bootstrap_image).
### Additional information
### Additional Information
After booting a node, the info will become available in `Your Nodes` table, including the status info along with the minting and fixup receipts.
@ -103,7 +103,7 @@ You can also download a single node's receipts using the `Download Receipts` but
as in farms table _Nodes_ table contains all your own nodes and its your entry point to manage your farm as in the following sections.
### Node information
### Node Information
Expand your node information by clicking on the expand button in the target node row.

Binary file not shown.

Before

Width:  |  Height:  |  Size: 188 KiB

After

Width:  |  Height:  |  Size: 183 KiB

Binary file not shown.

After

Width:  |  Height:  |  Size: 162 KiB

View File

@ -13,7 +13,7 @@
## Introduction
We cover the overall process to add a domain to a virtual machine running on the ThreeFold Grid.
We cover the overall process to add a domain to a virtual machine running on the ThreeFold Grid. This process works for both micro and full virtual machines.
## Preparation

View File

@ -5,7 +5,7 @@
- [Introduction](#introduction)
- [Requirements](#requirements)
- [Configs Tab](#configs-tab)
- [Admin and Workers Tabs](#admin-and-workers-tabs)
- [Leader and Workers Tabs](#leader-and-workers-tabs)
- [The Domain Name](#the-domain-name)
- [Domain Name Example](#domain-name-example)
- [How to Know the IP Address](#how-to-know-the-ip-address)
@ -31,10 +31,10 @@ Caprover is a very cool management app for containers based on Docker Swarm.
It has following benefits :
- easy to deploy apps (in seconds)
- easy to create new apps
- super good monitoring
- can be extended over the TFGrid
- Easy to deploy apps (in seconds)
- Easy to create new apps
- Super good monitoring
- Can be extended over the TFGrid
## Requirements
@ -46,23 +46,51 @@ It has following benefits :
![ ](./img/solutions_caprover.png)
- Enter domain for you Caprover instance, Be very careful about the domain name: it needs to be a wildcard domain name you can configure in your chosen domain name system.
- Enter domain for you Caprover instance.
- Be very careful about the domain name: it needs to be a wildcard domain name you can configure in your chosen domain name system.
- Enter password for you Caprover instance.
If you have more than one SSH keys set, you can click on `Manage SSH keys` to select which one to use for this deployment.
## Admin and Workers Tabs
## Leader and Workers Tabs
Each deployment will have one leader and there can be many workers. By default, CapRover is deployed on nodes with IPv4.
![ ](./img/solutions_caprover_leader.png)
![ ](./img/solutions_caprover_workers.png)
Use the Leader and Workers tabs to add nodes to your deployment.
- Enter a name for the deployment or keep the default name
- Select a capacity package:
- **Small**: {cpu: 1, memory: 2, diskSize: 25 }
- **Medium**: {cpu: 2, memory: 4, diskSize: 50 }
- **Large**: {cpu: 4, memory: 16, diskSize: 100 }
- Or choose a **Custom** plan
- Choose the network
- `Mycelium` flag gives the virtual machine a Mycelium address
- `Dedicated` flag to retrieve only dedicated nodes
- `Certified` flag to retrieve only certified nodes
- Choose the node
- Automated
- Choose the location of the node
- `Region`
- `Country`
- `Farm Name`
- Click on `Load Nodes`
- Click on the node you want to deploy on
- Manual selection
- Select a specific node ID
- Click `Deploy`
Note: Worker nodes only accept SSH keys of RSA format.
Deployment will take couple of minutes.
## The Domain Name
As per the [CapRover documentation](https://caprover.com/docs/get-started.html), you need to point a wildcard DNS entry to the VM IP address of your CapRover instance. You have to do this after having deployed the CapRover instance, otherwise you won't have access to the VM IP address.
As per the [CapRover documentation](https://caprover.com/docs/get-started.html), you need to point a wildcard DNS entry to the VM IP address of your CapRover Leader instance. You have to do this after having deployed the CapRover instance, otherwise you won't have access to the VM IP address.
Lets say your domain is **example.com** and your subdomain is **subdomain**. You can set **\*.subdomain.example.com** as an A record in your DNS settings to point to the VM IP address of the server hosting the CapRover instance, where **\*** acts as the wildcard. To do this, go to the DNS settings of your domain name registrar, and set a wild card A record entry.
@ -97,54 +125,14 @@ Go back to your CapRover weblet and go to the deployment list. Click on `Show De
![ ](./img/solution_caprover_list.png)
- The public IPv4 address is visible in here
![](./img/solutions_caprover_ipaddress.png)
- Now you can configure the domain name (see above, don't forget to point the wildcard domain to the public IP address)
Click on details if you want to see more details
Go to the `JSON` tab to see the Json ouput:
```json
{
"version": 0,
"name": "caprover_leader_cr_156e44f0",
"created": 1637843368,
"status": "ok",
"message": "",
"flist": "https://hub.grid.tf/samehabouelsaad.3bot/tf-caprover-main-a4f186da8d.flist",
"publicIP": {
"ip": "185.206.122.136/24",
"gateway": "185.206.122.1"
},
"planetary": false,
"yggIP": "",
"interfaces": [
{
"network": "caprover_network_cr_156e44f0",
"ip": "10.200.4.2"
}
],
"capacity": {
"cpu": 4,
"memory": 8192
},
"mounts": [
{
"name": "data0",
"mountPoint": "/var/lib/docker",
"size": 107374182400,
"state": "ok",
"message": ""
}
],
"env": {
"SWM_NODE_MODE": "leader",
"CAPROVER_ROOT_DOMAIN": "apps.openly.life",
"PUBLIC_KEY": "ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABAQC/9RNGKRjHvViunSOXhBF7EumrWvmqAAVJSrfGdLaVasgaYK6tkTRDzpZNplh3Tk1aowneXnZffygzIIZ82FWQYBo04IBWwFDOsCawjVbuAfcd9ZslYEYB3QnxV6ogQ4rvXnJ7IHgm3E3SZvt2l45WIyFn6ZKuFifK1aXhZkxHIPf31q68R2idJ764EsfqXfaf3q8H3u4G0NjfWmdPm9nwf/RJDZO+KYFLQ9wXeqRn6u/mRx+u7UD+Uo0xgjRQk1m8V+KuLAmqAosFdlAq0pBO8lEBpSebYdvRWxpM0QSdNrYQcMLVRX7IehizyTt+5sYYbp6f11WWcxLx0QDsUZ/J"
},
"entrypoint": "/sbin/zinit init",
"metadata": "",
"description": "caprover leader machine/node"
}
```
![](./img/solutions_caprover_json.png)
## How to Access the Admin Interface

View File

@ -30,8 +30,8 @@ __Process__ :
- Or choose a **Custom** plan
- Choose the network
- `Public IPv4` flag gives the virtual machine a Public IPv4
- `Dedicated` flag to retrieve only dedeicated nodes
- `Mycelium` flag gives the virtual machine a Mycelium address
- `Dedicated` flag to retrieve only dedicated nodes
- `Certified` flag to retrieve only certified nodes
- Choose the location of the node
- `Region`

View File

@ -27,8 +27,10 @@
- **Medium**: {cpu: 2, memory: 4, diskSize: 50 }
- **Large**: {cpu: 4, memory: 16, diskSize: 100 }
- Or choose a **Custom** plan
- `Dedicated` flag to retrieve only dedeicated nodes
- Choose the network
- `Public IPv4` flag gives the virtual machine a Public IPv4
- `Mycelium` flag gives the virtual machine a Mycelium address
- `Dedicated` flag to retrieve only dedicated nodes
- `Certified` flag to retrieve only certified nodes
- Choose the location of the node
- `Region`

View File

@ -0,0 +1,36 @@
<h1> Domains </h1>
<h2>Table of Contents </h2>
- [Introduction](#introduction)
- [Prerequisites](#prerequisites)
- [Deployment](#deployment)
---
## Introduction
The Domains solution allows users to securely expose servers hosted on local machines or VMs to the public internet. Users are required to specify the machine's IP, which can be a Mycelium IP, an Yggdrasil IP, or a public IP (IPv4 or IPv6).
## Prerequisites
- Make sure you have a [wallet](wallet_connector.md)
- From the sidebar click on **Applications**
- Click on **Domains**
## Deployment
![ ](./img/domains.png)
- Enter a subdomain to be added to your domain
- The `Custom Domain` flag allows the user to use a custom domain
- Choose a gateway node to deploy your static website
- Enter the port used to access the machine
- Enter the machine's IP
- It can be a Mycelium IP, an Yggdrasil IP, or a public IP (IPv4 or IPv6)
Once this is done, you can see a list of all of your deployed instances:
![ ](./img/domains_list.png)
Click on the button **Visit** under **Actions** to go to your exposed website!

View File

@ -40,7 +40,7 @@ Deploy a new full virtual machine on the Threefold Grid
- `Public IPv4` flag gives the virtual machine a Public IPv4
- `Public IPv6` flag gives the virtual machine a Public IPv6
- `Planetary Network` to connect the Virtual Machine to Planetary network
- `Myceluim` to enable mycelium on the virtual machine
- `Mycelium` to enable Mycelium on the virtual machine
- `Wireguard Access` to add a wireguard access to the Virtual Machine
- `GPU` flag to add GPU to the Virtual machine
- To deploy a Full VM with GPU, you first need to [rent a dedicated node](node_finder.md#dedicated-nodes)

View File

@ -37,8 +37,8 @@ __Process__ :
- Or choose a **Custom** plan
- Choose the network
- `Public IPv4` flag gives the virtual machine a Public IPv4
- `Dedicated` flag to retrieve only dedeicated nodes
- `Mycelium` flag gives the virtual machine a Mycelium address
- `Dedicated` flag to retrieve only dedicated nodes
- `Certified` flag to retrieve only certified nodes
- Choose the location of the node
- `Region`

View File

@ -0,0 +1,81 @@
<h1> Gitea </h1>
<h2>Table of Contents</h2>
- [Introduction](#introduction)
- [Prerequisites](#prerequisites)
- [Deployment](#deployment)
- [Setting Up Gitea](#setting-up-gitea)
***
## Introduction
[Gitea](https://about.gitea.com/) is a forge software package for hosting software development version control using Git as well as other collaborative features like bug tracking, code review, continuous integration, kanban boards, tickets, and wikis.
## Prerequisites
- Make sure you have a [wallet](wallet_connector.md)
- From the sidebar click on **Applications**
- Click on **Gitea**
## Deployment
![](./img/gitea_config.png)
![](./img/gitea_mail_server.png)
- Enter a name for your deployment or keep the default name
- Select a capacity package:
- **Small**: {cpu: 1, memory: 2, diskSize: 25 }
- **Medium**: {cpu: 2, memory: 4, diskSize: 50 }
- **Large**: {cpu: 4, memory: 16, diskSize: 100 }
- Or choose a **Custom** plan
- Choose the network
- `Public IPv4` flag gives the virtual machine a Public IPv4
- `Public IPv6` flag gives the virtual machine a Public IPv6
- `Planetary Network` flag gives the virtual machine an Yggdrasil address
- `Mycelium` flag gives the virtual machine a Mycelium address
- `Dedicated` flag to retrieve only dedicated nodes
- `Certified` flag to retrieve only certified nodes
- Choose the node
- Automated
- Choose the location of the node
- `Region`
- `Country`
- `Farm Name`
- Click on `Load Nodes`
- Click on the node you want to deploy on
- Manual selection
- Select a specific node ID
- Choose a domain
- Use a custom domain by enabling `Custom Domain`
- Choose a gateway domain for your instance
- Set the `Mail Server`
- In the tab `Mail Server` you can adjust the parameters or leave the default parameters
- The `Admin Email` by default is the email you used when logging into the Dashboard
- The `Admin Password` is generated automatically but you can modify it
- By default the SMTP server used for the `Hostname` is `smtp.gmail.com` with its associated port `587`
- Enable TLS by clicking on the `TLS` button
- Click `Deploy`
After deployment, you can see a list of all your deployed instances.
To access the Gitea instance, click on the `Actions` button called `Visit`.
![](./img/gitea_instances.png)
## Setting Up Gitea
Once you've accessed the Gitea instance, you can adjust the `Initial Configuration` if needed.
![](./img/gitea_initial_configuration.png)
Once you're ready, click on `Install Gitea`.
![](./img/gitea_install.png)
Once the configuration is completed, you can register to the Gitea instance and start using it.
![](./img/gitea_register.png)

View File

@ -0,0 +1 @@
solutions_jitsi_homepage.png

Binary file not shown.

After

Width:  |  Height:  |  Size: 49 KiB

Binary file not shown.

After

Width:  |  Height:  |  Size: 45 KiB

Binary file not shown.

After

Width:  |  Height:  |  Size: 47 KiB

Binary file not shown.

After

Width:  |  Height:  |  Size: 55 KiB

Binary file not shown.

After

Width:  |  Height:  |  Size: 82 KiB

Binary file not shown.

After

Width:  |  Height:  |  Size: 37 KiB

Binary file not shown.

After

Width:  |  Height:  |  Size: 39 KiB

Binary file not shown.

After

Width:  |  Height:  |  Size: 27 KiB

Binary file not shown.

After

Width:  |  Height:  |  Size: 94 KiB

Binary file not shown.

After

Width:  |  Height:  |  Size: 308 KiB

Binary file not shown.

After

Width:  |  Height:  |  Size: 48 KiB

Binary file not shown.

After

Width:  |  Height:  |  Size: 36 KiB

Binary file not shown.

Before

Width:  |  Height:  |  Size: 29 KiB

Binary file not shown.

Before

Width:  |  Height:  |  Size: 73 KiB

After

Width:  |  Height:  |  Size: 29 KiB

Binary file not shown.

After

Width:  |  Height:  |  Size: 107 KiB

Binary file not shown.

After

Width:  |  Height:  |  Size: 48 KiB

Binary file not shown.

After

Width:  |  Height:  |  Size: 353 KiB

Binary file not shown.

After

Width:  |  Height:  |  Size: 59 KiB

Binary file not shown.

After

Width:  |  Height:  |  Size: 42 KiB

Binary file not shown.

After

Width:  |  Height:  |  Size: 49 KiB

View File

@ -0,0 +1,65 @@
<h1> Jenkins </h1>
<h2>Table of Contents</h2>
- [Introduction](#introduction)
- [Prerequisites](#prerequisites)
- [Deployment](#deployment)
- [Troubleshooting](#troubleshooting)
***
## Introduction
[Jenkins](https://www.jenkins.io/) is a popular open-source automation server that enables developers to build, test, and deploy their applications continuously.
## Prerequisites
- Make sure you have a [wallet](wallet_connector.md)
- From the sidebar click on **Applications**
- Click on **Jenkins**
## Deployment
__Process__ :
![ ](./img/solutions_jenkins1.png)
- Enter an Application Name. It's used in generating a unique subdomain on one of the gateways on the network alongside your twin ID. Ex. ***fw100myfunk*.gent02.dev.grid.tf**
- Enter administrator information including **Username**, and **Password**. This admin user will have full permission on the deployed instance.
- Select a capacity package:
- **Small**: {cpu: 1, memory: 2, diskSize: 50 }
- **Medium**: {cpu: 2, memory: 4, diskSize: 100 }
- **Large**: {cpu: 4, memory: 16, diskSize: 250 }
- Or choose a **Custom** plan
- Choose the network
- `Public IPv4` flag gives the virtual machine a Public IPv4
- `Public IPv6` flag gives the virtual machine a Public IPv6
- `Planetary Network` to connect the Virtual Machine to Planetary network
- `Mycelium` flag gives the virtual machine a Mycelium address
- `Dedicated` flag to retrieve only dedicated nodes
- `Certified` flag to retrieve only certified nodes
- Choose the location of the node
- `Region`
- `Country`
- `Farm Name`
- Choose the node to deploy on
> Or you can select a specific node with manual selection.
- `Custom Domain` flag lets the user to use a custom domain
- Choose a gateway node to deploy your Jenkins instance on.
After that is done you can see a list of all of your deployed instances
![ ](./img/jenkins2.png)
Click on ***Visit*** to go to the homepage of your Jenkins instance!
![ ](./img/jenkins3.png)
## Troubleshooting
If you get a `Bad Gateway` while connecting to the website, you might simply need to wait for the deployment to complete.

View File

@ -0,0 +1,63 @@
<h1> Jitsi </h1>
<h2>Table of Contents</h2>
- [Introduction](#introduction)
- [Prerequisites](#prerequisites)
- [Deployment](#deployment)
- [Troubleshooting](#troubleshooting)
---
## Introduction
Jitsi Meet is a set of Open Source projects which empower users to use and deploy video conferencing platforms with state-of-the-art video quality and features. For more details, check [Jitsi documentation](https://jitsi.github.io/handbook/docs/intro/).
## Prerequisites
- Make sure you have a [wallet](wallet_connector.md)
- From the sidebar click on **Applications**
- Click on **Jitsi**
## Deployment
**Process** :
![ ](./img/solutions_jitsi.png)
- Enter an Application Name. It's used in generating a unique subdomain on one of the gateways on the network alongside your twin ID. Ex. **_fw100myfunk_.gent02.dev.grid.tf**
- Select a capacity package:
- **Small**: {cpu: 1, memory: 2, diskSize: 50 }
- **Medium**: {cpu: 2, memory: 4, diskSize: 100 }
- **Large**: {cpu: 4, memory: 16, diskSize: 250 }
- Or choose a **Custom** plan
- Choose the network
- `Public IPv4` flag gives the virtual machine a Public IPv4
- `Public IPv6` flag gives the virtual machine a Public IPv6
- `Planetary Network` flag gives the virtual machine an Yggdrasil address
- `Mycelium` flag gives the virtual machine a Mycelium address
- `Dedicated` flag to retrieve only dedicated nodes
- `Certified` flag to retrieve only certified nodes
- Choose the location of the node
- `Region`
- `Country`
- `Farm Name`
- Choose the node to deploy on
> Or you can select a specific node with manual selection.
- `Custom Domain` flag lets the user to use a custom domain
- Choose a gateway node to deploy your Jitsi instance on.
After that is done you can see a list of all of your deployed instances
![ ](./img/solutions_jitsi_instances.png)
Click on **_Visit_** to go to the homepage of your Jitsi instance!
![ ](./img/solutions_jitsi_homepage.png)
## Troubleshooting
If you get a `Bad Gateway` while connecting to the website, you might simply need to wait for the deployment to complete.

View File

@ -41,7 +41,7 @@ On the TF grid, Kubernetes clusters can be deployed out of the box. We have impl
## Kubeconfig
Once the cluster is ready, you can SSH into the cluster using `ssh root@IP`
> IP can be the public IP or the planetary network IP
> IP can be the public IP, Mycelium or the Planetary Network IP
Onced connected via SSH, you can execute commands on the cluster like `kubectl get nodes`, and to get the kubeconfig, you can find it in `/root/.kube/config`

View File

@ -29,7 +29,10 @@
- **Medium**: {cpu: 2, memory: 4, diskSize: 50 }
- **Large**: {cpu: 4, memory: 16, diskSize: 100 }
- Or choose a **Custom** plan
- `Dedicated` flag to retrieve only dedeicated nodes
- Choose the network
- `Public IPv4` flag gives the virtual machine a Public IPv4
- `Mycelium` flag gives the virtual machine a Mycelium address
- `Dedicated` flag to retrieve only dedicated nodes
- `Certified` flag to retrieve only certified nodes
- Choose the location of the node
- `Region`

View File

@ -16,13 +16,16 @@
- [TURN](#turn)
- [Use Talk](#use-talk)
- [Backups and Updates](#backups-and-updates)
- [Create a Backup](#create-a-backup)
- [Automatic Backups and Updates](#automatic-backups-and-updates)
- [BorgBackup](#borgbackup)
- [Create a BorgBackup](#create-a-borgbackup)
- [Automatic Borg Backups and Updates](#automatic-borg-backups-and-updates)
- [Secondary VM Backup](#secondary-vm-backup)
- [SSH Connection](#ssh-connection)
- [Secondary VM Backup Script](#secondary-vm-backup-script)
- [Troubleshooting](#troubleshooting)
- [Retrieve the Nextcloud AIO Password](#retrieve-the-nextcloud-aio-password)
- [Access the Nextcloud Interface Page](#access-the-nextcloud-interface-page)
- [Check the DNS Propagation](#check-the-dns-propagation)
- [Questions and Feedback](#questions-and-feedback)
***
@ -62,12 +65,15 @@ If you're not sure and just want the easiest, most affordable option, skip the p
* **Standard**: {cpu: 2, memory: 8gb, diskSize: 500gb }
* **Recommended**: {cpu: 4, memory: 16gb, diskSize: 1000gb }
* Or choose a **Custom** plan
* If want to reserve a public IPv4 address, click on Network then select **Public IPv4**
* If you want a [dedicated node](node_finder.md#dedicated-nodes) and/or a certified node, select the corresponding option
* Choose the location of the node
* `Country`
* `Farm Name`
* Select a node
- Choose the network
- `Public IPv4` flag gives the virtual machine a Public IPv4
- `Mycelium` flag gives the virtual machine a Mycelium address
- `Dedicated` flag to retrieve only dedicated nodes
- `Certified` flag to retrieve only certified nodes
- Choose the location of the node
- `Region`
- `Country`
- `Farm Name`
* If you want to use a custom domain, click on **Custom domain** under **Domain Name** and write your domain name
* Example: `nextcloudwebsite.com`
* The **Select gateway** box will be visible whenever a gateway is required. If so, click it and choose a gateway
@ -156,7 +162,11 @@ Note that the host of the video meeting might need to turn the VPN off before cr
# Backups and Updates
## Create a Backup
In this section, we cover how to make a BorgBackup on the Nextcloud VM and we also cover how to make a backup of the Nextcloud BorgBackup to a secondary VM for additional redundancy.
## BorgBackup
### Create a BorgBackup
In the section **Backup and restore**, you can set a [BorgBackup](https://www.borgbackup.org/) of your Nextcloud instance.
@ -166,7 +176,7 @@ In the section **Backup and restore**, you can set a [BorgBackup](https://www.bo
* This will stop all containers, run the backup container and create the backup.
* Once the backup is complete, you can click on **Start containers** to restart the Nextcloud instance.
## Automatic Backups and Updates
### Automatic Borg Backups and Updates
After the first manual backup of your Nextcloud instance is complete, you can set automatic backups and updates.
@ -174,7 +184,88 @@ After the first manual backup of your Nextcloud instance is complete, you can se
* In the section **Daily backup and automatic updates**, choose a time for your daily backup and click **Submit backup time**.
* To set automatic updates, make sure that the option **Automatically update all containers, the mastercontainer and on** is selected.
## Secondary VM Backup
To allow for another layer of redundancy, you can set a secondary VM on the grid and make a daily backup from the BorgBackup of your Nextcloud instance to the secondary VM. The following shows how to do this. It is based on the [File Transfer section](system_administrators@@file_transfer) of the manual.
For the following, we take into account that the BorgBackup is located at `/mnt/backup` on the VM running Nextcloud.
You will need to deploy a full VM on the TFGrid and SSH into this secondary VM.
### SSH Connection
We want to set an SSH connection between the Nextcloud VM and the secondary VM.
* Create SSH key pair on the secondary VM
```
ssh-keygen
```
* Install openssh-client on the secondary VM
```
apt install openssh-client
```
* Install openssh-server on the Nextcloud VM
```
apt install openssh-server
```
* Copy the public key of the secondary VM
```
cat ~/.ssh/id_rsa.pub
```
* Create the SSH directory on the Nextcloud VM
```
mkdir ~/.ssh
```
* Add the secondary VM public key in the file **authorized_keys** on the Nextcloud VM
```
nano ~/.ssh/authorized_keys
```
* Check the openssh-server status
```
service ssh status
```
* As a test, you can check if you can SSH into the Nextcloud VM from the secondary VM
```
ssh root@<Nextcloud_VM_IP_Address>
```
Once the SSH connection is set up, we need to prepare a backup script. This is covered in the next subsection.
### Secondary VM Backup Script
We now cover how to set up a backup script on the secondary VM.
On the secondary VM, do the following:
* Set a directory for your Nextcloud backup
```
mkdir -p /root/nextcloud_backup
```
* Create the script file
```
nano /root/rsync_nextcloud_backup.sh
```
* Write the following script. Here the log is saved in the same directory.
```
#!/bin/bash
sudo rsync -avz --progress --delete --log-file=/root/nextcloud_backup/rsync_nextcloud_storage.log /root/nextcloud_backup/ root@<Nextcloud_VM_IP_Address>:/mnt/backup
```
* Give permission to execute the script
```
sudo chmod +x /root/rsync_nextcloud_backup.sh
```
* Set a cron job to run the script periodically
* Open the cron file
```
sudo crontab -e
```
* At the bottom of the cron file, add the following to run the script everyday. For this example, we set the time at 18:00PM
```
0 18 * * * /root/rsync_nextcloud_backup.sh
```
Once this is set up, this secondary VM will make a daily backup of the Nextcloud BorgBackup. It is a good idea to set this secondary backup to run the script a few hours after the daily BorgBackup.
# Troubleshooting
@ -200,8 +291,3 @@ To access the Nextcloud interface page, follow those stepse
You can check if the DNS records are propagated globally with DNS propagation check services such as [DNS Checker](https://dnschecker.org/). You can use this tool to verify that your domain is properly pointing to the IPv4 address of the VM you deployed on.
# Questions and Feedback
If you have any questions, you can ask the ThreeFold community for help on the [ThreeFold Forum](http://forum.threefold.io/) or on the [ThreeFold Grid Tester Community](https://t.me/threefoldtesting) on Telegram.

View File

@ -24,9 +24,12 @@ This is a simple instance of upstream [Node Pilot](https://nodepilot.tech).
- Fill in the instance name: it's used to reference the node-pilot in the future.
- Minimum CPU allowed is 8 cores and minimum memory allowed is 8192.
- Minimum specs:
- 1 vcore
- 256 MB of memory
- 15 GB of storage
- `Dedicated` flag to retrieve only dedeicated nodes
- `Dedicated` flag to retrieve only dedicated nodes
- `Certified` flag to retrieve only certified nodes
- Choose the location of the node
@ -37,7 +40,7 @@ This is a simple instance of upstream [Node Pilot](https://nodepilot.tech).
> Or you can select a specific node with manual selection.
- When using the [flist](https://hub.grid.tf/tf-official-vms/node-pilot-zdbfs.flist) you get a node pilot instance ready out-of-box. You need to get a public ipv4 to get it to works.
- When using the [flist](https://hub.grid.tf/tf-official-vms/node-pilot-zdbfs.flist) you get a node pilot instance ready out-of-box. You need to get a public IPv4 to get it to works.
After that is done you can see a list of all of your deployed instances

View File

@ -0,0 +1,59 @@
<h1> Nostr </h1>
<h2>Table of Contents</h2>
- [Introduction](#introduction)
- [Prerequisites](#prerequisites)
- [Deployment](#deployment)
***
## Introduction
[Nostr](https://nostr.com/) is a decentralized network protocol for a distributed social networking system. The name is an acronym for "Notes and Other Stuff Transmitted by Relays". It was designed with goals of censorship-resistance in mind.
## Prerequisites
- Make sure you have a [wallet](wallet_connector.md)
- From the sidebar click on **Applications**
- Click on **Nostr**
## Deployment
![](./img/nostr_config.png)
- Enter a name for your deployment or keep the default name
- Select a capacity package:
- **Small**: {cpu: 1, memory: 2, diskSize: 25 }
- **Medium**: {cpu: 2, memory: 4, diskSize: 50 }
- **Large**: {cpu: 4, memory: 16, diskSize: 100 }
- Or choose a **Custom** plan
- Choose the network
- `Public IPv4` flag gives the virtual machine a Public IPv4
- `Public IPv6` flag gives the virtual machine a Public IPv6
- `Planetary Network` flag gives the virtual machine an Yggdrasil address
- `Mycelium` flag gives the virtual machine a Mycelium address
- `Dedicated` flag to retrieve only dedicated nodes
- `Certified` flag to retrieve only certified nodes
- Choose the node
- Automated
- Choose the location of the node
- `Region`
- `Country`
- `Farm Name`
- Click on `Load Nodes`
- Click on the node you want to deploy on
- Manual selection
- Select a specific node ID
- Choose a domain
- Use a custom domain by enabling `Custom Domain`
- Choose a gateway domain for your instance
- Click `Deploy`
After deployment, you can see a list of all your deployed instances.
To access the Nostr instance, click on the `Actions` button called `Visit`.
![](./img/nostr_instances.png)

View File

@ -30,12 +30,9 @@
- **Medium**: { cpu: 2, memory: 4, diskSize: 100 }
- **Large**: { cpu: 4, memory: 16, diskSize: 250 }
- Or choose a **Custom** plan
- `Public IPv4` flag gives the virtual machine a Public IPv4
- `Public IPv6` flag gives the virtual machine a Public IPv6
- `Planetary Network` to connect the Virtual Machine to Planetary network
- `Wiregaurd Access` to add a wiregaurd acces to the Virtual Machine
- `Dedicated` flag to retrieve only dedeicated nodes
- Choose the network
- `Mycelium` flag gives the virtual machine a Mycelium address
- `Dedicated` flag to retrieve only dedicated nodes
- `Certified` flag to retrieve only certified nodes
- Choose the location of the node
- `Region`

View File

@ -33,8 +33,9 @@
- Choose the network
- `Public IPv4` flag gives the virtual machine a Public IPv4
- `Planetary Network` to connect the Virtual Machine to Planetary network
- `Mycelium` flag gives the virtual machine a Mycelium address
- `Dedicated` flag to retrieve only dedeicated nodes
- `Dedicated` flag to retrieve only dedicated nodes
- `Certified` flag to retrieve only certified nodes
- Choose the location of the node
- `Region`

View File

@ -34,7 +34,9 @@ Static Website is an application where a user provides a GitHub repository URL f
- **Medium**: {cpu: 2, memory: 4, diskSize: 100 }
- **Large**: {cpu: 4, memory: 16, diskSize: 250 }
- Or choose a **Custom** plan
- Choose the network
- `Public IPv4` flag gives the virtual machine a Public IPv4
- `Mycelium` flag gives the virtual machine a Mycelium address
- `Dedicated` flag to retrieve only dedicated nodes
- `Certified` flag to retrieve only certified nodes
- Choose the location of the node

View File

@ -32,8 +32,9 @@
- **Medium**: {cpu: 2, memory: 4, diskSize: 100 }
- **Large**: {cpu: 4, memory: 16, diskSize: 250 }
- Or choose a **Custom** plan
- `Dedicated` flag to retrieve only dedeicated nodes
- Choose the network
- `Mycelium` flag gives the virtual machine a Mycelium address
- `Dedicated` flag to retrieve only dedicated nodes
- `Certified` flag to retrieve only certified nodes
- Choose the location of the node
- `Region`

View File

@ -11,7 +11,7 @@
## Introduction
[Taiga](https://www.taiga.io/) is the project management tool for multi-functional agile teams. It has a rich feature set and at the same time it is very simple to start with through its intuitive user interface.
[Taiga](https://taiga.io/) is the project management tool for multi-functional agile teams. It has a rich feature set and at the same time it is very simple to start with through its intuitive user interface.
## Prerequisites
@ -31,14 +31,16 @@
- **Medium**: {cpu: 4, memory: 8, diskSize: 150 }
- **Large**: {cpu: 4, memory: 16, diskSize: 250 }
- Or choose a **Custom** plan
- `Dedicated` flag to retrieve only dedeicated nodes
- Choose the network
- `Public IPv4` flag gives the virtual machine a Public IPv4
- `Mycelium` flag gives the virtual machine a Mycelium address
- `Dedicated` flag to retrieve only dedicated nodes
- `Certified` flag to retrieve only certified nodes
- Choose the location of the node
- `Region`
- `Country`
- `Farm Name`
- Choose the node to deploy the Tiaga instance on
- Choose the node to deploy the Taiga instance on
> Or you can select a specific node with manual selection.
- `Custom Domain` flag lets the user to use a custom domain
- Choose a gateway node to deploy your Funkwhale instance on.

View File

@ -0,0 +1,61 @@
<h1> TFRobot </h1>
<h2>Table of Contents</h2>
- [Introduction](#introduction)
- [Features](#features)
- [Prerequisites](#prerequisites)
- [Deployment](#deployment)
- [Deployed Instances Table](#deployed-instances-table)
***
## Introduction
[TFRobot](https://github.com/threefoldtech/tfgrid-sdk-go/blob/development/tfrobot/README.md) is tool designed to automate mass deployment of groups of VMs on the ThreeFold Grid, with support of multiple retries for failed deployments.
## Features
- **Mass Deployment:** Deploy groups of VMs on the grid simultaneously.
- **Mass Cancellation:** Cancel simultaneously all VMs on the grid defined in the configuration file.
- **Load Deployments:** Load simultaneously groups of VMs deployed with TFRobot.
- **Customizable Configurations:** Define node groups, VMs groups and other configurations through YAML or JSON files.
## Prerequisites
- Make sure you have a [wallet](wallet_connector.md)
- From the sidebar click on **Applications**
- Click on **TFRobot**
## Deployment
![ ](./img/solutions_tfrobot.png)
- Enter an Application Name.
- Select a capacity package:
- **Small**: {cpu: 1, memory: 2, diskSize: 25 }
- **Medium**: {cpu: 2, memory: 4, diskSize: 50 }
- **Large**: {cpu: 4, memory: 16, diskSize: 100 }
- Or choose a **Custom** plan
- Choose the network
- `Public IPv4` flag gives the virtual machine a Public IPv4
- `Public IPv6` flag gives the virtual machine a Public IPv6
- `Planetary Network` flag gives the virtual machine an Yggdrasil address
- `Mycelium` flag gives the virtual machine a Mycelium address
- `Wireguard Access` to add a WireGuard acces to the Virtual Machine
- `Dedicated` flag to retrieve only dedicated nodes
- `Certified` flag to retrieve only certified nodes
- Choose the location of the node
- `Region`
- `Country`
- `Farm Name`
- Click on `Load Nodes`
- Select the node you want to deploy on
- Click `Deploy`
## Deployed Instances Table
At all time, you can see a list of all of your deployed instances:
![ ](./img/tfrobot1.png)

View File

@ -31,8 +31,12 @@
- **Medium**: { cpu: 2, memory: 4 , diskSize: 50 }
- **Large**: { cpu: 4, memory: 16 , diskSize: 100 }
- Or choose a **Custom** plan
- `Dedicated` flag to retrieve only dedeicated nodes
- Choose the network
- `Public IPv4` flag gives the virtual machine a Public IPv4
- `Planetary Network` to connect the Virtual Machine to Planetary network
- `Mycelium` to enable Mycelium on the virtual machine
- `Wireguard Access` to add a wireguard acces to the Virtual Machine
- `Dedicated` flag to retrieve only dedicated nodes
- `Certified` flag to retrieve only certified nodes
- Choose the location of the node
- `Region`

View File

@ -4,8 +4,8 @@ On the TFGrid, you can deploy both micro and full virtual machines.
<h2> Table of Contents </h2>
- [Micro and Full VM Differences ](vm_differences.md)
- [Full Virtual Machine](fullvm.md)
- [Micro Virtual Machine](microvm.md)
- [Full Virtual Machine](fullVm.md)
- [Micro Virtual Machine](vm.md)
- [Nixos Micro VM](nixos_micro.md)
- [Add a Domain](add_domain.md)
- [Micro and Full VM Differences ](vm_differences.md)
- [Add a Domain to a VM](add_domain.md)

View File

@ -88,8 +88,8 @@ In this section, we cover the steps to deploy a WordPress instance on the Playgr
- Or choose a **Custom** plan
- Choose the network
- **Public IPv4** flag gives the virtual machine a Public IPv4
- `Public IPv4` flag gives the virtual machine a Public IPv4
- `Mycelium` to enable Mycelium on the virtual machine
- **Dedicated** flag to retrieve only dedicated nodes
- **Certified** flag to retrieve only certified nodes
- Choose the location of the node

View File

@ -7,86 +7,11 @@ For complementary information on the technology developed by ThreeFold, refer to
<h2> Table of Contents </h2>
- [Javascript Client](grid3_javascript_readme.md)
- [Installation](grid3_javascript_installation.md)
- [Loading Client](grid3_javascript_loadclient.md)
- [Deploy a VM](grid3_javascript_vm.md)
- [Capacity Planning](grid3_javascript_capacity_planning.md)
- [Deploy Multiple VMs](grid3_javascript_vms.md)
- [Deploy CapRover](grid3_javascript_caprover.md)
- [Gateways](grid3_javascript_vm_gateways.md)
- [Deploy a Kubernetes Cluster](grid3_javascript_kubernetes.md)
- [Deploy a ZDB](grid3_javascript_zdb.md)
- [Deploy ZDBs for QSFS](grid3_javascript_qsfs_zdbs.md)
- [QSFS](grid3_javascript_qsfs.md)
- [Key Value Store](grid3_javascript_kvstore.md)
- [VM with Wireguard and Gateway](grid3_wireguard_gateway.md)
- [GPU Support](grid3_javascript_gpu_support.md)
- [Go Client](grid3_go_readme.md)
- [Installation](grid3_go_installation.md)
- [Loading Client](grid3_go_load_client.md)
- [Deploy a VM](grid3_go_vm.md)
- [Deploy Multiple VMs](grid3_go_vms.md)
- [Deploy Gateways](grid3_go_gateways.md)
- [Deploy Kubernetes](grid3_go_kubernetes.md)
- [Deploy a QSFS](grid3_go_qsfs.md)
- [GPU Support](grid3_go_gpu.md)
- [TFCMD](tfcmd.md)
- [Getting Started](tfcmd_basics.md)
- [Deploy a VM](tfcmd_vm.md)
- [Deploy Kubernetes](tfcmd_kubernetes.md)
- [Deploy ZDB](tfcmd_zdbs.md)
- [Gateway FQDN](tfcmd_gateway_fqdn.md)
- [Gateway Name](tfcmd_gateway_name.md)
- [Contracts](tfcmd_contracts.md)
- [TFROBOT](tfrobot.md)
- [Installation](tfrobot_installation.md)
- [Configuration File](tfrobot_config.md)
- [Deployment](tfrobot_deploy.md)
- [Commands and Flags](tfrobot_commands_flags.md)
- [Supported Configurations](tfrobot_configurations.md)
- [ThreeFold Chain](tfchain.md)
- [Introduction](introduction.md)
- [Farming Policies](farming_policies.md)
- [External Service Contract](tfchain_external_service_contract.md)
- [Solution Provider](tfchain_solution_provider.md)
- [ThreeFold Chain](dev_tfchain.md)
- [Grid Proxy](proxy_readme.md)
- [Introducing Grid Proxy](proxy.md)
- [Setup](setup.md)
- [DB Testing](db_testing.md)
- [Commands](commands.md)
- [Contributions](contributions.md)
- [Explorer](explorer.md)
- [Database](database.md)
- [Production](production.md)
- [Release](release.md)
- [Flist](flist.md)
- [ThreeFold Hub Intro](zos_hub.md)
- [Generate an API Token](api_token.md)
- [Convert Docker Image Into Flist](convert_docker_image.md)
- [Supported Flists](grid3_supported_flists.md)
- [Flist Case Studies](flist_case_studies.md)
- [Case Study: Debian 12](flist_debian_case_study.md)
- [Case Study: Nextcloud AIO](flist_nextcloud_case_study.md)
- [Internals](internals.md)
- [Reliable Message Bus (RMB)](rmb_toc.md)
- [Introduction to RMB](rmb_intro.md)
- [RMB Specs](rmb_specs.md)
- [RMB Peer](peer.md)
- [RMB Relay](relay.md)
- [ZOS](index.md)
- [Manual](manual.md)
- [Workload Types](workload_types.md)
- [Internal Modules](zos_internals.md)
- [Capacity](capacity.md)
- [Performance Monitor Package](performance.md)
- [Public IPs Validation Task](publicips.md)
- [CPUBenchmark](cpubench.md)
- [IPerf](iperf.md)
- [Health Check](healthcheck.md)
- [API](api.md)
- [Grid Deployment](grid_deployment.md)
- [TFGrid Stacks](tfgrid_stacks.md)
- [Full VM Grid Deployment](grid_deployment_full_vm.md)
- [Grid Snapshots](snapshots.md)
- [Deploy the Dashboard](deploy_dashboard.md)
- [Application Case Study: Nextcloud](app_case_study_nextcloud.md)

Binary file not shown.

After

Width:  |  Height:  |  Size: 241 KiB

View File

@ -0,0 +1,200 @@
<h1>Zero-OS Boot Generator</h1>
<h2>Table of Contents</h2>
- [Introduction](#introduction)
- [Set the Environment](#set-the-environment)
- [With Docker Ubuntu 22.04](#with-docker-ubuntu-2204)
- [With Ubuntu 22.04 Micro VM on TFGrid](#with-ubuntu-2204-micro-vm-on-tfgrid)
- [Prepare the Machine](#prepare-the-machine)
- [Set a DNS A Record](#set-a-dns-a-record)
- [Set a Firewall](#set-a-firewall)
- [Set HTTPS with Caddy](#set-https-with-caddy)
- [Manually](#manually)
- [With Zinit](#with-zinit)
- [Run the Development Server](#run-the-development-server)
- [Visit the Boot Generator](#visit-the-boot-generator)
---
## Introduction
We cover how to deploy the development server of the Zero-OS Boot Generator Assistant.
Visit the [0-bootstrap repo](https://github.com/threefoldtech/0-bootstrap) for more information.
## Set the Environment
There are many ways to set your environment. Here we show with Docker and a micro VM.
Using Docker should only be used as a test before deploying on a micro VM with IPv4 and IPv6.
### With Docker Ubuntu 22.04
- Deploy Ubuntu 22.04 (Jammy) with Docker
```
sudo docker pull ubuntu:jammy
sudo docker run -it ubuntu:jammy /bin/bash
```
### With Ubuntu 22.04 Micro VM on TFGrid
- Deploy an Ubuntu 22.04 micro VM on the [Dashboard](https://dashboard.grid.tf/)
- Set IPv4 and IPv6 as `Network`
- SSH into the VM
- It is recommended to use VSCodium Explorer to facilitate the file management and editing
## Prepare the Machine
Set the machine to deploy the server
- Set the machine in i386
```
dpkg --add-architecture i386
```
- Update the packages
```
apt update
```
- Install python3-flask
```
echo "2" | apt install -y python3-flask
```
- Install the prerequisites
```
apt install -y mtools syslinux isolinux libc6-dev-i386 libc6-dbg:i386 git wget genisoimage liblzma-dev build-essential sqlite3 nano
```
## Set a DNS A Record
Set a DNS A Record pointing to the server hosting the micro VM.
* Go to your domain name registrar
* In the section **Advanced DNS**, add a **DNS A Record** to your domain and link it to the IP address of the VM you deployed on:
* Type: A Record
* Host: @
* Value: <IPv4_Address>
* TTL: Automatic
* It might take up to 30 minutes to set the DNS properly.
* To check if the A record has been registered, you can use a common DNS checker:
* ```
https://dnschecker.org/#A/example.com
```
## Set a Firewall
We set a firewall.
- Install ufw
```
apt install -y ufw
```
- Set the ports
```
ufw allow 80
ufw allow 443
ufw allow 22
```
- Enable and see the status
```
ufw enable
ufw status
```
## Set HTTPS with Caddy
We set HTTPS with Caddy. First, we test manually, then we set a zinit service.
### Manually
- Install Caddy
```
apt install -y debian-keyring debian-archive-keyring apt-transport-https curl
curl -1sLf 'https://dl.cloudsmith.io/public/caddy/stable/gpg.key' | gpg --dearmor -o /usr/share/keyrings/caddy-stable-archive-keyring.gpg
curl -1sLf 'https://dl.cloudsmith.io/public/caddy/stable/debian.deb.txt' > /etc/apt/sources.list.d/caddy-stable.list
apt update
apt install caddy
```
- Start Caddy
```
caddy reverse-proxy -r --from example.com --to :5555
```
You can also set Caddy to run with zinit, as shown below.
### With Zinit
We manage Caddy with zinit.
- Open the file for editing
```bash
nano /etc/zinit/caddy.yaml
```
- Insert the following line with your own domain and save the file
```
exec: caddy reverse-proxy -r --from example.com --to :80
```
- Add the new Caddy file to zinit
```bash
zinit monitor caddy
```
Zinit will start up Caddy immediately, restart it if it ever crashes, and start it up automatically after any reboots. Assuming you tested the Caddy invocation above and used the same form here, that should be all there is to it.
Here are some other Zinit commands that could be helpful to troubleshoot issues:
- See status of all services (same as "zinit list")
```
zinit
```
- Get logs for a service
```
zinit log caddy
```
- Restart a service (to test configuration changes, for example)
```
zinit stop caddy
zinit start caddy
```
Now that we set the domain and HTTPS, let's deploy the development server with Python.
## Run the Development Server
We show how to deploy the Boot Generator with Python.
- Clone the repository
```
git clone https://github.com/threefoldtech/0-bootstrap
cd 0-bootstrap
```
- Copy the sample file to config.py and add the proper info (e.g. set domain)
```
cp config.py.sample config.py
```
- In config.py, set the proper info, e.g. replace `http://default.tld` with your own domain, e.g. `https://example.com`
```
sed -i 's/http:\/\/default\.tld/https:\/\/example\.com/g' config.py
```
- Create the database
```
cat db/schema.sql | sqlite3 db/bootstrap.sqlite3
```
- Run the template script
```
bash setup/template.sh
```
- Run the development server with Python
```
python3 bootstrap.py
```
## Visit the Boot Generator
You can now access the boot generator on `https://example.com`.
![](./img/zos_boot_generator_main.png)

View File

@ -11,6 +11,7 @@ Please make sure to check the [basics](system_administrators@@tfgrid3_getstarted
- [Installation](grid3_javascript_installation.md)
- [Loading Client](grid3_javascript_loadclient.md)
- [Deploy a VM](grid3_javascript_vm.md)
- [Deploy a VM with Mycelium Network](grid3_javascript_vm_with_mycelium_network.md)
- [Capacity Planning](grid3_javascript_capacity_planning.md)
- [Deploy Multiple VMs](grid3_javascript_vms.md)
- [Deploy CapRover](grid3_javascript_caprover.md)

View File

@ -0,0 +1,202 @@
<h1> Deploying a VM with Mycelium Network</h1>
<h2>Table of Contents</h2>
- [Introduction](#introduction)
- [Example](#example)
- [Detailed Explanation](#detailed-explanation)
- [What is the Mycelium Network](#what-is-the-mycelium-network)
- [How to Deploy a Machine with the Mycelium Network](#how-to-deploy-a-machine-with-the-mycelium-network)
- [Summary](#summary)
- [Mycelium Flag Behavior](#mycelium-flag-behavior)
- [Mycelium Machine Seed](#mycelium-machine-seed)
- [Mycelium Network Seed](#mycelium-network-seed)
***
## Introduction
We present information on how to deploy a VM with `Mycelium network` by the Javascript client with concrete examples.
Consult the [official Mycelium repo](https://github.com/threefoldtech/mycelium) to learn more.
## Example
Here is a simple example on how to use Mycelium with the Javascript client:
```ts
import { generateRandomHexSeed, GridClient, MachinesDeleteModel, MachinesModel } from "../src";
import { config, getClient } from "./client_loader";
import { log } from "./utils";
async function deploy(client: GridClient, vms: MachinesModel) {
const res = await client.machines.deploy(vms);
log("================= Deploying VM =================");
log(res);
log("================= Deploying VM =================");
}
async function getDeployment(client: GridClient, name: string) {
const res = await client.machines.getObj(name);
log("================= Getting deployment information =================");
log(res);
log("================= Getting deployment information =================");
}
async function cancel(client: GridClient, options: MachinesDeleteModel) {
const res = await client.machines.delete(options);
log("================= Canceling the deployment =================");
log(res);
log("================= Canceling the deployment =================");
}
async function main() {
const name = "newMY";
const grid3 = await getClient(`vm/${name}`);
const vms: MachinesModel = {
name,
network: {
name: "hellotest",
ip_range: "10.249.0.0/16",
myceliumSeeds: [
{
nodeId: 168,
/**
* ### Mycelium Network Seed:
* - The `seed` is an optional field used to provide a specific seed for the Mycelium network.
* - If not provided, the `GridClient` will generate a seed automatically when the `mycelium` flag is enabled.
* - **Use Case:** If you need the new machine to have the same IP address as a previously deleted machine, you can reuse the old seed by setting the `myceliumSeed` field.
*/
seed: generateRandomHexSeed(32),
},
],
},
machines: [
{
name: "testvmMY",
node_id: 168,
disks: [
{
name: "wedDisk",
size: 8,
mountpoint: "/testdisk",
},
],
public_ip: false,
public_ip6: false,
planetary: true,
/**
* ### Mycelium Flag Behavior:
* - When the `mycelium` flag is enabled, theres no need to manually provide the `myceliumSeed` flag.
* - The `GridClient` will automatically generate the necessary seed for you.
* - **However**, if you have **an existing seed** from a previously deleted machine and wish to deploy a new machine that retains the same IP address,
* - **you can simply pass in the old seed during deployment instead of calling the `generateRandomHexSeed()` function**.
*/
mycelium: true,
/**
* ### Mycelium Seed:
* - The `myceliumSeed` is an optional field used to provide a specific seed for the Mycelium network.
* - If not provided, the `GridClient` will generate a seed automatically when the `mycelium` flag is enabled.
* - **Use Case:** If you need the new machine to have the same IP address as a previously deleted machine, you can reuse the old seed by setting the `myceliumSeed` field.
*/
myceliumSeed: generateRandomHexSeed(3), // (HexSeed of length 6)
cpu: 1,
memory: 1024 * 2,
rootfs_size: 0,
flist: "https://hub.grid.tf/tf-official-apps/base:latest.flist",
entrypoint: "/sbin/zinit init",
env: {
SSH_KEY: config.ssh_key,
},
},
],
metadata: "",
description: "test deploying single VM with mycelium via ts grid3 client",
};
//Deploy VMs
await deploy(grid3, vms);
//Get the deployment
await getDeployment(grid3, name);
//Uncomment the line below to cancel the deployment
// await cancel(grid3, { name });
await grid3.disconnect();
}
main();
```
## Detailed Explanation
### What is the Mycelium Network
Mycelium is an IPv6 overlay network written in Rust. Each node that joins the overlay network will receive an overlay network IP in the 400::/7 range.
### How to Deploy a Machine with the Mycelium Network
You just need to enable `mycelium`: set it to true as we did in the example above.
```ts
const machines = [
{
// Other attrs
mycelium: true,
}
]
```
## Summary
### Mycelium Flag Behavior
```ts
const machines = [
{
// Other attrs
mycelium: true,
}
]
```
- When the `mycelium` flag is enabled, theres no need to manually provide the `myceliumSeed` flag.
- The `GridClient` will automatically generate the necessary seed for you.
- **However**, if you have **an existing seed** from a previously deleted machine and wish to deploy a new machine that retains the same IP address,
- **you can simply pass in the old seed during deployment instead of calling the `generateRandomHexSeed()` function**.
### Mycelium Machine Seed
```ts
const machines = [
{
// Other attrs
myceliumSeed: generateRandomHexSeed(3),
}
]
```
- The `myceliumSeed` is an optional field used to provide a specific seed for the Mycelium network.
- If not provided, the `GridClient` will generate a seed automatically when the `mycelium` flag is enabled.
- **Use Case:** If you need the new machine to have the same IP address as a previously deleted machine, you can reuse the old seed by setting the `myceliumSeed` field.
### Mycelium Network Seed
```ts
const network = {
// Other attrs
myceliumSeeds: [
{
nodeId: 1,
seed: generateRandomHexSeed(32),
}
],
}
```
- The `seed` is an optional field used to provide a specific seed for the Mycelium network.
- If not provided, the `GridClient` will generate a seed automatically when the `mycelium` flag is enabled.
- **Use Case:** If you need the new machine to have the same IP address as a previously deleted machine, you can reuse the old seed by setting the `myceliumSeed` field.

View File

@ -99,7 +99,6 @@
- [I set up a 3Node in the middle of the month, does it affect uptime requirements and rewards?](#i-set-up-a-3node-in-the-middle-of-the-month-does-it-affect-uptime-requirements-and-rewards)
- [What is the difference between a certified and a non-certified 3Node?](#what-is-the-difference-between-a-certified-and-a-non-certified-3node)
- [What are the different certifications available for 3Node servers and farms? What are the Gold and Silver certifications?](#what-are-the-different-certifications-available-for-3node-servers-and-farms-what-are-the-gold-and-silver-certifications)
- [What is the difference between V2 and V3 minting?](#what-is-the-difference-between-v2-and-v3-minting)
- [What is the TFT minting address on Stellar Chain?](#what-is-the-tft-minting-address-on-stellar-chain)
- [Can Titans and DIY 3Nodes share the same farm?](#can-titans-and-diy-3nodes-share-the-same-farm)
- [Do I need one farm for each 3Node?](#do-i-need-one-farm-for-each-3node)
@ -116,7 +115,6 @@
- [Farming Rewards and Related Notions](#farming-rewards-and-related-notions)
- [What are the rewards of farming? Can I get more rewards when my 3Node is being utilized?](#what-are-the-rewards-of-farming-can-i-get-more-rewards-when-my-3node-is-being-utilized)
- [How can I know the potential farming rewards for Grid Utilization?](#how-can-i-know-the-potential-farming-rewards-for-grid-utilization)
- [What is the easiest way to farm ThreeFold tokens (TFT)?](#what-is-the-easiest-way-to-farm-threefold-tokens-tft)
- [When do I receive my rewards?](#when-do-i-receive-my-rewards)
- [Do farming rewards take into account the type of RAM, SSD, HDD and CPU of the 3Node server?](#do-farming-rewards-take-into-account-the-type-of-ram-ssd-hdd-and-cpu-of-the-3node-server)
- [Can I send my farming rewards directly to a crypto exchange?](#can-i-send-my-farming-rewards-directly-to-a-crypto-exchange)
@ -162,7 +160,6 @@
- [Farming Optimization](#farming-optimization)
- [What is the difference between a ThreeFold 3Node and a ThreeFold farm? What is the difference between the farm ID and the node ID?](#what-is-the-difference-between-a-threefold-3node-and-a-threefold-farm-what-is-the-difference-between-the-farm-id-and-the-node-id)
- [How can I know how many GB of SSD and RAM do I need?](#how-can-i-know-how-many-gb-of-ssd-and-ram-do-i-need)
- [What is the optimal ratio of virtual cores - vcores or threads - SSD storage and RAM memory? What is the best optimization scenario for a 3Node, in terms of TFT farming rewards?](#what-is-the-optimal-ratio-of-virtual-cores---vcores-or-threads---ssd-storage-and-ram-memory-what-is-the-best-optimization-scenario-for-a-3node-in-terms-of-tft-farming-rewards)
- [What does TBW mean? What is a good TBW level for a SSD disk?](#what-does-tbw-mean-what-is-a-good-tbw-level-for-a-ssd-disk)
- [Are SATA and SAS drives interchangeable?](#are-sata-and-sas-drives-interchangeable)
- [What is the speed difference between SAS and SATA disks?](#what-is-the-speed-difference-between-sas-and-sata-disks)
@ -1132,13 +1129,6 @@ Farms can also be certified. The certifications are: [gold certified farming](ht
Note that gold and silver certifications are still being discussed. Join the discussion on the [ThreeFold Forum](http://forum.threefold.io/).
### What is the difference between V2 and V3 minting?
V2 is being sunset. New miners should directly onboard to V3.
On the tokenomics side, V2 rewards decrease as the difficulty level increases. For the V3 rewards, the rewards are constant for 5 years. In short, V3 is more profitable in the long run. For more information, read [this post](https://forum.threefold.io/t/comparison-v2-vs-v3-minting/2122).
### What is the TFT minting address on Stellar Chain?
The TFT minting address on Stellar Chain is the following: GBOVQKJYHXRR3DX6NOX2RRYFRCUMSADGDESTDNBDS6CDVLGVESRTAC47
@ -1241,13 +1231,6 @@ By connecting a 3Node to the Grid, you get Farming Rewards. If you set a public
Go on the [ThreeFold simulator](https://simulator.grid.tf/), enter your 3Node resources, check the Public IP address. This will enable farming rewards from the parameter NU Required Per CU. Check the difference in the farming rewards per month. Note that you will need a Public IP address.
### What is the easiest way to farm ThreeFold tokens (TFT)?
Buy a [certified 3Node](https://marketplace.3node.global/index.php). This is more or less *plug n play*! You can also build a [DIY 3Node](#what-are-the-general-requirements-for-a-diy-3node-server). It's fun and there are many resources to help you along the way.
### When do I receive my rewards?
They are distributed once a month, around the 8th*. Distributions are not daily, or after a certain threshold. Note that upcoming minting rules may have a 24 month lockup or until 30% utilization for 3 months on your 3Node.
@ -1268,7 +1251,7 @@ This is not possible. When you send tokens to a crypto exchange, you need to inc
### Do I need collateral to farm ThreeFold tokens?
Many decentralized data projects require collateral, but not ThreeFold. There is an ongoing discussion on collateral. Join the discussion [here](https://forum.threefold.io/t/should-tft-collateral-be-required-for-3nodes/3724).
Many decentralized data projects require collateral, but not ThreeFold.
### Can I add external drives to the 3Nodes to increase rewards and resources available to the ThreeFold Grid?
@ -1575,19 +1558,6 @@ You need 50 GB of SSD per compute units (CU) and a minimum of 500 GB SSD and 2 G
A 3Node has, in general, 2 compute units (CU) per thread. Thus, for peak optimisation, you need 100 GB SSD and 8GB RAM per thread.
### What is the optimal ratio of virtual cores - vcores or threads - SSD storage and RAM memory? What is the best optimization scenario for a 3Node, in terms of TFT farming rewards?
In short, for peak optimization, aim for 100 GB SSD of storage and 8GB RAM of memory per virtual core (vcore or thread).
For example, a 32 threads (32 vcores) 3Nodes would need 3.2 TB SSD and 256GB RAM to be optimal, reward-wise.
That is: 32 * 100 = 3200 GB SSD = 3.2TB SSD, and 32 * 8 = 256 GB RAM total.
Adding more GB of RAM would not increase your TFT rewards. You would need more vcores if you want to expand.
NB: This is purely based on reward considerations. Some users might need different ratios for different specific uses of the Grid.
### What does TBW mean? What is a good TBW level for a SSD disk?
TBW means Terabytes Written. TBW directly measures how much you can write cumulatively into the drive over its lifetime. For your 3Node, it can be a good idea to prioritize a minimum ratio of 500 TBW per 1TB for SSD.

View File

@ -24,6 +24,8 @@
We will now learn how to create a Zero-OS bootstrap image in order to boot a DIY 3Node.
If you are building an HDD-only node, read [this section](hdd_only_nodes.md) too.
## Download the Zero-OS Bootstrap Image
Let's download the Zero-OS bootstrap image.

View File

@ -31,6 +31,8 @@ The essential information are available in the section [3Node Requirements Summa
You need a theoretical minimum of 500 GB of SSD and 2 GB of RAM on a mini pc, desktop or server. In short, for peak optimization, aim for 100 GB of SSD and 8GB of RAM per thread. (Thread is equivalent to virtual core or logical core.)
If you are building an HDD-only node, read [this section](hdd_only_nodes.md) too.
Also, TFDAO might implement a farming parameter based on [passmark](https://www.cpubenchmark.net/cpu_list.php). From the ongoing discussion on the Forum, you should aim at a CPU mark of 1000 and above per core.
> 3Node optimal farming hardware ratio -> 100 GB of SSD + 8 GB of RAM per Virtual Core
@ -58,6 +60,7 @@ Any computer with the following specifications can be used as a DIY 3Node.
*A passmark of 1000 per core is recommend and will be a minimum requirement in the future. This is not yet an official requirement. A 3Node with less than 1000 passmark per core of CPU would not be penalized if it is registered before the DAO settles the [Passmark Question](https://forum.threefold.io/t/cpu-benchmarking-for-reward-calculations/2479).
> Note: It is also possible to build an HDD-only node. Read more about this [here](hdd_only_nodes.md).
## Bandwidth Requirements

View File

@ -5,6 +5,7 @@
- [Introduction](#introduction)
- [1. Booting the 3Node with Zero-OS](#1-booting-the-3node-with-zero-os)
- [2. Check the 3Node Status Online](#2-check-the-3node-status-online)
- [Check Node Health](#check-node-health)
- [3. Receive the Farming Rewards](#3-receive-the-farming-rewards)
- [Advanced Booting Methods (Optional)](#advanced-booting-methods-optional)
- [PXE Booting with OPNsense](#pxe-booting-with-opnsense)
@ -42,6 +43,9 @@ You can use the ThreeFold [Node Finder](node_finder.md) to verify that your 3Nod
* [ThreeFold Dev Net Dashboard](https://dashboard.dev.grid.tf/)
* [ThreeFold QA Net Dashboard](https://dashboard.qa.grid.tf/)
### Check Node Health
It is also possible to check the node health via the Node Finder. Read [this section](dashboard@@node_finder) for more information.
## 3. Receive the Farming Rewards

View File

@ -74,13 +74,19 @@ included_nodes: [optional, if no nodes are added then the farmerbot will include
- "<your node ID to be included, required at least 2>"
excluded_nodes:
- "<your node ID to be excluded, optional>"
priority_nodes:
- "<your node ID to have a priority in nodes management, optional>"
never_shutdown_nodes:
- "<your node ID to be never shutdown, optional>"
power:
periodic_wake_up_start: "<daily time to wake up nodes for your farm, default is the time your run the command, format is 00:00AM or 00:00PM, optional>"
wake_up_threshold: "<the threshold number for resources usage that will need another node to be on, default is 80, optional>"
periodic_wake_up_limit: "<the number (limit) of nodes to be waken up everyday, default is 1, optional>"
overprovision_cpu: "<how much node allows over provisioning the CPU , default is 1, range: [1;4], optional>"
wake_up_threshold:
cru: "<the threshold number for cru usage that will need another node to be on, default is 80, optional>"
mru: "<the threshold number for mru usage that will need another node to be on, default is 80, optional>"
sru: "<the threshold number for sru usage that will need another node to be on, default is 80, optional>"
hru: "<the threshold number for hru usage that will need another node to be on, default is 80, optional>"
```
## Supported Commands and Flags

View File

@ -155,8 +155,6 @@ It is highly recommended to set a Ubuntu systemd service to keep the Farmerbot r
[Service]
Restart=always
RestartSec=5
StandardOutput=append:/root/farmerbotfiles/farmerbot.log
StandardError=append:/root/farmerbotfiles/farmerbot.log
ExecStart=/usr/local/bin/farmerbot run -e /root/farmerbotfiles/.env -c /root/farmerbotfiles/config.yml -d
[Install]
@ -175,25 +173,27 @@ It is highly recommended to set a Ubuntu systemd service to keep the Farmerbot r
### Check the Farmerbot Logs
Once you've set a Farmerbot systemd service [as show above](#set-a-systemd-service), the Farmerbot will start writing logs to the file `farmerbot.log` in the directory `farmerbotfiles`.
You can get more details on the operation of the Farmerbot by inspecting the log file.
Thus, you can get more details on the operation of the Farmerbot by inspecting the log file. This can also be used to see the **Farmerbot Report Table** as this table is printed in the Farmerbot log.
* See all logs so far
* View the log file
```
cat ~/farmerbotfiles/farmerbot.log
journalctl -u <unit file name>
```
* See the last ten lines and new logs as they are generated
* View the log file with reverse output
```
tail -f ~/farmerbotfiles/farmerbot.log
journalctl -ru <unit file name>
```
* See all logs and new lines as they are generated
* View X lines from the log file
```
tail -f -n +1 ~/farmerbotfiles/farmerbot.log
journalctl -u <unit file name> -n X
```
* See the last report table
* View the log file since a given date
```
tac ~/farmerbotfiles/farmerbot.log | grep -B5000 -m1 "Nodes report" | tac
journalctl -u <unit file name> --since <date>
```
* Create a log file with journalctl with the last 1000 lines
```
journalctl -u farmerbot -n 1000 > farmerbot.log
```
### Stop the Farmerbot

View File

@ -9,29 +9,10 @@ To buy a certified node from an official ThreeFold vendor, check the [ThreeFold
<h2> Table of Contents </h2>
- [Build a 3Node](3node_building.md)
- [1. Create a Farm](1_create_farm.md)
- [2. Create a Zero-OS Bootstrap Image](2_bootstrap_image.md)
- [3. Set the Hardware](3_set_hardware.md)
- [4. Wipe All the Disks](4_wipe_all_disks.md)
- [5. Set the BIOS/UEFI](5_set_bios_uefi.md)
- [6. Boot the 3Node](6_boot_3node.md)
- [Farming Optimization](farming_optimization.md)
- [GPU Farming](gpu_farming.md)
- [Set Additional Fees](set_additional_fees.md)
- [Minting Receipts](minting_receipts.md)
- [Minting Periods](minting_periods.md)
- [Room Parameters](farm_room_parameters.md)
- [Farming Costs](farming_costs.md)
- [Calculate Your ROI](calculate_roi.md)
- [Farming Requirements](farming_requirements.md)
- [Farming Optimization](farming_optimization.md)
- [Advanced Networking](advanced_networking_toc.md)
- [Networking Overview](networking_overview.md)
- [Network Considerations](network_considerations.md)
- [Network Setup](network_setup.md)
- [Farmerbot](farmerbot_intro.md)
- [Quick Guide](farmerbot_quick.md)
- [Additional Information](farmerbot_information.md)
- [Minting and the Farmerbot](farmerbot_minting.md)
- [Node Status Bot](node_status_bot.md)
- [Farming Troubleshooting](farming_troubleshooting.md)
- [Farmers FAQ](faq@@faq)

Some files were not shown because too many files have changed in this diff Show More