Training Node
Training nodes are responsible for training and fine-tuning the AI tasks initiated by the task creators. This mechanism ensures the integrity and health of the ecosystem, as nodes have vested interests via staking. In return, the nodes will be rewarded in proportion to their contributions. To become a training node, a user has to stake FLOCK.
0. Overview: reward drivers for training nodes
Put simply, a training nodeβs daily return from a task depends on three factors:
(1) the relative staking amount of this task against all tasks, meaning a training nodeβs stake in a particular task will indirectly affect itβs rewards from that task; and
(2) a training nodeβs stake in this task as well as stake delegated to this training node; and
(3) quality of the nodeβs submission, as shown by the nodeβs relative ranking. Specifically, it is a geometric series, along with its ranking, multiplied by the relative stake of this task.
Itβs important to note that rank is being used here to determine the quality of the training nodeβs work, not other metrics such as absolute scores, primarily because scores can come very closely between nodes. Such design decision is believed to make reward calculation fairer and easier.
Calculation of reward distributions for training nodes follow a three-step formula:
1. Reward distribution within a single AI Arena task
Within a single AI Arena task, the reward distribution between training nodes and validators is determined based on their relative stake amounts.
We assume there are π submissions from π training nodes with stakes , and π validators with stakes . Each validator evaluates the models submitted by the training nodes.
Let the total daily reward allocated to a task be denoted as β and the parameter controls the split rewards, defining the balance between fixed and stake-dependent reward components.
The total rewards for training nodes are:
2. Rewards for training nodes & their delegators
We can now compute the total rewards allocated for the training nodes as well as their delegators, which is based on the quality of their submission and their total amount of stake:
In which the total stake amount from the training node π as well as its respective delegators, is the scores of the submitted models from training node, whereas denotes a given training nodeβs rank amongst its peers in the same task. On the other hand, is a system parameter that determines the influence of the stake on the reward distribution.
3. Rewards for training nodes
If a training node βs stake in the task is and stakes delegated to training node is i.e. , in which refers to the effective delegation amount. Specifically, effective delegation amount adjusts how much of the delegated stake β is actually counted (i.e., how βeffectiveβ it is) when computing node 's total stake . When , then the delegated stake β is counted fullyβdelegations are treated just like the training nodeβs own stake. When , then the delegated stake is βdiscounted,β so training node only gets a fraction of when accounting for its total stake, and vice versa. Thus, the actual reward for training node is:
Note that in the front-end, you will see a βreward-sharing ratioβ, which refers to , which means when reward-sharing ratio is 60%, is 0.4. This ratio is set by training nodes and validators permissionlessly.
4. Example
Letβs assume daily total rewards for all AI Arena tasks for a given day is 309,157.68. We have 1 task with 2 nodes and 3 validators.
Nodes A and B stake 3,000 and 3,500 FLOCK respectively, while validators A, B and C stake 3,000, 6,000 and 3,000 respectively. Node A also receives an additional 1,000 FLOCK from its delegators, which brings the (total stake including delegated stake) to be 4,000 for Node A. For simplicity, we assume to be 0 and to be in this example.
First, for this given task, total rewards for *all* training nodes are:
We can then compute the rewards for *Node A and its delegators*. We are assuming that the scores for Node A and B are 0.501435 and 0.498565 respectively. Consider Ξ±tβ=1, rewards for Node A (together with delegators) are:
Finally, given π=0.4, the actual rewards for *Node A alone* is:
Last updated
Was this helpful?