Pricing strategy based on pvcg

In the role benefit distribution model of distributed machine learning, the benefits of the global verifier and local verifier are almost fixed like the miners in the blockchain, and will only change with the activity of the system. The aggregator is the task scheduler and coordinator, and its benefits are positively correlated with the trainer. The only more complicated issue is that during the local training process of the model, the trainer may use data from other data sources, and we need to give additional consideration to the gaming problem in the local training scenario of distributed machine learning. The trainer's operating environment will have some thresholds, while other network participants who do not fulfill the hardware requirements can participate in the training process by providing data. A game model is needed between the data providers and the real trainers to distribute their benefits fairly.

To incentivize data providers to contribute the best datasets to the training process, we need to pay sufficient rewards to data providers to cover their costs. The marginal monetary reward for contributing more data should be no less than the resulting marginal cost. In addition, our goal is to maintain a balanced budget and optimize social welfare. At least three sources of information asymmetry are intertwined in this problem: 1) The dataset owned by each data provider; 2) The cost of each data provider; 3) The valuation of trained distributed machine learning models by model users.

To overcome these information asymmetries and achieve the above objectives, we design rational incentives, i.e., a function that calculates participants' payoffs.

There exists a set of nn data providers, denoted by N=0,,n1N = {0, \ldots, n - 1} , and another set of mm model users, denoted by M=n,,n+m1M = {n, \ldots, n + m - 1}. Each data provider iNi \in N owns a dataset dˉi\bar{d}_i . It claims it owns a dataset d^i\hat{d}_i . The federation accepts a dataset did^id_i \oslash \hat{d}_i from this data provider. We call ηi=did^i\eta_i = \frac{d_i}{\hat{d}i} the acceptance ratio, where \oslash denotes element-wise division. Trained on datasets d=(d0,,dn1)d = (d_0, \ldots, d{n-1}) from all data providers, the usefulness of the distributed machine model is Q(d)Q(d) . Model users may be granted limited access to the distributed machine model such that the usefulness of the distributed machine model to model user jj is κjQ(d)\kappa_j Q(d), where κj\kappa_j is called the access permission. Each data provider iN i \in Nhas a cost type γiΓi\gamma_i \in \Gamma_i . Its cost of contributing data did_i is c(di,γi)c(d_i, \gamma_i). The collection of cost types of all data providers forms the cost type profile γ=(γ0,,γn)\gamma = (\gamma_0, \ldots, \gamma_n). Data provider ii may report a different cost type γ^i\hat{\gamma}_i.

Each model user jMj \in M has a valuation type θjΘj\theta_j \in \Theta_j . Its valuation on the trained distributed machine model is

w(κjQ(d),θj)=v(d,κj,θj)w(\kappa_j Q(d), \theta_j) = v(d, \kappa_j, \theta_j)

The collection of valuation types of all model users forms the valuation type profile θ=(θn,,θn+m1)\theta = (\theta_n, \ldots, \theta_{n+m-1}). Model user jj may report a different valuation type θ^j\hat{\theta}j . The payment to data provider iNi \in N is pi0p_i \geq 0 . The payment to model user jMj \in M is pj0p_j \leq 0. We denote ps=(p0,,pn1)p_s = (p_0, \ldots, p{n-1}) and pd=(pn,,pn+m1)p_d = (p_n, \ldots, p_{n+m-1}). The federation income is I=j=nn+m1pjI = - \sum_{j=n}^{n+m-1} p_j; the federation expenditure is E=i=0n1piE = \sum_{i=0}^{n-1} p_i; the federation profit is P=l=0n+m1plP = \sum_{l=0}^{n+m-1} p_l . Participants' preferences are represented by quasi-linear utility functions:

ui()=pi()ci() for iN;uj()=pj()+vj() for jMu_i(\cdot) = p_i(\cdot) - c_i(\cdot) \text{ for } i \in N; \\ u_j(\cdot) = p_j(\cdot) + v_j(\cdot) \text{ for } j \in M

The social effect of distributed machine learning is measured by social surplus, defined as

S()=j=nn+m1vj()i=0n1ci()S(\cdot) = \sum_{j=n}^{n+m-1} v_j(\cdot) - \sum_{i=0}^{n-1} c_i(\cdot)

which includes consumer surplus Sd=j=nn+m1vj()S_d = \sum_{j=n}^{n+m-1} v_j(\cdot) and producer surplus Sp=i=0n1ci()S_p = - \sum_{i=0}^{n-1} c_i(\cdot) . There are user-defined unfairness functions wˉs(ps,c)\bar{w}_s(p_s, c) and wˉd(ps,v)\bar{w}_d(p_s, v) that measure the unfairness among data providers and model users.

We set the data providers as the supply side and the trainers as the demand side, and use the PVCG model, which maximizes social welfare, as the supply-side auction offer model, and use the Cremer-McLean mechanism to maximize the demand-side utility.

Given that the federation Income I(Q)I(Q) and the model quality Q(d^,η)Q(\hat{d}, \eta) are exogenous functions, the supply-side distributed machine learning incentive mechanism design is to design the optimal. Specifically on the supply side, letting the data providers provide the maximum data efficiency ηi(d^,γ^)\eta_i(\hat{d}, \hat{\gamma}) and provide the optimal reward pi(d^,γ^)p_i(\hat{d}, \hat{\gamma}) for them, and on the demand side, letting the trainers produce the optimal model results so that the trainer outputs the maximum model validity κj(θ^)\kappa_j(\hat{\theta}) and provides the optimal reward pj(θ^)p_j(\hat{\theta}) for them.

Crémer-McLean Mechanism

The detailed process of the Crémer-McLean mechanism for the demand side involves the following steps:

Step 1: Consumer Valuation and Preferences

Each consumer ii privately holds a valuation θi\theta_i for the product or service being offered.

Step 2: Decision Rule

Consumers submit their valuations through a decision rule κ(θ^)\kappa(\hat{\theta}) , where θ^\hat{\theta} represents the reported valuations.

Step 3: Payment Rule Design

An interim incentive compatible and interim individually rational payment rule p(θ^)p(\hat{\theta}) is designed to extract the full consumer surplus.

Crémer-McLean Theorem Formulation

The Crémer-McLean Theorem states that for any decision rule κ(θ^)\kappa(\hat{\theta}) satisfying the Crémer-McLean condition and identifiability condition, there exists a payment rule p(θ^)p(\hat{\theta}) that extracts full consumer surplus:

j=nn+m1pj(θ^)=j=nn+m1w(κj(θ^)Q,θj)\sum_{j=n}^{n+m-1} p_j(\hat{\theta}) = \sum_{j=n}^{n+m-1} w(\kappa_j(\hat{\theta})Q, \theta_j)

Optimization Process

The payment rule p(θ^)p(\hat{\theta}) can be found by minimizing a loss function to ensure interim incentive compatibility, individual rationality, and full consumer surplus extraction.

Procurement-VCG (PVCG) Mechanism

As a counterpart of the Crémer-McLean mechanism, we create the PVCG on the supply side. The PVCG mechanism is designed to incentivize distributed machine learning participants to truthfully report their type parameters and offer their best datasets to the federation. This mechanism provides theoretical guarantees for incentive compatibility, allocative efficiency, individual rationality, and weak budget balancedness. The PVCG mechanism, along with the Crémer-McLean mechanism, aims to address the challenges of information asymmetry and free-riding in distributed machine learning by providing appropriate incentives to participants.

When designing the supply-side mechanism, we assume the federation income I(Q)I(Q) to be an exogenous function that depends on the quality of the distributed machine model QQ. This assumption allows us to focus on optimizing the supply-side distributed machine learning incentive mechanism without directly considering the intricacies of how the federation income is determined. By assuming I(Q)I(Q) as an exogenous parameter, we can streamline the design process and concentrate on factors such as dataset contributions, cost types, and acceptance ratios to achieve the desired objectives in distributed machine learning. The supply-side use of the PVCG mechanism gives us the revenue of the distributed machine learning training process as:

I(Q)=j=nn+m1pj(θ)=j=nn+m1w(κj(θ)Q,θj)I(Q) = - \sum_{j=n}^{n+m-1} p_j(\theta) = \sum_{j=n}^{n+m-1} w(\kappa_j(\theta)Q, \theta_j)

Procurement Auction Process of PVCG and Payment Calculation for Data Providers

Step 1: Data Providers Claim Datasets to Offer and Bid on Cost Types

As the first step, each data provider submits a sealed bid for their claimed datasets and cost types. The claimed dataset ( \hat{d}_i ) is the dataset that data provider ( i ) claims to offer for distributed machine learning. It may differ from the actual dataset ( \bar{d}_i ) owned by the data provider. Similarly, the reported cost type ( \hat{\gamma}_i ) may differ from the true cost type ( \gamma_i ).

Step 2: The Coordinator Chooses the Optimal Acceptance Ratios

The coordinator determines the optimal acceptance ratios ηi\eta_i for each data provider by maximizing the social surplus:

η=argmaxη[0,1]dim(di)×n{I(d^η)i=0n1ci(d^iηi,γ^i)}\eta^* = \arg\max_{\eta \in [0,1]^{dim(d_i) \times n}} \left\{ I(\hat{d} \cdot \eta) - \sum_{i=0}^{n-1} c_i(\hat{d}_i \cdot \eta_i, \hat{\gamma}_i) \right\}

Step 3: Data Providers Contribute Accepted Datasets to Distributed Lachine Learning

Data providers contribute their accepted datasets d^η\hat{d} \cdot \eta^* to distributed machine learning. If a data provider cannot contribute d^iηid^i\hat{d}_i \cdot \eta^*_i \leq \hat{d}_i, a high punishment is imposed. The income to the federation is I(d^η)I(\hat{d} \cdot \eta^*).

Step 4: The Coordinator Makes Transfer Payments to Data Providers According to the PVCG Sharing Rule

The PVCG payment pi()p_i(\cdot) consists of the VCG payment τi\tau_i and the optimal adjustment payment hih^*_i :

pi()=τi()+hi()p_i(\cdot) = \tau_i(\cdot) + h^*_i(\cdot)

The VCG payment τi\tau_i to data provider ii is calculated as:

τi=S(d^,γ^)Si(d^i,γ^i)+c(d^iηi) \tau_i = S^(\hat{d}, \hat{\gamma}) - S^{-i}(\hat{d}{-i}, \hat{\gamma}_{-i}) + c(\hat{d}_i \cdot \eta^*_i)

where SS^* represents the maximum producer surplus and SiS^*{-i} is the surplus without data provider ii. d^i\hat{d}{-i} and γ^i\hat{\gamma}_{-i} denote the claimed datasets and reported cost types excluding data provider ii.

Last updated