Table 1 Accuracy of introducing backdoor under draft attack.

From: Aggregation algorithm based on consensus verification

Symbols

Connotation

n

The total number of calculation nodes

m

The number of Byzantine

d

The number of dimensions (parameters) of the model

\(p_i\)

Calculate parameters for ith worker training

\((p_i)_j\)

jth dimension in \(p_i\) parameter vector

p

All calculate parameters

\(\epsilon\)

Constant variable

\(\theta\)

N-dimensional variable(\(p_i\))

\(<\nabla f_t(\theta _1),\theta _2-\theta _1>\)

Gradient from \(\theta _1\) to \(\theta _2\)

T

Training batch

t

tth batch

\(\theta ^(t)\)

Model parameters for batch t

\(f_t(\theta )\)

Optimization function for batch t

K

Sample size for one round of training (including batch T and batch size \(K_t\))

\(K_t\)

Batch size

\(x_k\)

Characteristics of the sample

\(y_k\)

Label of the sample

\(g(x_k;\theta )\)

The model

\(L(g(x_k;\theta )\)

The loss

\(min\sum _{k=1}^{K}L(g(x_k;\theta ),y_k)\)

Optimization problem of deep learning models

\(\sum _{t=1}^{T}f_t(\theta ^{(t)})\)

The optimal loss of T batch model

\(min\sum _{t=1}^{T}f_t(\theta )\)

Minimizing losses in training models

\(\sum _{t=1}^{T}f_t(\theta ^*)\)

Minimizing losses in training models with optimal parameters

\(a_t\)

Learning rate of batch t

\(\delta g_t\)

Training gradient for batch t

D

The upper bound of any variable

G

The upper bound of any gradient