Given N experts and a penalty parameter β (with a value ranging from 0 to 1), the initial weight of each expert is all 1. In each round of prediction, we listen to the opinions of all experts and randomly select one expert to adopt their opinions based on the weight. Compare this opinion with the facts and reduce the weight of all experts with incorrect predictions based on β.
Input: Penalty parameter β∈ (0,1)
1: Set
2: for :do
3: Receive
4: choose a according to the probability:
5: / ,
6: Receive
7: :
The number of mistakes made by the Randomized Multiplicative Weights Update Algorithm is bounded as:
where and .
Proof: we know that
and
While is the weighted average error rate among all experts at round ,
denotes the total weight of the experts who made a wrong prediction.We also suppose the best expert makes a total of mistakes, so its final weight is at least:
We can deduce that :
so:
Take the logarithms on both sides.:
By inequality:
(when ):
On the other hand,,
Comprehensively obtained:
Ultimately, the expected total number of errors of the algorithm satisfies:
Note that only the learning algorithm is randomized. The underlying assumption is that the examples and experts' predictions are not random. The only randomness is the randomness where the learner makes his own prediction.
In this randomized algorithm, if . Compared to weighted algorithm, this randomness halved the number of mistakes the algorithm is going to make.[1] However, it is important to note that in some research, people define in weighted majority algorithm and allow in randomized weighted majority algorithm.[2]
^Cite error: The named reference ref7 was invoked but never defined (see the help page).
^Cite error: The named reference ref2 was invoked but never defined (see the help page).