我正在阅读来自 Deepmind 研究员科学家 Sebastian Ruder blog-link的博客,并且在理解这个民主共同学习的伪代码时遇到了问题。有人可以帮助我了解发生了什么吗?具体来说,我只是想了解第 1-6 行。
博客中民主合作学习的描述:
Democratic Co-learning Rather than treating different feature sets as views, democratic co-learning (Zhou and Goldman, 2004) [18] employs models with different inductive biases. These can be different network architectures in the case of neural networks or completely different learning algorithms. Democratic co-learning first trains each model separately on the complete labelled data L. The models then make predictions on the unlabelled data U. If a majority of models confidently agree on the label of an example, the example is added to the labelled dataset. Confidence is measured in the original formulation by measuring if the sum of the mean confidence intervals w of the models, which agreed on the label is larger than the sum of the models that disagreed. This process is repeated until no more examples are added. The final prediction is made with a majority vote weighted with the confidence intervals of the models. The full algorithm can be seen below. M is the set of all models that predict the same label j for an example x.
我猜第 1-3 行是在标记数据上训练几个模型。我是标记数据集中的每个样本吗?mi 是使用标记数据集中所有样本的训练模型吗?我猜第 4-6 行正在遍历未标记数据集中的每个样本,但是 j 是什么?它是否遍历每个标签?
我特别困惑的这一行是 M <- {i | pi(x) = j}。我是否被分配了来自未标记样本的模型输出,然后被分配给 M,M 是什么?