Open Access iconOpen Access

ARTICLE

crossmark

Distributed Active Partial Label Learning

Zhen Xu1,2, Weibin Chen1,2,*

1 Key Laboratory of Intelligent Informatics for Safety & Emergency of Zhejiang Province, Wenzhou University, Wenzhou, 325006, China
2 College of Computer Science and Artificial Intelligence Engineering, Wenzhou University, Wenzhou, 325006, China

* Corresponding Author: Weibin Chen. Email: email

Intelligent Automation & Soft Computing 2023, 37(3), 2627-2650. https://doi.org/10.32604/iasc.2023.040497

Abstract

Active learning (AL) trains a high-precision predictor model from small numbers of labeled data by iteratively annotating the most valuable data sample from an unlabeled data pool with a class label throughout the learning process. However, most current AL methods start with the premise that the labels queried at AL rounds must be free of ambiguity, which may be unrealistic in some real-world applications where only a set of candidate labels can be obtained for selected data. Besides, most of the existing AL algorithms only consider the case of centralized processing, which necessitates gathering together all the unlabeled data in one fusion center for selection. Considering that data are collected/stored at different nodes over a network in many real-world scenarios, distributed processing is chosen here. In this paper, the issue of distributed classification of partially labeled (PL) data obtained by a fully decentralized AL method is focused on, and a distributed active partial label learning (dAPLL) algorithm is proposed. Our proposed algorithm is composed of a fully decentralized sample selection strategy and a distributed partial label learning (PLL) algorithm. During the sample selection process, both the uncertainty and representativeness of the data are measured based on the global cluster centers obtained by a distributed clustering method, and the valuable samples are chosen in turn. Meanwhile, using the disambiguation-free strategy, a series of binary classification problems can be constructed, and the corresponding cost-sensitive classifiers can be cooperatively trained in a distributed manner. The experiment results conducted on several datasets demonstrate that the performance of the dAPLL algorithm is comparable to that of the corresponding centralized method and is superior to the existing active PLL (APLL) method in different parameter configurations. Besides, our proposed algorithm outperforms several current PLL methods using the random selection strategy, especially when only small amounts of data are selected to be assigned with the candidate labels.

Keywords


Cite This Article

APA Style
Xu, Z., Chen, W. (2023). Distributed active partial label learning. Intelligent Automation & Soft Computing, 37(3), 2627-2650. https://doi.org/10.32604/iasc.2023.040497
Vancouver Style
Xu Z, Chen W. Distributed active partial label learning. Intell Automat Soft Comput . 2023;37(3):2627-2650 https://doi.org/10.32604/iasc.2023.040497
IEEE Style
Z. Xu and W. Chen, “Distributed Active Partial Label Learning,” Intell. Automat. Soft Comput. , vol. 37, no. 3, pp. 2627-2650, 2023. https://doi.org/10.32604/iasc.2023.040497



cc Copyright © 2023 The Author(s). Published by Tech Science Press.
This work is licensed under a Creative Commons Attribution 4.0 International License , which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.
  • 716

    View

  • 364

    Download

  • 0

    Like

Share Link