| چکیده انگلیسی مقاله |
The problem of classification can be divided into two categories: single-label and multi-label. Single-label classification consists of binary and multi-class classification. In binary classification, the task is to predict one in two possible classes, such as distinguishing between spam and non-spam emails. In multi-class classification, the goal is to classify instances into more than two classes, such as identifying different species of flowers based on petal measurements. In contrast to single-label classification, multi-label classification is more complex because each instance could belong to multiple categories simultaneously. In multi-label learning, instead of assigning a single label to each instance, a set of labels is assigned. This means that each sample may have zero, one, or more than one associated label. For example, in a text classification task, a news article about technology and business might be labeled as both "Technology" and "Business". To handle multi-label classification, several approaches have been developed. One of the simplest methods is Binary Relevance (BR), which transforms the multi-label problem into multiple independent binary classification tasks—one for each label. Although this approach is easy to implement, it treats each label independently and ignores possible relationships among them. However, in real-world applications, labels are often correlated; for instance, in medical diagnosis, certain diseases frequently appear together. In another approach, Label Powerset (LP), considers label dependencies by treating each unique combination of labels as a separate class. While this method captures relationships between labels, it suffers from scalability issues while dealing with a large number of labels, as the number of possible label combinations increases exponentially. To address these challenges, the proposed method incorporates k-means constraint clustering to group both labels and features prior to applying classification. In the first step, clustering is performed to group similar labels together, ensuring that label correlations are preserved. This also helps to mitigate the issue of imbalanced classification, where certain labels may be underrepresented in the dataset. Once the labels are being clustered, a separate multi-layer neural network would be assigned to each cluster. Instead of using a single large neural network for all labels, multiple smaller networks would be trained for different label clusters. This approach enhances learning efficiency and improves accuracy by focusing on relevant label groups. However, using multiple classifiers increases computational costs and training time. To mitigate this issue, a scatter-add dimension reduction technique is applied. Using scatter-add, attributes are efficiently assigned to the input of each neural network, ensuring that each classifier receives only the relevant feature subset. Each neural network then predicts labels within its designated cluster. Eventually, the predictions from all classifiers are combined to generate the final multi-label output for each instance. To evaluate the effectiveness of the proposed method, experiments were conducted on various text datasets. The results were compared with traditional multi-label classification methods, including Binary Relevance and Label Powerset. The evaluation has been based on several performance metrics, such as accuracy, precision, and hamming-loss. The results demonstrated that the proposed approach achieved superior performance across multiple datasets, ranking first in several evaluation criteria. Notably, it outperformed existing methods by a margin of approximately 1% in accuracy. These findings suggest that clustering-based multi-label classification using k-means constraint clustering and multi-layer neural networks is a promising approach. By leveraging label correlations and reducing dimensionality, the proposed method effectively improves classification performance while addressing issues such as label imbalance and computational inefficiency. Future research may further explore optimization techniques to reduce training time while maintaining high accuracy. |