Gai Y, Liu J. Interpretable unsupervised neural network structure for data clustering via differentiable reconstruction of ONMF and sparse autoencoder.
Neural Netw 2025;
188:107504. [PMID:
40318423 DOI:
10.1016/j.neunet.2025.107504]
[Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/25/2024] [Revised: 03/20/2025] [Accepted: 04/16/2025] [Indexed: 05/07/2025]
Abstract
Neural networks, while powerful, often face significant challenges in terms of interpretability, particularly in clustering tasks. Traditional methods typically rely on post-hoc explanations or supervised learning, which limit their ability to provide transparent, understandable results. The innovation of this method is the integration of Orthogonal Non-negative Matrix Factorization (ONMF) into a clustering layer, which is directly incorporated into the OSINN neural network model and enhanced by a Sparse Autoencoder (SAE). This integration allows for end-to-end model training, which improves clustering performance and provides interpretability by differentiably reconstructing ONMF and extracting sparse features with SAE. OSINN's main innovation is the differentiable reconstruction of ONMF, which creates a transparent clustering layer that is both end-to-end trainable and easily interpretable. Firstly, OSINN, as a transparent neural network with an embedded clustering layer, allows for pre-training model decision predictions, where the model is first trained on a large set of unlabeled data to learn useful features, and then makes clustering decisions based on these learned features. Secondly, by using unsupervised neural networks for clustering, OSINN can handle more complex data, where the network automatically groups similar data points without needing labeled examples, which is especially useful for tasks involving large, unlabeled datasets. Thirdly, as a differentiable version of ONMF, OSINN excels at data clustering by transforming the ONMF method into a differentiable process, making it easy to integrate into the neural network architecture for end-to-end learning, which in turn enhances both interpretability and performance. Experimental results on MNIST, CIFAR-10, Fashion-MNIST, and CIFAR-100 datasets show that OSINN outperforms existing methods, achieving clustering accuracies of 90%, 24%, 64%, and 44%, respectively. Compared to traditional clustering algorithms, OSINN improves accuracy by over 10%, outperforms deep clustering methods by more than 1.2%, and surpasses Non-negative Matrix Factorization (NMF) and Autoencoder (AE) variants by over 1.5%.
Collapse