Abstract
In recent years, graph neural network-based community detection methods have integrated local structure and node attributes, incorporating various optimization strategies with notable progress. However, most current algorithms require predefining the number of communities, introducing human bias, and rely on contrastive objectives or data augmentation, leading to extra hyperparameters and complexity. To address these issues without sacrificing detection quality, we propose an adaptive community detection framework that eliminates contrastive learning and the need for pre-specified community numbers, simplifying training and reducing prior dependency. First, the adaptive detection method is introduced to ensure the identification of high-quality structural communities as reliable global references. Then, a novel mechanism for modeling node-community relationships is proposed, integrating global structure, local structure, and attribute information into a unified space. Finally, a reconstructed soft modularity loss is applied to optimize node-community relationships end-to-end, enhancing community structure without data augmentation or contrastive learning. The proposed approach is efficient to train and computationally lightweight, demonstrating superior detection efficiency and competitive accuracy across multiple graph datasets compared to traditional and recent deep learning methods. The code is available at https://github.com/wuanghoong/Less-is-More.git.
Similar content being viewed by others
Introduction
Numerous complex systems in the real world are composed of interconnected entities. To uncover their underlying principles, such systems are often abstracted as complex network models1, including social networks (Facebook2, Twitter3), computer networks (the Internet, LAN4), and biological networks (gene regulatory networks5, protein-protein interaction networks6), among others. In a complex network, entities are represented as nodes, and the relationships between them correspond to edges. A set of nodes with dense internal connections often shares similar attributes or functions in reality. Such node clusters are referred to as “communities”. Communities exhibit cohesion and homogeneity internally, while connections between different communities are sparse and heterogeneous. Therefore, community detection has become a crucial approach for analyzing complex network structures and uncovering group behaviors and functional patterns.
Community detection aims to identify clusters of nodes within a network that are densely interconnected and share similar attributes, known as community structures, which are characterized by dense internal links and sparser connections between communities7. Traditional community detection methods primarily include techniques such as modularity optimization, spectral clustering, random walks, and label propagation8. While effective in small to medium-sized networks, these methods reveal significant limitations when applied to today’s large-scale complex networks. Many traditional approaches rely solely on topological structure, resulting in high computational complexity that hinders scalability. Moreover, they often overlook rich node attributes, limiting their ability to capture complex relational patterns in real-world networks.
In recent years, deep learning methods have emerged as a new research focus in community detection, leveraging strong nonlinear representation capabilities and efficient large-scale data processing9. The application of deep learning in this field has evolved from graph embedding to graph neural networks (GNNs). Early graph embedding methods10,11,12,13 mapped network nodes into low-dimensional vectors so that structurally similar nodes are close in the embedding space, after which standard clustering algorithms14,15 were applied to group these vectors. However, the inherent disconnection between the embedding and clustering stages in this two-step paradigm motivated the shift toward end-to-end GNN frameworks, which have become the mainstream architecture for community detection. Typical GNN models offer diverse information aggregation schemes for community detection. As a foundational work, GCN (Graph Convolutional Network)16 employs spectral graph convolution to aggregate neighborhood features, yet its equal-weight aggregation struggles to distinguish the importance of different neighbors. GAT (Graph Attention Network)17 addresses this by introducing an attention mechanism, enabling dynamic weighting of neighbor contributions. In contrast, GAE (Graph Autoencoder)18 adopts an unsupervised approach, using an encoder (e.g., GCN) to learn node representations and a decoder to reconstruct the adjacency matrix, which helps preserve community structure in low-dimensional embeddings.
Building on these classic models, researchers have integrated advanced learning strategies with GNN architectures to enhance the perception and utilization of complex network information19. For instance, by designing contrastive tasks20,21,22,23,24,25,26 at the node, subgraph, or community level, GNNs can learn intrinsic structural patterns without relying on ground-truth labels, enabling the identification of stable community features across varying augmented views. Another direction involves guiding the optimization process through differentiable objective functions27,28,29,30,31,32. For example, some methods allow GNNs to jointly optimize both representation learning and community partitioning in an end-to-end manner, by transforming modularity maximization into a differentiable loss function33. These strategies not only expand the expressive power of GNNs in community detection but also promote the development of more adaptive, interpretable, and integrated solutions.
Despite significant progress in GNN-based community detection, overall performance and practical utility remain constrained by several core issues. First, there is an excessive reliance on specific learning strategies: current methods heavily depend on carefully designed graph data augmentation or contrastive learning mechanisms. Their performance is sensitive to the choice of augmentation strategies and the quality of sample pairs, lacking inherent robustness. Second, the optimization process is often overcomplicated. Many advanced models enhance performance by stacking multiple loss functions or introducing complex contrastive mechanisms, resulting in bloated model architectures, increased training difficulty, and challenges in convergence. Finally, adaptive capability remains insufficient: although end-to-end learning has become mainstream, practical GNN models that can fully adaptively determine the number of communities without any prior knowledge are still relatively scarce.
To address the above challenges, the main contributions of this paper are summarized as follows:
-
1.
Adaptive Detection: By leveraging the Louvain algorithm combined with a size-based filtering strategy, the method adaptively extracts high-quality structural communities without pre-defined cluster numbers to provide reliable global structural information.
-
2.
Global-Local-Attribute Integrated Node-Community Relationship Modeling: Within a unified shared embedding space, a single-layer GCN integrates local topology and node attributes to generate node representations, while structural communities are projected as structural center vectors. A soft relational matrix, constructed based on node-center similarity, serves as the optimization objective for community refinement and interpretation.
-
3.
Reconstructed Soft Modularity Objective Based on Node-Community Relations: Using the soft relational matrix, a reconstructed soft modularity loss is formulated, enabling end-to-end optimization of node-community affiliation probabilities without relying on contrastive learning or data augmentation. This results in a more streamlined and robust training process.
-
4.
Efficiency and Practical Applicability: The framework is lightweight and easy to train. Experiments on multiple real-world network datasets demonstrate higher detection efficiency and competitive accuracy compared to both traditional and recent deep learning approaches.
Related work
GNN-based community detection
GNN-based community detection methods learn node embeddings to identify community structures. These approaches leverage the powerful capability of GNNs in modeling graph data, incorporating both topological information and node features to significantly improve detection performance. Compared to traditional methods, GNN-based techniques can more accurately capture complex community patterns, particularly in graphs with rich node attributes. Recent studies have integrated various learning strategies to optimize node representations, thereby enhancing the accuracy and robustness of community detection. These methods not only strengthen the modeling of graph structure but also exploit underlying relationships among node features to further refine community partitioning. CommDGI21 introduces a dual objective combining “community mutual information” and modularity, coupled with a differentiable soft K-means clustering layer, enabling end-to-end joint optimization of GNN representation learning and community detection. SGCMC22 employs a graph attentional autoencoder with a self-supervised mechanism to co-optimize node representations and affinity matrix learning, achieving end-to-end training of multi-view GNNs for soft clustering and significantly enhancing joint modeling of graph structure and nonlinear semantics. DCGL23 addresses generic clustering scenarios without prior graph structures by proposing a pseudo-siamese network that parallelizes GCN and autoencoder. It applies centroid-guided contrastive loss at the feature level and local-global graph contrast at the cluster level to explicitly optimize cluster compactness, significantly improving discriminativity and robustness. DCLN24 incorporates a dual-level contrastive learning mechanism, introducing high-order neighborhood similarity constraints at the node level and dimension decorrelation constraints at the feature level, effectively alleviating representation collapse and enhancing structural and feature discrimination. SCGC25 replaces GNN with a lightweight MLP, condensing multi-hop structures into “influence scores,” and uses an augmentation-free IAC loss to dynamically guide embedding learning, achieving highly efficient and scalable deep graph clustering. CPGCL26 enables a GCN to simultaneously output node embeddings and community distributions, uses community probabilities to weight contrastive loss dynamically, and reinforces high-confidence samples in a self-supervised manner to suppress false negatives and co-optimize community assignment and representation.
Despite the notable progress achieved by GNN-based community detection methods, their performance often heavily relies on data augmentation or contrastive learning strategies. However, improperly designed data augmentation may disrupt the inherent semantic structure of graph data, while the effectiveness of contrastive learning is highly dependent on the quality of positive and negative sample pairs. If poorly constructed, these strategies can mislead the model into learning spurious correlations instead of essential community structures and compromise its generalization capability as a result.
Modularity maximization
Modularity, as one of the standard metrics for evaluating the quality of network community division, primarily assesses the density of connections within communities and the sparsity of links between them34. In the context of modularity maximization, Ulrik Brandes35 demonstrated that maximizing modularity is an NP-complete problem, a finding that spurred the development of heuristic approaches such as spectral relaxation36 and greedy algorithms37. However, these earlier studies focused predominantly on network topology, often overlooking the interrelationships between node attributes. With the rapid advancement of deep learning, integrating deep learning techniques with modularity maximization has emerged as a mainstream optimization strategy in community detection, leading to more accurate and robust partitioning solutions. Yang et al.27 proposed a Deep Nonlinear Reconstruction (DNR) method, which uses stacked autoencoders to perform nonlinear low-dimensional embedding and reconstruction of the modularity matrix, overcoming the limitations of traditional linear methods in representation capability. Alexandre Hollocou et al.28 introduced a soft clustering relaxation method based on modularity maximization, along with an efficient local sparsification algorithm, allowing nodes to probabilistically belong to multiple communities. Guillaume Salha-Galvan et al.29 developed a modularity-aware graph autoencoder that incorporates community-preserving message passing and a modularity-inspired regularization loss, effectively integrating graph structure and community information during encoding to significantly improve detection performance. DGCluster30 proposed a deep graph clustering framework based on differentiable modularity maximization. By softening the modularity objective and combining it with GNN-based community similarity, it achieves efficient clustering without predefining the number of communities. MAGI31 reformulated modularity maximization from a contrastive learning perspective, showing its equivalence to graph contrastive learning guided by modularity coefficients, and proposed a community-aware self-supervised pretraining task that captures high-order proximity without graph augmentation. MOMCD32 introduced a motif-weighted modularity optimization model, integrating high-order motif structures and low-order edge information into a unified weighting scheme. By constructing a motif adjacency matrix and defining a weighted modularity metric, it uses heuristic algorithms to maximize modularity, enabling higher-quality and higher-order community detection.
The integration of deep learning with modularity not only enables effective fusion of node features and network structure from complex networks but also injects community-level semantic information into node representation learning through the objective of modularity maximization. Furthermore, end-to-end deep learning frameworks avoid the complex iterative processes of traditional optimization algorithms, significantly reducing computational overhead.
Table 1 provides a comparative summary of the aforementioned relevant algorithms and the method proposed in this paper, offering a clear overview of the design differences among them. Although the MOMCD method in Table 1 appears to be consistent with our method at first glance, there are essential differences between them. Both MOMCD and Louvain belong to traditional methods and do not take into account the attribute features of nodes.
Preliminaries
Definition 1
Undirected Attributed Graph. An undirected attributed graph can be formally defined as a triple \(G = (V,E,X)\). Where \(V=\{v_1, v_2, v_3, \ldots , v_n\}\) is the set of nodes, and \(n = |V|\) denotes the total number of nodes in the network. \(E \subseteq V \times V\) is the set of edges, representing pairwise relationships between nodes. An edge \(e_{ij} = (v_i, v_j) \in E\) indicates a connection between nodes \(v_i\) and \(v_j\). Let \(M=|E|\) denote the total number of edges. \(X = \left\{ x_1, x_2, x_3, \ldots , x_n \right\} \in \mathbb {R}^{n \times d}\) is the node attribute matrix. The i-th row vector \(x_i \in \mathbb {R}^d\) corresponds to the d-dimensional feature representation of node \(v_i\). The topological structure of the graph is characterized by the adjacency matrix \(A \in \left\{ 0, 1 \right\} ^{n \times n}\) , where \(a_{ij}=1\) if and only if \(\left( v_i, v_j \right) \in E\) , otherwise \(a_{ij}=0\). Based on the adjacency matrix, the degree matrix \(D=diag(d_1,d_2,d_3,...,d_n)\) is defined as a diagonal matrix whose elements \(d_i = \sum _{j=1}^{n} a_{ij}\) represent the degree of node \(v_i\), i.e., the number of neighbors directly connected to it.
Definition 2
Graph convolutional networks. The core idea of GCN originates from spectral filtering of graph signals in spectral graph theory. Its layer-wise propagation rule represents a first-order Chebyshev polynomial approximation of the graph Laplacian operator, achieving efficient spatial-domain neighborhood aggregation. The propagation rule for a single-layer GCN is given by:
where \(Z^{(l)} = \left\{ z_1^{(l)}, z_2^{(l)}, z_3^{(l)}, \ldots , z_n^{(l)} \right\} \in \mathbb {R}^{n \times d'}\) denotes the node representation matrix at the \(l\)-th layer, with the initial input \(Z^{(0)} = X\). \(\widetilde{A} = \hat{D}^{-\frac{1}{2}} \hat{A} \hat{D}^{-\frac{1}{2}}\) is the normalized adjacency matrix, which stabilizes the training process and mitigates gradient explosion or vanishing. \(\hat{A} = A + I\) is the adjacency matrix with self-loops added, where \(I\) is the identity matrix, ensuring that each node retains its own features during aggregation. \(\hat{D}\) is the degree matrix of \(\hat{A}\), i.e., \({\hat{D}}_{ii} = \sum _{j} {\hat{A}}_{ij}\). \(W^{(l)}\) is the trainable weight parameter matrix of the l-th layer. \(\sigma (\cdot )\) is the nonlinear activation function (PReLU), enhancing the model’s nonlinear expressive capacity and feature discrimination ability.
The objective of node representation learning is to obtain a low-dimensional, dense vector representation \(z_i\) for each node through stacked GCN layers, which simultaneously encodes both its local topological structure and intrinsic attribute features.
Definition 3
Structural community centers. The purpose of structural community centers is to learn a prototype vector for each potential community in the graph, which represents the core characteristics of that community in the feature space. Given the learned embedding matrix \(H \in \mathbb {R}^{n \times d'}\) from representation learning and the pre-detected structural communities \(\{ C_1, C_2, \dots , C_k \}\), the structural center \(u_j\) of the j-th community can be computed by aggregating the representations of its member nodes. A common approach is to compute the mean vector:
where \(C_j\) denotes the set of nodes assigned to the j-th community. The set of all structural community centers \(U = \{ u_1, u_2, \dots , u_k \}\) forms a compact representation of the global community structure in the graph. In community detection, these centers serve as reference points, enabling the inference of node-community assignments by computing the similarity between node representations and each center.
Definition 4
Modularity. Modularity Q is a widely used metric in network science for quantifying the quality of a given community partition by measuring its deviation from a random connectivity null model with the same degree distribution. Its mathematical definition is as follows:
where \(M\) represents the total number of edges in the network, \(a_{ij}\) is an element of the adjacency matrix A, and \(d_i\) denotes the degree of node \(v_i\). The term \(\frac{d_i d_j}{2M}\) indicates the expected number of edges between nodes \(v_i\) and \(v_j\). The variable \(c_i\) represents the community label of node \(v_i\), and \(\delta (\cdot )\) is the Kronecker delta function, which equals 1 if two nodes belong to the same community and 0 otherwise. Today, modularity maximization is often directly employed as the optimization objective in community detection algorithms.
Methods
This section outlines the framework of the proposed method, as illustrated in Fig. 1. First, in the global structure extraction phase, the Louvain algorithm combined with a size-based filtering strategy is employed to adaptively identify structural communities with meaningful global topology, while effectively filtering out noisy clusters. This yields reliable global structural guidance. Second, a lightweight single-layer GCN is utilized to integrate node attributes and local structural information, generating low-dimensional embeddings that preserve essential relational patterns. Subsequently, through a mean aggregation operation, each structural community is mapped to a structural center within the same embedding space. A fine-grained node-community relational matrix is then constructed based on the similarity between node embeddings and these structural centers. Finally, a reconstructed soft modularity loss, derived from the modeled relationships, is optimized to directly refine the community membership associations among closely connected nodes, thereby deeply excavating the latent community structure.
Adaptive structural community pre-detection based on global structure
In real-world networks, structure can be analyzed from both local and global perspectives. The global structural perspective helps capture the overall distribution pattern of communities but is often sensitive to noisy clusters. Conversely, the local structural perspective effectively identifies high-connectivity patterns among nearby nodes but is susceptible to interference from inter-community connections. As shown in the top-right corner of Fig. 1, the differences between the two are clearly noticeable.
To overcome the limitations of the local perspective and achieve adaptive community discovery, this paper proposes a global structure-guided approach for adaptive community identification and high-quality global structure extraction. The core idea is to leverage the macro-level distribution information embedded in the global network structure to provide boundary constraints and validation for partitioning locally dense regions, thereby dynamically guiding the delineation of community boundaries. The Louvain algorithm accomplishes adaptive community detection through iterative optimization of the network topology, continuously merging adjacent nodes and communities. Under the constraint of modularity, this method ensures both clear separation between communities and high cohesion within them, aligning well with our design principles of adaptive partitioning and global structure discovery.
Using the Louvain algorithm, we first extract preliminary structural communities \(C=\{c_1,c_2,c_3,\ldots ,c_t \}\), with the number of communities denoted as \(t\). Although these communities preserve the overall structural information of the graph, many of them are small in size and isolated, contributing little to the global structure and potentially introducing noise. To address this, we propose an adaptive size-based filtering mechanism. The threshold \(T\) is determined by calculating the mean \(\mu\) and standard deviation \(\sigma\) of community sizes, and only communities exceeding this threshold are retained. The threshold \(T\) is calculated as follows:
Finally, the structural communities that satisfy the condition of having a community node count \(|c_i |\ge T\) are denoted as \(S=\{s_1,s_2,s_3,\ldots ,s_k \}\), where \(k\) represents the number of structural communities that meet the requirements.
The threshold defined using the mean and standard deviation can automatically adapt to the size of the dataset without manual tuning. When community sizes vary greatly, the mean may decrease; however, the standard deviation will increase substantially due to the extreme differences. This effect raises the threshold, ensuring the retention of medium- and large-sized core communities that are more representative of the overall structure, thereby providing a stable and high-quality structural foundation.
Fusion learning of local structural information and feature information
In attributed graphs, node attributes provide crucial feature information for community detection, effectively compensating for the limitations of relying solely on topological structure. Specifically, attribute features can transcend the constraints of topological connectivity, enabling nodes that are not directly connected but share similar attributes to form meaningful clusters in the feature space. This offers important clues for resolving the ambiguity in community assignments for nodes located at community boundaries.
Original node attributes typically contain only node-specific features and lack the capacity to capture relationships between related nodes, making it challenging to identify potentially similar nodes. To address this limitation, we employ GCNs to explicitly integrate node features with local structural information. GCNs utilize a symmetrically normalized adjacency matrix with self-connections, ensuring numerical stability during aggregation and mitigating the influence of node degree disparities. Through this aggregation process, GCNs combine each node’s features with those of its neighbors, enhance representational power and discriminative ability via nonlinear activation functions, and produce low-dimensional embeddings \(Z\in \mathbb {R}^{n\times d'}\) that jointly encode attribute information and local structural patterns, thereby providing a robust foundation for subsequent community partitioning.
L2 normalization achieves feature scaling by dividing each vector by its Euclidean norm, thereby eliminating the influence of scale variation on similarity calculations. To this end, we apply L2 normalization to the node representations \(Z\) learned by the GCN, obtaining the normalized embedding vectors \(H\in \mathbb {R}^{n\times d'}\). The final node representation learning process is formulated as follows:
Normalized vectors exhibit geometric relationships determined solely by their directions, independent of their magnitudes. This property not only enhances computational efficiency but also ensures gradient stability during propagation.
Node-community relationship modeling mechanism
We propose a multi-source information fusion mechanism for node-community relationship modeling, aiming to reconcile structural information and node features within a unified perspective, thereby directly capturing and revealing associations between nodes and communities. The core of this mechanism lies in mapping global structure, local topology, and node attributes into a unified embedding space, achieving alignment and integration of heterogeneous information sources. Within this space, soft membership relationships between nodes and communities are directly modeled, effectively leveraging the complementary advantages of multi-source information.
In this mechanism, the first step involves mapping global structural information into the continuous embedding space. Using the high-quality structural communities S with reference to the node representation matrix H, we compute a center matrix \(U\in \mathbb {R}^{k\times d'}\), where the representation of the j-th community \(s_j\) is derived by averaging the representations of all its member nodes:
where \(|s_j |\) represents the number of nodes in the j-th community, and \(h_i\) represents the representation of j-th node belonging to the \(s_j\) community. This operation materializes the abstract community structure \(s_j\) into a point \(u_j\), effectively mapping global structural information into a continuous vector space and laying the foundation for subsequent relationship modeling.
Following the global structure mapping, the second step infers the relationships between nodes and communities within the same space. Specifically, we quantify the association strength between a node and a community by measuring the similarity between the node representation \(h_i\) and each structural community center \(u_j\). Thanks to the normalization of the feature representations, using cosine similarity to accurately measure the similarity between node \(h_i\) and structural center \(u_j\) in the embedding space is equivalent to performing a vector dot product operation.
To make this similarity more intuitively reflect node-to-community assignments, the similarities are normalized using a Softmax function, yielding a node-community affiliation matrix \(P\in \mathbb {R}^{n\times k}\),
where \(\delta\) is a temperature hyperparameter that controls the sharpness of the community distribution. By integrating global structure, local topology, and node attributes within a unified embedding space, this mechanism effectively leverages the complementary strengths of multi-source information. It not only offers more accurate membership determination for nodes at structural boundaries but also identifies and associates topologically disconnected nodes with high attribute similarity, ultimately enhancing the accuracy and robustness of community detection. Finally, the community to which a node belongs is determined according to the principle of maximum membership probability,
Soft modularity-based community optimization
Modularity serves as a core metric for evaluating the quality of network community partitions, and its optimization process directly determines the rationality of community discovery. Traditional modularity functions rely on hard assignments, using discrete indicator functions to determine whether nodes belong to the same community. This approach fails to capture the strength of node-to-community affiliations and lacks flexibility in handling boundary nodes. To address this limitation, this study proposes a soft modularity function that incorporates a node-community affiliation probability matrix. This matrix transforms discrete community assignments into continuous probabilistic representations, enabling fine-grained optimization of node-community membership. By replacing the hard-assignment indicator function in traditional modularity with an inner product form of affiliation probabilities, we derive the soft modularity function \(Q'\),
where \(a_{ij}\) is an element of the adjacency matrix A, \(d_i=\sum _{j} a_{ij}\) denotes the degree of node \(v_i\), and \(p_{im}\) represents the affiliation probability of node i to community m. This design ensures that a node’s contribution to a community is proportional to its affiliation probability, allowing nodes to participate in the structural optimization of multiple communities through soft assignments. To simplify computation and improve optimization efficiency, the soft modularity function is converted into matrix form:
where \(B\) denotes the modularity matrix, and \(tr()\) denotes the trace of a matrix. This transformation not only reduces computational complexity but, more importantly, facilitates subsequent gradient-based optimization, enabling the modularity maximization process to be embedded into an end-to-end neural network training framework. The final loss function is:
\(\alpha\) is the scaling factor. During the optimization process, the magnitude of a single loss value affects the gradient’s variation. Excessive changes in gradient magnitude can lead to an unstable training process. By setting an appropriate loss scaling factor \(\alpha\), the loss values can be scaled to enhance the stability of the optimization, ensuring a smoother optimization process and preventing the algorithm from getting stuck in local optima.
Thus, unlike contrastive learning methods that rely on selecting positive and negative samples to make closely connected nodes more similar in the feature space, our approach leverages the node’s local structural connections, global structural position, and feature information to discover effective community-level information that enhances the correlation between nodes and their communities. This means that nodes with close connections within the same community will become increasingly similar. The specific process of the proposed method is shown in Algorithm 1.
Are contrastive learning strategies necessary
In recent years, a large number of community detection methods have been developed that optimize node representations by incorporating contrastive learning strategies, thereby improving clustering performance. These methods typically construct positive and negative sample pairs based on structural proximity or semantic similarity, and enhance the separability of different communities in the embedding space by maximizing the consistency of positive sample pair representations and minimizing the consistency of negative sample pair representations.
In this study, we attempt to integrate three representative contrast strategies into the proposed method: contrast loss designs from CommDGI21, SupCon40, and MAGI31. In CommDGI, nodes within the same community as the current center are selected as positive samples, while nodes drawn from the nearest different community are used as negative samples. For SupCon and MAGI, the positive samples are defined as first-order neighbors within the same community, and the negative samples are nodes drawn from the nearest different community. This sampling scheme pulls together representations of nodes from the same class and pushes apart those from different classes in the embedding space, thereby learning more discriminative embeddings.
The CommDGI contrastive loss is computed as follows:
where \(M_+\) denotes the positive sample set, and \(v_+\) represents a positive sample node. Similarly, \(M_-\) represents the negative sample set, and \(v_-\) represents a negative sample node. \(D()\) denotes a discriminator, computed by applying a sigmoid function to the dot product of two vectors.
The SupCon contrastive loss is computed as follows:
The MAGI contrastive loss is computed as follows:
Finally, all the aforementioned contrastive losses are collectively denoted as \(L_{contarst}\), and incorporated into the total loss function as:
By introducing different contrastive losses within the same framework, we can systematically evaluate the performance gains brought by contrastive strategies to our method and further verify the effectiveness of our model in the absence of such strategies. This design not only helps quantify the contribution of each contrastive strategy but also reveals the advantages of our model in simplifying the training process, reducing computational complexity, and enhancing generalization capability.
Experimentation
The experiments were conducted on a computer with an Intel i9 processor, 128GB of RAM, and the Windows 11 operating system, using a Python 3.8 environment for programming and computation.
Datasets
The datasets used in our experiments can be categorized as follows:
Citation Networks41: Including Citeseer, Cora, and Pubmed, where nodes represent publications, edges represent citation relationships, node features are either bag-of-words or TF-IDF features, and labels correspond to publication topics.
Co-authorship Networks: ACM42 is a paper network where edges indicate shared authorship, while CoCS is a scholar network where edges represent co-authorship. Node features for both are bag-of-words from keywords, and labels denote research fields or subject categories, respectively.
Product Co-purchasing Networks: Amazon-Computers (Amac) and Amazon-Photo (Amap)43 contain computers and photography-related products, respectively, with edges indicating frequent co-purchase. Features are derived from review bag-of-words, and labels are product categories. Electronics-Photo (Ele-photo)44 is a network of electronics products based on co-purchase or co-view relationships.
Social Networks: Film45 is an actor network where edges denote co-appearance on the same Wikipedia page; UAT46 is an airport network where edges represent commercial flight routes. Labels indicate actor genres and passenger traffic levels, respectively.
The dataset is processed and provided by47, with a detailed description of the dataset shown in Table 2.
Comparison models
This paper selects eight representative community detection algorithms for comparison to comprehensively evaluate the performance of the proposed method. These algorithms cover three main paradigms: attribute-based, structure-based, and methods that integrate both attributes and structure.
K-means15 is a classic attribute-based clustering algorithm. It partitions nodes solely based on the distribution of their feature vectors in the latent space, serving as a baseline for attribute-only clustering.
Louvain38 is a heuristic structure-based algorithm. It iteratively merges nodes to maximize modularity, serving as a benchmark for pure structural methods.
Six baseline algorithms that integrate attributes and structure are selected:
CommDGI21 learns node embeddings by maximizing the mutual information between local node representations and a global graph summary.
DGCluster30 transforms modularity maximization into a differentiable loss function, enabling end-to-end joint optimization with GNN-based representation learning.
DCGL23 employs a pseudo-siamese network architecture to extract features from structural and attribute perspectives separately, enhancing the representations through cross-view contrastive learning.
MAGI31 uses the modularity matrix as an anchor for contrastive learning, capturing high-order structural similarity without requiring graph augmentation.
MGCN39 designs multi-hop graph convolution to adaptively fuse information from higher-order neighborhoods, learning more comprehensive node representations.
CPGCL26 jointly learns node representations and soft community assignments, dynamically refining sample pairs in contrastive learning to alleviate the false negative problem.
Evaluation metrics and parameter settings
In this experiment, the community detection task in attribute graphs will be the main focus, and the performance of all community detection methods will be compared. To evaluate the quality of the predicted communities, we use eight evaluation metrics: DBI, Q, NMI, ACC, F1-score, ARI, FMI and SC to assess the effectiveness of the community detection results. The DBI metric primarily measures the similarity and separation between the detected communities, aiming to make communities more compact internally and more separated from each other. A smaller DBI value is preferable. For the other metrics, higher values are better, that is
where k represents the number of communities, \(avg(c_i )\) represents the average distance of all nodes in the i-th community to its center, and \(d_{cen} (u_i+u_j)\) represents the distance between the centers of the i-th and j-th communities.
The NMI metric is normalized based on the concept of mutual information from information theory, and is used to measure the similarity between the community detection results and the ground truth, that is
where \(G\) represents the ground truth, and \(C\) represents the community detection results. \(P(c,g)\) denotes the joint probability distribution of a node being in both the true community g and the detected community c. \(P(g)\) represents the probability of a node being in the true community g. \(MI()\) represents mutual information, and \(H()\) represents entropy.
The ACC metric is used to measure the consistency between the community detection results and the ground truth, that is
where \(y_i\) represents the true community label of i-th node, and \(\hat{y}_i\) represents the community detection label of the node. \(\rho\) is an indicator function that takes the value 1 if the true label and the detected label are the same, and 0 otherwise. \(map(\hat{y}_i )\) denotes the mapping of the detection label of node i to the true label.
The F1-score provides a comprehensive evaluation of the model’s precision and recall, measuring the consistency between the communities detected by the algorithm and the true communities, that is
where TP represents the number of nodes predicted to belong to community c and actually belong to c, FP represents the number of nodes predicted to belong to community c but do not belong to c, FN represents the number of nodes that actually belong to community c but are predicted not to belong to c.
The ARI is a metric used to measure the similarity between detection results and true labels, that is
where TN represents the number of nodes that do not actually belong to community c and are predicted not to belong to c, and E[] denotes the expected value.
The FMI metric measures the geometric mean of precision and recall for community detection results, balancing the trade-off between false positives and false negatives, that is
The SC metric evaluates the compactness within communities and the separation between communities, that is
where SC(x) is the Silhouette Score of node x. a(x) is the mean intra-cluster distance of node x, and b(x) is the minimum mean distance from x to any other cluster. SC ranges from \(-1\) to 1, with higher values indicating more coherent and well-separated clusters.
In our work, a single-layer convolutional network is used for node representation learning. The hyperparameter \(\delta\) is set to 30, and the loss coefficient \(\alpha\) is set to 0.001. The Adam optimizer is used for 300 iterations of model training, with a learning rate of 0.001 and weight decay set to 0.005. In all experiments, the dimension of node representations is fixed at 512. For comparison experiments, the number of communities to be specified is set to the number of communities detected in this experiment. All other parameters for baseline methods follow their original papers to ensure optimal performance. For example, CommDGI uses a learning rate of 0.001 for 500 iterations; DGCluster uses 0.001 for 300 iterations; DCGL uses 0.001 for 300 iterations; MAGI uses 0.0005 for 400 iterations; MGCN uses 0.003 for 700 iterations; and CPGCL uses 0.0007 for 600 iterations.
Experiment result
This paper compares the performance of seven community detection methods for the community detection task. Table 3 summarizes the results of the community performance comparison across different algorithms. Bold numbers indicate the best performance, while underlined numbers indicate the runner-up performance.
As shown in Table 3, the proposed algorithm achieves the best or runner-up performance in most evaluation metrics across different datasets. Compared to other deep learning-based methods, our approach incorporates global, local structural, and feature information for community detection, enhancing modularity loss by leveraging community membership probabilities derived from these three information sources. By maximizing the improved modularity loss, the algorithm effectively uncovers the community memberships of nodes, leading to strong experimental results in the community detection task.
Comprehensive algorithm evaluation and statistical validation
Experiment 1: Wilcoxon signed-rank test for algorithm comparison
The Wilcoxon signed-rank test48 is a non-parametric statistical method used to compare paired or related samples. This study adopted a significance level of 0.05 to assess the differences between the proposed algorithm and comparative algorithms. The results are recorded in Table 4. The terms R+ and R- represent the sum of ranks where the proposed algorithm performed superior or inferior to its competing algorithms, respectively. A p-value below 0.05 indicates a statistically significant difference between the proposed algorithm and the compared algorithm. Specifically, when the p-value is less than 0.05, it signifies a significant difference between the proposed algorithm and the comparative algorithm, which is highlighted in bold in the table.
Table 4 presents the Wilcoxon test results for the ACC, F1, ARI, and SC metrics of the algorithms. The data results demonstrate that the metrics of the proposed algorithm differ significantly from those of the comparison algorithms, thus validating its strong performance across the 10 datasets.
Experiment 2: Comprehensive ranking experiment based on multi-criteria decision-making
To comprehensively evaluate the overall performance of different algorithms, this study adopts the TOPSIS49 method as a multi-criteria decision-making (MCDM) approach50. Considering that the eight evaluation metrics used in this experiment are of diverse types and that some of them are highly correlated, the CRITIC method51 is employed to automatically compute objective weights, thereby avoiding the subjectivity of manual weight assignment.
The experimental procedure is as follows: first, missing values and special entries (e.g., OM, N/A) in the original evaluation metrics are processed, followed by orientation unification and normalization to eliminate differences in scale and direction; next, the CRITIC method is applied to determine the weights by jointly considering the discrimination power and information independence of each metric; finally, the TOPSIS method is used to calculate the closeness coefficient of each algorithm to the positive and negative ideal solutions, and the algorithms are ranked in descending order of the coefficient, yielding an objective and unified performance ranking in the multi-dimensional evaluation system.
As shown in Table 5, our method consistently captures the overall performance differences among algorithms across multiple real-world datasets. It achieves the top rank on most datasets, thereby demonstrating its superiority in terms of comprehensive performance across multiple evaluation metrics.
Ablation experiments
Experiment 1: Effectiveness analysis of the adaptive structural community extraction module
The first ablation experiment aims to remove the Louvain-based adaptive structural community extraction module to verify the effectiveness of global structure information. Since the proposed method relies on global centroids during the fusion stage to compute node-community memberships, in the absence of Louvain, we directly perform K-means clustering on the learned node embedding matrix \(H\), using the ground-truth number of communities \(K_R\) as the number of clusters to generate global centroids. These centroids then replace the structural community centers extracted by Louvain in the subsequent steps, allowing us to evaluate the contribution and necessity of Louvain pre-detection to the quality of global structural information.
The experimental results are shown in Table 6. Specifically, the model incorporating global structural information outperforms the counterpart without global structural information across multiple evaluation metrics, indicating that global structural information plays a significant role in improving the accuracy and stability of community partitioning, thereby highlighting its necessity in the proposed method.
Experiment 2: Effectiveness analysis of the contrastive strategy
To evaluate the impact of different contrastive learning strategies on the performance of our method, we incorporated three representative contrastive losses–\(L_{CommDGI}\), \(L_{SupCon}\) and \(L_{MAGI}\)–into the same framework and compared them with a baseline model without contrastive optimization. All experiments were conducted on the same datasets (Cora, Acm, Amap, and Uat) to ensure comparability, with the contrastive loss coefficient \(\beta\) tested at values of 1, 0.1, 0.01, and 0.001, and the best-performing setting (\(\beta\)=0.001) adopted for reporting results. Performance was quantitatively assessed using six metrics: DBI, Q, NMI, ACC, F1-score, and ARI.
The detailed results are presented in Tables 7, 8 and 9, where Table 7 compares the results with and without \(L_{CommDGI}\), Table 8 compares the results with and without \(L_{SupCon}\), and Table 9 compares the results with and without \(L_{MAGI}\). These comparisons intuitively reveal the actual performance gains of different contrastive strategies and validate the effectiveness of our method in the absence of contrastive optimization.
Parameter analysis
Experiment 1: Sensitivity Analysis of the Threshold Filtering Mechanism
This experiment analyzes the threshold filtering mechanism applied after Louvain-based community partitioning. To evaluate the sensitivity of this mechanism to threshold selection, we vary the standard deviation coefficient from 0.1 to 1.0 and systematically examine the impact of different threshold strategies on both the fidelity of global information and the final detection performance. This process helps verify the effectiveness and robustness of the filtering mechanism in balancing noise suppression with information preservation.
We analyzed the trends of the Q, ACC, and F1 metrics on the Acm, Amap, Uat, and Cocs datasets. As shown in Fig. 2, when the standard deviation coefficient is in the range of 0.4-0.6, most metrics achieve both superior and stable performance. In particular, a coefficient of 0.5 achieves a good balance among Q, ACC, and F1.
Experiment 2: Sensitivity Analysis of the \(\alpha\)Parameter
In this experiment, we investigate the impact of different values of \(\alpha\) on the performance of the experiment across various datasets. Specifically, we set \(\alpha\) = 1.0, 0.1, 0.01, 0.001 and perform experiments on the Cora, Citeseer, ACM, and AMAP datasets, observing the Q, NMI, and ACC metrics. The experimental results shown in Fig. 3 demonstrate that when \(\alpha\) is set to 0.001, the performance is optimal. This is because a smaller modularity loss helps mitigate the problem of modularity optimization getting stuck in local optima, preventing overfitting caused by forcing the algorithm to rigidly determine community memberships. A more comprehensive analysis of the relationships between all communities allows for better extraction of community membership information.
Depth sensitivity analysis of GCN
To further investigate whether increasing the number of GCN layers can enhance the model’s ability to capture deep local features and attribute associations, we designed a depth sensitivity experiment using single-layer, two-layer, and three-layer GCN encoders. All parameters other than the number of layers were kept consistent with the baseline to ensure result comparability.
The experimental results, as shown in Table 10, indicate that excessively deep GCNs suffer from over-smoothing, leading to a decline in accuracy. Since the proposed method already incorporates rich global structural information, a single-layer GCN is sufficient to capture the necessary local structural features, thereby achieving a better balance between performance and computational efficiency.
Runtime comparison
In this experiment, the runtime comparison of various deep learning-based algorithms is conducted on four different datasets: Cora, Citeseer, Acm, and Uat. As shown in the experimental comparison in Fig. 4, our algorithm achieves the fastest runtime across different datasets. The reason for this is that the community detection framework designed in this paper employs the fast Louvain algorithm for pre-community detection to obtain global structural information, and uses a basic GCN to integrate local structural and feature information for learning node representations. These useful pieces of information are then fused to compute the community membership probabilities of nodes. Finally, modularity-based community optimization is performed using the membership probabilities to mine the community information, which is faster and more effective compared to other deep learning-based algorithms. Our community detection framework reduces the extra view generation for data augmentation, the construction of positive and negative samples for contrastive learning, and the joint optimization of multiple objectives, achieving rapid and effective detection with a simple yet efficient framework and optimization approach.
Visualization comparison of community detection results
To intuitively verify the effectiveness of the algorithm presented in this paper, we use the T-distributed stochastic neighbor embedding (T-SNE) algorithm52 to visualize the final node representations and community partition results in a two-dimensional space, as shown in Fig. 5. In this figure, we compare the community detection results after learning node representations based on the original features using K-means and deep learning on the Cora dataset. Our method more effectively distinguishes community differences and ensures community cohesion.
Conclusion
In this study, we proposed a straightforward and effective approach for community detection. Our method features adaptive detection, identifying high-quality structural communities without needing a predefined number of communities. It also includes node-community relationship modeling, which integrates local topology and node attributes into a shared embedding space, allowing us to model the soft relationships between nodes and community centers. Additionally, we implement soft modularity reconstruction, which optimizes community partitioning in an end-to-end manner using the soft relational matrix, without relying on contrastive learning or data augmentation. Our approach offers a new perspective on the research landscape, hoping to inspire researchers to develop improved algorithms for community detection.
We validate the effectiveness of our proposed algorithm through comparative experiments using various metrics, including DBI, Q, NMI, ACC, F1, ARI, FMI, and SC. In future work, we plan to explore different community detection techniques and social network analysis methods, focusing on optimizing the acquisition of global structural information.
Data availability
The datasets generated during and/or analysed during the current study are available in the Github repository, https://github.com/wuanghoong/Less-is-More.git.
References
Gao, Y. & Zhang, H. Community detection based on graph diffusion: A survey of definitions, challenges, and models. Expert Syst. Appl. 290, 128396 (2025).
Gangarde, R., Sharma, A. & Pawar, A. Clustering approach to anonymize online social network data. In Proceedings of the International Conference on Sustainable Computing and Data Communication Systems (ICSCDS), 1070–1076 (2022).
Zhang, Y., Wu, Y. & Yang, Q. Community discovery in twitter based on user interests. J. Comput. Inf. Syst. 8, 991–1000 (2012).
Fortunato, S. Community detection in graphs. Phys. Rep. 486, 75–174 (2010).
Gupta, S. & Singh, D. P. Recent trends on community detection algorithms: A survey. Modern Phys. Lett. B 34, 2050408 (2020).
Akhter, N. & Shehu, A. From extraction of local structures of protein energy landscapes to improved decoy selection in template-free protein structure. Molecules 23(1), 216 (2018).
Singh, D. & Garg, R. A survey of community detection algorithms and its comparative performance analysis. Comput. Sci. Rev. 58, 100799 (2025).
Dang, T. A. & Viennet, E. Community detection based on structural and attribute similarities. In Proceedings of the International Conference on Digital Society (ICDS), 7–12 (2012).
Su, X. et al. A comprehensive survey on community detection with deep learning. IEEE Trans. Neural Netw. Learn. Syst. 35, 4682–4702 (2024).
Perozzi, B., Al-Rfou, R. & Skiena, S. Deepwalk: Online learning of social representations. In Proceedings of the 20th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, 701–710 (2014).
Grover, A. & Leskovec, J. node2vec: Scalable feature learning for networks. In Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, 855–864 (2016).
Ahmed, A., Shervashidze, N., Narayanamurthy, S., Josifovski, V. & Smola, A. J. Distributed large-scale natural graph factorization. In Proceedings of the 22nd International Conference on World Wide Web, 37–48 (2013).
Ou, M., Cui, P., Pei, J., Zhang, Z. & Zhu, W. Asymmetric transitivity preserving graph embedding. In Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, 1105–1114 (2016).
Girvan, M. & Newman, M. E. J. Community structure in social and biological networks. Proc. Natl. Acad. of Sci. 99, 7821–7826 (2002).
Lloyd, S. Least squares quantization in pcm. IEEE Trans. Inf. Theory 28, 129–137 (1982).
Kipf, T. N. & Welling, M. Semi-supervised classification with graph convolutional networks. In Proceedings of the International Conference on Learning Representations (ICLR), 1–14 (2017).
Veličković, P. et al. Graph attention networks. In Proceedings of the International Conference on Learning Representations (ICLR), 1–12 (2018).
Kipf, T. N. & Welling, M. Variational graph auto-encoders. In Proceedings of the 30th International Conference on Neural Information Processing Systems (NeurIPS), 1–9 (2016).
Zhou, S. et al. A comprehensive survey on deep clustering: Taxonomy, challenges, and future directions. ACM Comput. Surv. 57, 38 (2024).
Veličković, P. et al. Deep graph infomax. In Proceedings of the International Conference on Learning Representations (ICLR), 1–12 (2019).
Zhang, T. et al. Commdgi: Community detection oriented deep graph infomax. In Proceedings of the 29th ACM International Conference on Information and Knowledge Management, 1843–1852 (2020).
Xia, W., Wang, Q., Gao, Q., Zhang, X. & Gao, X. Self-supervised graph convolutional network for multi-view clustering. IEEE Trans. Multim. 24, 3182–3192 (2021).
Chen, M., Wang, B. & Li, X. Deep contrastive graph learning with clustering-oriented guidance. Proceedings of the AAAI Conference on Artificial Intelligence 38, 11364–11372 (2024).
Peng, X., Cheng, J., Tang, X., Liu, J. & Wu, J. Dual contrastive learning network for graph clustering. IEEE Trans. Neural Netw. Learn. Syst. 35, 10846–10856 (2024).
Kulatilleke, G. K., Portmann, M. & Chandra, S. S. Scgc: Self-supervised contrastive graph clustering. Neurocomputing 611, 128629 (2025).
Li, H., Han, F. & Wang, W. Cluster-perceptive graph contrastive learning for community detection. In Proceedings of the IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), 1–5 (2025).
Yang, L. et al. Modularity based community detection with deep learning. In Proceedings of the 25th International Joint Conference on Artificial Intelligence (IJCAI), 2252–2258 (2016).
Hollocou, A., Bonald, T. & Lelarge, M. Modularity-based sparse soft graph clustering. In Proceedings of the Twenty-Second International Conference on Artificial Intelligence and Statistics, 323–332 (2019).
Salha-Galvan, G., Lutzeyer, J. F., Dasoulas, G., Hennequin, R. & Vazirgiannis, M. Modularity-aware graph autoencoders for joint community detection and link prediction. Neural Netw. 153, 474–495 (2022).
Bhowmick, A., Kosan, M., Huang, Z., Singh, A. & Medya, S. Dgcluster: A neural framework for attributed graph clustering via modularity maximization. Proceedings of the AAAI Conference on Artificial Intelligence 38, 11069–11077 (2024).
Liu, Y. et al. Revisiting modularity maximization for graph clustering: A contrastive learning perspective. In Proceedings of the 30th ACM SIGKDD Conference on Knowledge Discovery and Data Mining, 1968–1979 (2024).
Xiao, J., Zou, Y.-C. & Xu, X.-K. Higher-order community detection by motif-based modularity optimization. IEEE Trans. Big Data 11, 2529–2544 (2025).
Aref, S. & Mostajabdaveh, M. Analyzing modularity maximization in approximation, heuristic, and graph neural network algorithms for community detection. J. Comput. Sci. 78, 102283 (2024).
Newman, M. E. J. & Girvan, M. Finding and evaluating community structure in networks. Phys. Rev. E 69, 026113 (2004).
Brandes, U. et al. On modularity clustering. IEEE Trans. Knowl. Data Eng. 20, 172–188 (2008).
Newman, M. E. J. Finding community structure in networks using the eigenvectors of matrices. Phys. Rev. E 74, 036104 (2006).
Newman, M. E. J. Fast algorithm for detecting community structure in networks. Phys. cal Rev. E 69, 066133 (2004).
Blondel, V. D., Guillaume, J.-L., Lambiotte, R. & Lefebvre, E. Fast unfolding of communities in large networks. J. Stat. Mech.: Theory Exp. 2008, P10008 (2008).
Li, X., Wu, W., Zhang, B. & Peng, X. Multi-scale graph clustering network. Inf. Sci. 678, 121023 (2024).
Khosla, P. et al. Supervised contrastive learning. In Proceedings of the 33rd International Conference on Neural Information Processing Systems (NeurIPS), 18661–18673 (2020).
Yang, Z., Cohen, W. & Salakhudinov, R. Revisiting semi-supervised learning with graph embeddings. In Proceedings of the 33rd International Conference on Machine Learning (ICML), 40–48 (2016).
Xu, Y.-K., Huang, D., Wang, C.-D. & Lai, J.-H. Glac-gcn: Global and local topology-aware contrastive graph clustering network. IEEE Trans. Artif. Intell. 6, 1448–1459 (2025).
Liu, Y. et al. Reinforcement graph clustering with unknown cluster number. In Proceedings of the 31st ACM international conference on multimedia, 3528–3537 (2023).
Yan, H. et al. A comprehensive study on text-attributed graphs: Benchmarking and rethinking. In Proceedings of the 37th International Conference on Neural Information Processing Systems (NeurIPS), 17238–17264 (2023).
Tang, J., Sun, J., Wang, C. & Yang, Z. Social influence analysis in large-scale networks. In Proceedings of the 15th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, 807–816 (2009).
Ribeiro, L. F. R., Saverese, P. H. P. & Figueiredo, D. R. struc2vec: Learning node representations from structural identity. In Proceedings of the 23rd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, 385–394 (2017).
Liu, Y. et al. A survey of deep graph clustering: Taxonomy, challenge, application, and open resource. arXiv preprint arXiv:2211.12875 (2022).
Derrac, J., García, S., Molina, D. & Herrera, F. A practical tutorial on the use of nonparametric statistical tests as a methodology for comparing evolutionary and swarm intelligence algorithms. Swarm Evolution. Comput. 1, 3–18 (2011).
Chakraborty, S. Topsis and modified topsis: A comparative analysis. Decision Anal. J. 2, 100021 (2022).
Kumar, R. A comprehensive review of mcdm methods, applications, and emerging trends. Decision Making Adv. 3, 185–199 (2025).
Diakoulaki, D., Mavrotas, G. & Papayannakis, L. Determining objective weights in multiple criteria problems: The critic method. Computers Op. Res. 22, 763–770 (1995).
Maaten, L. & Hinton, G. Visualizing data using t-sne. J. Mach. Learn. Res. 9, 2579–2605 (2008).
Acknowledgements
This work was supported by Fujian Provincial Natural Science Foundation of China, No. 2023J01922; Advanced Training Program of Minnan Normal University, No. MSGJB2023015; Headmaster Fund of Minnan Normal University No. KJ19009; Zhangzhou City’s Project for Introducing High-level Talents; the National Natural Science Foundation of China, No. 61762036, 62163016.
Author information
Authors and Affiliations
Contributions
Hong Wang contributed to the conceptualization, methodology, data curation, software development, and the writing of the original draft, as well as the review and editing of the manuscript. Yinglong Zhang was responsible for the conceptualization, methodology, data curation, supervision, and the writing of the original draft, in addition to the review and editing of the manuscript. Zhangqi Zhao contributed to data curation and software development, as well as the review and revision of the manuscript. Zhicong Cai also contributed to in data curation and software development, along with reviewing and revising the manuscript. Xuewen Xia provided supervision and participated in the manuscript’s review and revision. Xing Xu also provided supervision and participated in the manuscript’s review and revision.
Corresponding author
Ethics declarations
Competing interests
No conflict of interest exists in the submission of this manuscript, and the manuscript is approved by all authors for publication. The authors declare no competing interests.
Additional information
Publisher’s note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Rights and permissions
Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article’s Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article’s Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/.
About this article
Cite this article
Wang, H., Zhang, Y., Zhao, Z. et al. Simple yet effective heuristic community detection with graph convolution network. Sci Rep 15, 39249 (2025). https://doi.org/10.1038/s41598-025-22860-z
Received:
Accepted:
Published:
Version of record:
DOI: https://doi.org/10.1038/s41598-025-22860-z








