Abstract

Motivation

Recent advances in spatial transcriptomics allow spatially resolved gene expression measurements with cellular or even sub-cellular resolution, directly characterizing the complex spatiotemporal gene expression landscape and cell-to-cell interactions in their native microenvironments. Due to technology limitations, most spatial transcriptomic technologies still yield incomplete expression measurements with excessive missing values. Therefore, gene imputation is critical to filling in missing data, enhancing resolution, and improving overall interpretability. However, existing methods either require additional matched single-cell RNA-seq data, which is rarely available, or ignore spatial proximity or expression similarity information.

Results

To address these issues, we introduce Impeller, a path-based heterogeneous graph learning method for spatial transcriptomic data imputation. Impeller has two unique characteristics distinct from existing approaches. First, it builds a heterogeneous graph with two types of edges representing spatial proximity and expression similarity. Therefore, Impeller can simultaneously model smooth gene expression changes across spatial dimensions and capture similar gene expression signatures of faraway cells from the same type. Moreover, Impeller incorporates both short- and long-range cell-to-cell interactions (e.g. via paracrine and endocrine) by stacking multiple GNN layers. We use a learnable path operator in Impeller to avoid the over-smoothing issue of the traditional Laplacian matrices. Extensive experiments on diverse datasets from three popular platforms and two species demonstrate the superiority of Impeller over various state-of-the-art imputation methods.

Availability and implementation

The code and preprocessed data used in this study are available at https://github.com/aicb-ZhangLabs/Impeller and https://zenodo.org/records/11212604.

1 Introduction

The orchestration of cellular life hinges on the precise control of when and where genes are activated or silenced. Characterizing such spatiotemporal gene expression patterns is crucial for a better understanding of life, from development to disease to adaptation (Mantri et al. 2021). While single-cell RNA sequencing (scRNA-seq) is a revolutionary and widely available technology that enables simultaneous gene expression profiling over thousands of cells, it usually needs to dissociate cells from their native tissue and thus loses the spatial context (Lähnemann et al. 2020). Recent advances in spatial transcriptomics (Ståhl et al. 2016) allow spatially resolved gene expression measurements at a single-cell or even sub-cellular resolution, providing unprecedented opportunities to characterize the complex landscape of spatiotemporal gene expression and understand the intricate interplay between cells in their native microenvironments (Strell et al. 2019). However, due to technical and biological limitations, most spatial transcriptomic profiling technologies still yield incomplete datasets with excessive missing gene expression values, hindering our biological interpretation of such valuable datasets (Choe et al. 2023). Therefore, gene imputation is a critical task to enrich spatial transcriptomics by filling in missing data, enhancing resolution, and improving the overall quality and interpretability of the datasets.

Several methods have been successfully developed for gene imputation in spatial transcriptomics, which can be broadly summarized into two categories—reference-based and reference-free approaches. Since scRNA-seq data usually offer a deeper dive into transcriptome profiling, reference-based methods integrated spatial transcriptomic data with matched scRNA-seq data from the same sample for accurate imputation. While promising, these referenced-based methods usually suffer from two limitations. First, most studies do not always have matched scRNA-seq data, especially those using valuable and rare samples. Second, even with matched data, there can be significant gene expression distribution shifts due to sequencing protocol differences (e.g. single nuclei RNA-seq versus whole cell spatial transcriptomics) (Zeng et al. 2022).

Researchers also used reference-free methods for direct gene expression imputation. For instance, traditional gene imputation methods designed for scRNA-seq data, such as scVI (Lopez et al. 2018), ALRA (Linderman et al. 2018), Magic (van Dijk et al. 2018), and scGNN (Wang et al. 2021), have been adapted for spatial transcriptomic data imputation. While effectively capturing cell-type-specific gene expression signatures, these methods completely ignored the rich spatial information, resulting in suboptimal results. Later, scientists emphasized the importance of spatial context for cell-to-cell interaction (CCI) in modulating expression changes in response to external stimuli (Armingol et al. 2021). Therefore, Graph Neural Network (GNN) based methods have been developed to mimic CCIs for imputation tasks with improved performance. However, different types of CCI involve distinct cell signaling mechanisms with varying interaction ranges. Existing GNN-based methods used very shallow convolutional layers for computational convenience, successfully modeling short-range CCI (e.g. via autocrine and juxtacrine) but ignoring long-range interactions (e.g. via paracrine and endocrine). As a result, they cannot fully exploit the spatial information for gene expression imputation.

To address the abovementioned issues, we propose Impeller, a path-based heterogeneous graph learning method for accurate spatial transcriptomic data imputation. Impeller contains two unique components to exploit both transcriptomic and spatial information. First, it builds a heterogeneous graph with nodes representing cells and two types of edges describing expression similarity and spatial proximity. Therefore, the expression-based edges allow it to capture cell-type-specific expression signatures of faraway cells from the same type, and the proximity-based edges incorporate CCI effects in the spatial context. Second, Impeller models long-range CCI by stacking multiple GNN layers and uses a learnable path operator instead of the traditional Laplacian matrices to avoid the over-smoothing problem. Extensive experiments on diverse datasets from three popular platforms and two species demonstrate the superiority of Impeller over various state-of-the-art imputation methods.

Our main contributions are summarized below:

  • We propose a graph neural network, Impeller, for reference-free spatial transcriptomic data imputation. Impeller incorporates cell-type-specific expression signatures and CCI via a heterogeneous graph with edges representing transcriptomic similarity and spatial proximity.

  • Impeller stacks multiple GNN layers to include both short- and long-range cell-to-cell interactions in the spatial context. Moreover, it uses a learnable path-based operator to avoid over-smoothing.

  • To the best of our knowledge, this is the first paper to combine cell-type-specific expression signatures with spatial short- and long-range CCI for gene expression imputation.

  • We extensively evaluate Impeller alongside state-of-the-art competitive methods on datasets from three sequencing platforms and two species. The results demonstrate that Impeller outperforms all of the baselines.

2 Related work

2.1 Imputation methods ignoring spatial information

Earlier spatial transcriptomic data imputation methods adapted the computational strategies originally developed for scRNA-seq data, overlooking the spatial coordinate information of each spot. For instance, eKNN (expression-based K nearest neighbor), and eSNN (expression-based Shared nearest neighbor) are methods implemented using the Seurat R-package that rely on gene expressions of nearest neighbors. MAGIC adopted data diffusion across similar cells to impute missing transcriptomic data. ALARA used low-rank approximation to distinguish genuine nonexpression from technical dropouts, thus preserving true gene absence in samples. scVI used a deep variational autoencoder for gene imputation by assuming the read counts per gene follow a zero-inflated negative binomial distribution. However, these methods completely ignored the rich spatial information, resulting in sub-optimal performance.

2.1 Imputation methods utilizing spatial information

Later on, several methods were developed to exploit the spatial coordinate information to improve imputation accuracy. Since scRNA-seq data are usually sequenced deeper to provide more accurate expression measurements, several methods incorporated additional scRNA-seq data during the imputation process. For instance, gimVI used a low-rank approximation and included scRNA reference (Lopez et al. 2019). Tangram mapped scRNA-seq data onto spatial transcriptomics data to facilitate imputation by fitting expression values on the shared genes (Biancalani et al. 2021). STLearn used gene expression data, spatial distance, and tissue morphology data for imputing absent gene reads (Pham et al. 2020). However, additional scRNA-seq data are not always available and there can be large gene expression distribution shifts between these datasets due to differences in sequencing protocols (e.g. single-cell versus single-nuclei), resulting in limited applications for reference-based methods.

On the other hand, several reference-free methods have been developed for more generalized settings. For example, the seKNN (spatial-expression-based K nearest neighbor) and seSNN (spatial-expression-based shared nearest neighbor) models (Satija et al. 2015, Butler et al. 2018, Stuart et al. 2019, Hao et al. 2021) incorporated cell-to-cell distance when defining the KNN for imputation tasks. Recently, STAGATE (Dong and Zhang 2022) is a graph attention auto-encoder framework that effectively imputes genes by integrating spatial data and cell type labels. Overall, these methods did not deeply integrate and exploit the full potential of combining expression and spatial data.

3 Materials and methods

3.1 Problem definition

Here, we aim to impute the excessive missing gene expression values in spatial transcriptomics data without matched reference scRNA-seq data. Formally, given a sparse cell-by-gene count matrix XobsRn×m which represents observations for n cells across m genes, and the spatial coordinates CRn×2 of these cells, our goal is to impute the gene expression matrix X^Rn×m. Xobs is derived from the ground truth matrix XgtRn×m, which contains the observed nonzero entries pre-masking. To simulate real-world data conditions, 10% of the nonzero entries in Xgt are masked to form a test set and another 10% for validation, thus creating Xobs. This matrix serves as the input for our imputation model. The major challenge is to generate X^ that is as close as possible to the ground truth gene expression Xgt, using both the observed gene expressions in Xobs and the spatial information in C.

3.2 Heterogeneous graph construction

As shown in Fig. 1, we build our Impeller model based on two widely accepted biological insights–(i) gene expression can be modulated by surrounding cells via CCI; (ii) faraway cells of the same cell type may share stable gene expression signatures. Therefore, Impeller first builds a heterogeneous graph G to fully exploit both spatial and cell-type information, with nodes and edges representing cells and their relationships.

The overview of Impeller. (A) Given the observed matrix Xobs∈Rn×m of n cells and m genes, and the cells’ spatial coordinates C∈Rn×2, we build the spatial graph Gs and the gene similarity graph Gg. The learned spatial and gene similarity path operators ops and opg are obtained through paths and pathg, respectively. Convoluting cell features with path operators yields spatial/gene similarity embeddings, which are concatenated and fed into a multilayer perceptron for final gene imputation. (B) and (C) Comparison of neighbor aggregation methods in GNNs. (B) Traditional GNN stacks multiple layers to gather information from distant nodes. (C) The path-based GNN, Impeller samples a path to the target node.
Figure 1.

The overview of Impeller. (A) Given the observed matrix XobsRn×m of n cells and m genes, and the cells’ spatial coordinates CRn×2, we build the spatial graph Gs and the gene similarity graph Gg. The learned spatial and gene similarity path operators ops and opg are obtained through paths and pathg, respectively. Convoluting cell features with path operators yields spatial/gene similarity embeddings, which are concatenated and fed into a multilayer perceptron for final gene imputation. (B) and (C) Comparison of neighbor aggregation methods in GNNs. (B) Traditional GNN stacks multiple layers to gather information from distant nodes. (C) The path-based GNN, Impeller samples a path to the target node.

Specifically, G contains two complementary graphs: a spatial graph (Gs) and a gene similarity graph (Gg). Edges in Gs represent the cell’s spatial proximity to model CCI, while edges in Gg denote the cell’s transcriptomic similarity to capture the cell-type-specific expression signatures.

3.2.1 Spatial graph construction

The spatial graph Gs(Vs,Es) is created based on the spatial distance between cells, with nodes Vs representing the cells and edges in Es connecting nearby cells. Specifically, an edge es,{ij} in Gs is established between vi,vjVs if and only if their Euclidean distance di,j is less than a predefined threshold dthr, which can be represented as:
(1)
where Ci=[Ci,0,Ci,1] and Cj=[Cj,0,Cj,1] are 2D spatial coordinates of cell i and j, respectively.

3.2.2 Gene similarity graph construction

Impeller also builds a gene expression similarity graph Gg similar to that in scRNA-seq analysis. Specifically, we first extract the highly variable genes (default 3100). Then, for each target cell, we select its top K most similar cells. Mathematically,
(2)
where Xih is the expression vector of highly variable genes in cell i, Kg(Xih) returns the top kg cells most similar to cell i (e.g. using the Euclidean distance as the similarity metric), and eg,{ij} is the edge between cells i and j in Gg.

3.3 GNN model on heterogeneous graph

With the heterogeneous graph built, Impeller uses a path-based heterogeneous GNN to synthesize the impacts of spatial CCI (Gs) and cell-type-specific expression signatures (Gg) for the imputation task. We introduce the problem of traditional GNN, our learnable path operator, and the overall architecture of Impeller as follows.

3.3.1 Problem of traditional GNN

We aim to impute the missing gene expression values in spatial transcriptomics data by incorporating its physical and transcriptional neighbors via a heterogeneous graph. By treating expression profiles as initial cell embeddings (f(0)=Xobs), the lth (l{1,2,,L1}) GNN layer follows a message passing form (Duan et al. 2022a,b,c, Wang et al. 2019, 2022, Xu et al. 2020, 2021, Duan et al. 2023, 2024) to generate cell i’s embedding in layer l as follows:
(3)
where fi(l)Rdemb(l) is the embedding of cell i at lth layer, demb(l) is the embedding dimension at lth layer, and Ns(i) and Ng(i) are neighboring cell i in Gs and Gg. denotes a differentiable, permutation invariant function, e.g. sum, mean, and γΘ,ϕΘ, and ψΘ denote differentiable functions such as MLPs. After L layers, we obtain the imputed gene expressions, denoted as X^=f(L)Rn×m.

In order to capture long-range CCI interaction, we have to include relatively far away cells by stacking multiple GNN layers via a larger L. Traditional Laplacian matrices-based GNN suffers from over-smoothing, resulting in deteriorated performance as L increases (Eliasof et al. 2022). Therefore, we introduce a learnable path operator to overcome this issue and better capture the long-range CCI.

3.3.2 Learnable path operator

We first define path Ps=(s1,s2,,sks) on Gs of length ks and path Pg=(g1,g2,,gkg) on Gg of length kg, where si and gi are node (cell) indexes. Node embeddings at lth layer are denoted by fsi(l)Rdemb(l) and fgi(l)Rdemb(l). Then, ops(l)Rks×demb(l) and opg(l)Rkg×demb(l) are two learnable path operators which allow us to convolve node embeddings along paths:
(4)
where * denotes the convolution operation, and · symbol is the multiplication operation between two scalars. Here ops,i(l)[j],opg,i(l)[j],fsi(l)[j] and fgi(l)[j] represent the jth scalars of the demb(l)-dimensional vector ops,i(l),opg,i(l),fsi(l) and fgi(l), respectively. Starting from each node, we generate multiple paths on Gs and Gg and aggregate results for a more expressive representation:
(5)
where Ps and Pg are sets of paths sampled from the Gs and Gg, each containing Ts and Tg paths. Each path PsPs and PgPg are separately convolved using ops(l) or opg(l), and the results are averaged to acquire the node embeddings.

3.3.3 The overall architecture of impeller

After convolving both spatial and gene similarity paths, we concatenate their embeddings to form the overall node embeddings, as in
(6)
where σ(·) denotes the ReLU activation function, W(l)Rdemb(l+1)×2*demb(l) is the learnable weight matrix, demb(l) is the embedding dimension at lth layer, and [·,·] denotes concatenation operation. Then, Impeller tries to minimize the Mean Squared Error (MSE) between X^ and Xgt:
(7)
where 1[·] is an indicator function that equals 1 if the condition inside brackets is met (Xgt,(i,j)0), and 0 otherwise. The loss is computed only over nonzero entries of Xgt.

3.4 Computational complexity analysis

3.4.1 k-hop complexity analysis

Traditional GNNs need to gather information from k-hop neighbor nodes after stacking of k layers. Given the complexity of each layer as O(n×dt), where n is the number of nodes and dt is the average node degree, the overall complexity becomes O(n×dt×k). In contrast, Impeller can directly access neighbors up to k-hop distance via a single layer by setting ks=kg=k. The computational complexity per layer for Impeller is O(n×(Ts×ks+Tg×kg)), with Ts and Tg representing the number of paths in Gs and Gg, ks and kg denoting path lengths. As a result, when Ts<dt (a condition satisfied in our task), Impeller offers superior computational efficiency.

3.4.2 The number of parameters

Traditional GNNs have O(demb(l)×demb(l+1)) parameters per layer while the path operator of Impeller adds (ks+kg)*demb(l) parameters. Since ks+kg is typically much smaller than demb(l+1), Impeller’s number of parameters remains on par with traditional GNNs.

4 Experiments

4.1 Detailed experimental setup

4.1.1 Data sources and preprocessing

In our study, we tested Impeller using diverse datasets from three popular sequencing platforms and two organisms. Specifically, we included 10X Visium datasets from the human dorsolateral prefrontal cortex (DLPFC), (Maynard et al. 2021), Steroseq datasets from mouse olfactory bulb (Chen et al. 2021), and Slide-seqV2 from mouse olfactory bulb (Stickels et al. 2021) in our analyses. Detailed attributes of these datasets are summarized in Table 1 (for filter details and visualizations, see the Supplementary Appendix). After standard pre-processing and normalization procedures, we downsampled the data according to scGNN, where 10% of nonzero entries in the dataset were used as a test set, and another 10% of nonzero entries were reserved for validation. For a fair comparison, we repeat ten times with different mask configurations.

Table 1.

Summary of datasets.

PlatformOrganismSample IDRaw matrix (cell, gene)Raw densityFilter matrix (cell, gene)Filter densityNo. of imputed entries
10xVisiumHuman dorsolateral prefrontal cortex (DLPFC)1515074226, 33 5380.0424117, 40280.261437 240
1515084384, 33 5380.0364148, 33420.258358 184
1515094789, 33 5380.0434700, 41880.258508 186
1515104643, 33 5380.0414547, 39080.259461 112
1516693661, 33 5380.0543617, 52460.277525 930
1516703498, 33 5380.0503433, 49090.272457 770
1516714110, 33 5380.0553988, 55390.278615 111
1516724015, 33 5380.0523809, 52730.279561 166
1516733639, 33 5380.0663628, 65380.286677 473
1516743673, 33 5380.0803668, 77960.305871 032
1516753592, 33 5380.0543565, 54540.267518 515
1516763460, 33 5380.0583449, 57840.274545 920
StereoseqMouse/19 109, 14 3760.0244036, 15810.193123 444
SlideseqV2Mouse/20 139, 11 7500.0315161, 26110.217292 418
PlatformOrganismSample IDRaw matrix (cell, gene)Raw densityFilter matrix (cell, gene)Filter densityNo. of imputed entries
10xVisiumHuman dorsolateral prefrontal cortex (DLPFC)1515074226, 33 5380.0424117, 40280.261437 240
1515084384, 33 5380.0364148, 33420.258358 184
1515094789, 33 5380.0434700, 41880.258508 186
1515104643, 33 5380.0414547, 39080.259461 112
1516693661, 33 5380.0543617, 52460.277525 930
1516703498, 33 5380.0503433, 49090.272457 770
1516714110, 33 5380.0553988, 55390.278615 111
1516724015, 33 5380.0523809, 52730.279561 166
1516733639, 33 5380.0663628, 65380.286677 473
1516743673, 33 5380.0803668, 77960.305871 032
1516753592, 33 5380.0543565, 54540.267518 515
1516763460, 33 5380.0583449, 57840.274545 920
StereoseqMouse/19 109, 14 3760.0244036, 15810.193123 444
SlideseqV2Mouse/20 139, 11 7500.0315161, 26110.217292 418
Table 1.

Summary of datasets.

PlatformOrganismSample IDRaw matrix (cell, gene)Raw densityFilter matrix (cell, gene)Filter densityNo. of imputed entries
10xVisiumHuman dorsolateral prefrontal cortex (DLPFC)1515074226, 33 5380.0424117, 40280.261437 240
1515084384, 33 5380.0364148, 33420.258358 184
1515094789, 33 5380.0434700, 41880.258508 186
1515104643, 33 5380.0414547, 39080.259461 112
1516693661, 33 5380.0543617, 52460.277525 930
1516703498, 33 5380.0503433, 49090.272457 770
1516714110, 33 5380.0553988, 55390.278615 111
1516724015, 33 5380.0523809, 52730.279561 166
1516733639, 33 5380.0663628, 65380.286677 473
1516743673, 33 5380.0803668, 77960.305871 032
1516753592, 33 5380.0543565, 54540.267518 515
1516763460, 33 5380.0583449, 57840.274545 920
StereoseqMouse/19 109, 14 3760.0244036, 15810.193123 444
SlideseqV2Mouse/20 139, 11 7500.0315161, 26110.217292 418
PlatformOrganismSample IDRaw matrix (cell, gene)Raw densityFilter matrix (cell, gene)Filter densityNo. of imputed entries
10xVisiumHuman dorsolateral prefrontal cortex (DLPFC)1515074226, 33 5380.0424117, 40280.261437 240
1515084384, 33 5380.0364148, 33420.258358 184
1515094789, 33 5380.0434700, 41880.258508 186
1515104643, 33 5380.0414547, 39080.259461 112
1516693661, 33 5380.0543617, 52460.277525 930
1516703498, 33 5380.0503433, 49090.272457 770
1516714110, 33 5380.0553988, 55390.278615 111
1516724015, 33 5380.0523809, 52730.279561 166
1516733639, 33 5380.0663628, 65380.286677 473
1516743673, 33 5380.0803668, 77960.305871 032
1516753592, 33 5380.0543565, 54540.267518 515
1516763460, 33 5380.0583449, 57840.274545 920
StereoseqMouse/19 109, 14 3760.0244036, 15810.193123 444
SlideseqV2Mouse/20 139, 11 7500.0315161, 26110.217292 418

4.1.2 Baseline methods for benchmarking

We conducted a comparative study utilizing 12 state-of-the-art methods, including reference-free and reference-based methods that originally required additional scRNA-seq data. However, in our analysis, we did not use any additional scRNA-seq data for a fair comparison.

First, we included methods directly adapted from scRNA-seq data imputation and completely ignored the rich spatial information, including a deep generative model scVI, a low-rank approximation model ALRA, nearest neighbors-based models eKNN and eSNN, a diffusion-based model MAGIC, and a GNN-based model scGNN. Furthermore, we used several imputation methods specifically designed for spatial transcriptomic data, such as seKNN (spatial-expression-based K nearest neighbor), and seSNN (spatial-expression-based shared nearest neighbor). gimVI and Tangram need additional scRNA-seq from matched samples, so we used a reference-free implementation available through their website for a fair comparison. Lastly, we included STAGATE a graph attention auto-encoder framework by amalgamating spatial data and gene expression profiles. We use default parameters in most baseline methods (for details, see the Supplementary Appendix).

4.1.3 Evaluation metrics

We first define a test mask MRn×m where the entries to be imputed are marked as 1 and the others as 0. Then we extract the relevant entries from both the imputed matrix X^ and the ground truth matrix Xgt to form two vectors: x^ (from X^) and xgt (from Xgt), each of length N, where N is the total number of entries to be imputed. Following scGNN settings, we use L1 Distance, Cosine Similarity, and Root-Mean-Square Error (RMSE) to compare imputed gene expressions x^ with the ground truth xgt. Mathematically:
(8)
 
(9)
 
(10)

4.2 Experimental results

4.2.1 Improved imputation accuracy

We benchmarked our performance against 12 leading methods by assessing imputation accuracy across 14 datasets. These datasets span three prominent sequencing platforms (10x Visium, Stereoseq, and Slideseq) and two species (human and mouse). Table 2 summarizes the performance of Impellers and other baselines (for results of the other six samples of the DLPFC dataset, please see the Supplementary Appendix). For a fair comparison, we did not include any additional scRNA-seq data to facilitate the imputation task. Overall, Impeller consistently outperforms others in all datasets using L1 distance, Cosine Similarity, and RMSE, indicating the effectiveness and robustness of our strategy.

Table 2.

Gene imputation benchmark.a

MetricMethodPlatform and dataset
10xVisium
Stereoseq
SlideseqV2
DLPFC
Mouse
Mouse
151507151508151509151510151669151670//
L1 distancewoscVI0.794 ± 0.0040.838 ± 0.0060.800 ± 0.0020.670 ± 0.0030.810 ± 0.0030.696 ± 0.0051.442 ± 0.0051.127 ± 0.006
ALRA0.499 ± 0.0030.512 ± 0.0010.490 ± 0.0010.496 ± 0.0010.467 ± 0.0020.472 ± 0.0020.406 ± 0.0130.649 ± 0.066
eKNN0.274 ± 0.0010.281 ± 0.0010.275 ± 0.0000.275 ± 0.0010.269 ± 0.0000.272 ± 0.0000.205 ± 0.0010.294 ± 0.001
eSNN1.254 ± 0.0011.373 ± 0.0011.266 ± 0.0011.294 ± 0.0001.017 ± 0.0011.071 ± 0.0012.802 ± 0.0022.071 ± 0.002
Magic0.779 ± 0.0010.825 ± 0.0010.787 ± 0.0000.664 ± 0.0010.795 ± 0.0010.692 ± 0.0001.324 ± 0.0011.080 ± 0.000
scGNN0.583 ± 0.0110.665 ± 0.0850.589 ± 0.0110.584 ± 0.0040.550 ± 0.0060.532 ± 0.0090.819 ± 0.2400.664 ± 0.018
wgimVI0.838 ± 0.0030.890 ± 0.0030.835 ± 0.0010.737 ± 0.0020.863 ± 0.0030.765 ± 0.0011.325 ± 0.0011.153 ± 0.002
seKNN0.306 ± 0.0000.309 ± 0.0010.307 ± 0.0000.307 ± 0.0000.281 ± 0.0000.289 ± 0.0000.263 ± 0.0010.876 ± 0.001
seSNN1.254 ± 0.0011.371 ± 0.0011.266 ± 0.0001.294 ± 0.0001.017 ± 0.0011.072 ± 0.0012.775 ± 0.0021.998 ± 0.001
Tangram1.691 ± 0.0011.811 ± 0.0011.689 ± 0.0001.420 ± 0.0001.728 ± 0.0011.474 ± 0.0002.899 ± 0.0012.185 ± 0.000
STLearn1.333 ± 0.0011.423 ± 0.0011.332 ± 0.0011.148 ± 0.0011.369 ± 0.0021.206 ± 0.001NANA
STAGATE0.297 ± 0.0010.300 ± 0.0020.295 ± 0.0050.294 ± 0.0040.274 ± 0.0050.278 ± 0.0020.289 ± 0.0060.502 ± 0.007
Impeller0.248 ± 0.0010.252 ± 0.0030.247 ± 0.0030.254 ± 0.0030.242 ± 0.0040.237 ± 0.0010.190 ± 0.0050.292 ± 0.005
Cosine similaritywoscVI0.907 ± 0.0010.913 ± 0.0010.906 ± 0.0010.903 ± 0.0010.909 ± 0.0010.904 ± 0.0010.941 ± 0.0010.919 ± 0.002
ALRA0.948 ± 0.0020.952 ± 0.0020.952 ± 0.0010.952 ± 0.0010.938 ± 0.0060.944 ± 0.0030.980 ± 0.0020.927 ± 0.018
eKNN0.983 ± 0.0000.984 ± 0.0000.983 ± 0.0000.984 ± 0.0000.979 ± 0.0000.979 ± 0.0000.993 ± 0.0000.989 ± 0.000
eSNN0.842 ± 0.0000.841 ± 0.0000.839 ± 0.0000.840 ± 0.0000.846 ± 0.0000.843 ± 0.0000.777 ± 0.0010.838 ± 0.000
Magic0.915 ± 0.0000.920 ± 0.0000.914 ± 0.0000.909 ± 0.0000.916 ± 0.0000.910 ± 0.0000.968 ± 0.0020.936 ± 0.000
scGNN0.933 ± 0.0040.927 ± 0.0160.932 ± 0.0020.936 ± 0.0000.917 ± 0.0020.929 ± 0.0020.948 ± 0.0350.953 ± 0.002
wgimVI0.957 ± 0.0000.965 ± 0.0010.955 ± 0.0010.947 ± 0.0010.962 ± 0.0010.948 ± 0.0020.964 ± 0.0000.936 ± 0.001
seKNN0.982 ± 0.0000.985 ± 0.0000.982 ± 0.0000.983 ± 0.0000.979 ± 0.0000.980 ± 0.0000.995 ± 0.0000.982 ± 0.000
seSNN0.843 ± 0.0000.841 ± 0.0000.840 ± 0.0000.841 ± 0.0000.851 ± 0.0000.847 ± 0.0000.768 ± 0.0000.817 ± 0.000
Tangram0.713 ± 0.0010.725 ± 0.0010.717 ± 0.0010.716 ± 0.0010.717 ± 0.0010.715 ± 0.0000.772 ± 0.0010.763 ± 0.001
STLearn0.718 ± 0.0000.718 ± 0.0000.715 ± 0.0010.724 ± 0.0000.715 ± 0.0010.717 ± 0.000NANA
STAGATE0.983 ± 0.0000.985 ± 0.0000.983 ± 0.0010.984 ± 0.0010.980 ± 0.0010.980 ± 0.0000.990 ± 0.0000.961 ± 0.000
Impeller0.987 ± 0.0000.988 ± 0.0000.987 ± 0.0000.987 ± 0.0000.983 ± 0.0010.985 ± 0.0000.997 ± 0.0000.990 ± 0.000
RMSEwoscVI0.940 ± 0.0050.993 ± 0.0060.949 ± 0.0030.803 ± 0.0030.959 ± 0.0030.834 ± 0.0051.628 ± 0.0051.307 ± 0.007
ALRA0.784 ± 0.0030.810 ± 0.0050.766 ± 0.0010.777 ± 0.0010.735 ± 0.0040.743 ± 0.0030.723 ± 0.0361.061 ± 0.107
eKNN0.380 ± 0.0010.395 ± 0.0020.382 ± 0.0010.384 ± 0.0010.368 ± 0.0010.374 ± 0.0010.402 ± 0.0080.416 ± 0.003
eSNN1.378 ± 0.0011.503 ± 0.0001.393 ± 0.0001.419 ± 0.0011.143 ± 0.0021.199 ± 0.0012.778 ± 0.0012.177 ± 0.001
Magic0.917 ± 0.0010.972 ± 0.0010.929 ± 0.0000.792 ± 0.0000.936 ± 0.0010.824 ± 0.0001.453 ± 0.0011.238 ± 0.001
scGNN0.755 ± 0.0160.850 ± 0.0960.762 ± 0.0110.755 ± 0.0020.717 ± 0.0070.686 ± 0.0101.051 ± 0.3070.842 ± 0.021
wgimVI0.955 ± 0.0021.002 ± 0.0010.957 ± 0.0010.858 ± 0.0010.970 ± 0.0020.890 ± 0.0021.448 ± 0.0011.217 ± 0.004
seKNN0.392 ± 0.0010.395 ± 0.0000.392 ± 0.0000.392 ± 0.0000.361 ± 0.0010.370 ± 0.0000.361 ± 0.0010.523 ± 0.000
seSNN1.354 ± 0.0011.474 ± 0.0001.370 ± 0.0001.395 ± 0.0011.119 ± 0.0011.175 ± 0.0012.770 ± 0.0022.087 ± 0.001
Tangram1.768 ± 0.0011.889 ± 0.0011.767 ± 0.0001.503 ± 0.0001.804 ± 0.0011.557 ± 0.0002.970 ± 0.0012.284 ± 0.000
STLearn1.516 ± 0.0011.629 ± 0.0011.521 ± 0.0011.300 ± 0.0011.556 ± 0.0021.362 ± 0.001NANA
STAGATE0.384 ± 0.0020.393 ± 0.0020.379 ± 0.0070.380 ± 0.0070.357 ± 0.0070.365 ± 0.0040.485 ± 0.0080.765 ± 0.005
Impeller0.337 ± 0.0010.341 ± 0.0000.336 ± 0.0000.340 ± 0.0040.327 ± 0.0070.323 ± 0.0000.277 ± 0.0020.391 ± 0.006
MetricMethodPlatform and dataset
10xVisium
Stereoseq
SlideseqV2
DLPFC
Mouse
Mouse
151507151508151509151510151669151670//
L1 distancewoscVI0.794 ± 0.0040.838 ± 0.0060.800 ± 0.0020.670 ± 0.0030.810 ± 0.0030.696 ± 0.0051.442 ± 0.0051.127 ± 0.006
ALRA0.499 ± 0.0030.512 ± 0.0010.490 ± 0.0010.496 ± 0.0010.467 ± 0.0020.472 ± 0.0020.406 ± 0.0130.649 ± 0.066
eKNN0.274 ± 0.0010.281 ± 0.0010.275 ± 0.0000.275 ± 0.0010.269 ± 0.0000.272 ± 0.0000.205 ± 0.0010.294 ± 0.001
eSNN1.254 ± 0.0011.373 ± 0.0011.266 ± 0.0011.294 ± 0.0001.017 ± 0.0011.071 ± 0.0012.802 ± 0.0022.071 ± 0.002
Magic0.779 ± 0.0010.825 ± 0.0010.787 ± 0.0000.664 ± 0.0010.795 ± 0.0010.692 ± 0.0001.324 ± 0.0011.080 ± 0.000
scGNN0.583 ± 0.0110.665 ± 0.0850.589 ± 0.0110.584 ± 0.0040.550 ± 0.0060.532 ± 0.0090.819 ± 0.2400.664 ± 0.018
wgimVI0.838 ± 0.0030.890 ± 0.0030.835 ± 0.0010.737 ± 0.0020.863 ± 0.0030.765 ± 0.0011.325 ± 0.0011.153 ± 0.002
seKNN0.306 ± 0.0000.309 ± 0.0010.307 ± 0.0000.307 ± 0.0000.281 ± 0.0000.289 ± 0.0000.263 ± 0.0010.876 ± 0.001
seSNN1.254 ± 0.0011.371 ± 0.0011.266 ± 0.0001.294 ± 0.0001.017 ± 0.0011.072 ± 0.0012.775 ± 0.0021.998 ± 0.001
Tangram1.691 ± 0.0011.811 ± 0.0011.689 ± 0.0001.420 ± 0.0001.728 ± 0.0011.474 ± 0.0002.899 ± 0.0012.185 ± 0.000
STLearn1.333 ± 0.0011.423 ± 0.0011.332 ± 0.0011.148 ± 0.0011.369 ± 0.0021.206 ± 0.001NANA
STAGATE0.297 ± 0.0010.300 ± 0.0020.295 ± 0.0050.294 ± 0.0040.274 ± 0.0050.278 ± 0.0020.289 ± 0.0060.502 ± 0.007
Impeller0.248 ± 0.0010.252 ± 0.0030.247 ± 0.0030.254 ± 0.0030.242 ± 0.0040.237 ± 0.0010.190 ± 0.0050.292 ± 0.005
Cosine similaritywoscVI0.907 ± 0.0010.913 ± 0.0010.906 ± 0.0010.903 ± 0.0010.909 ± 0.0010.904 ± 0.0010.941 ± 0.0010.919 ± 0.002
ALRA0.948 ± 0.0020.952 ± 0.0020.952 ± 0.0010.952 ± 0.0010.938 ± 0.0060.944 ± 0.0030.980 ± 0.0020.927 ± 0.018
eKNN0.983 ± 0.0000.984 ± 0.0000.983 ± 0.0000.984 ± 0.0000.979 ± 0.0000.979 ± 0.0000.993 ± 0.0000.989 ± 0.000
eSNN0.842 ± 0.0000.841 ± 0.0000.839 ± 0.0000.840 ± 0.0000.846 ± 0.0000.843 ± 0.0000.777 ± 0.0010.838 ± 0.000
Magic0.915 ± 0.0000.920 ± 0.0000.914 ± 0.0000.909 ± 0.0000.916 ± 0.0000.910 ± 0.0000.968 ± 0.0020.936 ± 0.000
scGNN0.933 ± 0.0040.927 ± 0.0160.932 ± 0.0020.936 ± 0.0000.917 ± 0.0020.929 ± 0.0020.948 ± 0.0350.953 ± 0.002
wgimVI0.957 ± 0.0000.965 ± 0.0010.955 ± 0.0010.947 ± 0.0010.962 ± 0.0010.948 ± 0.0020.964 ± 0.0000.936 ± 0.001
seKNN0.982 ± 0.0000.985 ± 0.0000.982 ± 0.0000.983 ± 0.0000.979 ± 0.0000.980 ± 0.0000.995 ± 0.0000.982 ± 0.000
seSNN0.843 ± 0.0000.841 ± 0.0000.840 ± 0.0000.841 ± 0.0000.851 ± 0.0000.847 ± 0.0000.768 ± 0.0000.817 ± 0.000
Tangram0.713 ± 0.0010.725 ± 0.0010.717 ± 0.0010.716 ± 0.0010.717 ± 0.0010.715 ± 0.0000.772 ± 0.0010.763 ± 0.001
STLearn0.718 ± 0.0000.718 ± 0.0000.715 ± 0.0010.724 ± 0.0000.715 ± 0.0010.717 ± 0.000NANA
STAGATE0.983 ± 0.0000.985 ± 0.0000.983 ± 0.0010.984 ± 0.0010.980 ± 0.0010.980 ± 0.0000.990 ± 0.0000.961 ± 0.000
Impeller0.987 ± 0.0000.988 ± 0.0000.987 ± 0.0000.987 ± 0.0000.983 ± 0.0010.985 ± 0.0000.997 ± 0.0000.990 ± 0.000
RMSEwoscVI0.940 ± 0.0050.993 ± 0.0060.949 ± 0.0030.803 ± 0.0030.959 ± 0.0030.834 ± 0.0051.628 ± 0.0051.307 ± 0.007
ALRA0.784 ± 0.0030.810 ± 0.0050.766 ± 0.0010.777 ± 0.0010.735 ± 0.0040.743 ± 0.0030.723 ± 0.0361.061 ± 0.107
eKNN0.380 ± 0.0010.395 ± 0.0020.382 ± 0.0010.384 ± 0.0010.368 ± 0.0010.374 ± 0.0010.402 ± 0.0080.416 ± 0.003
eSNN1.378 ± 0.0011.503 ± 0.0001.393 ± 0.0001.419 ± 0.0011.143 ± 0.0021.199 ± 0.0012.778 ± 0.0012.177 ± 0.001
Magic0.917 ± 0.0010.972 ± 0.0010.929 ± 0.0000.792 ± 0.0000.936 ± 0.0010.824 ± 0.0001.453 ± 0.0011.238 ± 0.001
scGNN0.755 ± 0.0160.850 ± 0.0960.762 ± 0.0110.755 ± 0.0020.717 ± 0.0070.686 ± 0.0101.051 ± 0.3070.842 ± 0.021
wgimVI0.955 ± 0.0021.002 ± 0.0010.957 ± 0.0010.858 ± 0.0010.970 ± 0.0020.890 ± 0.0021.448 ± 0.0011.217 ± 0.004
seKNN0.392 ± 0.0010.395 ± 0.0000.392 ± 0.0000.392 ± 0.0000.361 ± 0.0010.370 ± 0.0000.361 ± 0.0010.523 ± 0.000
seSNN1.354 ± 0.0011.474 ± 0.0001.370 ± 0.0001.395 ± 0.0011.119 ± 0.0011.175 ± 0.0012.770 ± 0.0022.087 ± 0.001
Tangram1.768 ± 0.0011.889 ± 0.0011.767 ± 0.0001.503 ± 0.0001.804 ± 0.0011.557 ± 0.0002.970 ± 0.0012.284 ± 0.000
STLearn1.516 ± 0.0011.629 ± 0.0011.521 ± 0.0011.300 ± 0.0011.556 ± 0.0021.362 ± 0.001NANA
STAGATE0.384 ± 0.0020.393 ± 0.0020.379 ± 0.0070.380 ± 0.0070.357 ± 0.0070.365 ± 0.0040.485 ± 0.0080.765 ± 0.005
Impeller0.337 ± 0.0010.341 ± 0.0000.336 ± 0.0000.340 ± 0.0040.327 ± 0.0070.323 ± 0.0000.277 ± 0.0020.391 ± 0.006
a

The best results are bolded. Results marked “NA” for stLearn indicate unavailable HE stained images required by the method.

Table 2.

Gene imputation benchmark.a

MetricMethodPlatform and dataset
10xVisium
Stereoseq
SlideseqV2
DLPFC
Mouse
Mouse
151507151508151509151510151669151670//
L1 distancewoscVI0.794 ± 0.0040.838 ± 0.0060.800 ± 0.0020.670 ± 0.0030.810 ± 0.0030.696 ± 0.0051.442 ± 0.0051.127 ± 0.006
ALRA0.499 ± 0.0030.512 ± 0.0010.490 ± 0.0010.496 ± 0.0010.467 ± 0.0020.472 ± 0.0020.406 ± 0.0130.649 ± 0.066
eKNN0.274 ± 0.0010.281 ± 0.0010.275 ± 0.0000.275 ± 0.0010.269 ± 0.0000.272 ± 0.0000.205 ± 0.0010.294 ± 0.001
eSNN1.254 ± 0.0011.373 ± 0.0011.266 ± 0.0011.294 ± 0.0001.017 ± 0.0011.071 ± 0.0012.802 ± 0.0022.071 ± 0.002
Magic0.779 ± 0.0010.825 ± 0.0010.787 ± 0.0000.664 ± 0.0010.795 ± 0.0010.692 ± 0.0001.324 ± 0.0011.080 ± 0.000
scGNN0.583 ± 0.0110.665 ± 0.0850.589 ± 0.0110.584 ± 0.0040.550 ± 0.0060.532 ± 0.0090.819 ± 0.2400.664 ± 0.018
wgimVI0.838 ± 0.0030.890 ± 0.0030.835 ± 0.0010.737 ± 0.0020.863 ± 0.0030.765 ± 0.0011.325 ± 0.0011.153 ± 0.002
seKNN0.306 ± 0.0000.309 ± 0.0010.307 ± 0.0000.307 ± 0.0000.281 ± 0.0000.289 ± 0.0000.263 ± 0.0010.876 ± 0.001
seSNN1.254 ± 0.0011.371 ± 0.0011.266 ± 0.0001.294 ± 0.0001.017 ± 0.0011.072 ± 0.0012.775 ± 0.0021.998 ± 0.001
Tangram1.691 ± 0.0011.811 ± 0.0011.689 ± 0.0001.420 ± 0.0001.728 ± 0.0011.474 ± 0.0002.899 ± 0.0012.185 ± 0.000
STLearn1.333 ± 0.0011.423 ± 0.0011.332 ± 0.0011.148 ± 0.0011.369 ± 0.0021.206 ± 0.001NANA
STAGATE0.297 ± 0.0010.300 ± 0.0020.295 ± 0.0050.294 ± 0.0040.274 ± 0.0050.278 ± 0.0020.289 ± 0.0060.502 ± 0.007
Impeller0.248 ± 0.0010.252 ± 0.0030.247 ± 0.0030.254 ± 0.0030.242 ± 0.0040.237 ± 0.0010.190 ± 0.0050.292 ± 0.005
Cosine similaritywoscVI0.907 ± 0.0010.913 ± 0.0010.906 ± 0.0010.903 ± 0.0010.909 ± 0.0010.904 ± 0.0010.941 ± 0.0010.919 ± 0.002
ALRA0.948 ± 0.0020.952 ± 0.0020.952 ± 0.0010.952 ± 0.0010.938 ± 0.0060.944 ± 0.0030.980 ± 0.0020.927 ± 0.018
eKNN0.983 ± 0.0000.984 ± 0.0000.983 ± 0.0000.984 ± 0.0000.979 ± 0.0000.979 ± 0.0000.993 ± 0.0000.989 ± 0.000
eSNN0.842 ± 0.0000.841 ± 0.0000.839 ± 0.0000.840 ± 0.0000.846 ± 0.0000.843 ± 0.0000.777 ± 0.0010.838 ± 0.000
Magic0.915 ± 0.0000.920 ± 0.0000.914 ± 0.0000.909 ± 0.0000.916 ± 0.0000.910 ± 0.0000.968 ± 0.0020.936 ± 0.000
scGNN0.933 ± 0.0040.927 ± 0.0160.932 ± 0.0020.936 ± 0.0000.917 ± 0.0020.929 ± 0.0020.948 ± 0.0350.953 ± 0.002
wgimVI0.957 ± 0.0000.965 ± 0.0010.955 ± 0.0010.947 ± 0.0010.962 ± 0.0010.948 ± 0.0020.964 ± 0.0000.936 ± 0.001
seKNN0.982 ± 0.0000.985 ± 0.0000.982 ± 0.0000.983 ± 0.0000.979 ± 0.0000.980 ± 0.0000.995 ± 0.0000.982 ± 0.000
seSNN0.843 ± 0.0000.841 ± 0.0000.840 ± 0.0000.841 ± 0.0000.851 ± 0.0000.847 ± 0.0000.768 ± 0.0000.817 ± 0.000
Tangram0.713 ± 0.0010.725 ± 0.0010.717 ± 0.0010.716 ± 0.0010.717 ± 0.0010.715 ± 0.0000.772 ± 0.0010.763 ± 0.001
STLearn0.718 ± 0.0000.718 ± 0.0000.715 ± 0.0010.724 ± 0.0000.715 ± 0.0010.717 ± 0.000NANA
STAGATE0.983 ± 0.0000.985 ± 0.0000.983 ± 0.0010.984 ± 0.0010.980 ± 0.0010.980 ± 0.0000.990 ± 0.0000.961 ± 0.000
Impeller0.987 ± 0.0000.988 ± 0.0000.987 ± 0.0000.987 ± 0.0000.983 ± 0.0010.985 ± 0.0000.997 ± 0.0000.990 ± 0.000
RMSEwoscVI0.940 ± 0.0050.993 ± 0.0060.949 ± 0.0030.803 ± 0.0030.959 ± 0.0030.834 ± 0.0051.628 ± 0.0051.307 ± 0.007
ALRA0.784 ± 0.0030.810 ± 0.0050.766 ± 0.0010.777 ± 0.0010.735 ± 0.0040.743 ± 0.0030.723 ± 0.0361.061 ± 0.107
eKNN0.380 ± 0.0010.395 ± 0.0020.382 ± 0.0010.384 ± 0.0010.368 ± 0.0010.374 ± 0.0010.402 ± 0.0080.416 ± 0.003
eSNN1.378 ± 0.0011.503 ± 0.0001.393 ± 0.0001.419 ± 0.0011.143 ± 0.0021.199 ± 0.0012.778 ± 0.0012.177 ± 0.001
Magic0.917 ± 0.0010.972 ± 0.0010.929 ± 0.0000.792 ± 0.0000.936 ± 0.0010.824 ± 0.0001.453 ± 0.0011.238 ± 0.001
scGNN0.755 ± 0.0160.850 ± 0.0960.762 ± 0.0110.755 ± 0.0020.717 ± 0.0070.686 ± 0.0101.051 ± 0.3070.842 ± 0.021
wgimVI0.955 ± 0.0021.002 ± 0.0010.957 ± 0.0010.858 ± 0.0010.970 ± 0.0020.890 ± 0.0021.448 ± 0.0011.217 ± 0.004
seKNN0.392 ± 0.0010.395 ± 0.0000.392 ± 0.0000.392 ± 0.0000.361 ± 0.0010.370 ± 0.0000.361 ± 0.0010.523 ± 0.000
seSNN1.354 ± 0.0011.474 ± 0.0001.370 ± 0.0001.395 ± 0.0011.119 ± 0.0011.175 ± 0.0012.770 ± 0.0022.087 ± 0.001
Tangram1.768 ± 0.0011.889 ± 0.0011.767 ± 0.0001.503 ± 0.0001.804 ± 0.0011.557 ± 0.0002.970 ± 0.0012.284 ± 0.000
STLearn1.516 ± 0.0011.629 ± 0.0011.521 ± 0.0011.300 ± 0.0011.556 ± 0.0021.362 ± 0.001NANA
STAGATE0.384 ± 0.0020.393 ± 0.0020.379 ± 0.0070.380 ± 0.0070.357 ± 0.0070.365 ± 0.0040.485 ± 0.0080.765 ± 0.005
Impeller0.337 ± 0.0010.341 ± 0.0000.336 ± 0.0000.340 ± 0.0040.327 ± 0.0070.323 ± 0.0000.277 ± 0.0020.391 ± 0.006
MetricMethodPlatform and dataset
10xVisium
Stereoseq
SlideseqV2
DLPFC
Mouse
Mouse
151507151508151509151510151669151670//
L1 distancewoscVI0.794 ± 0.0040.838 ± 0.0060.800 ± 0.0020.670 ± 0.0030.810 ± 0.0030.696 ± 0.0051.442 ± 0.0051.127 ± 0.006
ALRA0.499 ± 0.0030.512 ± 0.0010.490 ± 0.0010.496 ± 0.0010.467 ± 0.0020.472 ± 0.0020.406 ± 0.0130.649 ± 0.066
eKNN0.274 ± 0.0010.281 ± 0.0010.275 ± 0.0000.275 ± 0.0010.269 ± 0.0000.272 ± 0.0000.205 ± 0.0010.294 ± 0.001
eSNN1.254 ± 0.0011.373 ± 0.0011.266 ± 0.0011.294 ± 0.0001.017 ± 0.0011.071 ± 0.0012.802 ± 0.0022.071 ± 0.002
Magic0.779 ± 0.0010.825 ± 0.0010.787 ± 0.0000.664 ± 0.0010.795 ± 0.0010.692 ± 0.0001.324 ± 0.0011.080 ± 0.000
scGNN0.583 ± 0.0110.665 ± 0.0850.589 ± 0.0110.584 ± 0.0040.550 ± 0.0060.532 ± 0.0090.819 ± 0.2400.664 ± 0.018
wgimVI0.838 ± 0.0030.890 ± 0.0030.835 ± 0.0010.737 ± 0.0020.863 ± 0.0030.765 ± 0.0011.325 ± 0.0011.153 ± 0.002
seKNN0.306 ± 0.0000.309 ± 0.0010.307 ± 0.0000.307 ± 0.0000.281 ± 0.0000.289 ± 0.0000.263 ± 0.0010.876 ± 0.001
seSNN1.254 ± 0.0011.371 ± 0.0011.266 ± 0.0001.294 ± 0.0001.017 ± 0.0011.072 ± 0.0012.775 ± 0.0021.998 ± 0.001
Tangram1.691 ± 0.0011.811 ± 0.0011.689 ± 0.0001.420 ± 0.0001.728 ± 0.0011.474 ± 0.0002.899 ± 0.0012.185 ± 0.000
STLearn1.333 ± 0.0011.423 ± 0.0011.332 ± 0.0011.148 ± 0.0011.369 ± 0.0021.206 ± 0.001NANA
STAGATE0.297 ± 0.0010.300 ± 0.0020.295 ± 0.0050.294 ± 0.0040.274 ± 0.0050.278 ± 0.0020.289 ± 0.0060.502 ± 0.007
Impeller0.248 ± 0.0010.252 ± 0.0030.247 ± 0.0030.254 ± 0.0030.242 ± 0.0040.237 ± 0.0010.190 ± 0.0050.292 ± 0.005
Cosine similaritywoscVI0.907 ± 0.0010.913 ± 0.0010.906 ± 0.0010.903 ± 0.0010.909 ± 0.0010.904 ± 0.0010.941 ± 0.0010.919 ± 0.002
ALRA0.948 ± 0.0020.952 ± 0.0020.952 ± 0.0010.952 ± 0.0010.938 ± 0.0060.944 ± 0.0030.980 ± 0.0020.927 ± 0.018
eKNN0.983 ± 0.0000.984 ± 0.0000.983 ± 0.0000.984 ± 0.0000.979 ± 0.0000.979 ± 0.0000.993 ± 0.0000.989 ± 0.000
eSNN0.842 ± 0.0000.841 ± 0.0000.839 ± 0.0000.840 ± 0.0000.846 ± 0.0000.843 ± 0.0000.777 ± 0.0010.838 ± 0.000
Magic0.915 ± 0.0000.920 ± 0.0000.914 ± 0.0000.909 ± 0.0000.916 ± 0.0000.910 ± 0.0000.968 ± 0.0020.936 ± 0.000
scGNN0.933 ± 0.0040.927 ± 0.0160.932 ± 0.0020.936 ± 0.0000.917 ± 0.0020.929 ± 0.0020.948 ± 0.0350.953 ± 0.002
wgimVI0.957 ± 0.0000.965 ± 0.0010.955 ± 0.0010.947 ± 0.0010.962 ± 0.0010.948 ± 0.0020.964 ± 0.0000.936 ± 0.001
seKNN0.982 ± 0.0000.985 ± 0.0000.982 ± 0.0000.983 ± 0.0000.979 ± 0.0000.980 ± 0.0000.995 ± 0.0000.982 ± 0.000
seSNN0.843 ± 0.0000.841 ± 0.0000.840 ± 0.0000.841 ± 0.0000.851 ± 0.0000.847 ± 0.0000.768 ± 0.0000.817 ± 0.000
Tangram0.713 ± 0.0010.725 ± 0.0010.717 ± 0.0010.716 ± 0.0010.717 ± 0.0010.715 ± 0.0000.772 ± 0.0010.763 ± 0.001
STLearn0.718 ± 0.0000.718 ± 0.0000.715 ± 0.0010.724 ± 0.0000.715 ± 0.0010.717 ± 0.000NANA
STAGATE0.983 ± 0.0000.985 ± 0.0000.983 ± 0.0010.984 ± 0.0010.980 ± 0.0010.980 ± 0.0000.990 ± 0.0000.961 ± 0.000
Impeller0.987 ± 0.0000.988 ± 0.0000.987 ± 0.0000.987 ± 0.0000.983 ± 0.0010.985 ± 0.0000.997 ± 0.0000.990 ± 0.000
RMSEwoscVI0.940 ± 0.0050.993 ± 0.0060.949 ± 0.0030.803 ± 0.0030.959 ± 0.0030.834 ± 0.0051.628 ± 0.0051.307 ± 0.007
ALRA0.784 ± 0.0030.810 ± 0.0050.766 ± 0.0010.777 ± 0.0010.735 ± 0.0040.743 ± 0.0030.723 ± 0.0361.061 ± 0.107
eKNN0.380 ± 0.0010.395 ± 0.0020.382 ± 0.0010.384 ± 0.0010.368 ± 0.0010.374 ± 0.0010.402 ± 0.0080.416 ± 0.003
eSNN1.378 ± 0.0011.503 ± 0.0001.393 ± 0.0001.419 ± 0.0011.143 ± 0.0021.199 ± 0.0012.778 ± 0.0012.177 ± 0.001
Magic0.917 ± 0.0010.972 ± 0.0010.929 ± 0.0000.792 ± 0.0000.936 ± 0.0010.824 ± 0.0001.453 ± 0.0011.238 ± 0.001
scGNN0.755 ± 0.0160.850 ± 0.0960.762 ± 0.0110.755 ± 0.0020.717 ± 0.0070.686 ± 0.0101.051 ± 0.3070.842 ± 0.021
wgimVI0.955 ± 0.0021.002 ± 0.0010.957 ± 0.0010.858 ± 0.0010.970 ± 0.0020.890 ± 0.0021.448 ± 0.0011.217 ± 0.004
seKNN0.392 ± 0.0010.395 ± 0.0000.392 ± 0.0000.392 ± 0.0000.361 ± 0.0010.370 ± 0.0000.361 ± 0.0010.523 ± 0.000
seSNN1.354 ± 0.0011.474 ± 0.0001.370 ± 0.0001.395 ± 0.0011.119 ± 0.0011.175 ± 0.0012.770 ± 0.0022.087 ± 0.001
Tangram1.768 ± 0.0011.889 ± 0.0011.767 ± 0.0001.503 ± 0.0001.804 ± 0.0011.557 ± 0.0002.970 ± 0.0012.284 ± 0.000
STLearn1.516 ± 0.0011.629 ± 0.0011.521 ± 0.0011.300 ± 0.0011.556 ± 0.0021.362 ± 0.001NANA
STAGATE0.384 ± 0.0020.393 ± 0.0020.379 ± 0.0070.380 ± 0.0070.357 ± 0.0070.365 ± 0.0040.485 ± 0.0080.765 ± 0.005
Impeller0.337 ± 0.0010.341 ± 0.0000.336 ± 0.0000.340 ± 0.0040.327 ± 0.0070.323 ± 0.0000.277 ± 0.0020.391 ± 0.006
a

The best results are bolded. Results marked “NA” for stLearn indicate unavailable HE stained images required by the method.

In addition, we found that most methods utilizing spatial information (w* group in Table 2) demonstrated higher imputation accuracy than those ignoring spatial information (wo* group in Table 2), validating the presence of rich information in the spatial context. Notably, Impeller surpasses even the best gene expression-only method, eKNN, with improvements of 11.32% on 10xVisium DLPFC, 31.09% on Stereoseq, and 6.01% on SlideseqV2 Mouse. Furthermore, compared to uniform averaging using KNN, GNN allows for more flexible neighbor information aggregation for better imputation accuracy, as reflected by the noticeably improved performance of Impeller and STAGATE.

4.2.2 Impact of long-range CCI

To probe disparities between Impeller and traditional GNNs in capturing long-range cell dependencies, we examined several models–Impeller, GCN (Kipf and Welling 2016), GraphSAGE (Hamilton et al. 2017), GAT (Veličković et al. 2017), and GraphTransformer (Shi et al. 2020)–across varying receptive fields in the Stereoseq dataset.

In Table 3, GAT and GraphSAGE suffer from gradient vanishing/exploding issues as more layers are added to capture long-range CCI, resulting in quickly degraded performance. GCN works best initially, but its performance drops with more layers added. This could be because the number of neighbors grows fast as we increase the receptive field, leaving it difficult for the target cell to understand the influence of each neighbor. Furthermore, GraphTransformer starts with high errors at a receptive field of 2. It works best at a receptive field of 8, but the error goes up again at 32. This increase in error is similar to the problem of GCN, as all cells start to look too similar to make useful representations. On the other hand, Impeller effectively tackles these challenges by the path operator, as reflected by the consistently improved results until the receptive field of 32. As the receptive field continues to grow, Impeller’s performance slightly declines, likely because distant information becomes less relevant for the target cell’s gene imputation. An additional perturbation study, demonstrating the effectiveness of Impeller in capturing CCI, is shown in the Supplementary Appendix.

Table 3.

Performance of different receptive fields (RMSE).

Receptive fieldGCNGraphSAGEGATGraphTransformerImpeller
20.339 ± 0.0000.352 ± 0.0080.360 ± 0.0051.058 ± 0.4630.310±0.016
40.348 ± 0.0010.362 ± 0.0130.372 ± 0.0220.424 ± 0.0310.286±0.009
80.386 ± 0.0120.496 ± 0.0330.454 ± 0.0240.351 ± 0.0020.279±0.001
160.403 ± 0.0010.617 ± 0.0540.506 ± 0.0610.435 ± 0.0170.286±0.010
320.418 ± 0.0151.466 ± 0.0311.458 ± 0.0370.420 ± 0.0000.277±0.002
640.430 ± 0.0021.615 ± 0.0101.621 ± 0.0040.420 ± 0.0010.302±0.028
1280.429 ± 0.0011.629 ± 0.0031.614 ± 0.0220.420 ± 0.0010.357±0.001
Receptive fieldGCNGraphSAGEGATGraphTransformerImpeller
20.339 ± 0.0000.352 ± 0.0080.360 ± 0.0051.058 ± 0.4630.310±0.016
40.348 ± 0.0010.362 ± 0.0130.372 ± 0.0220.424 ± 0.0310.286±0.009
80.386 ± 0.0120.496 ± 0.0330.454 ± 0.0240.351 ± 0.0020.279±0.001
160.403 ± 0.0010.617 ± 0.0540.506 ± 0.0610.435 ± 0.0170.286±0.010
320.418 ± 0.0151.466 ± 0.0311.458 ± 0.0370.420 ± 0.0000.277±0.002
640.430 ± 0.0021.615 ± 0.0101.621 ± 0.0040.420 ± 0.0010.302±0.028
1280.429 ± 0.0011.629 ± 0.0031.614 ± 0.0220.420 ± 0.0010.357±0.001

The best imputation performance is highlighted in bold.

Table 3.

Performance of different receptive fields (RMSE).

Receptive fieldGCNGraphSAGEGATGraphTransformerImpeller
20.339 ± 0.0000.352 ± 0.0080.360 ± 0.0051.058 ± 0.4630.310±0.016
40.348 ± 0.0010.362 ± 0.0130.372 ± 0.0220.424 ± 0.0310.286±0.009
80.386 ± 0.0120.496 ± 0.0330.454 ± 0.0240.351 ± 0.0020.279±0.001
160.403 ± 0.0010.617 ± 0.0540.506 ± 0.0610.435 ± 0.0170.286±0.010
320.418 ± 0.0151.466 ± 0.0311.458 ± 0.0370.420 ± 0.0000.277±0.002
640.430 ± 0.0021.615 ± 0.0101.621 ± 0.0040.420 ± 0.0010.302±0.028
1280.429 ± 0.0011.629 ± 0.0031.614 ± 0.0220.420 ± 0.0010.357±0.001
Receptive fieldGCNGraphSAGEGATGraphTransformerImpeller
20.339 ± 0.0000.352 ± 0.0080.360 ± 0.0051.058 ± 0.4630.310±0.016
40.348 ± 0.0010.362 ± 0.0130.372 ± 0.0220.424 ± 0.0310.286±0.009
80.386 ± 0.0120.496 ± 0.0330.454 ± 0.0240.351 ± 0.0020.279±0.001
160.403 ± 0.0010.617 ± 0.0540.506 ± 0.0610.435 ± 0.0170.286±0.010
320.418 ± 0.0151.466 ± 0.0311.458 ± 0.0370.420 ± 0.0000.277±0.002
640.430 ± 0.0021.615 ± 0.0101.621 ± 0.0040.420 ± 0.0010.302±0.028
1280.429 ± 0.0011.629 ± 0.0031.614 ± 0.0220.420 ± 0.0010.357±0.001

The best imputation performance is highlighted in bold.

4.2.3 Advantage of heterogeneous graph

In our study, we explored the influence of graph modalities on imputation accuracy by assessing three key variants: vars, using solely the spatial graph; varg, utilizing only the gene similarity graph; varh, integrating both graphs. We then calculated the performance improvement from adding Gg by comparing varh with vars, and the improvement from adding Gs by comparing varh with varg. As shown in Fig. 2, the majority of the cases (22 out of 24) exhibit positive improvements. Specifically, in the DLPFC sample 151674, the inclusion of the gene similarity graph yields a 17.3% improvement, and the 13.6% enhancement is achieved by adding the spatial graph alone. Similarly, in sample 151508, the gene similarity graph and the spatial graph contribute to improvements of 3.6% and 9.9%, respectively. These results underscore the efficacy of our approach, particularly in scenarios where the complex interaction between spatial and gene expression data is pivotal for enhancing gene imputation accuracy.

RMSE improvement by adding different graph modalities.
Figure 2.

RMSE improvement by adding different graph modalities.

4.3 Ablation study

We conducted an ablation study to evaluate the performance of four primary path operator variants of Impeller: opglo, where all Impeller layers and channels (each channel representing 1D of f(l)) share one path operator; opcha, where channels share an operator but layers have distinct ones; oplay, where all layers share one, but channels have individual operators; and opind where every layer and channel possesses an independent path operator. As depicted in Fig. 3, both opglo and opcha performed poorly on the DLPFC dataset, indicating the importance of distinct operators for each channel. Notably, oplay and opind showed comparable results, suggesting that layer-specific operators might be optional, depending on the specific application. Another ablation study regarding different path construction and graph construction methods is shown in the Supplementary Appendix.

RMSE w.r.t. different path operators.
Figure 3.

RMSE w.r.t. different path operators.

4.4 Parameter analysis

To investigate the influence of Impeller’s various hyper-parameters, we conducted extensive experiments using the DLPFC dataset (Sample ID: 151507) and reported the mean and standard deviation of the imputation accuracy over ten repetitions.

First, we studied the impact of qs and qg on the RMSE of a random walk on Gs and Gg following the Node2Vec mechanism. Higher values of q (i.e. qs and qg) encourage the walk to sample more distant nodes, enhancing the exploration of the global graph structure, while lower values bias the walk toward neighboring nodes, facilitating local exploration. As shown in Fig. 4A, Impeller exhibits strong robustness with RMSE from 0.33 to 0.36 when qs and qg varied from 0.1 to 5. However, higher values of qs and qg tend to induce larger errors. For generality, we selected 1 as the default value for qs and qg.

Parameter analysis. (A) The mean RMSE w.r.t. different qs and qg for generating random walks in Gs and Gg. (B) The mean RMSE w.r.t. different path lengths ks and kg, and the number of Impeller layers L. (C) The L1 distance, cosine similarity, and RMSE w.r.t. different number of paths Ts and Tg. (D) The L1 distance, cosine similarity and RMSE w.r.t. different embedding dimensions demb(l).
Figure 4.

Parameter analysis. (A) The mean RMSE w.r.t. different qs and qg for generating random walks in Gs and Gg. (B) The mean RMSE w.r.t. different path lengths ks and kg, and the number of Impeller layers L. (C) The L1 distance, cosine similarity, and RMSE w.r.t. different number of paths Ts and Tg. (D) The L1 distance, cosine similarity and RMSE w.r.t. different embedding dimensions demb(l).

We investigated the impact of random walk length (ks and kg) and layer number (L), shown in Fig. 4B. A path length of 2 with 10 layers results in maximum errors, reducing our model to a standard ten-layer GCN. This is because, at this path length, the model focuses on immediate neighbors, akin to how traditional GCNs operate. Such a setup, while deep, limits neighborhood exploration and increases over-smoothing risk. Conversely, a path length of 8 with 4 layers allows for capturing broader interactions (up to 28 hops), balancing extended reach and computational efficiency, thus avoiding over-smoothing and optimizing long-range CCI capture.

Next, we examined the impact of the number of random walks (Ts and Tg). As shown in Fig. 4C, Ts and Tg appeared to have a minimal effect on results, due to the robustness of Impeller which resamples at each epoch during training. We chose 8 as the default number of random walks.

Lastly, we evaluated how the embedding dimension demb(l) affects Impeller’s performance. As shown in Fig. 4D, smaller demb(l) (such as 2, 4, 8) leads to limited expressive power and larger imputation errors. As demb(l) increases to 16, 32, 64, or 128, Impeller’s expressive power improves and operations converge well in each run. Due to our early stopping criterion, we cease training if the validation RMSE does not improve for 50 consecutive epochs. When demb(l) was set to 256 or 512, it’s hard for Impeller to converge quickly at these dimensions. To strike a balance between complexity and representational power, we opted for demb(l) of 64.

In summary, these comprehensive parameter analyses reveal that Impeller is robust across a wide range of parameter settings, while still providing tunable options for balancing computational efficiency and prediction accuracy. These results further substantiate the effectiveness and practicality of our proposed model for gene imputation tasks.

4.5 Neighbor visualization

To better understand the differences between traditional GNNs and our path-based GNN, Impeller, we turned to a visual example (sample 151507 from the DLPFC dataset). Figure 1B shows how the typical GNN gathers information from far-away neighbors. The center node (red sphere) stacks five GNN layers to gather information from distant nodes like the one shown in yellow. But this method sometimes pulls in extra information from different tissue layers that is not needed. On the other hand, Fig. 1C shows our Impeller model. Instead of stacking GNN layers, Impeller samples a direct path from the center node to the target node. While using this direct path method, Impeller offers better gene imputation performance by capturing the relevant long-range CCI.

4.6 Running time analysis

As shown in Table 4, we conducted a comparative model parameter and runtime analysis with popular graph-based models (GCN, GAT, GraphSAGE, and Transformer) on the DLPFC dataset. As discussed in Section 3.4.2, our model maintains a parameter count comparable to traditional GNNs, with the complexity per layer defined as O(demb(l)×demb(l+1)). Specifically, our model introduces only a 3.5% increase in parameters for GCN and a 2.7% increase for GAT. In contrast, it achieves a 48.0% reduction in parameters for GraphSAGE and a 74.1% reduction for GraphTransformer (Table 4). Despite its additional path sampling step, Impeller remarkably outperformed the others in training and inference efficiency. This can be partially credited to leveraging the DGL library’s optimized implementation for path sampling (https://docs.dgl.ai/en/0.8.x/api/python/dgl.sampling.html) and the inherently faster multiplication process used in path-based convolution compared to edge-wise information aggregation in traditional GNNs. In addition, Impeller showed the lowest RMSE, indicating superior prediction accuracy. Hence, Impeller offers a balanced blend of efficiency and precision for spatial transcriptomic data imputation, outperforming other graph-based models.

Table 4.

Running time summary of graph-based models.

ModelNo. of parametersPath (ms)Training (ms)Inference (ms)RMSE
GCN519 67621.73 ± 10.4421.04 ± 10.840.37 ± 0.02
GAT523 76823.84 ± 13.2524.99 ± 11.410.36 ± 0.02
GraphSAGE1 035 26018.77 ± 11.0616.97 ± 12.160.38 ± 0.01
Transformer2 078 70433.94 ± 16.1638.13 ± 8.710.36 ± 0.01
Impeller538 1081.61 ± 0.606.35 ± 0.308.43 ± 0.250.34 ± 0.00
ModelNo. of parametersPath (ms)Training (ms)Inference (ms)RMSE
GCN519 67621.73 ± 10.4421.04 ± 10.840.37 ± 0.02
GAT523 76823.84 ± 13.2524.99 ± 11.410.36 ± 0.02
GraphSAGE1 035 26018.77 ± 11.0616.97 ± 12.160.38 ± 0.01
Transformer2 078 70433.94 ± 16.1638.13 ± 8.710.36 ± 0.01
Impeller538 1081.61 ± 0.606.35 ± 0.308.43 ± 0.250.34 ± 0.00

Training and inference times with the fastest performance, as well as the best imputation performance (RMSE), are highlighted in bold.

Table 4.

Running time summary of graph-based models.

ModelNo. of parametersPath (ms)Training (ms)Inference (ms)RMSE
GCN519 67621.73 ± 10.4421.04 ± 10.840.37 ± 0.02
GAT523 76823.84 ± 13.2524.99 ± 11.410.36 ± 0.02
GraphSAGE1 035 26018.77 ± 11.0616.97 ± 12.160.38 ± 0.01
Transformer2 078 70433.94 ± 16.1638.13 ± 8.710.36 ± 0.01
Impeller538 1081.61 ± 0.606.35 ± 0.308.43 ± 0.250.34 ± 0.00
ModelNo. of parametersPath (ms)Training (ms)Inference (ms)RMSE
GCN519 67621.73 ± 10.4421.04 ± 10.840.37 ± 0.02
GAT523 76823.84 ± 13.2524.99 ± 11.410.36 ± 0.02
GraphSAGE1 035 26018.77 ± 11.0616.97 ± 12.160.38 ± 0.01
Transformer2 078 70433.94 ± 16.1638.13 ± 8.710.36 ± 0.01
Impeller538 1081.61 ± 0.606.35 ± 0.308.43 ± 0.250.34 ± 0.00

Training and inference times with the fastest performance, as well as the best imputation performance (RMSE), are highlighted in bold.

5 Conclusion

In this study, we introduced Impeller, a path-based heterogeneous graph learning approach tailored for spatial transcriptomic data imputation. By constructing a heterogeneous graph capturing both spatial proximity and gene expression similarity, Impeller offers a refined representation of cellular landscapes. Further, its integration of multiple GNN layers, coupled with a learnable path operator, ensures comprehensive modeling of both short and long-range cellular interactions while effectively averting over-smoothing issues. Benchmark tests across diverse datasets spanning various platforms and species underscore Impeller’s superior performance compared to state-of-the-art imputation methods. This work not only establishes Impeller’s prowess in spatial transcriptomic imputation but also underscores its potential to model both short- and long-range cell-cell interactions.

Acknowledgements

The authors thank the anonymous reviewers for their valuable suggestions and the UCI OIT department for GPU resources.

Supplementary data

Supplementary data are available at Bioinformatics online.

Conflict of interest

None declared.

Funding

This work was supported by the National Institutes of Health [R01HG012572, R01NS128523].

Data availability

The code and preprocessed data used in this study are available at https://github.com/aicb-ZhangLabs/Impeller and https://zenodo.org/records/11212604.

References

Armingol
E
,
Officer
A
,
Harismendy
O
 et al.  
Deciphering cell–cell interactions and communication from gene expression
.
Nat Rev Genet
 
2021
;
22
:
71
88
.

Biancalani
T
,
Scalia
G
,
Buffoni
L
 et al.  
Deep learning and alignment of spatially resolved single-cell transcriptomes with tangram
.
Nat Methods
 
2021
;
18
:
1352
62
. https://doi.org/10.1038/s41592-021-01264-7.

Butler
A
,
Hoffman
P
,
Smibert
P
 et al.  
Integrating single-cell transcriptomic data across different conditions, technologies, and species
.
Nat Biotechnol
 
2018
;
36
:
411
20
. https://doi.org/10.1038/nbt.4096.

Chen
A
,
Liao
S
,
Cheng
M
 et al. Large field of view-spatially resolved transcriptomics at nanoscale resolution. bioRxiv, https://doi.org/10.1101/2021.01.17.427004,
2021
, preprint: not peer reviewed.

Choe
K
,
Pak
U
,
Pang
Y
 et al.  
Advances and challenges in spatial transcriptomics for developmental biology
.
Biomolecules
 
2023
;
13
:
156
.

Dong
K
,
Zhang
S.
 
Deciphering spatial domains from spatially resolved transcriptomics with an adaptive graph attention auto-encoder
.
Nat Commun
 
2022
;
13
:
1739
.

Duan
Z
,
Lee
C
,
Zhang
J.
 
EXAD-GNN: explainable graph neural network for Alzheimer’s disease state prediction from single-cell data
.
SIP
 
2023
;
12
:e201.

Duan
Z
,
Wang
Y
,
Ye
W
 et al.  
Connecting latent relationships over heterogeneous attributed network for recommendation
.
Appl Intell
 
2022a
;
52
:
16214
32
.

Duan
Z
,
Xu
H
,
Huang
Y
 et al.  
Multivariate time series forecasting with transfer entropy graph
.
Tsinghua Sci Technol
 
2022b
;
28
:
141
9
.

Duan
Z
,
Xu
H
,
Wang
Y
 et al.  
Multivariate time-series classification with hierarchical variational graph pooling
.
Neural Netw
 
2022c
;
154
:
481
90
.

Duan
Z
,
Dai
Y
,
Hwang
A
 et al.  
iherd: an integrative hierarchical graph representation learning framework to quantify network changes and prioritize risk genes in disease
.
PLoS Comput Biol
 
2023
;
19
:
e1011444
.

Duan
Z
,
Xu
S
,
Sai Srinivasan
S
 et al.  
scencore: leveraging single-cell epigenetic data to predict chromatin conformation using graph embedding
.
Brief Bioinform
 
2024
;
25
:
bbae096
.

Eliasof
M
,
Haber
E
,
Treister
E.
pathgcn: learning general graph spatial operators from paths. In: International Conference on Machine Learning. PMLR,
2022
,
5878
91
.

Hamilton
W
,
Ying
Z
,
Leskovec
J.
 
Inductive representation learning on large graphs
.
Adv Neural Inf Process Syst
 
2017
;
30
.

Hao
Y
,
Hao
S
,
Andersen-Nissen
E
 et al.  
Integrated analysis of multimodal single-cell data
.
Cell
 
2021
;
184
:
3573
87.e29
. https://doi.org/10.1016/j.cell.2021.04.048.

Kipf
TN
,
Welling
M.
Semi-supervised classification with graph convolutional networks. arXiv, arXiv:1609.02907,
2016
, preprint: not peer reviewed.

Lähnemann
D
,
Köster
J
,
Szczurek
E
 et al.  
Eleven grand challenges in single-cell data science
.
Genome Biol
 
2020
;
21
:
31
5
.

Linderman
GC
,
Zhao
J
,
Roulis
M
et al.  
Zero-preserving imputation of single-cell RNA-seq data[J]
.
Nature communications
,
2022
,
13
(
1
):
192
.

Lopez
R
,
Regier
J
,
Cole
MB
 et al.  
Deep generative modeling for single-cell transcriptomics
.
Nat Methods
 
2018
;
15
:
1053
8
. https://doi.org/10.1038/s41592-018-0229-2.

Lopez
R
,
Nazaret
A
,
Langevin
M
 et al. A joint model of unpaired data from scrna-seq and spatial transcriptomics for imputing missing gene expression measurements. CoRR, CoRR:abs/1905.02269,
2019
, preprint: not peer reviewed.

Mantri
M
,
Scuderi
GJ
,
Abedini-Nassab
R
 et al.  
Spatiotemporal single-cell RNA sequencing of developing chicken hearts identifies interplay between cellular differentiation and morphogenesis
.
Nat Commun
 
2021
;
12
:
1771
. https://doi.org/10.1038/s41467-021-21892-z.

Maynard
KR
,
Collado-Torres
L
,
Weber
LM
 et al.  
Transcriptome-scale spatial gene expression in the human dorsolateral prefrontal cortex
.
Nat Neurosci
 
2021
;
24
:
425
36
.

Pham
D
,
Tan
X
,
Balderson
B
 et al.  
Robust mapping of spatiotemporal trajectories and cell–cell interactions in healthy and diseased tissues[J]
.
Nature communications
,
2023
,
14
(
1
):
7739
.

Satija
R
,
Farrell
JA
,
Gennert
D
 et al.  
Spatial reconstruction of single-cell gene expression data
.
Nat Biotechnol
 
2015
;
33
:
495
502
. https://doi.org/10.1038/nbt.3192.

Shi
Y
,
Huang
Z
,
Feng
S
 et al. Masked label prediction: Unified message passing model for semi-supervised classification. arXiv, arXiv:2009.03509,
2020
, preprint: not peer reviewed.

Ståhl
PL
,
Salmén
F
,
Vickovic
S
 et al.  
Visualization and analysis of gene expression in tissue sections by spatial transcriptomics
.
Science
 
2016
;
353
:
78
82
.

Stickels
RR
,
Murray
E
,
Kumar
P
 et al.  
Highly sensitive spatial transcriptomics at near-cellular resolution with slide-seqv2
.
Nat Biotechnol
 
2021
;
39
:
313
9
.

Strell
C
,
Hilscher
MM
,
Laxman
N
 et al.  
Placing RNA in context and space–methods for spatially resolved transcriptomics
.
FEBS J
 
2019
;
286
:
1468
81
.

Stuart
T
,
Butler
A
,
Hoffman
P
III
, et al.  
Comprehensive integration of single-cell data
.
Cell
 
2019
;
177
:
1888
902.e21
. https://doi.org/10.1016/j.cell.2019.05.031.

van Dijk
D
,
Sharma
R
,
Nainys
J
 et al.  
Recovering gene interactions from single-cell data using data diffusion
.
Cell
 
2018
;
174
:
716
29.e27
. https://doi.org/10.1016/j.cell.2018.05.061.

Veličković
P
,
Cucurull
G
,
Casanova
A
 et al. Graph attention networks. arXiv, arXiv:1710.10903,
2017
, preprint: not peer reviewed.

Wang
J
,
Ma
A
,
Chang
Y
 et al.  
scgnn is a novel graph neural network framework for single-cell RNA-seq analyses
.
Nat Commun
 
2021
;
12
:
1882
.

Wang
Y
,
Duan
Z
,
Liao
B
 et al.  
Heterogeneous attributed network embedding with graph convolutional networks
.
AAAI
 
2019
;
33
:
10061
2
.

Wang
Y
,
Duan
Z
,
Huang
Y
 et al.  
Mthetgnn: a heterogeneous graph embedding framework for multivariate time series forecasting
.
Pattern Recognition Letters
 
2022
;
153
:
151
8
.

Xu
H
,
Chen
R
,
Wang
Y
 et al. Cosimgnn: Towards large-scale graph similarity computation. arXiv, arXiv:2005.07115,
2020
, preprint: not peer reviewed.

Xu
H
,
Duan
Z
,
Wang
Y
 et al.  
Graph partitioning and graph neural network based hierarchical graph matching for graph similarity computation
.
Neurocomputing
 
2021
;
439
:
348
62
.

Zeng
Z
,
Li
Y
,
Li
Y
 et al.  
Statistical and machine learning methods for spatially resolved transcriptomics data analysis
.
Genome Biol
 
2022
;
23
:
83
23
.

This is an Open Access article distributed under the terms of the Creative Commons Attribution License (https://creativecommons.org/licenses/by/4.0/), which permits unrestricted reuse, distribution, and reproduction in any medium, provided the original work is properly cited.
Associate Editor: Anthony Mathelier
Anthony Mathelier
Associate Editor
Search for other works by this author on:

Supplementary data