Skip to main page content
U.S. flag

An official website of the United States government

Dot gov

The .gov means it’s official.
Federal government websites often end in .gov or .mil. Before sharing sensitive information, make sure you’re on a federal government site.

Https

The site is secure.
The https:// ensures that you are connecting to the official website and that any information you provide is encrypted and transmitted securely.

Access keys NCBI Homepage MyNCBI Homepage Main Content Main Navigation
[Preprint]. 2023 Apr 4:2023.04.03.535322.
doi: 10.1101/2023.04.03.535322.

MANGEM: a web app for Multimodal Analysis of Neuronal Gene expression, Electrophysiology and Morphology

Affiliations

MANGEM: a web app for Multimodal Analysis of Neuronal Gene expression, Electrophysiology and Morphology

Robert Hermod Olson et al. bioRxiv. .

Update in

Abstract

Single-cell techniques have enabled the acquisition of multi-modal data, particularly for neurons, to characterize cellular functions. Patch-seq, for example, combines patch-clamp recording, cell imaging, and single-cell RNA-seq to obtain electrophysiology, morphology, and gene expression data from a single neuron. While these multi-modal data offer potential insights into neuronal functions, they can be heterogeneous and noisy. To address this, machine-learning methods have been used to align cells from different modalities onto a low-dimensional latent space, revealing multi-modal cell clusters. However, the use of those methods can be challenging for biologists and neuroscientists without computational expertise and also requires suitable computing infrastructure for computationally expensive methods. To address these issues, we developed a cloud-based web application, MANGEM (Multimodal Analysis of Neuronal Gene expression, Electrophysiology, and Morphology) at https://ctc.waisman.wisc.edu/mangem. MANGEM provides a step-by-step accessible and user-friendly interface to machine-learning alignment methods of neuronal multi-modal data while enabling real-time visualization of characteristics of raw and aligned cells. It can be run asynchronously for large-scale data alignment, provides users with various downstream analyses of aligned cells and visualizes the analytic results such as identifying multi-modal cell clusters of cells and detecting correlated genes with electrophysiological and morphological features. We demonstrated the usage of MANGEM by aligning Patch-seq multimodal data of neuronal cells in the mouse visual cortex.

PubMed Disclaimer

Conflict of interest statement

Competing interests The authors declare no competing interests.

Figures

Figure 1.
Figure 1.. Overview of MANGEM (Multimodal Analysis of Neuronal Gene Expression, Electrophysiology and Morphology).
User input to MANGEM includes multimodal single-cell data together with cell metadata. Within MANGEM, the multimodal data are aligned using machine learning methods, projecting disparate modalities into a low-dimensional common latent space. Clustering algorithms are applied within the latent space to identify cell clusters, and then analysis methods are provided in MANGEM to characterize the clusters by differential feature expression and correlation of features with the latent space. In addition to interactive plots generated at each step of the workflow, downloadable output includes tabular data files (cell coordinates in latent space, cluster annotations, top features for each cluster) and images depicting alignment, cross-modal cell clusters, and cluster analyses.
Figure 2.
Figure 2.. Cloud implementation of MANGEM using AWS infrastructure.
The application runs on Amazon Cloud Services using Elastic Beanstalk to provision an EC2 instance. The web server nginx serves as a reverse proxy to the gunicorn WSGI server. MANGEM is written in Python using the Plotly Dash framework. Long-running tasks are run in the background by Celery workers, with Redis acting as the message broker between MANGEM and Celery. Uploaded and processed data files are stored in a filesystem cache in AWS S3.
Figure 3.
Figure 3.. Data flow through MANGEM web application.
Input data passes into an alignment process, which will either run in the main process or in the background, depending on the method. In the case of background (asynchronous) alignment, a URL will be supplied to the user which will allow them to check on the job’s status and access the results upon completion. Aligned data feed into a clustering algorithm, and then data analysis methods can be applied to the cell clusters. Data visualization output can be produced at each stage of the process, and tabular data files of aligned data, cell clusters, and analysis results can be downloaded.
Figure 4.
Figure 4.. MANGEM analysis and visualization of neuronal gene expression and electrophysiological features in mouse visual cortex.
a.) Measures of alignment error and 3-d plot of superimposed aligned data in latent space are shown for the preloaded mouse visual cortex dataset after nonlinear manifold alignment. b.) Cross-modal clusters, obtained by Gaussian mixture model, are indicated by color in plots of aligned data for each modality. c.) Feature levels across all cells for the top 5 features for each cross-modal cluster. Normalized feature magnitude was ranked using the Wilcox Rank Sum test. Cross-modal clusters are identified by the colored bar at the top of each plot. d.) Biplots for Gene Expression and Electrophysiological features using dimensions 1 and 2 of the latent space. The top 15 features by correlation with the latent space are shown plotted as radial lines where the length is the value of correlation (max value 1).
Figure 5.
Figure 5.. MANGEM analysis and visualization of neuronal gene expression and morphological features in mouse visual cortex.
a.) Measures of alignment error and 3-d plot of superimposed aligned data in latent space are shown for the mouse morphology cortex dataset after nonlinear manifold alignment. b.) Cross-modal clusters, obtained by Gaussian mixture model, are indicated by color in plots of aligned data for each modality. c.) Feature expression levels across all cells for the top 10 differentially expressed features for each cross-modal cluster. Normalized feature expression was ranked using the Wilcox Rank Sum test. Cross-modal clusters are identified by the colored bar at the top of each plot. d.) Biplots for Gene Expression and Electrophysiological features using dimensions 1 and 2 of the latent space. The top 15 features by correlation with the latent space are shown plotted as radial lines where the length is the value of correlation (max value 1).

References

    1. Marx V. Patch-seq takes neuroscience to a multimodal place. Nat Methods. 2022. Nov;19(11):1340–4. - PubMed
    1. Institute for Genome Sciences. NeMO Archive - Home [Internet]. The Neuroscience Multi-omic Archive. [cited 2023 Mar 27]. Available from: https://nemoarchive.org/
    1. DANDI Archive [Internet]. [cited 2023 Mar 27]. Available from: https://dandiarchive.org/
    1. Rübel O, Tritt A, Ly R, Dichter BK, Ghosh S, Niu L, et al. The Neurodata Without Borders ecosystem for neurophysiological data science. Colgin LL, Jadhav SP, editors. eLife. 2022. Oct 4;11:e78362. - PMC - PubMed
    1. Efremova M, Teichmann SA. Computational methods for single-cell omics across modalities. Nat Methods. 2020. Jan;17(1):14–7. - PubMed

Publication types