Pretrained deep learning model for cell type annotation
Top 85.7% on sourcepulse
scBERT is a deep learning model designed for cell type annotation of single-cell RNA sequencing (scRNA-seq) data. It addresses limitations in existing methods by leveraging a pre-trained Transformer architecture to understand gene-gene interactions, handle batch effects, and utilize latent information, benefiting researchers in genomics and bioinformatics.
How It Works
scBERT employs a pre-train and fine-tune paradigm. It is first pre-trained on large unlabeled scRNA-seq datasets to learn general gene-gene interaction patterns. This pre-trained model is then fine-tuned on user-specific data for supervised cell annotation tasks. The architecture is based on the Performer encoder, a variant of Transformers, which is advantageous for its ability to capture complex relationships within the high-dimensional scRNA-seq data.
Quick Start & Requirements
sc.pp.normalize_total
, sc.pp.log1p
).preprocess.py
.Highlighted Details
Maintenance & Community
Developed by Tencent AI Lab. The project is associated with a publication in Nature Machine Intelligence. Contact email fionafyang@tencent.com
is provided for questions.
Licensing & Compatibility
All rights reserved by Tencent AI Lab. The tool is for research purposes only and not approved for clinical use.
Limitations & Caveats
The tool is explicitly stated as being for research purposes and not for clinical use. Gene symbols must be revised according to a specific NCBI Gene database version (Jan 10, 2020), and data requires specific normalization steps using scanpy.
1 year ago
Inactive