Importance of normalisation method for Reference Mapping #4627
Unanswered
chris-rands
asked this question in
Q&A
Replies: 1 comment 1 reply
-
While you may get reasonable results when using log-normalization for query and SCTransform for reference, we don't recommend this. The following script can be used to convert a log_normalized UMI matrix into a counts matrix. It only works if the original dataset was a UMI matrix, and therefore the smallest non-zero value in each cell vector represents 1 UMI. You should be able to use this to convert the data in your AnnData object to a counts matrix, and then can map as described in the vignette.
|
Beta Was this translation helpful? Give feedback.
1 reply
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
-
I'm testing the nice reference mapping feature. It says "The reference was normalized using SCTransform(), so we use the same approach to normalize the query here." My question: is that important? I have a scanpy AnnData object that was processed with a scandard scanpy workflow (normalize_total, log1p, no regression/scaling), which I convert to h5seurat format- is it okay to use this to compare to the Seurat PBMC reference, which was normalised differently? The scanpy object does not hold the raw counts so I cannot re-normalize. The results look promising, thanks
(Cross-posted with issue #4625, feel free to delete one of these threads)
Beta Was this translation helpful? Give feedback.
All reactions