Thank you for your interest in the Infinity Flow approach. This vignette describes how to apply the package to your massively parallel cytometry (e.g. LEGENDScreen or Lyoplates kits) experiment. Massively parallel cytometry experiments are cytometry experiments where a sample is aliquoted in n subsamples, each stained with a fixed panel of “Backbone” antibodies. Each aliquot is in addition stained with a unique “Infinity” exploratory antibody. The goal of the infinityFlow package is to use information from the ubiquitous Backbone staining to predict the expression of sparsely-measured Infinity antibodies across the entire dataset. To learn more about this type of experiments and details about the Infinity Flow approach, please consult Becht et al, 2020. In this vignette we achieve this by using the XGBoost machine-learning framework implemented in the xgboost R package. This vignette aims at explaining how to apply a basic infinityFlow analysis. Advanced usages, including different machine learning models and custom hyperparameters values, are covered in a dedicated vignette.
This vignette will cover:
You can install the package from Bioconductor using
Now that the package is installed, we load the package and attach its example data. We also load flowCore that will be used to manipulate FCS files in R.
The example data is a subset of a massively parallel cytometry experiment of the mouse lung at steady state. The example data contains 10 FCS files. To mimick real world-conditions, we will write this set of FCS files to disk. In this vignette we will use a temporary directory, but you can use the directory of your choice.
dir <- file.path(tempdir(), "infinity_flow_example") print(dir) #>  "/tmp/Rtmp6OLoDr/infinity_flow_example" input_dir <- file.path(dir, "fcs") write.flowSet(steady_state_lung, outdir = input_dir) ## Omit this if you already have FCS files #>  "/tmp/Rtmp6OLoDr/infinity_flow_example/fcs" list.files(input_dir) #>  "Plate1_Specimen_001_B12_B12_024.fcs" "Plate1_Specimen_001_D8_D08_044.fcs" #>  "Plate1_Specimen_001_H11_H11_095.fcs" "Plate3_Specimen_001_A3_A03_003.fcs" #>  "Plate3_Specimen_001_A9_A09_009.fcs" "Plate3_Specimen_001_C11_C11_035.fcs" #>  "Plate3_Specimen_001_C6_C06_030.fcs" "Plate3_Specimen_001_E5_E05_053.fcs" #>  "Plate3_Specimen_001_F5_F05_065.fcs" "Plate3_Specimen_001_G2_G02_074.fcs" #>  "annotation.txt"
The second input we have to manually produce is the annotation of the experiment. In the context of a massively parallel cytometry experiment, we need to know what is the protein target of each Infinity (usually PE-conjugated or APC-conjugated) antibody, and what is its isotype. For the example dataset, the annotation is provided in the package and looks like this:
data(steady_state_lung_annotation) print(steady_state_lung_annotation) #> Infinity_target Infinity_isotype #> Plate1_Specimen_001_B12_B12_024.fcs CD28 SHIgG #> Plate1_Specimen_001_D8_D08_044.fcs CD49b(pan-NK) rIgM #> Plate1_Specimen_001_H11_H11_095.fcs CD137 SHIgG #> Plate3_Specimen_001_A3_A03_003.fcs KLRG1 SHIgG #> Plate3_Specimen_001_A9_A09_009.fcs Ly-49c/F/I/H SHIgG #> Plate3_Specimen_001_C6_C06_030.fcs Podoplanin SHIgG #> Plate3_Specimen_001_C11_C11_035.fcs SSEA-3 rIgM #> Plate3_Specimen_001_E5_E05_053.fcs TCR Vg3 SHIgG #> Plate3_Specimen_001_F5_F05_065.fcs SHIgG SHIgG #> Plate3_Specimen_001_G2_G02_074.fcs rIgM rIgM
steady_state_lung_annotation data.frame contains one line per FCS file, with
rownames(steady_state_lung_annotation) == sampleNames(steady_state_lung). The first column specifies the proteins targeted by the Infinity antibody in each FCS file, and the second column specifies its isotype (species and constant region of the antibody). If you load an annotation file from disk, use this command:
steady_state_lung_annotation = read.csv("path/to/targets/and/isotypes/annotation/file", row.names = 1, stringsAsFactors = FALSE)
That is all we need in terms of inputs! To recap you only need
Now that we have our input data, we need to specify which antibodies are part of the Backbone, and which one is the Infinity antibody. We provide an interactive function to specify this directly in R. This function can be run once, its output saved for future use by downstream functions in the infinityFlow package. The
select_backbone_and_exploratory_markers function will parse an FCS file in the input directory, and for each acquisition channel, ask the user whether it should be used as a predictor (Backbone), exploratory target (Infinity antibodies), or omitted (e.g. Time or Event ID columns…).
Below is an example of the interactive execution of the
select_backbone_and_exploratory_markers function for the example data. The resulting data.frame is printed too.
For each data channel, enter either: backbone, exploratory or discard (can be abbreviated) FSC-A (FSC-A):discard FSC-H (FSC-H):backbone FSC-W (FSC-W):b SSC-A (SSC-A):d SSC-H (SSC-H):b SSC-W (SSC-W):b CD69-CD301b (FJComp-APC-A):b Zombie (FJComp-APC-eFlour780-A):b MHCII (FJComp-Alexa Fluor 700-A):b CD4 (FJComp-BUV395-A):b CD44 (FJComp-BUV737-A):b CD8 (FJComp-BV421-A):b CD11c (FJComp-BV510-A):b CD11b (FJComp-BV605-A):b F480 (FJComp-BV650-A):b Ly6C (FJComp-BV711-A):b Lineage (FJComp-BV786-A):b CD45a488 (FJComp-GFP-A):b Legend (FJComp-PE(yg)-A):exploratory CD24 (FJComp-PE-Cy7(yg)-A):b CD103 (FJComp-PerCP-Cy5-5-A):b Time (Time):d name desc type $P1 FSC-A <NA> discard $P2 FSC-H <NA> backbone $P3 FSC-W <NA> backbone $P4 SSC-A <NA> discard $P5 SSC-H <NA> backbone $P6 SSC-W <NA> backbone $P7 FJComp-APC-A CD69-CD301b backbone $P8 FJComp-APC-eFlour780-A Zombie backbone $P9 FJComp-Alexa Fluor 700-A MHCII backbone $P10 FJComp-BUV395-A CD4 backbone $P11 FJComp-BUV737-A CD44 backbone $P12 FJComp-BV421-A CD8 backbone $P13 FJComp-BV510-A CD11c backbone $P14 FJComp-BV605-A CD11b backbone $P15 FJComp-BV650-A F480 backbone $P16 FJComp-BV711-A Ly6C backbone $P17 FJComp-BV786-A Lineage backbone $P18 FJComp-GFP-A CD45a488 backbone $P19 FJComp-PE(yg)-A Legend exploratory $P20 FJComp-PE-Cy7(yg)-A CD24 backbone $P21 FJComp-PerCP-Cy5-5-A CD103 backbone $P22 Time <NA> discard Is selection correct? (yes/no): yes
We cannot run this function interactively from this vignette, so we load the result from the package instead:
You need to save this backbone specification file as a CSV file for future use.
Now that we have our input data, FCS files annotation and specification of the Backbone and Infinity antibodies, we have everything we need to run the pipeline.
All the pipeline is packaged into a single function,
Here is a description of the basic arguments it requires:
We have everything we need in our input folder to fill these arguments:
First, input FCS files:
path_to_fcs <- file.path(dir, "fcs") head(list.files(path_to_fcs, pattern = ".fcs")) #>  "Plate1_Specimen_001_B12_B12_024.fcs" "Plate1_Specimen_001_D8_D08_044.fcs" #>  "Plate1_Specimen_001_H11_H11_095.fcs" "Plate3_Specimen_001_A3_A03_003.fcs" #>  "Plate3_Specimen_001_A9_A09_009.fcs" "Plate3_Specimen_001_C11_C11_035.fcs"
Output directory. It will be created if it doesn’t already exist
Backbone selection file:
list.files(dir) #>  "backbone_selection_file.csv" "fcs" backbone_selection_file <- file.path(dir, "backbone_selection_file.csv") head(read.csv(backbone_selection_file)) #> name desc type #> 1 FSC-A <NA> discard #> 2 FSC-H <NA> backbone #> 3 FSC-W <NA> backbone #> 4 SSC-A <NA> discard #> 5 SSC-H <NA> backbone #> 6 SSC-W <NA> backbone
Annotation of Infinity antibody targets and isotypes:
targets <- steady_state_lung_annotation$Infinity_target names(targets) <- rownames(steady_state_lung_annotation) isotypes <- steady_state_lung_annotation$Infinity_isotype names(isotypes) <- rownames(steady_state_lung_annotation) head(targets) #> Plate1_Specimen_001_B12_B12_024.fcs Plate1_Specimen_001_D8_D08_044.fcs #> "CD28" "CD49b(pan-NK)" #> Plate1_Specimen_001_H11_H11_095.fcs Plate3_Specimen_001_A3_A03_003.fcs #> "CD137" "KLRG1" #> Plate3_Specimen_001_A9_A09_009.fcs Plate3_Specimen_001_C6_C06_030.fcs #> "Ly-49c/F/I/H" "Podoplanin" head(isotypes) #> Plate1_Specimen_001_B12_B12_024.fcs Plate1_Specimen_001_D8_D08_044.fcs #> "SHIgG" "rIgM" #> Plate1_Specimen_001_H11_H11_095.fcs Plate3_Specimen_001_A3_A03_003.fcs #> "SHIgG" "SHIgG" #> Plate3_Specimen_001_A9_A09_009.fcs Plate3_Specimen_001_C6_C06_030.fcs #> "SHIgG" "SHIgG"
Other arguments are optional, but it is notably worth considering the number of input cells and the number of output cells. This will notably be important if you are using a computer with limited RAM. For the example data it does not matter as we only have access to 2,000 cells per well, but if you run the pipeline on your own data I suggest you start by low values, and ramp up (to 20,000 or 50,000 input cells, and e.g. 10,000 output cells per well) once everything is setup. Another optional argument is
cores which controls multicore computing, which can speed up execution at the cost of memory usage. In this vignette we use cores = 1, but you probably want to increase this to 4 or 8 or more if your computer can accomodate it.
There is also an argument to store temporary files, which can be useful to further analyze the data in R. If missing, this argument will default to a temporary directory.
At last, now let us execute the pipeline:
imputed_data <- infinity_flow( path_to_fcs = path_to_fcs, path_to_output = path_to_output, path_to_intermediary_results = path_to_intermediary_results, backbone_selection_file = backbone_selection_file, annotation = targets, isotype = isotypes, input_events_downsampling = input_events_downsampling, prediction_events_downsampling = prediction_events_downsampling, verbose = TRUE, cores = cores ) #> /tmp/Rtmp6OLoDr/infinity_flow_example/tmp and /tmp/Rtmp6OLoDr/infinity_flow_example/tmp/subsetted_fcs and /tmp/Rtmp6OLoDr/infinity_flow_example/tmp/rds and /tmp/Rtmp6OLoDr/infinity_flow_example/output: directories not found, creating directory(ies) #> Using directories... #> input: /tmp/Rtmp6OLoDr/infinity_flow_example/fcs #> intermediary: /tmp/Rtmp6OLoDr/infinity_flow_example/tmp #> subset: /tmp/Rtmp6OLoDr/infinity_flow_example/tmp/subsetted_fcs #> rds: /tmp/Rtmp6OLoDr/infinity_flow_example/tmp/rds #> annotation: /tmp/Rtmp6OLoDr/infinity_flow_example/tmp/annotation.csv #> output: /tmp/Rtmp6OLoDr/infinity_flow_example/output #> Parsing and subsampling input data #> Downsampling to 1000 events per input file #> Concatenating expression matrices #> Writing to disk #> Logicle-transforming the data #> Backbone data #> Exploratory data #> Writing to disk #> Transforming expression matrix #> Writing to disk #> Harmonizing backbone data #> Scaling expression matrices #> Writing to disk #> Fitting regression models #> Randomly selecting 50% of the subsetted input files to fit models #> Fitting... #> XGBoost #> #> 12.53386 seconds #> Imputing missing measurements #> Randomly drawing events to predict from the test set #> Imputing... #> XGBoost #> #> 2.537451 seconds #> Concatenating predictions #> Writing to disk #> Performing dimensionality reduction #> 21:02:27 UMAP embedding parameters a = 1.262 b = 1.003 #> 21:02:27 Read 5000 rows and found 17 numeric columns #> 21:02:27 Using Annoy for neighbor search, n_neighbors = 15 #> 21:02:28 Building Annoy index with metric = euclidean, n_trees = 50 #> 0% 10 20 30 40 50 60 70 80 90 100% #> [----|----|----|----|----|----|----|----|----|----| #> **************************************************| #> 21:02:29 Writing NN index file to temp file /tmp/Rtmp6OLoDr/file15001ed9f58b #> 21:02:29 Searching Annoy index using 1 thread, search_k = 1500 #> 21:02:31 Annoy recall = 100% #> 21:02:31 Commencing smooth kNN distance calibration using 1 thread #> 21:02:32 Initializing from normalized Laplacian + noise #> 21:02:32 Commencing optimization for 1000 epochs, with 102372 positive edges using 1 thread #> 21:02:45 Optimization finished #> Exporting results #> Transforming predictions back to a linear scale #> Exporting FCS files (1 per well) #> Plotting #> Chopping off the top and bottom 0.005 quantiles #> Shuffling the order of cells (rows) #> Producing plot #> Background correcting #> Transforming background-corrected predictions. (Use logarithm to visualize) #> Exporting FCS files (1 per well) #> Plotting #> Chopping off the top and bottom 0.005 quantiles #> Shuffling the order of cells (rows) #> Producing plot
The above command populated our output directory with new sets of files which we describe in the next section.
The output mainly consists of
Each of the above comes in two flavours, either raw or background-corrected.
At the end of the pipeline, input FCS files are augmented with imputed data. Feel free to explore these files in whatever flow cytometry software you are comfortable with! They should look pretty much like regular FCS files, although they are computationnally derived. You can find the output FCS files in the
path_to_output directory, specifically:
head(list.files(path_to_fcs)) ## Input files #>  "Plate1_Specimen_001_B12_B12_024.fcs" "Plate1_Specimen_001_D8_D08_044.fcs" #>  "Plate1_Specimen_001_H11_H11_095.fcs" "Plate3_Specimen_001_A3_A03_003.fcs" #>  "Plate3_Specimen_001_A9_A09_009.fcs" "Plate3_Specimen_001_C11_C11_035.fcs" fcs_raw <- file.path(path_to_output, "FCS", "split") head(list.files(fcs_raw)) ## Raw output FCS files #>  "Plate1_Specimen_001_B12_B12_024_target_CD28.fcs" #>  "Plate1_Specimen_001_D8_D08_044_target_CD49b(pan-NK).fcs" #>  "Plate1_Specimen_001_H11_H11_095_target_CD137.fcs" #>  "Plate3_Specimen_001_A3_A03_003_target_KLRG1.fcs" #>  "Plate3_Specimen_001_A9_A09_009_target_Ly-49c-F-I-H.fcs" #>  "Plate3_Specimen_001_C11_C11_035_target_SSEA-3.fcs" fcs_bgc <- file.path(path_to_output, "FCS_background_corrected", "split") ## Background-corrected output FCS files head(list.files(fcs_bgc)) ## Background-corrected output FCS files #>  "Plate1_Specimen_001_B12_B12_024_target_CD28.fcs" #>  "Plate1_Specimen_001_D8_D08_044_target_CD49b(pan-NK).fcs" #>  "Plate1_Specimen_001_H11_H11_095_target_CD137.fcs" #>  "Plate3_Specimen_001_A3_A03_003_target_KLRG1.fcs" #>  "Plate3_Specimen_001_A9_A09_009_target_Ly-49c-F-I-H.fcs" #>  "Plate3_Specimen_001_C11_C11_035_target_SSEA-3.fcs"
Finally, the pipeline produces two PDF files, with a UMAP embedding of the backbone data color-coded by the imputed data. This is a very informative output and a good way to start analyzing your data. These files are present in the
path_to_output directory. The example dataset is very small but feel free to look at the result for illustration purposes. This PDF is available at
file.path(path_to_output, "umap_plot_annotated.pdf") ## Raw plot #>  "/tmp/Rtmp6OLoDr/infinity_flow_example/output/umap_plot_annotated.pdf" file.path(path_to_output, "umap_plot_annotated_backgroundcorrected.pdf") ## Background-corrected plot #>  "/tmp/Rtmp6OLoDr/infinity_flow_example/output/umap_plot_annotated_backgroundcorrected.pdf"
Thank you for following this vignette, I hope you made it through the end without too much headache and that it was informative. General questions about proper usage of the package are best asked on the Bioconductor support site to maximize visibility for future users. If you encounter bugs, feel free to raise an issue on infinityFlow’s github.
sessionInfo() #> R version 4.0.3 (2020-10-10) #> Platform: x86_64-pc-linux-gnu (64-bit) #> Running under: Ubuntu 18.04.5 LTS #> #> Matrix products: default #> BLAS: /home/biocbuild/bbs-3.12-bioc/R/lib/libRblas.so #> LAPACK: /home/biocbuild/bbs-3.12-bioc/R/lib/libRlapack.so #> #> Random number generation: #> RNG: L'Ecuyer-CMRG #> Normal: Inversion #> Sample: Rejection #> #> locale: #>  LC_CTYPE=en_US.UTF-8 LC_NUMERIC=C #>  LC_TIME=en_US.UTF-8 LC_COLLATE=C #>  LC_MONETARY=en_US.UTF-8 LC_MESSAGES=en_US.UTF-8 #>  LC_PAPER=en_US.UTF-8 LC_NAME=C #>  LC_ADDRESS=C LC_TELEPHONE=C #>  LC_MEASUREMENT=en_US.UTF-8 LC_IDENTIFICATION=C #> #> attached base packages: #>  stats graphics grDevices utils datasets methods base #> #> other attached packages: #>  infinityFlow_1.0.0 flowCore_2.2.0 #> #> loaded via a namespace (and not attached): #>  Rcpp_1.0.5 raster_3.3-13 knitr_1.30 #>  magrittr_1.5 BiocGenerics_0.36.0 lattice_0.20-41 #>  rlang_0.4.8 pbapply_1.4-3 matlab_1.0.2 #>  stringr_1.4.0 tools_4.0.3 parallel_4.0.3 #>  grid_4.0.3 data.table_1.13.2 Biobase_2.50.0 #>  xfun_0.18 png_0.1-7 RSpectra_0.16-0 #>  htmltools_0.5.0 matrixStats_0.57.0 xgboost_220.127.116.11 #>  yaml_2.2.1 RcppParallel_5.0.2 digest_0.6.27 #>  Matrix_1.2-18 RcppAnnoy_0.0.16 cytolib_2.2.0 #>  RProtoBufLib_2.2.0 codetools_0.2-16 S4Vectors_0.28.0 #>  uwot_0.1.8 evaluate_0.14 rmarkdown_2.5 #>  sp_1.4-4 stringi_1.5.3 compiler_4.0.3 #>  generics_0.0.2 stats4_4.0.3