## ----setup, include=FALSE------------------------------------------------
knitr::opts_chunk$set(dpi = 300)
knitr::opts_chunk$set(cache=FALSE)

## ---- message=FALSE------------------------------------------------------
require(flowAI)

## ---- collapse = TRUE----------------------------------------------------
data(Bcells)
Bcells

## ---- eval=FALSE---------------------------------------------------------
#  setwd(...)
#  fcsfiles <- dir(".", pattern="*fcs$")

## ---- eval=FALSE---------------------------------------------------------
#  resQC <- flow_auto_qc(Bcells)  # using a flowSet
#  resQC <- flow_auto_qc(Bcells[[1]]) # using a flowFrame
#  resQC <- flow_auto_qc(fcsfiles) # using a character vector

## ---- eval=FALSE---------------------------------------------------------
#  GbLimit <- 2    # decide the limit in gigabyte for your batches of FCS files
#  size_fcs <- file.size(fcsfiles)/1024/1024/1024    # it calculates the size in gigabytes for each FCS file
#  groups <- ceiling(sum(size_fcs)/GbLimit)
#  cums <- cumsum(size_fcs)
#  batches <- cut(cums, groups)

## ---- eval=FALSE---------------------------------------------------------
#  for(i in 1:groups){
#      flow_auto_qc(fcsfiles[which(batches == levels(batches)[i])], output = 0)
#  }