Manual vs. Automated Gating: A Decision Framework
The gating problem
Gating is the central analytical act in flow cytometry. Every downstream result — population frequencies, MFI values, functional readouts — depends on where you draw the boundaries. And yet gating remains largely manual, subjective, and surprisingly variable between analysts.
Published studies have quantified the problem. The FlowCAP challenges (Aghaeepour et al., "Critical assessment of automated flow cytometry data analysis techniques," Nature Methods 10, 228–238, 2013) showed that inter-analyst variability in manual gating can reach 20–30% for difficult populations. Even experienced cytometrists disagree on gate boundaries, particularly for populations without clear separation from the negative.
This guide is a practical framework for deciding when to gate manually, when to use automated or AI-assisted methods, and how to combine the two.
How manual gating works
Manual gating means drawing geometric regions on bivariate dot plots or univariate histograms to define cell populations. The standard gate types are:
Polygon gates — the workhorse. You click vertices around an irregular population boundary. Best for populations with non-axis-aligned shapes, which is most real data. Most cytometrists use polygons for 80%+ of their gates.
Rectangle gates — axis-aligned boxes. Faster to draw but only appropriate when population boundaries align cleanly with both axes. Useful for broad cleanup gates (scatter gating for debris removal) or when populations have clear gaps.
Ellipse gates — smooth oval boundaries. Good for normally distributed populations, but less common in practice because real populations are rarely symmetrical after compensation.
Quadrant gates — a single crosshair dividing a 2D plot into four mutually exclusive regions. Standard for co-expression analysis (e.g., CD4 vs CD8). The limitation: quadrant boundaries are always axis-aligned, which may not match the actual population boundaries after compensation spread.
Boolean gates — logical combinations (AND, OR, NOT) of existing gates. Essential for polyfunctional analysis (e.g., IFNg+ AND TNFa+ AND IL-2+ cells).
The standard gating hierarchy flows from cleanup to biology: time gate → scatter gate (FSC-A/SSC-A) → singlet gate (FSC-A/FSC-H) → live/dead → lineage markers → functional markers. Order matters — dead cells produce non-specific antibody binding, and doublets create false double-positives.
Sources of manual gating variability
Where does the subjectivity come in? Several places:
Boundary placement on continuous distributions. When there is no clear valley between positive and negative populations, two analysts will draw different boundaries. This is the dominant source of inter-analyst variability. FMO (Fluorescence Minus One) controls define the theoretical boundary, but even FMO interpretation involves judgment.
Gate shape choice. A polygon drawn tightly around a population gives different results than a generous rectangle. Neither is objectively wrong — the "correct" gate depends on whether you prioritize purity or recovery.
Compensation and transformation effects. Gate coordinates are display-dependent. The same gate drawn on biexponential vs. log display captures different events, because the transform changes where events appear visually.
Fatigue and inconsistency. An analyst gating 200 samples will drift over the course of a session. The gates drawn on sample 1 may differ from sample 150, even with the best intentions.
Maecker et al. ("Standardizing immunophenotyping for the Human Immunology Project," Nature Reviews Immunology, 2012) documented that standardized protocols and centralized gating reduced variability from ~20% to ~5% CV in multi-site HIV vaccine trials. The problem is tractable — but only with deliberate effort.
How automated gating approaches work
Automated gating algorithms fall into several categories:
Model-based methods (flowClust, flowMerge) fit statistical distributions (e.g., multivariate t-distributions) to the data and define populations by cluster membership. These work well on clean, well-separated populations but struggle with irregularly shaped clusters.
Density-based methods (flowDensity, OpenCyto) find populations by identifying density peaks and valleys. They handle irregular shapes better but require careful parameter tuning per panel.
Supervised/template methods (FLOCK, DAFi) learn from manually gated examples and apply similar logic to new data. These preserve the analyst's gating philosophy while reducing manual effort.
k-NN (k-Nearest Neighbor) approaches — this is what Cytomaton uses. The system records the gate coordinates from your manual gates on a per-panel basis. When you have enough examples (typically 3+ gates on the same panel type), it uses k-NN on the gate vertex coordinates to suggest boundaries for new samples. The suggestions are personalized — they learn your gating conventions, not a generic model.
The key distinction: model-based and density-based methods define populations from statistical properties of the data. Template and k-NN methods define populations based on what the analyst has previously drawn. The first approach discovers structure; the second reproduces a human decision.
When manual gating is the right choice
Novel experiments. When you are developing a new panel or looking at a cell type for the first time, you need to see the data, understand its structure, and make judgment calls. No algorithm can substitute for scientific reasoning about what a population "should" look like.
Small sample numbers (< 10 files). The overhead of setting up automated gating may exceed the time saved. If you are gating a pilot experiment with 5 tubes, just gate them.
Unusual or pathological samples. Clinical samples with abnormal populations (leukemias, severely immunocompromised patients) may confuse automated methods trained on healthy controls.
Publication-critical analyses. For key figures in a paper, many reviewers expect to see manually defined gating strategies. Automated gates should be validated against manual gates before publication.
Teaching and training. Learning to gate manually is essential for developing cytometric intuition. Automated methods should augment expertise, not replace it.
When automated gating is the right choice
Large batch processing (> 20 files). When you are applying the same gating strategy to 50–200 files from the same panel, automated or AI-assisted gating saves hours and improves consistency. This is the strongest use case.
Longitudinal studies. Multi-timepoint studies where you need consistent gating across months of acquisitions. Instrument drift between sessions means fixed gate coordinates may not work, but adaptive AI suggestions can track population shifts.
Multi-analyst standardization. When multiple analysts gate the same panel (core facilities, multi-site trials), automated suggestions provide a consistent starting point that reduces inter-analyst variability.
High-parameter exploratory analysis. With 30+ parameter spectral panels, manual gating of every possible combination is impractical. Clustering algorithms (FlowSOM, PhenoGraph) and dimensionality reduction (UMAP, tSNE) are the primary discovery tools, with manual gating used for validation.
Routine panels with established gating strategies. If you have gated the same immunophenotyping panel hundreds of times, the gates are highly predictable. AI suggestions that learn from your history can apply them in seconds.
Combining manual and automated: a practical workflow
The most effective approach is usually hybrid. Here is a workflow that works well in practice:
1. Gate a representative subset manually. Choose 3–5 samples that span the biological variability in your experiment (not just the cleanest ones). Gate them carefully, using FMO controls to set boundaries.
2. Let the AI suggest gates on remaining samples. Review the suggestions visually. Look for samples where the population has shifted relative to your training set — instrument drift, staining variation, or genuine biology.
3. Adjust where needed. Accept suggestions that match your judgment. Modify the ones that do not. Each adjustment improves future suggestions (in systems like Cytomaton that learn continuously).
4. Spot-check batch results. After batch processing, review summary statistics. Flag samples with unusual frequencies (e.g., CD3+ percentage far outside the expected range) for manual re-gating.
The goal is not to eliminate human judgment — it is to focus it where it matters. Let the algorithm handle the repetitive parts while you make the scientific decisions.
What Cytomaton can and cannot do
Cytomaton's AI gating uses a per-user k-NN model that learns from your gate history. It activates after you have drawn 3+ gates on the same panel type. Suggestions appear as dashed overlays with confidence scores — you always have the final say.
What it does well: Reproducing your gating conventions on new samples of the same panel type. Handling routine panels (immunophenotyping, T cell subsets) where population locations are predictable. Reducing batch processing time from hours to minutes.
What it cannot do: Gate populations it has never seen in your data. Handle novel panel types without training examples. Make scientific decisions about ambiguous populations — that is still your job. It also requires cloud connectivity and does not work offline.
Concordance: Internal testing shows ≥80% IoU (Intersection over Union) with the analyst's manual gates when ≥5 training gates are available. This is comparable to inter-analyst agreement on well-defined populations. On difficult populations (dim markers, continuous distributions), both human and AI variability increase.
Try Cytomaton
AI-assisted flow cytometry analysis that learns your gating style. Free during beta.
Join the beta