Sample barcoding-associated technical variation in probe-based single-cell RNA sequencing

This study reveals that probe set barcoding in the 10x Genomics Flex v1 single-cell RNA sequencing assay introduces substantial, reproducible technical variation that can generate false positive differential expression results when confounded with biological samples, a limitation effectively mitigated by the improved Flex v2 design.

Weir, J. A., Krebs, Y., Chen, F.

Published 2026-04-08
📖 4 min read☕ Coffee break read
⚕️

This is an AI-generated explanation of a preprint that has not been peer-reviewed. It is not medical advice. Do not make health decisions based on this content. Read full disclaimer

Imagine you are a detective trying to solve a mystery by interviewing hundreds of witnesses (the cells) to see what they saw (their gene activity). You want to compare two groups: the "Good Guys" and the "Bad Guys."

In the past, interviewing this many witnesses was slow and expensive. But a new technology called 10x Genomics Flex arrived like a high-tech, super-fast interrogation room. It can interview thousands of witnesses at once, even from old, dusty case files (archival tissues).

The "Labeling" Glitch

Here is how the new system works:
To keep track of which witness belongs to which group, the system uses special colored stickers (barcodes) on the interview forms.

In the first version of this system (Flex v1), the scientists made a shortcut to save money. Instead of giving every single witness a unique sticker, they grouped the interviews into 16 different batches. Each batch had its own type of sticker.

  • Batch A got "Red Stickers."
  • Batch B got "Blue Stickers."
  • And so on.

The idea was: "If we interview all the 'Good Guys' in Batch A and all the 'Bad Guys' in Batch B, we can just look at the stickers later to know who is who."

The Problem: The Sticker Itself Was Lying

The paper discovered a major flaw in this shortcut. It turns out that the Red Stickers and Blue Stickers didn't just label the groups; they actually changed how the interview forms were filled out!

  • Maybe the "Red Sticker" batch naturally wrote down more words.
  • Maybe the "Blue Sticker" batch missed a few details.

It wasn't because the witnesses were different; it was because the sticker itself made the notes look different.

The Analogy:
Imagine you are grading two classes of students.

  • Class A takes a test on Blue Paper.
  • Class B takes a test on Red Paper.

If the Blue Paper makes the ink look darker and the Red Paper makes it look lighter, you might accidentally think Class A is smarter just because their answers look "bolder." In reality, the students are the same; the paper is the problem.

In the Flex v1 system, if you accidentally put all your "Bad Guys" on Red Paper and all your "Good Guys" on Blue Paper, the computer would think the groups are totally different. It would find hundreds of "differences" that are actually just fake clues caused by the paper color, not the students.

The Solution: A Better System

The scientists found that the new version of the system (Flex v2) fixed this.

  • Flex v1: The sticker type = The sample group. (Bad idea, because the sticker causes the error).
  • Flex v2: The sticker is mixed up randomly. The sample group is tracked separately from the sticker type.

It's like giving every student a random colored paper, but writing their name clearly on the top. Now, even if the Blue Paper makes ink look darker, you know exactly which student wrote it, so you don't get fooled.

Why This Matters

This paper is a warning to all scientists using this technology: Be careful how you set up your experiment.

If you aren't careful, you might think you've discovered a huge difference between two groups of people (like a disease vs. healthy), when in reality, you just discovered that one group was tested with "Red Stickers" and the other with "Blue Stickers."

The Takeaway:
Just because a new, fast, and cheap technology exists doesn't mean it's perfect. Sometimes, the way we organize the data (the "sticker") can create fake results. Scientists need to design their experiments carefully to make sure they are measuring the people, not the paper.

Get papers like this in your inbox

Personalized daily or weekly digests matching your interests. Gists or technical summaries, in your language.

Try Digest →