TopicENA: Enabling Epistemic Network Analysis at Scale through Automated Topic-Based Coding

This paper introduces TopicENA, a scalable framework that integrates BERTopic with Epistemic Network Analysis to automate concept coding, thereby enabling the structural analysis of large text corpora while providing practical guidance on optimizing topic granularity and inclusion thresholds.

Owen H. T. Lu, Tiffany T. Y. Hsu

Published 2026-03-05
📖 5 min read🧠 Deep dive

Imagine you are a detective trying to solve a mystery hidden inside a massive library containing millions of books. Your goal isn't just to count how many times the word "cat" appears; you want to understand how ideas connect. Do "cats" usually appear near "milk"? Do they appear near "dogs" or "space rockets"?

This is what Epistemic Network Analysis (ENA) does. It maps out how concepts in text (like student essays or chat logs) hang out together, creating a visual "friendship map" of ideas.

The Old Problem: The Manual Bottleneck

Traditionally, to create these maps, a team of expert detectives had to read every single sentence and manually tag the ideas.

  • The Analogy: Imagine trying to map the entire internet by having one person read every single webpage and write down the topics by hand. It would take forever. You could only analyze a tiny, tiny slice of the data before you got exhausted. This meant ENA was great for small studies but useless for huge datasets.

The New Solution: TopicENA

The authors of this paper, Owen Lu and Tiffany Hsu, built a new tool called TopicENA. Think of it as hiring a super-fast, AI-powered robot librarian to do the tagging for you.

Instead of humans reading every sentence, they use a smart AI system called BERTopic to automatically group words into "topics" (like "Electric Cars," "Pollution," or "Space Exploration"). Then, they feed these AI-generated topics into the ENA system to build the maps.

The paper tests this new robot librarian in three different scenarios to see how to get the best results.


The Three Test Cases (The "Goldilocks" Experiments)

The researchers ran three experiments to figure out the "settings" for their robot. They realized that if you set the robot's brain too high or too low, the maps break.

Case 1: The Size of the "Bins" (Topic Granularity)

Imagine you are sorting a pile of mixed LEGO bricks into bins.

  • Coarse Granularity (Big Bins): You have a few huge bins. One bin is "Red Stuff," another is "Blue Stuff."
    • Result: If you have a tiny pile of bricks, big bins are useless because everything ends up in the same bin, and you can't see any patterns. But if you have a mountain of bricks, big bins work great because they keep the chaos organized.
  • Fine Granularity (Tiny Bins): You have hundreds of tiny bins. One is "Red 2x4 Brick," another is "Red 1x2 Brick."
    • Result: If you have a mountain of bricks, you end up with thousands of empty bins, and the map looks messy and broken. But if you have a tiny pile, tiny bins help you see the specific differences.

The Lesson: For huge datasets, use "big bins" (coarse topics). For small datasets, use "small bins" (fine topics).

Case 2: The Strictness of the "Guest List" (Topic Inclusion Threshold)

Imagine a party where a guest can belong to multiple groups (e.g., "The Science Club" and "The Art Club").

  • Low Threshold (Too Lenient): You let everyone into every club. Suddenly, the "Science Club" and "Art Club" are the exact same group of people. The map becomes a giant, blurry blob where you can't tell anyone apart.
  • High Threshold (Too Strict): You are so picky that you only let people in if they are 100% sure they belong. Most people get kicked out. The party is empty, and you can't see any connections.
  • Medium Threshold (Just Right): You let people in if they have a decent connection. Now you can see that the "Science Club" has a few artists, and the "Art Club" has a few scientists. The map is clear and interesting.

The Lesson: You need to find a "sweet spot" for how strict you are about assigning topics. If you are too loose or too strict, the map becomes useless.

Case 3: The Stress Test (Scale)

Finally, they threw the entire library (24,000 essays, nearly half a million sentences) at the robot.

  • The Result: The robot didn't crash! It successfully identified the 7 main themes of the 7 different writing assignments (like "Electoral College" or "Driverless Cars") without a human ever telling it what those themes were.
  • The Discovery: It even found subtle differences between high-scoring students and low-scoring students. For example, high-scoring students connected "driverless cars" and "pollution" more often in their thinking than low-scoring students did.

Why This Matters (The Big Picture)

Before this, analyzing the "thinking patterns" of thousands of students was impossible because it required armies of human coders.

TopicENA changes the game:

  1. It's Scalable: It can handle massive amounts of data (like a whole school district's essays) in a fraction of the time.
  2. It Shifts the Human Role: Instead of humans doing the boring, repetitive work of tagging every sentence, humans now act as conductors. They set the parameters (like the bin sizes and guest list rules) and then interpret the beautiful, complex maps the AI generates.
  3. It's More Accurate: Unlike older methods that just counted words, this AI understands context. It knows that "bank" in a river context is different from "bank" in a money context.

The Bottom Line

This paper introduces a way to use AI to map the "thought structures" of huge groups of people. It's like upgrading from a hand-drawn sketch of a single neighborhood to a satellite view of an entire continent, allowing educators and researchers to finally see the big picture of how people learn and think.