Learning Hierarchical Knowledge in Text-Rich Networks with Taxonomy-Informed Representation Learning

The paper proposes TIER, a novel framework for Text-Rich Networks that constructs an implicit hierarchical taxonomy through similarity-guided contrastive learning and LLM refinement, then integrates this structure into node representations via a cophenetic correlation-based regularization loss to achieve superior, interpretable modeling of hierarchical semantics.

Yunhui Liu, Yongchao Liu, Yinfeng Chen, Chuntao Hong, Tao Zheng, Tieke He

Published 2026-03-10
📖 5 min read🧠 Deep dive

Imagine you walk into a massive, chaotic library. This library doesn't just have books; every book is a "node" in a giant web, connected to other books by invisible strings (like citations or co-purchases). Some books are about "Science," others about "Cooking," and some are about "How to bake a cake."

The problem? The librarians (current AI models) are great at reading the words on the page, but they treat every book as a flat, isolated item. They don't realize that "Baking a Cake" is a tiny branch of "Cooking," which is a branch of "Food," which is a branch of "Life." They miss the hierarchy.

This paper introduces a new system called TIER (Taxonomy-Informed Representation Learning) that teaches AI to understand this library not just as a pile of books, but as a structured tree with roots, branches, and leaves.

Here is how TIER works, broken down into simple steps:

1. The Problem: The "Flat" Library

Current AI models are like a student who memorizes every book title but doesn't understand the Dewey Decimal System. If you ask them, "What is similar to a book on 'Sourdough Bread'?", they might say "A book on 'Sourdough Starter'" (good!) but also "A book on 'Bread Machines'" (okay) and maybe "A book on 'The History of France'" (because they both have the word 'history' in the title). They miss the coarse-to-fine structure. They don't know that "Bread" is a sub-category of "Food."

2. The Solution: TIER's Three-Step Magic

TIER acts like a super-smart librarian who reorganizes the library from scratch, using a mix of math and a "Super-Brain" (a Large Language Model or LLM).

Step A: The "Similarity Magnet" (Clustering)

First, TIER uses a technique called Contrastive Learning. Imagine you have a pile of mixed-up socks. You want to group them.

  • The Old Way: You just look at the color.
  • The TIER Way: You look at the color and you notice that socks that are often found in the same drawer (connected by edges in the network) probably belong together.
    TIER pulls similar documents closer together in a digital space and pushes different ones apart, creating a "clustering-friendly" map.

Step B: The "Smart Sorter" (LLM Refinement)

This is the coolest part. Standard math (K-Means) is good at grouping things that look alike, but it's bad at understanding meaning. It might group "Apple the Fruit" and "Apple the Tech Company" together just because they both have the word "Apple."

TIER brings in an LLM (like a super-smart AI assistant) to act as a quality control inspector.

  • The Split: If a group is too messy (e.g., it has both "Apple Fruit" and "Apple Tech"), the LLM says, "Hey, split this group in two!"
  • The Merge: If two groups are actually about the same thing (e.g., "Baking Cakes" and "Making Pies"), the LLM says, "Combine these!"
  • The Labeling: The LLM reads the books in the group and gives them a human-readable name, like "Desserts" or "Electronics."

This creates a beautiful, hierarchical tree: Food → Desserts → Cakes.

Step C: The "Ruler" (Regularization)

Now that TIER has built this perfect tree, it needs to make sure the AI's internal brain (the embeddings) actually looks like that tree.

  • Imagine the tree is a blueprint.
  • TIER uses a mathematical ruler called the Cophenetic Correlation Coefficient.
  • It checks: "If two items are close cousins on the tree (like 'Cakes' and 'Pies'), are they also close neighbors in the AI's brain?"
  • If the AI puts them far apart, TIER gently nudges them closer. If the AI puts distant cousins (like 'Cakes' and 'Cars') too close, TIER pushes them apart.

3. The Result: A Smarter, Faster AI

By teaching the AI to respect this hierarchy, TIER achieves two amazing things:

  1. Better Understanding: The AI now knows that "Natural Language Understanding" is a specific type of "Computer Science," and it treats them with the right level of importance. It's not just memorizing words; it's understanding the structure of knowledge.
  2. Efficiency: Because TIER builds the tree once and uses it to guide the learning, it doesn't need to ask the "Super-Brain" (LLM) to read every single book in the library. It only asks the LLM to help organize the groups. This makes it much faster and cheaper than other methods that try to use a giant LLM for every single task.

The Analogy: Organizing a Wardrobe

  • Old AI: Throws all your clothes into one giant pile. To find a "Red Shirt," it has to dig through everything.
  • TIER:
    1. Sorts clothes into broad piles (Shirts, Pants, Shoes).
    2. Asks a smart friend (LLM) to check the piles: "Wait, this pile has both winter coats and summer t-shirts; let's split them."
    3. Creates a hierarchy: Clothing → Tops → Shirts → Red Shirts.
    4. Trains you to always put your Red Shirt in the "Red Shirts" bin, not just anywhere.

Why This Matters

In the real world, data is rarely flat. Scientific papers, medical records, and product catalogs are all organized in trees. TIER is the first tool that effectively teaches AI to navigate these trees automatically, without needing humans to manually label every single branch. It makes AI smarter, more organized, and much more efficient at finding what you're looking for.