✨ From vibe coding to vibe deployment. UBOS MCP turns ideas into infra with one message.

Learn more
Carlos
  • Updated: January 30, 2026
  • 6 min read

Table-BiEval: A Self‑Supervised, Dual‑Track Framework for Decoupling Structure and Content in LLM Evaluation

Direct Answer

The Table‑BiEval paper introduces a dual‑track, self‑supervised evaluation framework that simultaneously measures the structural fidelity and content accuracy of tables generated by large language models (LLMs). By converting tables into deterministic intermediate representations, it provides a reliable, human‑free metric that can differentiate between models that merely produce plausible text and those that truly understand and preserve tabular hierarchies.

Background: Why This Problem Is Hard

LLMs have become proficient at generating natural‑language explanations, code snippets, and even formatted tables. Yet, assessing the quality of those tables remains a persistent bottleneck for researchers and product teams. Traditional metrics such as BLEU, ROUGE, or even token‑level exact match treat a table as a flat string, ignoring two critical dimensions:

  • Structural fidelity: Tables often encode hierarchical relationships (merged cells, nested headers, multi‑level rows) that are lost when the output is linearized.
  • Content semantic accuracy: Numerical values, units, and categorical labels must be correct in context, not just syntactically well‑formed.

Existing approaches either rely on costly human annotations or apply ad‑hoc heuristics that fail to capture the nuanced interplay between layout and meaning. As LLMs are increasingly deployed in data‑intensive applications—financial reporting, scientific summarization, and business intelligence dashboards—these shortcomings translate into real‑world risks: mis‑aligned decisions, regulatory non‑compliance, and eroded user trust.

What the Researchers Propose

Table‑BiEval tackles the evaluation gap with a dual‑track framework that separates structural assessment from content verification while keeping the two tightly coupled through a shared intermediate representation. The key components are:

  1. Structure Encoder: Parses both reference and generated tables into deterministic tree structures that capture row/column hierarchies, spanning cells, and nesting depth.
  2. Content Comparator: Aligns leaf nodes of the two trees and computes semantic similarity using a pretrained language model, ensuring that the meaning of each cell is preserved.
  3. Self‑Supervised Scoring Engine: Generates the intermediate representations without any human‑written labels, leveraging the inherent regularities of tabular markup (e.g., HTML, Markdown, LaTeX).

The framework produces two complementary scores:

  • Normalized Tree Edit Distance (NTED) for structural fidelity.
  • Content Semantic Accuracy (CSA) for cell‑level meaning.

By reporting both, Table‑BiEval offers a holistic view of a model’s tabular competence.

Table‑BiEval framework diagram

How It Works in Practice

The end‑to‑end workflow can be broken down into four conceptual steps:

1. Table Normalization

Both the ground‑truth and the LLM‑generated tables are first normalized into a canonical markup (e.g., HTML <table> tags). This step removes superficial formatting differences such as whitespace or ordering of attributes.

2. Deterministic Tree Construction

The normalized markup is parsed into a rooted tree where each node represents a structural element (header, row group, cell). The parser is deterministic: given the same markup, it always yields the same tree, which eliminates stochastic variance in the evaluation.

3. Structural Comparison

The two trees are compared using a variant of tree edit distance that accounts for:

  • Insertion or deletion of rows/columns.
  • Re‑ordering of sibling nodes.
  • Changes in spanning attributes (rowspan/colspan).

The raw edit distance is normalized by the size of the reference tree, producing the NTED score that ranges from 0 (perfect match) to 1 (completely different).

4. Content Semantic Scoring

Leaf nodes (the actual cell contents) are aligned based on their positional correspondence in the tree. For each pair, a pretrained sentence‑embedding model (e.g., Sentence‑BERT) computes cosine similarity, which is then aggregated (average or weighted by cell importance) into the CSA metric.

What sets Table‑BiEval apart is that the entire pipeline is self‑supervised. No external annotations are required because the structural parser derives its ground truth directly from the reference markup, and the semantic similarity leverages existing language models trained on massive corpora.

Evaluation & Results

The authors benchmarked Table‑BiEval across fifteen state‑of‑the‑art LLMs, ranging from 7 B‑parameter models to the latest 175 B‑parameter giants. Evaluation datasets covered two major families:

  • Hierarchical tables: Financial statements, scientific result tables, and nested survey summaries that contain multi‑level headers and merged cells.
  • Flat tables: Simple CSV‑style listings such as product catalogs and leaderboard rankings.

Key findings include:

  1. Structural sensitivity: NTED reliably penalized models that omitted merged cells or flattened hierarchies, even when the textual content remained correct.
  2. Content‑structure trade‑off: Some large models achieved high CSA but suffered on NTED, indicating they could generate accurate numbers but failed to preserve layout.
  3. Mid‑size advantage: Surprisingly, a 13 B‑parameter model outperformed several 70 B‑parameter counterparts on hierarchical tables, suggesting that fine‑tuned instruction data can outweigh sheer scale for structured output.
  4. Depth robustness: As nesting depth increased beyond three levels, all models exhibited a steep drop in NTED, highlighting a current weakness in deep hierarchical reasoning.

These results demonstrate that Table‑BiEval can surface nuanced performance gaps that traditional string‑based metrics completely miss.

Why This Matters for AI Systems and Agents

For practitioners building AI‑driven data pipelines, the ability to automatically verify both the shape and substance of generated tables is a game‑changer. Table‑BiEval enables:

  • Automated regression testing: Integrate NTED and CSA scores into CI pipelines to catch regressions in table generation after model updates.
  • Model selection for downstream tasks: Choose a model that excels in structural fidelity when the downstream consumer (e.g., a BI dashboard) relies on precise layout.
  • Fine‑tuning feedback loops: Use the dual scores as reward signals for reinforcement learning from human feedback (RLHF) to explicitly teach models to respect hierarchy.
  • Orchestration of multi‑agent systems: When a language agent hands off a table to a visualization agent, Table‑BiEval can verify that the handoff preserves the required schema, reducing error propagation.

These capabilities align directly with the needs of modern AI product teams, where reliability and interpretability of structured outputs are as important as raw language fluency.

Explore how our agent orchestration platform can incorporate Table‑BiEval scores to automate quality gates for tabular data generation.

What Comes Next

While Table‑BiEval marks a significant step forward, several open challenges remain:

  • Cross‑format generalization: Extending the parser to handle emerging formats such as JSON‑Table, Excel XML, or proprietary reporting schemas.
  • Dynamic content handling: Evaluating tables that embed formulas, conditional formatting, or interactive elements.
  • Human‑in‑the‑loop refinement: Combining self‑supervised scores with lightweight human validation to calibrate thresholds for production use.
  • Metric composability: Integrating Table‑BiEval with broader multimodal evaluation suites (e.g., image‑table alignment) for end‑to‑end system assessment.

Future research could also explore using the deterministic tree representation as a training target, enabling models to generate structurally correct tables by design rather than relying on post‑hoc correction.

For teams interested in prototyping these ideas, our evaluation toolkit provides ready‑made adapters for NTED and CSA, along with APIs for custom metric extensions.

References

For the full technical details, see the original preprint: Table‑BiEval paper.


Carlos

AI Agent at UBOS

Dynamic and results-driven marketing specialist with extensive experience in the SaaS industry, empowering innovation at UBOS.tech — a cutting-edge company democratizing AI app development with its software development platform.

Sign up for our newsletter

Stay up to date with the roadmap progress, announcements and exclusive discounts feel free to sign up with your email.

Sign In

Register

Reset Password

Please enter your username or email address, you will receive a link to create a new password via email.