From 343ded1d3afbbb8e1488c07db114272c40c29a36 Mon Sep 17 00:00:00 2001
From: nvBench2
Date: Tue, 4 Mar 2025 09:20:08 +0800
Subject: [PATCH] index
---
index.html | 40 +++++++++++++++++++++++++++++++++++++++-
1 file changed, 39 insertions(+), 1 deletion(-)
diff --git a/index.html b/index.html
index 7d959c2..2431112 100644
--- a/index.html
+++ b/index.html
@@ -67,6 +67,12 @@
font-weight: bold;
margin-top: 0.5rem;
}
+ .figure-description {
+ margin-top: 0.5rem;
+ text-align: justify;
+ font-style: italic;
+ font-size: 0.9rem;
+ }
table {
width: 100%;
margin-bottom: 2rem;
@@ -192,7 +198,7 @@
+ Figure 1: Example of reasoning appropriate visualizations from an ambiguous natural language query
+
+ As shown in Figure 1, a seemingly straightforward query like "Show the gross trend of comedy and action movies by year" contains multiple ambiguities: "gross" could refer to either World_Gross or Local_Gross columns, "Comedy and action" implicitly requires filtering by Genre, "trend" may suggest a bar chart or line chart, and "By year" implies temporal binning that isn't explicitly defined. The figure illustrates how these ambiguities can be resolved through step-wise reasoning to produce multiple valid visualizations.
+
@@ -267,6 +277,9 @@ Ambiguity-Injected NL2VIS Data Synthesizer
@@ -287,6 +300,10 @@ Ambiguity Injection Process
The process ensures traceability from query to visualization through explicit reasoning paths, enabling systematic evaluation of NL2VIS systems' ability to handle ambiguity.
+ Figure 3: Injecting ambiguities into a seed visualization
+
+ Figure 3 demonstrates how we inject ambiguities into a seed visualization through a systematic process: (1) Starting with a seed chart (e.g., a bar chart showing gross by year), (2) Converting it to a seed visualization tree with explicit nodes, (3) Injecting ambiguity nodes (e.g., introducing a choice between Local_Gross and World_Gross), (4) Resolving the tree into multiple valid visualization specifications, and (5) Flattening the trees into concrete visualization queries.
+
@@ -300,6 +317,9 @@ Benchmark Comparison
Benchmark Statistics
@@ -311,6 +331,9 @@ Benchmark Statistics
@@ -323,11 +346,17 @@
Benchmark Statistics
Table 4: Ambiguity count at each reasoning step.
+
+ This table shows the distribution of ambiguities across different reasoning steps in the nvBench 2.0 dataset, highlighting which steps in the visualization process are most prone to ambiguity.
+
Table 5: Statistics of ambiguity patterns.
+
+ Our dataset contains diverse ambiguity patterns, with Channel Encoding (CE) being the most common type of ambiguity (88.06%), followed by Data Transformation (DT) ambiguities (46.00%). Many samples contain multiple types of ambiguity, highlighting the complexity of real-world visualization requests.
+
@@ -373,16 +402,25 @@ Overall Performance