Towards natural language interfaces for data visualization: A survey

L Shen, E Shen, Y Luo, X Yang, X Hu… - IEEE transactions on …, 2022 - ieeexplore.ieee.org
Utilizing Visualization-oriented Natural Language Interfaces (V-NLI) as a complementary
input modality to direct manipulation for visual analytics can provide an engaging user …

Ai4vis: Survey on artificial intelligence approaches for data visualization

A Wu, Y Wang, X Shu, D Moritz, W Cui… - … on Visualization and …, 2021 - ieeexplore.ieee.org
Visualizations themselves have become a data format. Akin to other data formats such as
text and images, visualizations are increasingly created, stored, shared, and (re-) used with …

Accessible visualization via natural language descriptions: A four-level model of semantic content

A Lundgard, A Satyanarayan - IEEE transactions on …, 2021 - ieeexplore.ieee.org
Natural language descriptions sometimes accompany visualizations to better communicate
and contextualize their insights, and to improve their accessibility for readers with …

Chart-to-text: A large-scale benchmark for chart summarization

S Kantharaj, RTK Leong, X Lin, A Masry… - arxiv preprint arxiv …, 2022 - arxiv.org
Charts are commonly used for exploring data and communicating insights. Generating
natural language summaries from charts can be very helpful for people in inferring key …

Expanding performance boundaries of open-source multimodal models with model, data, and test-time scaling

Z Chen, W Wang, Y Cao, Y Liu, Z Gao, E Cui… - arxiv preprint arxiv …, 2024 - arxiv.org
We introduce InternVL 2.5, an advanced multimodal large language model (MLLM) series
that builds upon InternVL 2.0, maintaining its core model architecture while introducing …

Vistext: A benchmark for semantically rich chart captioning

BJ Tang, A Boggust, A Satyanarayan - arxiv preprint arxiv:2307.05356, 2023 - arxiv.org
Captions that describe or explain charts help improve recall and comprehension of the
depicted data and provide a more accessible medium for people with visual disabilities …

Nvlm: Open frontier-class multimodal llms

W Dai, N Lee, B Wang, Z Yang, Z Liu, J Barker… - arxiv preprint arxiv …, 2024 - arxiv.org
We introduce NVLM 1.0, a family of frontier-class multimodal large language models (LLMs)
that achieve state-of-the-art results on vision-language tasks, rivaling the leading proprietary …

Mmt-bench: A comprehensive multimodal benchmark for evaluating large vision-language models towards multitask agi

K Ying, F Meng, J Wang, Z Li, H Lin, Y Yang… - arxiv preprint arxiv …, 2024 - arxiv.org
Large Vision-Language Models (LVLMs) show significant strides in general-purpose
multimodal applications such as visual dialogue and embodied navigation. However …

Unichart: A universal vision-language pretrained model for chart comprehension and reasoning

A Masry, P Kavehzadeh, XL Do, E Hoque… - arxiv preprint arxiv …, 2023 - arxiv.org
Charts are very popular for analyzing data, visualizing key insights and answering complex
reasoning questions about data. To facilitate chart-based data analysis using natural …

Communicating visualizations without visuals: Investigation of visualization alternative text for people with visual impairments

C Jung, S Mehta, A Kulkarni, Y Zhao… - IEEE transactions on …, 2021 - ieeexplore.ieee.org
Alternative text is critical in communicating graphics to people who are blind or have low
vision. Especially for graphics that contain rich information, such as visualizations, poorly …