AutoCaption: An Approach to Generate Natural Language Description from Visualization Automatically

In this paper, we propose a novel approach to generate captions for visualization charts automatically. In the proposed method, visual marks and visual channels, together with the associated text information in the original charts, are first extracted and identified with a multilayer perceptron classifier. Meanwhile, data information can also be retrieved by parsing visual marks with extracted mapping relationships. Then a 1-D convolutional residual network is employed to analyze the relationship between visual elements, and recognize significant features of the visualization charts, with both data and visual information as input. In the final step, the full description of the visual charts can be generated through a template-based approach. The generated captions can effectively cover the main visual features of the visual charts and support major feature types in commons charts. We further demonstrate the effectiveness of our approach through several cases.

[1]  Kwan-Liu Ma,et al.  Temporal Summary Images: An Approach to Narrative Visualization via Interactive Annotation Generation and Placement , 2017, IEEE Transactions on Visualization and Computer Graphics.

[2]  Robert Dale,et al.  Building applied natural language generation systems , 1997, Natural Language Engineering.

[3]  C. Lee Giles,et al.  Automatic Extraction of Data from Bar Charts , 2015, K-CAP.

[4]  Lotfi A. Zadeh A prototype-centered approach to adding deduction capability to search engines-the concept of protoform , 2002, 2002 Annual Meeting of the North American Fuzzy Information Processing Society Proceedings. NAFIPS-FLINT 2002 (Cat. No. 02TH8622).

[5]  Kathleen F. McCoy,et al.  Interactive SIGHT: textual access to simple bar charts , 2010, New Rev. Hypermedia Multim..

[6]  Samy Bengio,et al.  Show and tell: A neural image caption generator , 2014, 2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).

[7]  Kim Marriott,et al.  Tactile chart generation tool , 2008, Assets '08.

[8]  Peng Wu,et al.  Recognizing the Intended Message of Line Graphs , 2010, Diagrams.

[9]  Nancy Green,et al.  A Model of Perceptual Task Effort for Bar Charts and its Role in Recognizing Intention , 2006, User Modeling and User-Adapted Interaction.

[10]  Johanna D. Moore,et al.  Describing Complex Charts in Natural Language: A Caption Generation System , 1998, CL.

[11]  Man Lung Yiu,et al.  Extracting Top-K Insights from Multi-dimensional Data , 2017, SIGMOD Conference.

[12]  Nicholas Diakopoulos,et al.  Contextifier: automatic generation of annotated stock visualizations , 2013, CHI.

[13]  Peter J. Haas,et al.  Foresight: Recommending Visual Insights , 2017, Proc. VLDB Endow..

[14]  Alex Endert,et al.  Augmenting Visualizations with Interactive Data Facts to Facilitate Interpretation and Communication , 2019, IEEE Transactions on Visualization and Computer Graphics.

[15]  Jeffrey Heer,et al.  ReVision: automated classification, analysis and redesign of chart images , 2011, UIST.

[16]  Anna Wilbik,et al.  Linguistic summarization of sensor data for eldercare , 2011, 2011 IEEE International Conference on Systems, Man, and Cybernetics.

[17]  Jian Sun,et al.  Deep Residual Learning for Image Recognition , 2015, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).

[18]  Yoshua Bengio,et al.  Show, Attend and Tell: Neural Image Caption Generation with Visual Attention , 2015, ICML.

[19]  Stephanie Elzer Schwartz,et al.  Information graphics: an untapped resource for digital libraries , 2006, SIGIR.

[20]  Jeffrey Heer,et al.  Reverse‐Engineering Visualizations: Recovering Visual Encodings from Chart Images , 2017, Comput. Graph. Forum.