Data Standards Challenges for Interoperable and Quality Data

Data standards are agreed-on specifications about data objects and their relationships used to enable semantic interoperability of data originated from multiple sources and to help improve data quality. Despite their importance and the large number of data standards being created by standards development organizations [Cargill and Bolin 2007], little is understood regarding the quality of data standards and the costly and complex process of developing, maintaining, and using them [Lyytinen and King 2006]. It is common to see failures of standards efforts in practice [Bernstein and Haas 2008; Rosenthal et al. 2004]. A twodecade-old call for re-theorizing data standards still applies today [Wybo and Goodhue 1995]. Recent studies present important findings and opportunities for future research. We identify four representative works that (1) confirmed empirically the value of data standards for interoperability and business performance [Zhao and Xia 2014], (2) presented rules to identify and exclude suboptimal standards approaches under certain circumstances [Rosenthal et al. 2014], (3) explained the difficulties and the gap between standards development and implementation in the U.S. mortgage industry [Markus et al. 2006], and (4) proposed a set of characteristics of quality of data standards [Folmer 2012]. While we benefit from the above and other work on data standards, many questions remain unanswered. What is a “good” data standard? How do we measure its quality? What are the best processes and mechanisms for developing and maintaining standards that optimally address multiple objectives? How do we best manage the evolution of data standards? What kinds of data standards are most effective, or, more generally, what are the effects of data standards? Addressing these questions will reduce failures and improve the ability of data standards to produce interoperable and quality data.

[1]  Arnon Rosenthal,et al.  From semantic integration to semantics management: case studies and a way forward , 2004, SGMD.

[2]  Richard Y. Wang,et al.  A product perspective on total data quality management , 1998, CACM.

[3]  Richard Y. Wang,et al.  Journey to Data Quality , 2006 .

[4]  Carl Cargill,et al.  Standards and Public Policy: Standardization: a failing paradigm , 2006 .

[5]  Les Gasser,et al.  A framework for information quality assessment , 2007, J. Assoc. Inf. Sci. Technol..

[6]  Xitong Li,et al.  Data Standardization and Quality Degradation of Human-readable Data , 2015, ICIS.

[7]  K. Lyytinen,et al.  Standard making: a critical research frontier for information systems research , 2006 .

[8]  Harris Wu,et al.  Assessing the quality of large-scale data standards: A case of XBRL GAAP Taxonomy , 2014, Decis. Support Syst..

[9]  Dale Goodhue,et al.  Using interdependence as a predictor of data standards. Theoretical and measurement issues , 1995, Inf. Manag..

[10]  Diane M. Strong,et al.  Beyond Accuracy: What Data Quality Means to Data Consumers , 1996, J. Manag. Inf. Syst..

[11]  M. Lynne Markus,et al.  Industry-Wide Information Systems Standardization as Collective Action: The Case of the U.S. Residential Mortgage Industry , 2006, MIS Q..

[12]  Yang W. Lee,et al.  Crafting Rules: Context-Reflective Data Quality Problem Solving , 2003, J. Manag. Inf. Syst..

[13]  Hongwei Zhu,et al.  Firm Clustering Based on Financial Statements , 2012 .

[14]  Stuart E. Madnick,et al.  Overview and Framework for Data and Information Quality Research , 2009, JDIQ.

[15]  Laura M. Haas,et al.  Information integration in the enterprise , 2008, CACM.

[16]  Kexin Zhao,et al.  Forming Interoperability Through Interorganizational Systems Standards , 2014, J. Manag. Inf. Syst..

[17]  Adriane Chapman,et al.  Fit for purpose: engineering principles for selecting an appropriate type of data exchange standard , 2014, Inf. Syst. E Bus. Manag..

[18]  Erwin Folmer,et al.  Quality of semantic standards , 2012 .