Improved performance of data warehouse

Data quality is the most important variables for data warehousing. Numerous data warehouse ventures fall flat because of low quality of the data. It is trusted that the issues can be altered later and hence, a great deal of the reality of the situation will become obvious eventually spent to settle the errors. In the event that low-quality data nourished in the data warehouse, the outcome will be not precise if these data are utilized as a part of the decision making. ETL is the center procedure of data integration and normally connected with the data warehousing. ETL is the lead procedure to fetch all the data in the form of homogenous, standard environment. In the event that source data obtained from different sources is not cleansed, extracted perfectly, converted and incorporated in the best possible way then we use Query Cache method to enhance the performance of ETL processing and minimize response time. Later if data is not cleansed properly after coming out data from ETL then there is algorithm which detects errors and dirty data inside the data warehouse. By using data coalition rule which applies in data warehouse to detect errors, dirty data and faults that could be

[1]  Jennifer Widom,et al.  Performance Issues in Incremental Warehouse Maintenance , 2000, VLDB.

[2]  Huamin Wang,et al.  An ETL Services Framework Based on Metadata , 2010, 2010 2nd International Workshop on Intelligent Systems and Applications.

[3]  Hongming Cai,et al.  An automatic method of data warehouses multi-dimension modeling for distributed information systems , 2011, Proceedings of the 2011 15th International Conference on Computer Supported Cooperative Work in Design (CSCWD).

[4]  Weiming Shen,et al.  Computer Supported Cooperative Work in Design , 2003, Trans. SDPS.

[5]  Prayag Tiwari Improvement of ETL through integration of query cache and scripting method , 2016, 2016 International Conference on Data Science and Engineering (ICDSE).

[6]  Hector Garcia-Molina,et al.  Efficient resumption of interrupted warehouse loads , 2000, SIGMOD '00.

[7]  Prayag Tiwari Advanced ETL (AETL) by integration of PERL and scripting method , 2016, 2016 International Conference on Inventive Computation Technologies (ICICT).

[8]  Dennis Shasha,et al.  AJAX: an extensible data cleaning tool , 2000, SIGMOD '00.

[9]  Anany Levitin,et al.  The Notion of Data and Its Quality Dimensions , 1994, Inf. Process. Manag..

[10]  Hector Garcia-Molina,et al.  Efficient resumption of interrupted warehouse loads , 2000, SIGMOD 2000.

[11]  Gerald Weber,et al.  An Event-Based Near Real-Time Data Integration Architecture , 2008, 2008 12th Enterprise Distributed Object Computing Conference Workshops.

[12]  Ralph Kimball,et al.  Dealing with dirty data , 1996 .

[13]  Payal Pahwa,et al.  Alliance Rules for Data Warehouse Cleansing , 2009, 2009 International Conference on Signal Processing Systems.

[14]  Prayag Tiwari Comparative Analysis of Big Data , 2016 .

[15]  Praveen Sharma Advanced Applications of Data Warehousing Using 3-tier Architecture , 2009 .