AScale: automatically scaling the ETL+Q process for performance Online publication date: Tue, 15-Dec-2015
by Pedro Martins; Maryam Abbasi; Pedro Furtado
International Journal of Business Process Integration and Management (IJBPIM), Vol. 7, No. 4, 2015
Abstract: We investigate the problem of providing automatic scalability and data freshness to data warehouses, at the same time dealing with high-rate data efficiently. In general, data freshness is not guaranteed in those contexts, since data loading, transformation and integration are heavy tasks that are performed only periodically, instead of row by row. Many current data warehouses are designed to be deployed and work in a single server. However, for many applications, problems related to data volume processing times, data rates and requirements for fresh and fast responses, require new solutions to be found. The solution is to use/build parallel architectures and mechanisms to speed-up data integration and to handle fresh data efficiently. We propose a universal data warehouse parallelisation solution, that is, an approach that enables the automatic scalability and freshness of any data warehouse and ETL process. Our results show that the proposed system can handle scalablity to provide the desired processing speed and data freshness.
Existing subscribers:
Go to Inderscience Online Journals to access the Full Text of this article.
If you are not a subscriber and you just want to read the full contents of this article, buy online access here.Complimentary Subscribers, Editors or Members of the Editorial Board of the International Journal of Business Process Integration and Management (IJBPIM):
Login with your Inderscience username and password:
Want to subscribe?
A subscription gives you complete access to all articles in the current issue, as well as to all articles in the previous three years (where applicable). See our Orders page to subscribe.
If you still need assistance, please email subs@inderscience.com