EXTRACT DATA IN LARGE DATABASE WITH HADOOP
S.P. Siddique Ibrahim
Computer Science and Engineering Kumaraguru College of Technology Coimbatore, India
Published Online : 2014-11-30
Download Full Article : PDF Check for Updates
S.P. Siddique Ibrahim
Computer Science and Engineering Kumaraguru College of Technology Coimbatore, India
Published Online : 2014-11-30
Download Full Article : PDF Check for Updates
Abstract:
Data is basic building block of any organization and extracting useful information from raw available data is the big task and high complexity task. Data are the patterns which are used to develop or enhance knowledge. The rapid growth in the size of datasets that are collected from different resources has made capturing, managing and analyzing the datasets beyond the ability of most software tools. The current methodologies and data mining software tools cannot manage the current growth of datasets and extracting knowledge. With the advancement in information technology starting from file system to Object Oriented database, now it has reached to Data Warehouse and Data Marts. But every piece of data stored in these databases may not be useful for the decision purpose. Organizations need to filter the useful data from bulk of data which can be used for decision making, reporting or analysis. Big Data mining is the capacity of extracting useful information from these large datasets or social networking datasets, that due to its volume, variability, and velocity, it was not answer with the available methodology. Hadoop is an open source project, and pioneered a fundamentally a new way of storing and processing data.
Keywords: Big Data, Hadoop, Data Warehouse