In
information technology, big data is a collection of data sets so large
and complex that it becomes difficult to process using on-hand database
management tools. The challenges include capture, storage, search, sharing,
analysis, and visualization. The trend to larger data sets is due to the
additional information derivable from analysis of a single large set of related
data, as compared to separate smaller sets with the same total amount of data,
allowing correlations to be found to "spot business trends, determine
quality of research, prevent diseases, link legal citations, combat crime, and
determine real-time roadway traffic conditions."
Big data usually includes data sets with sizes
beyond the ability of commonly-used software tools to capture, curate, manage,
and process the data within a tolerable elapsed time. Big data sizes are a
constantly moving target, as of 2012 ranging from a few dozen terabytes to many
petabytes of data in a single data set. With this difficulty, a new platform of
"big data" tools has arisen to handle sense making over large
quantities of data.
Examples
Examples include web logs, RFID, sensor networks,
social networks, social data (due to the social data revolution), Internet text
and documents, Internet search indexing, call detail records, astronomy,
atmospheric science, genomics, biogeochemical, biological, and other complex
and often interdisciplinary scientific research, military surveillance, medical
records, photography archives, video archives, and large-scale e-commerce.
Market
"Big data" has increased the demand of
information management specialists. IT companies are spending billions of
dollars on software firms only specializing in data management and analytics.
This industry on its own is worth more than $100 billion and growing at almost
10% a year which is roughly twice as fast as the software business as a whole.
Technologies
Big data requires exceptional technologies to
efficiently process large quantities of data within tolerable elapsed times. A
2011 McKinsey report suggests suitable technologies include A/B testing,
association rule learning, classification, cluster analysis, crowd sourcing,
data fusion and integration, ensemble learning, genetic algorithms, machine
learning, natural language processing, neural networks, pattern recognition,
predictive modeling, regression, sentiment analysis, signal processing,
supervised and unsupervised learning, simulation, time series analysis and
visualization.
Real or near-real time information delivery is
one of the defining characteristics of big data analytics. Latency is therefore
avoided whenever and wherever possible. Data in memory is good—data on spinning
disk at the other end of a FC SAN connection is not. The cost of a SAN at the
scale needed for analytics applications is very much higher than other storage
techniques.
No comments:
Post a Comment