Академический Документы
Профессиональный Документы
Культура Документы
12/3/2013
Introduction
Identifying the data you need Understanding the fundamentals of big data integration Using Hadoop as ETL Knowing best practices for data integration
12/3/2013
need to take stock of the type of data you are dealing with. By leveraging new tools, organizations are gaining new insight from previously untapped sources of unstructured data in e-mails, customer service records, sensor data, and security logs. As you begin your big data analysis, you probably do not know exactly what you will find. Your analysis will go through several stages. Exploratory stage Codifying stage Integration and incorporation stage
12/3/2013
BIG DATA FOR DUMMIES
Exploratory Stage
In the early stages of your analysis, you will want to search for
patterns in the data. It is only by examining very large volumes (terabytes and petabytes) of data that new and unexpected relationships and correlations among elements may become apparent. You will need a platform such as Hadoop for organizing your big data to look for these patterns. In the exploratory stage, you are not so concerned about integration with operational data.
12/3/2013
distributed servers. Flume is designed for scalability and can continually add more resources to a system to handle extremely large amounts of data in an efficient way. Flumes output can be integrated with Hadoop and Hive for analysis of the data. Flume also has transformation elements to use on the data and can turn your Hadoop infrastructure into a streaming source of unstructured data.
12/3/2013
through huge amounts of streaming data and pull out the trending patterns that relate to specific products or customers. As companies search for patterns in big data, the huge data volumes are narrowed down as if they are passed through a funnel. You may start with petabytes of data and then, as you look for data with similar characteristics or data that forms a particular pattern, you eliminate data that does not match up.
12/3/2013
Codifying stage
After you find something interesting in your big data analysis,
12/3/2013
will allow to integrate or incorporate the results of big data analysis into business process and real-time business actions. Technologies for high-speed transport of very large and fast data are a requirement for integrating across distributed big data sources and between big data and operational data. A company that uses big data to predict customer interest in new products needs to make a connection between the big data and the operational data on customers and products to take action. If the company wants to use this information to buy new products or change pricing it needs to integrate its operational data with the results of its big data analysis.
12/3/2013
BIG DATA FOR DUMMIES
12/3/2013
10
warehouse environments.
ETL tools are used to transform the data into the format required
more than traditional data warehouses. ETL can support integration across transactional systems, operational data stores, BI platforms, MDM hubs, the cloud, and Hadoop platforms
12/3/2013
11
Extract
from
the
source
Transform
Convert the format of the extracted data so that it conform to the requirements of the target database.
Load
12/3/2013
12
Data transformation
Data transformation is the process of changing the format of data
so that it can be used by different applications. This process also includes mapping instructions so that applications are told how to get the data they need to process. The process of data transformation is made far more complex because of the staggering growth in the amount of unstructured data. Data transformation tools are not designed to work well with unstructured data. As a result, companies faced with a significant amount of manual coding to accomplish the required data integration.
12/3/2013
BIG DATA FOR DUMMIES
13
quality. Phase 2: After you locate your patterns and establish results that are important to the business, apply the same data quality standards that you apply to your traditional data sources. You want to avoid collecting and managing big data that is not important to the business and will potentially corrupt other data elements in Hadoop or other big data platforms.
12/3/2013
14
and to otherwise improve on the ETL and data-staging processes. You can speed up the data integration process by loading both unstructured data and traditional operational and transactional data directly into Hadoop, regardless of the initial structure of the data. After the data is loaded into Hadoop, it can be further integrated using traditional ETL tools. When Hadoop is used as an aid to the ETL process, it speeds the analytics process.
12/3/2013
15
12/3/2013
16
17
Introduction
Explaining Streaming Data
Business
12/3/2013
BIG DATA FOR DUMMIES
18
Data Streaming
MEANING :Streaming data is an analytic computing platform that is focused on speed. This is because these applications require a continuous stream of often unstructured data to be processed. o Therefore, data is continuously analyzed and transformed in memory before it is stored on a disk. o Processing streams of data works by processing time windows of data in memory across a cluster of servers. o It is a single-pass analysis i.e the analyst cannot reanalyze the data after it is streamed. o Streaming data is useful when analytics need to be done in real time while the data is in motion.
12/3/2013
BIG DATA FOR DUMMIES
19
PRINCIPLES :-
12/3/2013
USES :A power plant A power plant needs to be a highly secure environment. Companies often place sensors around the perimeter of a site to detect movement. Therefore, the vast amount of data coming from these sensors needs to be analyzed in real time so that an alarm is sounded only when an actual threat exists.
It is a highly competitive market. Communications systems generate huge volumes of data that have to be analyzed in real time to take th appropriate action. A delay in detecting an error can seriously impact customer satisfaction.
BIG DATA FOR DUMMIES
12/3/2013
21
Continued.
This needs to know exactly the sources of oil, environmental factors impacting their Operations, water depth, temperature, ice flows etc. This massive amount of data needs to be analyzed and computed so that mistakes are avoided.
These are required to be able to take massive amounts of data from brain scans and analyze the results in real time to determine where the source of a problem is and what type of action needed to be taken to help the patient.
22
PRODUCTS FOR STREAMING DATA : IBM Infosphere Streams InfoSphere Streams provides continuous analysis of massive data volumes. It is intended to perform complex analytics of heterogeneous data types. It can perform real-time and look-ahead analysis of regularly generated data, using digital filtering, pattern/correlation analysis, and decomposition as well as geospacial analysis.
12/3/2013
BIG DATA FOR DUMMIES
23
Twitters Storm Twitters Storm is an open source real-time analytics engine. Twitter uses Storm internally. It is still available as open source and has been gaining significant
traction among emerging companies. It can be used with any programming language for applications Storm is designed to work with existing queuing and database technologies.
12/3/2013
24
Apache S4 The four Ss in S4 stand for Simple Scalable Streaming System. It allows programmers to easily develop applications for
processing continuous streams of data. S4 is designed as a highly distributed system. The S4 design is best suited for large-scale applications for data mining and machine learning in a production environment.
12/3/2013
25
12/3/2013
26
It creates a tiered loyalty program to increase repeat sales. Using a CEP platform, the system triggers a process that offers the customer an extra discount on a related product.
These uses CEP to better manage fraud.. The underlying system will correlate the incoming transactions, track the stream of event data, and trigger a process.
CEP is also implemented in financial trading applications, weather-reporting applications, and sales management Applications.
27
BIG DATA FOR DUMMIES
VENDORS OF CEP : Esper (open source vendor), IBM with IBM Operational Decision Manager, Informatica with RulePoint, Oracle with its Complex Event Processing Solution, Microsofts StreamInsights, SAS DataFlux Event Stream Processing Engine, Streambases CEP
12/3/2013
28
computing is typically applied to analyzing vast amounts of data in real time, while CEP is focused on solving a specific use case based on events and actions. In many situations CEP is dependent on data streams; however, CEP is not required for streaming data. Streaming computing is used to handle unstructured data, while CEP deals with variables correlated with specific business process. Streaming data is managed in a highly distributed clustered environment, while CEP often run on less complex hardware.
12/3/2013
29
leverage a business process engine to apply business rules to the results of that streaming data analysis.
12/3/2013
30
12/3/2013
31