Вы находитесь на странице: 1из 2

DATA FABRIC FOR 360 DATA HUB

SOLUTION BRIEF

TURNING DATA INTO BUSINESS OUTCOME WITH A 360 DATA HUB


dfhglkshjglkjshgTION BRIEF
Data is the lifeblood of digital transformations. The challenge for a successful digital transformation doesn’t stop at collecting, connecting and
turning data into insights through data warehouses orDATASHEET
lakes. It also requires delivering those data-driven insights into everything you do. Your data
brings much more value once it is actionable, flowing in real time in a controlled way across people, applications, analytics, and machine learning.
This is particularly true for data about your most precious assets: your customers, products, and employees.
Modern data architectures don’t stop at creating a single source of trusted data and storing
it in a central place. They provide an API-first approach for connecting data providers and
data consumers, while allowing data controllers to orchestrate data flows in a governed
way. The result is more actionable and trustworthy data, delivered in real-time, both
internally and externally to the edge.
Once you’ve overcome the challenge of delivering trusted data in real-time at the edge,
you can reap the business benefits:
• Sales can grow through personalized offers and recommendations
• Products get smarter and adapt to their usage context
• Compliance is under control in a holistic way, and
• Employees can constantly learn and boost efficiency with smart guidance

3600 Data Hub – Frictionless connection across any data points, from collection to consumption
Increasingly, organizations are turning to data hubs to accomplish these and similar results by improving the exchange of information between data
sources and data consumers. A data hub is a data exchange with frictionless data flow at its core. Endpoints, which can be applications, processes,
people or algorithms interact with the hub, potentially in real time, to provision data into the hub or receive data from it. When data enters the
data hub, it gets connected to other data points and reconciled into a 360° view. The data hub also provides a point of governance for the data, as
well as visibility into how the data is flowing throughout the enterprise. Ultimately, trustworthy data can be consumed and reused in an easy and
consistent way by anyone or any application that can benefit.
Primarily, a 360 data hub is a set of rules for data mediation and governance that helps organizations:
• Expand and speed up the sharing of information – between own applications, employees as well as with/from customers and partners
• Create a trusted, but continuously evolving “Golden Record” of their customers, vendors and partners
• Drive business outcomes with data across the operations, for improved customer experiences, streamlined operations, or risk control.
Most data architectures have been designed as centralized data stores. This design pattern has proven success in analytics, for delivering
structured analytics to the mass – the data warehouses – or to discover hidden insights from big data and learn from it - the data lakes. But when
the goal is data exchange — I.e. data flows rather than data stores — a 3600 data hub should be at the core of the architecture because it is more
agile and flexible than legacy architectures.

Enabling any application with unified metadata, master data, operational data, and analytics
A data hub is designed for rapid exchange of information needed by today’s organizations. It captures any data, including metadata, master data,
operational data, and analytical data and reconciles them. Then it delivers the data in multiple desired formats without necessarily storing it
physically in a central place. Using search-based applications and API Services, data consumers can easily discover and get instant access to data
they can trust. A data hub also establishes a single
point of trust for data flows, with end to end DATA WAREHOUSE DATA LAKE DATA HUB
visibility and lineage.
A 360° data hub is a design pattern, rather than
something that comes out of the box from a
commercial product. And it preferably runs on the
cloud to reach any potential data points and • STRUCTURED FOR ANALYTICS • (UN)STRUCTURED FOR DISCOVERY • STRUCTURED FOR DATA PORTABILITY
• CONSUMED BY PEOPLE AS A SELF- • CONSUMED BY DATA PROFESSIONALS • CONSUMED BY PEOPLE AND APPS
process data at scale. • FOCUSED ON DATA INTEGRITY AND
SERVICE AND ALGORITHMS
• FOCUSED ON DECISION MAKING • FOCUSED ON DEEP LEARNING, AI SPEED FOR SHARING
Enacting the 360° Data Hub in 6 steps
A data hub is designed and customized to fit the needs of an organization, and can be architected with a six-step approach that combines the most
common discipline of data modeling, data integration, application integration, and data governance:
1. The first step is to connect the data sources that need to be shared with the data hub. At this stage, depending on an organization’s
needs, the data in its raw format could be stored in the hub or simply pass through.
2. The incoming data flows are then put into a canonical data model so that they can easily be connected.
3. The data sets then get inventoried in a data catalog and organized for tasks like data categorization, curation, protection and
remediation.
4. Data quality techniques, such as matching, survivorship, and deduplication are then applied to create a golden record. The golden record
is stored in the data hub using the data store that best meets the context.
5. Analytics can then be deployed to augment the data – for segmentation, forecasting, predictive behavior, recommendations, etc.
6. Finally, the data can then be shared with data consumers – applications, systems, business users and data professionals, 3rd parties, etc.

Raw Conformed Curated Reconciled Augmented Shared

CRM CRM

Batch or Real Time


Batch or Real Time

ERP ERP

Logs Logs

IOT IOT

External External
Data Data

Analytics Analytics

Data Lineage

Talend Data Fabric – The Foundation For A Successful 360° Data Hub Implementation
Talend Data Fabric is a unified platform for data ingestion, integration, governance and sharing, which can simplify the process of building a data
hub. Using Talend Data Fabric, organizations can collect, govern, transform and share any data pipeline as part of a data hub:
1. Collect: with 1,000+ built-in components and
connectors, datasets can be captured,
standardized, and cleaned so that they can be
connected altogether. Data cataloging techniques
are applied to discover and automatically profile,
document, and categorize the incoming data.
2. Govern: Data governance is then applied for data
ownership, data certification, data stewardship,
and data remediation. You can also track and
trace all your data flows, with end to end lineage
that shows where the data came from and where
it is going.
3. Transform: Talend Data Fabric can create a golden record for reconciling and cross referencing your data using data quality, matching, and
survivorship. Eventually, the golden record can be persisted in potentially any storage layer for reuse. Once it has been quality proofed and
reconciled, data can also be enriched through analytics, leveraging Spark®, Python®, Databricks®, RapidMiner®, Qubole®, etc.
4. Share: Because they have been automatically documented in a data catalog, datasets can be easily discovered and consumed by anyone –
from data engineers to business users and application developers. Data can also be easily consumed into any applications through Talend’s
Cloud API Services.
Contact Talend today to find out more about how Talend Data Fabric can help you build a Customer 3600 Data Hub.

© 2019 Talend All rights reserved WP277-EN

Вам также может понравиться