Вы находитесь на странице: 1из 14

Talend Open Studio

Talend Open Studio is a versatile set of open source products for developing, testing, deploying
and administrating data management and application integration projects. Talend delivers a
platform that makes data management and application integration easier by providing a unified
environment for managing the entire lifecycle across enterprise boundaries. For ETL projects,
Talend Open Studio for Data Integration delivers a rich feature set including a graphical
integrated development environment with an intuitive Eclipse-based interface. Drag-and-drop
job design and a unified repository for storing and reusing metadata. The broadest data
connectivity support of any data integration platform, with more than 400 built-in connector
components that let you quickly bridge between databases, mainframes, file systems, web
services, packaged enterprise applications, data warehouses, OLAP applications, Software-as-a-
Service and Cloud-based applications, and more. The advanced ETL functionality including
string manipulations, automatic lookup handling, and management of slowly changing
dimensions and support for ELT (extract, load, and transform) as well as ETL, even within a
single job.

Try Sisense Free

Sisense lets you analyze complex data with complete agility. From easy data preparation to
beautiful dashboards and interactive analytics, Sisense gives your business a simple way to
manage, analyze and visualize all your data.

Sisense Free
Talend Open Studio

Talend Open Studio

2.GeoKettle ETL

GeoKettle is a powerful, metadata-driven spatial ETL (Extract, Transform and Load) tool
dedicated to the integration of different data sources for building and updating geospatial
databases, data warehouses and services. GeoKettle enables the Extraction of data from data
sources, the Transformation of data in order to correct errors, make some data cleansing,
change the data structure, make them compliant to defined standards, and the Loading of
transformed data into a target DataBase Management System (DBMS) in OLTP or OLAP/SOLAP
mode, GIS file or Geospatial Web Service.

GeoKettle
GeoKettle

3.Dataiku Data Science Studio (DSS) Community Edition

Dataiku Data Science Studio (DSS) is a software platform that aggregates all the steps and big
data tools necessary to get from raw data to production ready application. It provides Visual
interactive data preparation (80+ processors), Visual transformations (Group, join, union, split,
sampling, …), Smart incremental rebuild, Concurrent jobs, Builtin engines (Streaming and in-
memory), In-database processing. Provides Interactive data cleaning and enrichment with easy
access to over 80 built-in visual processors for code-free data wrangling, automatically
suggested contextual transformations and perform mass actions on your data.

Data Science Studio (DSS) Community Edition


4.Jaspersoft ETL

Jaspersoft ETL is easy to deploy and out-performs many proprietary and open source ETL
systems. It is used to extract data from your transactional system to create a consolidated data
warehouse or data mart for reporting and analysis. Features include business modeler to access
a non-technical view of the information workflow, display and edit the ETL process with Job
Designer, a graphical editing tool, define complex mappings and transformations with
Transformation Mapper and other transformation components and generate portable Perl or
Java code that can be executed on any machine. Also the ability to track ETL statistics from start
to finish with real-time debugging, allow simultaneous output from and input to multiple
sources including flat files, XML files, databases, web services, POP and FTP servers with
hundreds of available connectors and use of the Activity Monitoring Console (AMC) to monitor
job events (successes, failures, warnings, etc.), execution times, and data volumes.

Jasepersoft ETL

Jaspersoft ETL

5.HPCC Systems

HPCC Systems is an Open-source platform for Big Data analysis with a Data Refinery engine
called Thor. Thor clean, link, transform and analyze Big Data. Thor supports ETL (Extraction,
Transformation and Loading) functions like ingesting unstructured/structured data out, data
profiling, data hygiene, and data linking out of the box. The Thor processed data can be
accessed by a large number of users concurrently in real time fashion using the Roxie, which is a
Data Delivery engine. Roxie provides highly concurrent and low latency real time query
capability.

HPCC Systems
HPCC Systems

6. Jedox

Jedox is an Open-Source BI solution for Performance Management including Planning, Analysis,


Reporting and ETL. The Open Core consist of an in-memory OLAP Server, ETL Server and OLAP
client libraries. Powerfully supporting Jedox OLAP server as a source and target system, Jedox
ETL is specifically designed to meet the challenges of OLAP analysis. Working with cubes and
dimensions couldn’t be easier. Flexibly generate frequently-needed time hierarchies and
efficiently transform the relational model of source systems into an OLAP model – with JEDOX
ETL.

Jedox

7. Pentaho ETL

Pentaho ETL is an intuitive, graphical, drag and drop design environment and a proven, scalable,
standards-based architecture. Pentaho Data Integration also called Kettle is the component of
Pentaho responsible for the Extract, Transform and Load (ETL) processes. Features include
migrating data between applications or databases, exporting data from databases to flat files,
loading data massively into databases, data cleansing and integrating applications.
Pentaho ETL

Pentaho ETL

8. No frills transformation

“No frills transformation” (NFT) is intended to be a lightweight transformation engine, having


an extensible interface which makes it simple to extend with Source Readers, extend with
Target Writers and extend with additional Operators (if you can’t do with the Custom
Operators)

Out of the box, NFT will read from CSV files in any encoding Salesforce SOQL queries, SQLite
Databases, MySql Databases, Oracle Databases, SQL Server Databases and from SAP RFCs if
they have a TABLE as output value and write to CSV files in any encoding (including with or
without UTF-8 BOMs), Salesforce Objects (including Upserts and using External IDs), Oracle
Databases and Rudimentary XML files.

No frills transformation
9. EplSite ETL

EplSite ETL is a tool to do easy the data migrations and fact table creation, doing extraction,
transformation, validation and load in a very fast way. EplSite ETL is low resource consuming,
has a Web interface, and very easy to customize it because it is developed in Perl. It is possible
to run transformations using cron jobs on Linux or task manager on Windows.

EplSite ETL
10. GETL ETL

GETL, automates the work of loading and transforming data. GETL is a set of libraries of pre-
built classes and objects that can be used to solve problems unpacking, transform and load data
into programs written in Groovy, or Java, as well as from any software that supports the work
with Java classes. GETL features include simpler the class hierarchy, the easier solution, the
data structures tend to change over time, or not be known in advance, working with them must
be maintained. All routine work ETL should be automated wherever possible, compiling the
code on the fly bail speed and reserve for the optimization, sophisticated class hierarchy
guarantee easy connection of other open source solutions.

GETL

GETL

11. Scriptella ETL


Scriptella is an open source ETL (Extract-Transform-Load) and script execution tool written in
Java. Its primary focus is simplicity. The features include executing scripts written in SQL,
JavaScript, JEXL, Velocity .Database migration, interoperability with LDAP, JDBC, XML and other
data sources. Cros database ETL operations, import/export from/to CSV, text and XML and
other formats.

Scriptella

12. KETL(tm)

KETL(tm) is a production ready ETL platform. The engine is built upon an open, multi-threaded,
XML-based architecture. The data integration platform is built with portable, java-based
architecture and open, XML-based configuration and job language. KETL major features include
support for integration of security and data management tools, proven scalability across
multiple servers and CPU’s and any volume of data and no additional need for third party
schedule, dependency, and notification tools.

KETL(tm)

13. Apatar ETL

Apatar ETL brings a set of unmatched capabilities in an open source package. Features include
connectivity to Oracle, MS SQL, MySQL, Sybase, DB2, MS Access, PostgreSQL, XML, InstantDB,
Paradox, BorlandJDataStore, Csv, MS Excel, Qed, HSQL, Compiere ERP, SalesForce.Com,
SugarCRM, Goldmine, any JDBC data sources. There is a single interface to manage all
integration projects, flexible deployment options, bi-directional integration, platform-
independent, runs from Windows, Linux, Mac; 100% Java- based, no coding, visual job designer
and mapping enable non-developers to design and perform transformations.

Apatar ETL
Apatar

14. RapidMiner

RapidMiner is one of the leading data mining software suites. RapidMiner supports all steps of
the data mining process from data loading, pre-processing, visualization, interactive data
mining process design and inspection, automated modeling, automated parameter and process
optimization, automated feature construction and feature selection, evaluation, and
deployment. RapidMiner can be used as stand-alone program on the desktop with its graphical
user interface (GUI), on a server via its command line version.

RapidMiner
RapidMiner

15. Anatella

Anatella is an ETL tool built especially for analytical purposes and predictive datamining. It
includes some features such as data transformations and meta-data transformations that are
unique and extremely valuable in this field. Anatella enables to use the flexible JavaScript
language to easily create new, extremely complex data transformations. Using, creating and
debugging new data manipulation scripts is simple and intuitive.

Anatella

16. Apache Falcon

Falcon is a feed processing and feed management system aimed at making it easier for end
consumers to onboard their feed processing and feed management on Hadoop clusters. Falcon
establishes relationship between various data and processing elements on a Hadoop
environment. Feed management services such as feed retention, replications across clusters,
archival etc. Easy to onboard new workflows/pipelines, with support for late data handling,
retry policies. Integration with megastore/catalog such as Hive/HCatalog and provide
notification to end customer based on availability of feed groups.

Apache Falcon

17. Apache Crunch

Crunch, is a Java library that aims to make writing, testing, and running MapReduce pipelines
easy, efficient. Running on top of Hadoop MapReduce and Apache Spark, the Apache Crunch
library is a simple Java API for tasks like joining and data aggregation that are tedious to
implement on plain MapReduce. The APIs are especially useful when processing data that does
not fit naturally into relational model, such as time series, serialized object formats like protocol
buffers or Avro records, and HBase rows and columns. For Scala users, there is the Scrunch API,
which is built on top of the Java APIs and includes a REPL (read-eval-print loop) for creating
MapReduce pipelines.

Apache Crunch

18. Cascading

Cascading is a Java library and does not require installation. The data processing APIs define
data processing flows. The APIs exposed provide a rich set of capabilities that allow you to think
in terms of the data and the business problem such as sort, average, filter, merge etc. The data
integration API allows you to isolate your integration dependencies from your business logic.
You can easily read/write from a variety of external systems to Hadoop, and then write those
results to another system. Taps and Schemes enable read/write capabilities between any
source and in any format. Cascading comes with several pre-built taps and schemes and also
provides you the flexibility to quickly build your own.

Cascading

19. Apache Oozie

Apache Oozie is a Java Web application used to schedule Apache Hadoop jobs. Oozie combines
multiple jobs sequentially into one logical unit of work. It is integrated with the Hadoop stack,
with YARN as its architectural center, and supports Hadoop jobs for Apache MapReduce,
Apache Pig, Apache Hive, and Apache Sqoop. Oozie can also schedule jobs specific to a system,
like Java programs or shell scripts.
Apache Oozie

Вам также может понравиться