Академический Документы
Профессиональный Документы
Культура Документы
Aggregate cache: The Integration Service stores data in the aggregate cache until it completes
aggregate calculations. It stores group values in an index cache and row data in the data cache.
Aggregate expression: Enter an expression in an output port. The expression can include nonaggregate expressions and conditional clauses.
Group by port: Indicate how to create groups. The port can be any input, input/output, output, or
variable port. When grouping data, the Aggregator transformation outputs the last row of each
group unless otherwise specified.
Sorted input: Select this option to improve session performance. To use sorted input, you must
pass data to the Aggregator transformation sorted by group by port, in ascending or descending
order.
You can configure the Aggregator transformation components and options on the Properties and Ports
tab.
Property
Cache Directory
Tracing Level
Sorted Input
Aggregator Data
Cache Size
Aggregator Index
Cache Size
Description
Local directory where the Integration Service creates the index and data cache files. By
default, the Integration Service uses the directory entered in the Workflow Manager for
the process variable $PMCacheDir. If you enter a new directory, make sure the directory
exists and contains enough disk space for the aggregate caches. If you have enabled
incremental aggregation, the Integration Service creates a backup of the files each time
you run the session. The cache directory must contain enough disk space for two sets of
the files
Amount of detail displayed in the session log for this transformation.
Indicates input data is presorted by groups. Select this option only if the mapping passes
sorted data to the Aggregator transformation
Data cache size for the transformation. Default cache size is 2,000,000 bytes. If the total
configured session cache size is 2 GB (2,147,483,648 bytes) or greater, you must run the
session on a 64-bit Integration Service. You can configure the Integration Service to
determine the cache size at run time, or you can configure a numeric value. If you
configure the Integration Service to determine the cache size, you can also configure a
maximum amount of memory for the Integration Service to allocate to the cache
Index cache size for the transformation. Default cache size is 1,000,000 bytes. If the total
configured session cache size is 2 GB (2,147,483,648 bytes) or greater, you must run the
session on a 64-bit Integration Service. You can configure the Integration Service to
determine the cache size at run time, or you can configure a numeric value. If you
configure the Integration Service to determine the cache size, you can also configure a
maximum amount of memory for the Integration Service to
allocate to the cache.
Transformation
Scope
Specifies how the Integration Service applies the transformation logic to incoming data:
Transaction. Applies the transformation logic to all rows in a transaction.
Choose Transaction when a row of data depends on all rows in the same
transaction, but does not depend on rows in other transactions.
All Input. Applies the transformation logic on all incoming data. When you choose
All Input, the PowerCenter drops incoming transaction boundaries. Choose All
Input when a row of data depends on all rows in the source.
Solution:
You can also select Transformation by clicking function button on Informatica Designer
Drag input port from EMPLOYEE_ID, SALARY, DEPARTMENT_ID from SQ_EMPLOYEES (Source
Qualifier ) to aggr_emp_Salary
Also add additional below output port (by Clicking on Create port button)
o COUNT
o MIN_SAL
o MAX_SAL
o AVG_SAL
o SUM_SAL
Edit the expression for AVG_SAL ( by clicking on expression editor ) and add below expression
AVG(SALARY)
Overall Mapping:
FROM
(SELECT
FROM
UNION ALL
SELECT
Problem Statement: Remove the duplicate rows coming from the union transformation before
loading into the target table.
Solution:
Use the mapping that weve previously created for Union transformation example:
Disconnect the links from the union transformation to the target table and insert an Aggregator
transformation between them. Link the output ports of the union transformation to the
Aggregator tranformation:
In the Ports tab of the aggregator transformation select all the columns as Group by columns:
Click OK
Drag all the ports from aggr_Employees to the target table Employees:
Click on Mapping ( from tool bar) -> then Validate ( to validate the mapping)
Now save the mapping ( by clicking on Repository-> Save or you can also press Ctrl+S)
Generate the workflow and run it