Documenti di Didattica
Documenti di Professioni
Documenti di Cultura
The Browse In-DB tool lets you view your data at any point in an In-DB workflow. Use the Browse In-DB tool
as you build an In-DB workflow to ensure data is passing through the way you intend it to. It allows you to
view results at any point in the In-DB workflow.
In-Database enables blending and analysis against large sets of data without moving the data out of a database
and can provide significant performance improvements over traditional analysis methods. For more about the
In-Database tool category, see In-Database Overview.
Each Browse In-DB tool triggers a database query and can impact performance.
Enable caching: The Browse In-DB tool will cache the number of records returned when the workflow is run.
This option will default checked, but can be turned off. If the database connection or query (including the
number of records to browse) does not change then the query will not be re-run, but instead the records that
display in the window are pulled from the cache.
Clear Cache: Click the Clear Cache button to clear the data from the cache.
The tool has a display limitation of 2 billion records. If there are more than 2 billion records in the data stream,
there will be a message in the Data View statistics. All records will be written to the desired file type when
exporting from the Browse.
In-Database enables blending and analysis against large sets of data without moving the data out of a database
and can provide significant performance improvements over traditional analysis methods. For more about the
In-Database tool category, see In-Database Overview.
2. Once the connection is configured, Table or Query displays the name of the selected database table.
3. (Optional) Click Query Builder to select tables and construct queries. See Choose Table or Specify Query
Window.
In-Database enables blending and analysis against large sets of data without moving the data out of a database
and can provide significant performance improvements over traditional analysis methods. For more about the
In-Database tool category, see In-Database Overview.
Tool complexity
While most In-DB tools do not require SQL commands, this tool requires SQL for more advanced processing.
o Basic Filter: Use the basic filter to construct a simple query on a single field in the database.
Operator Meaning
= Equals
!= Does not equal
> Is greater than
>= Is greater than or equal to
< Is less than
<= Is less than or equal to
IS NULL Is a missing or unknown value
IS NOT NULL Is not a missing or unknown value
Operator Meaning
LIKE Is similar to a specified pattern in a column
o Custom Filter: The Custom Filter acts as a SQL WHERE query. Use the custom filter to
construct a more complex expression or to query from multiple fields in the database.
1. Use Insert Fields to pick from available fields to construct your expression.
2. Type the rest of the query in the box using the native language of the database.
If a query is constructed using the Basic Filter, a read-only query displays in the Custom Filter
area. If the Custom Filter option is then selected, the query becomes editable.
2. Validate expression here at runtime: An option that sends a query to the database to report errors
contained in the expression in the results window of this tool.
In-Database enables blending and analysis against large sets of data without moving the data out of a database
and can provide significant performance improvements over traditional analysis methods. For more about the
In-Database tool category, see In-Database Overview.
2. If multiple join fields are desired, an additional row of join fields can be configured. Simply click
on the drop down to choose additional join field per input.
3. To delete a join field, click on the number on the left hand side and click the delete button on
the right.
The Join tool restricts what field types can be joined together. Mismatching data types can result in error
messages.
Inner Join: Contains only the records from the Left input that joined to
records in the Right input.
Left Outer Join: Contains all records from the Left input including the
records that joined with the Right input.
Right Outer Join: Contains all records from the Right input including the
records that joined with the Left input.
Full Outer Join: Contains all records from both the Left and Right inputs.
In-Database enables blending and analysis against large sets of data without moving the data out of a database
and can provide significant performance improvements over traditional analysis methods. For more about the
In-Database tool category, see In-Database Overview.
Creation Mode: Select the appropriate option for writing the data. Choices include:
Create Temporary Table: Writes to a temporary table that is available until the end of the session. This
option is useful for building In-DB predictive macros because it holds the metadata in place temporarily.
If this option is selected, the Table Name field is disabled and displays “[a unique temporary table name
is generated on each run]".
Create New Table: Creates a new table. Will not overwrite an existing table.
If an HDFS Avro option is selected, the avro.schema.literal fails at 4000 characters and an error will
occur on the table creation. Try reducing the character length of the column names, or selecting fewer
columns.
Overwrite Table (Drop): Completely drops the existing table and creates a new one.
Table Name: Enter the name of the database table to create or update. If Create Temporary Table is selected, the
Table Name field is disabled and displays “[a unique temporary table name is generated on each run]".
Oracle pemissions
To use this tool with Oracle, you must have permissions to write to the tempspace assigned to GLOBAL
TEMPORARY. Contact your Oracle database administrator.
In-Database enables blending and analysis against large sets of data without moving the data out of a database
and can provide significant performance improvements over traditional analysis methods. For more about the
In-Database tool category, see In-Database Overview.
3. Manipulate the sorting order by using the up, down, and delete buttons on the right.
Connect a standard tool to the output of the Data Stream Out tool.
The Dynamic Input In-DB tool is used in conjunction with the Dynamic Output In-DB Tool when creating an
In-DB macro for predictive analysis. The Dynamic Input In-DB tool can take In-DB Connection Name and
Query fields from a standard data stream and input them back into an In-DB data stream.
In-Database enables blending and analysis against large sets of data without moving the data out of a database
and can provide significant performance improvements over traditional analysis methods. For more about the
In-Database tool category, see In-Database Overview.
Query / Query Alias List Field: The In-DB query created at this point in the workflow.
The Dynamic Output In-DB tool is used in conjunction with the Dynamic Input In-DB Tool when creating an
In-DB macro for predictive analysis. The Dynamic Output In-DB tool can take the metadata from the In-DB
query and pass it into a standard workflow with predictive tools.
In-Database enables blending and analysis against large sets of data without moving the data out of a database
and can provide significant performance improvements over traditional analysis methods. For more about the
In-Database tool category, see In-Database Overview.
o Input Connection String: The information about the database that is needed to be able to
establish a connection to it.
o Output Connection String: The information about the database that is needed to be able to
establish a connection to it.
o Query Alias List: Displays each segment of the query in the form of a common table expression.
o Last Query Alias: The last alias from the Query Alias List.
Possible Errors
If either the Input Connection String, or the Output Connection String fields are selected for output, the
following error may display:
To use this tool select an appropriate data source and select the "Allow Decryption of Password" Password
Encryption option in Manage In-DB Connections.
To resolve the error, modify the original connection string by changing the Password Encryption option to
"Allow Decryption of Password" so that the password is decrypted in the metadata.
An input anchor will appear on the macro tool for each Macro Input In-DB tool used in the macro workflow in
the order they are brought onto the canvas (left to right, or up to down). The input anchors can be re-ordered in
the Interface Designer Window from the Tree view.
In-Database enables blending and analysis against large sets of data without moving the data out of a database
and can provide significant performance improvements over traditional analysis methods. For more about the
In-Database tool category, see In-Database Overview.
Manage Connections: Click to make edits to a connection that has already been set up or to
create a new connection. See Manage In-DB Connections .
o Table or Query: Displays the name of the selected table in the database once the connection has been
configured. Click Query Builder to easily select tables and construct queries in the Choose Table or
Specify Query Window.
2. Input Name: The Input Name will be visible to the end user when they configure the macro tool.
3. Anchor Abbreviation: This optional parameter will display an abbreviation on the input anchor of the macro tool.
4. Show Field Map: When this option is checked, and the macro tool is added to a workflow, the end user will be
asked to select the fields that match up with the selected template Input.
An output anchor will appear on the macro tool for each Macro Output In-DB tool used in the macro workflow
in the order they are brought onto the canvas (left to right, or up to down). The output anchors can be re-ordered
in the Interface Designer Window from the Tree view.
In-Database enables blending and analysis against large sets of data without moving the data out of a database
and can provide significant performance improvements over traditional analysis methods. For more about the
In-Database tool category, see In-Database Overview.
2. Anchor Abbreviation: This optional parameter will display an abbreviation on the output anchor of the macro
tool.
o Percent: Returns the percent of records specified. Selecting this option requires the data to pass thru
the tool twice: once to calculate the count of records and then to return the specified percent of
records.
3. Sample records based on order: When checked, the records will be sorted in-database before the number or
percent of records is chosen to produce the results. To configure the order, the data should be sorted using the
Fields table.
o Manipulate the sorting order by using the up, down, and delete buttons on the right.
Use the Select menu to the right to Select to make multiple field selections Choices include:
o Numeric: only numeric fields are selected (integers, fixed decimals, floats, doubles) to apply to Actions.
Make the selection and it will appear in the Actions section. Different summary functions are available
depending on the type of data field selected.
Group by: Combines database records with identical values in a specified field into a single record. All
of the resulting data from the records in a group are then summarized. (any non-blob or spatial object
has this option)
Count Non Null: Count of unique records in the group that are not [Null]. A Null field means there is no value set
for this field; this is different from having a zero or an empty string.
Count Null: Count of unique records on the group that are [Null].
o Sum: Returns the sum value for the group. The sum is calculated by adding all of the values of a group.
o Average: Calculates an average value for the group. The average is calculated by taking the sum of all
values divided by the total number of values.
o Standard Deviation: Calculates the standard deviation for the group. Standard Deviation is a
measurement variability used in statistics.
o Variance: Calculates the Variance for the group. The variance is calculated by taking the Standard
Deviation and multiplying it times itself, StdDev^2.
From the Actions section, you can select the field and use the up, down, and delete buttons to specify field order
for the output.
Rename a field by typing a new field name into the Output Field Name column.
Properties: Additional properties need specification for certain actions. Actions with additional properties
specification include: Concatenate Strings and Finance actions.
In-Database enables blending and analysis against large sets of data without moving the data out of a database
and can provide significant performance improvements over traditional analysis methods. For more about the
In-Database tool category, see In-Database Overview.
o Auto Config by Position: Aligns fields by their field order in the stream.
2. When Fields are Different: Select how to handle nonconforming data fields from the dropdown. Choices are:
o Error - Stop Processing: will throw an error in the Results window, and end the schema.
o Output All Fields: All fields will be included. Null values will populate empty fields.
o Output Common Subset of Fields: Only the fields that each input has in common will be outputted.
o Append Existing: Appends all the data to an existing table. Output will consist of Records Before +
Records After.
o Delete Data & Append: Deletes all the original records from the table and then appends the data into
the existing table.
o Overwrite Table (Drop): Completely drops the existing table and creates a new one.
o Create New Table: Creates a new table. Will not overwrite an existing table.
o Create Temporary Table: Writes to a temporary table that is available until the end of the session. This
option is useful for building In-DB predictive macros because it holds the metadata in place temporarily.
If this option is selected, the Table Name field is disabled and displays “[a unique temporary table name
is generated on each run]".
2. Table Name: Enter the name of the database table to create or update.
3. Append Fields Mapping: this area becomes active when Append Existing or Delete Data & Append is chosen
above.
Auto Config by Position: Aligns fields by their field order in the stream.
o When Fields are Different: Select how to handle nonconforming data fields from the options using the
drop-down.
Error - Stop Processing: will throw an error in the Results window, and end processing.
Output Applicable Fields: Applicable fields will be included. Null values will populate empty
fields.